program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.7.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "9.0b1"}})] { func main(tensor audio_length, tensor audio_signal) { tensor var_21 = const()[name = tensor("op_21"), val = tensor(160)]; tensor var_22 = const()[name = tensor("op_22"), val = tensor(1)]; tensor var_32 = const()[name = tensor("op_32"), val = tensor(512)]; tensor var_33 = add(x = audio_length, y = var_32)[name = tensor("op_33")]; tensor var_34 = const()[name = tensor("op_34"), val = tensor(512)]; tensor var_35 = sub(x = var_33, y = var_34)[name = tensor("op_35")]; tensor floor_div_0 = floor_div(x = var_35, y = var_21)[name = tensor("floor_div_0")]; tensor var_36_to_fp16_dtype_0 = const()[name = tensor("op_36_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_37_promoted_to_fp16 = const()[name = tensor("op_37_promoted_to_fp16"), val = tensor(0x1p+0)]; tensor floor_div_0_to_fp16 = cast(dtype = var_36_to_fp16_dtype_0, x = floor_div_0)[name = tensor("cast_181")]; tensor seq_len_1_cast_fp16 = add(x = floor_div_0_to_fp16, y = var_37_promoted_to_fp16)[name = tensor("seq_len_1_cast_fp16")]; tensor seq_len_dtype_0 = const()[name = tensor("seq_len_dtype_0"), val = tensor("int32")]; tensor var_41_begin_0 = const()[name = tensor("op_41_begin_0"), val = tensor([0, 0])]; tensor var_41_end_0 = const()[name = tensor("op_41_end_0"), val = tensor([1, 1])]; tensor var_41_end_mask_0 = const()[name = tensor("op_41_end_mask_0"), val = tensor([true, false])]; tensor var_41_squeeze_mask_0 = const()[name = tensor("op_41_squeeze_mask_0"), val = tensor([false, true])]; tensor audio_signal_to_fp16_dtype_0 = const()[name = tensor("audio_signal_to_fp16_dtype_0"), val = tensor("fp16")]; tensor audio_signal_to_fp16 = cast(dtype = audio_signal_to_fp16_dtype_0, x = audio_signal)[name = tensor("cast_179")]; tensor var_41_cast_fp16 = slice_by_index(begin = var_41_begin_0, end = var_41_end_0, end_mask = var_41_end_mask_0, squeeze_mask = var_41_squeeze_mask_0, x = audio_signal_to_fp16)[name = tensor("op_41_cast_fp16")]; tensor var_42_axes_0 = const()[name = tensor("op_42_axes_0"), val = tensor([1])]; tensor var_42_cast_fp16 = expand_dims(axes = var_42_axes_0, x = var_41_cast_fp16)[name = tensor("op_42_cast_fp16")]; tensor var_44_begin_0 = const()[name = tensor("op_44_begin_0"), val = tensor([0, 1])]; tensor var_44_end_0 = const()[name = tensor("op_44_end_0"), val = tensor([1, 240000])]; tensor var_44_end_mask_0 = const()[name = tensor("op_44_end_mask_0"), val = tensor([true, true])]; tensor var_44_cast_fp16 = slice_by_index(begin = var_44_begin_0, end = var_44_end_0, end_mask = var_44_end_mask_0, x = audio_signal_to_fp16)[name = tensor("op_44_cast_fp16")]; tensor var_46_begin_0 = const()[name = tensor("op_46_begin_0"), val = tensor([0, 0])]; tensor var_46_end_0 = const()[name = tensor("op_46_end_0"), val = tensor([1, 239999])]; tensor var_46_end_mask_0 = const()[name = tensor("op_46_end_mask_0"), val = tensor([true, false])]; tensor var_46_cast_fp16 = slice_by_index(begin = var_46_begin_0, end = var_46_end_0, end_mask = var_46_end_mask_0, x = audio_signal_to_fp16)[name = tensor("op_46_cast_fp16")]; tensor var_47_to_fp16 = const()[name = tensor("op_47_to_fp16"), val = tensor(0x1.f0cp-1)]; tensor var_48_cast_fp16 = mul(x = var_46_cast_fp16, y = var_47_to_fp16)[name = tensor("op_48_cast_fp16")]; tensor var_49_cast_fp16 = sub(x = var_44_cast_fp16, y = var_48_cast_fp16)[name = tensor("op_49_cast_fp16")]; tensor input_1_interleave_0 = const()[name = tensor("input_1_interleave_0"), val = tensor(false)]; tensor input_1_cast_fp16 = concat(axis = var_22, interleave = input_1_interleave_0, values = (var_42_cast_fp16, var_49_cast_fp16))[name = tensor("input_1_cast_fp16")]; tensor var_55 = const()[name = tensor("op_55"), val = tensor([1, 1, 240000])]; tensor input_3_cast_fp16 = reshape(shape = var_55, x = input_1_cast_fp16)[name = tensor("input_3_cast_fp16")]; tensor input_5_pad_0 = const()[name = tensor("input_5_pad_0"), val = tensor([0, 0, 0, 0, 256, 256])]; tensor input_5_mode_0 = const()[name = tensor("input_5_mode_0"), val = tensor("reflect")]; tensor const_3_to_fp16 = const()[name = tensor("const_3_to_fp16"), val = tensor(0x0p+0)]; tensor input_5_cast_fp16 = pad(constant_val = const_3_to_fp16, mode = input_5_mode_0, pad = input_5_pad_0, x = input_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; tensor var_61 = const()[name = tensor("op_61"), val = tensor([1, 240512])]; tensor input_7_cast_fp16 = reshape(shape = var_61, x = input_5_cast_fp16)[name = tensor("input_7_cast_fp16")]; tensor expand_dims_6 = const()[name = tensor("expand_dims_6"), val = tensor([160])]; tensor expand_dims_7_axes_0 = const()[name = tensor("expand_dims_7_axes_0"), val = tensor([1])]; tensor expand_dims_7_cast_fp16 = expand_dims(axes = expand_dims_7_axes_0, x = input_7_cast_fp16)[name = tensor("expand_dims_7_cast_fp16")]; tensor conv_0_pad_type_0 = const()[name = tensor("conv_0_pad_type_0"), val = tensor("valid")]; tensor conv_0_pad_0 = const()[name = tensor("conv_0_pad_0"), val = tensor([0, 0])]; tensor conv_0_dilations_0 = const()[name = tensor("conv_0_dilations_0"), val = tensor([1])]; tensor conv_0_groups_0 = const()[name = tensor("conv_0_groups_0"), val = tensor(1)]; tensor expand_dims_4_to_fp16 = const()[name = tensor("expand_dims_4_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor conv_0_cast_fp16 = conv(dilations = conv_0_dilations_0, groups = conv_0_groups_0, pad = conv_0_pad_0, pad_type = conv_0_pad_type_0, strides = expand_dims_6, weight = expand_dims_4_to_fp16, x = expand_dims_7_cast_fp16)[name = tensor("conv_0_cast_fp16")]; tensor conv_1_pad_type_0 = const()[name = tensor("conv_1_pad_type_0"), val = tensor("valid")]; tensor conv_1_pad_0 = const()[name = tensor("conv_1_pad_0"), val = tensor([0, 0])]; tensor conv_1_dilations_0 = const()[name = tensor("conv_1_dilations_0"), val = tensor([1])]; tensor conv_1_groups_0 = const()[name = tensor("conv_1_groups_0"), val = tensor(1)]; tensor expand_dims_5_to_fp16 = const()[name = tensor("expand_dims_5_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(263296)))]; tensor conv_1_cast_fp16 = conv(dilations = conv_1_dilations_0, groups = conv_1_groups_0, pad = conv_1_pad_0, pad_type = conv_1_pad_type_0, strides = expand_dims_6, weight = expand_dims_5_to_fp16, x = expand_dims_7_cast_fp16)[name = tensor("conv_1_cast_fp16")]; tensor stack_0_axis_0 = const()[name = tensor("stack_0_axis_0"), val = tensor(-1)]; tensor stack_0_cast_fp16 = stack(axis = stack_0_axis_0, values = (conv_0_cast_fp16, conv_1_cast_fp16))[name = tensor("stack_0_cast_fp16")]; tensor var_15_promoted_to_fp16 = const()[name = tensor("op_15_promoted_to_fp16"), val = tensor(0x1p+1)]; tensor var_65_cast_fp16 = pow(x = stack_0_cast_fp16, y = var_15_promoted_to_fp16)[name = tensor("op_65_cast_fp16")]; tensor var_67_axes_0 = const()[name = tensor("op_67_axes_0"), val = tensor([-1])]; tensor var_67_keep_dims_0 = const()[name = tensor("op_67_keep_dims_0"), val = tensor(false)]; tensor var_67_cast_fp16 = reduce_sum(axes = var_67_axes_0, keep_dims = var_67_keep_dims_0, x = var_65_cast_fp16)[name = tensor("op_67_cast_fp16")]; tensor x_9_cast_fp16 = identity(x = var_67_cast_fp16)[name = tensor("x_9_cast_fp16")]; tensor x_11_transpose_x_0 = const()[name = tensor("x_11_transpose_x_0"), val = tensor(false)]; tensor x_11_transpose_y_0 = const()[name = tensor("x_11_transpose_y_0"), val = tensor(false)]; tensor const_6_to_fp16 = const()[name = tensor("const_6_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(526528)))]; tensor x_11_cast_fp16 = matmul(transpose_x = x_11_transpose_x_0, transpose_y = x_11_transpose_y_0, x = const_6_to_fp16, y = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; tensor var_74_to_fp16 = const()[name = tensor("op_74_to_fp16"), val = tensor(0x1p-24)]; tensor var_75_cast_fp16 = add(x = x_11_cast_fp16, y = var_74_to_fp16)[name = tensor("op_75_cast_fp16")]; tensor x_13_epsilon_0 = const()[name = tensor("x_13_epsilon_0"), val = tensor(0x1p-149)]; tensor x_13_cast_fp16 = log(epsilon = x_13_epsilon_0, x = var_75_cast_fp16)[name = tensor("x_13_cast_fp16")]; tensor var_80 = const()[name = tensor("op_80"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750, 751, 752, 753, 754, 755, 756, 757, 758, 759, 760, 761, 762, 763, 764, 765, 766, 767, 768, 769, 770, 771, 772, 773, 774, 775, 776, 777, 778, 779, 780, 781, 782, 783, 784, 785, 786, 787, 788, 789, 790, 791, 792, 793, 794, 795, 796, 797, 798, 799, 800, 801, 802, 803, 804, 805, 806, 807, 808, 809, 810, 811, 812, 813, 814, 815, 816, 817, 818, 819, 820, 821, 822, 823, 824, 825, 826, 827, 828, 829, 830, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 856, 857, 858, 859, 860, 861, 862, 863, 864, 865, 866, 867, 868, 869, 870, 871, 872, 873, 874, 875, 876, 877, 878, 879, 880, 881, 882, 883, 884, 885, 886, 887, 888, 889, 890, 891, 892, 893, 894, 895, 896, 897, 898, 899, 900, 901, 902, 903, 904, 905, 906, 907, 908, 909, 910, 911, 912, 913, 914, 915, 916, 917, 918, 919, 920, 921, 922, 923, 924, 925, 926, 927, 928, 929, 930, 931, 932, 933, 934, 935, 936, 937, 938, 939, 940, 941, 942, 943, 944, 945, 946, 947, 948, 949, 950, 951, 952, 953, 954, 955, 956, 957, 958, 959, 960, 961, 962, 963, 964, 965, 966, 967, 968, 969, 970, 971, 972, 973, 974, 975, 976, 977, 978, 979, 980, 981, 982, 983, 984, 985, 986, 987, 988, 989, 990, 991, 992, 993, 994, 995, 996, 997, 998, 999, 1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009, 1010, 1011, 1012, 1013, 1014, 1015, 1016, 1017, 1018, 1019, 1020, 1021, 1022, 1023, 1024, 1025, 1026, 1027, 1028, 1029, 1030, 1031, 1032, 1033, 1034, 1035, 1036, 1037, 1038, 1039, 1040, 1041, 1042, 1043, 1044, 1045, 1046, 1047, 1048, 1049, 1050, 1051, 1052, 1053, 1054, 1055, 1056, 1057, 1058, 1059, 1060, 1061, 1062, 1063, 1064, 1065, 1066, 1067, 1068, 1069, 1070, 1071, 1072, 1073, 1074, 1075, 1076, 1077, 1078, 1079, 1080, 1081, 1082, 1083, 1084, 1085, 1086, 1087, 1088, 1089, 1090, 1091, 1092, 1093, 1094, 1095, 1096, 1097, 1098, 1099, 1100, 1101, 1102, 1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114, 1115, 1116, 1117, 1118, 1119, 1120, 1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128, 1129, 1130, 1131, 1132, 1133, 1134, 1135, 1136, 1137, 1138, 1139, 1140, 1141, 1142, 1143, 1144, 1145, 1146, 1147, 1148, 1149, 1150, 1151, 1152, 1153, 1154, 1155, 1156, 1157, 1158, 1159, 1160, 1161, 1162, 1163, 1164, 1165, 1166, 1167, 1168, 1169, 1170, 1171, 1172, 1173, 1174, 1175, 1176, 1177, 1178, 1179, 1180, 1181, 1182, 1183, 1184, 1185, 1186, 1187, 1188, 1189, 1190, 1191, 1192, 1193, 1194, 1195, 1196, 1197, 1198, 1199, 1200, 1201, 1202, 1203, 1204, 1205, 1206, 1207, 1208, 1209, 1210, 1211, 1212, 1213, 1214, 1215, 1216, 1217, 1218, 1219, 1220, 1221, 1222, 1223, 1224, 1225, 1226, 1227, 1228, 1229, 1230, 1231, 1232, 1233, 1234, 1235, 1236, 1237, 1238, 1239, 1240, 1241, 1242, 1243, 1244, 1245, 1246, 1247, 1248, 1249, 1250, 1251, 1252, 1253, 1254, 1255, 1256, 1257, 1258, 1259, 1260, 1261, 1262, 1263, 1264, 1265, 1266, 1267, 1268, 1269, 1270, 1271, 1272, 1273, 1274, 1275, 1276, 1277, 1278, 1279, 1280, 1281, 1282, 1283, 1284, 1285, 1286, 1287, 1288, 1289, 1290, 1291, 1292, 1293, 1294, 1295, 1296, 1297, 1298, 1299, 1300, 1301, 1302, 1303, 1304, 1305, 1306, 1307, 1308, 1309, 1310, 1311, 1312, 1313, 1314, 1315, 1316, 1317, 1318, 1319, 1320, 1321, 1322, 1323, 1324, 1325, 1326, 1327, 1328, 1329, 1330, 1331, 1332, 1333, 1334, 1335, 1336, 1337, 1338, 1339, 1340, 1341, 1342, 1343, 1344, 1345, 1346, 1347, 1348, 1349, 1350, 1351, 1352, 1353, 1354, 1355, 1356, 1357, 1358, 1359, 1360, 1361, 1362, 1363, 1364, 1365, 1366, 1367, 1368, 1369, 1370, 1371, 1372, 1373, 1374, 1375, 1376, 1377, 1378, 1379, 1380, 1381, 1382, 1383, 1384, 1385, 1386, 1387, 1388, 1389, 1390, 1391, 1392, 1393, 1394, 1395, 1396, 1397, 1398, 1399, 1400, 1401, 1402, 1403, 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, 1413, 1414, 1415, 1416, 1417, 1418, 1419, 1420, 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, 1429, 1430, 1431, 1432, 1433, 1434, 1435, 1436, 1437, 1438, 1439, 1440, 1441, 1442, 1443, 1444, 1445, 1446, 1447, 1448, 1449, 1450, 1451, 1452, 1453, 1454, 1455, 1456, 1457, 1458, 1459, 1460, 1461, 1462, 1463, 1464, 1465, 1466, 1467, 1468, 1469, 1470, 1471, 1472, 1473, 1474, 1475, 1476, 1477, 1478, 1479, 1480, 1481, 1482, 1483, 1484, 1485, 1486, 1487, 1488, 1489, 1490, 1491, 1492, 1493, 1494, 1495, 1496, 1497, 1498, 1499, 1500]])]; tensor var_83_axes_0 = const()[name = tensor("op_83_axes_0"), val = tensor([1])]; tensor seq_len_1_cast_fp16_to_int32 = cast(dtype = seq_len_dtype_0, x = seq_len_1_cast_fp16)[name = tensor("cast_180")]; tensor var_83 = expand_dims(axes = var_83_axes_0, x = seq_len_1_cast_fp16_to_int32)[name = tensor("op_83")]; tensor valid_mask = less(x = var_80, y = var_83)[name = tensor("valid_mask")]; tensor var_85_axes_0 = const()[name = tensor("op_85_axes_0"), val = tensor([1])]; tensor var_85 = expand_dims(axes = var_85_axes_0, x = valid_mask)[name = tensor("op_85")]; tensor var_85_after_broadcast_reps_0 = const()[name = tensor("op_85_after_broadcast_reps_0"), val = tensor([1, 80, 1])]; tensor var_85_after_broadcast = tile(reps = var_85_after_broadcast_reps_0, x = var_85)[name = tensor("op_85_after_broadcast")]; tensor var_8_after_broadcast_to_fp16 = const()[name = tensor("op_8_after_broadcast_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(567744)))]; tensor var_86_cast_fp16 = select(a = x_13_cast_fp16, b = var_8_after_broadcast_to_fp16, cond = var_85_after_broadcast)[name = tensor("op_86_cast_fp16")]; tensor x_mean_numerator_axes_0 = const()[name = tensor("x_mean_numerator_axes_0"), val = tensor([2])]; tensor x_mean_numerator_keep_dims_0 = const()[name = tensor("x_mean_numerator_keep_dims_0"), val = tensor(false)]; tensor x_mean_numerator_cast_fp16 = reduce_sum(axes = x_mean_numerator_axes_0, keep_dims = x_mean_numerator_keep_dims_0, x = var_86_cast_fp16)[name = tensor("x_mean_numerator_cast_fp16")]; tensor x_mean_denominator_axes_0 = const()[name = tensor("x_mean_denominator_axes_0"), val = tensor([1])]; tensor x_mean_denominator_keep_dims_0 = const()[name = tensor("x_mean_denominator_keep_dims_0"), val = tensor(false)]; tensor cast_2_to_fp16_dtype_0 = const()[name = tensor("cast_2_to_fp16_dtype_0"), val = tensor("fp16")]; tensor valid_mask_to_fp16 = cast(dtype = cast_2_to_fp16_dtype_0, x = valid_mask)[name = tensor("cast_178")]; tensor x_mean_denominator_cast_fp16 = reduce_sum(axes = x_mean_denominator_axes_0, keep_dims = x_mean_denominator_keep_dims_0, x = valid_mask_to_fp16)[name = tensor("x_mean_denominator_cast_fp16")]; tensor var_91_axes_0 = const()[name = tensor("op_91_axes_0"), val = tensor([1])]; tensor var_91_cast_fp16 = expand_dims(axes = var_91_axes_0, x = x_mean_denominator_cast_fp16)[name = tensor("op_91_cast_fp16")]; tensor x_mean_cast_fp16 = real_div(x = x_mean_numerator_cast_fp16, y = var_91_cast_fp16)[name = tensor("x_mean_cast_fp16")]; tensor var_94_axes_0 = const()[name = tensor("op_94_axes_0"), val = tensor([2])]; tensor var_94_cast_fp16 = expand_dims(axes = var_94_axes_0, x = x_mean_cast_fp16)[name = tensor("op_94_cast_fp16")]; tensor var_95_cast_fp16 = sub(x = x_13_cast_fp16, y = var_94_cast_fp16)[name = tensor("op_95_cast_fp16")]; tensor var_96_cast_fp16 = select(a = var_95_cast_fp16, b = var_8_after_broadcast_to_fp16, cond = var_85_after_broadcast)[name = tensor("op_96_cast_fp16")]; tensor var_15_promoted_1_to_fp16 = const()[name = tensor("op_15_promoted_1_to_fp16"), val = tensor(0x1p+1)]; tensor var_97_cast_fp16 = pow(x = var_96_cast_fp16, y = var_15_promoted_1_to_fp16)[name = tensor("op_97_cast_fp16")]; tensor var_99_axes_0 = const()[name = tensor("op_99_axes_0"), val = tensor([2])]; tensor var_99_keep_dims_0 = const()[name = tensor("op_99_keep_dims_0"), val = tensor(false)]; tensor var_99_cast_fp16 = reduce_sum(axes = var_99_axes_0, keep_dims = var_99_keep_dims_0, x = var_97_cast_fp16)[name = tensor("op_99_cast_fp16")]; tensor var_101_to_fp16 = const()[name = tensor("op_101_to_fp16"), val = tensor(0x1p+0)]; tensor var_102_cast_fp16 = sub(x = var_91_cast_fp16, y = var_101_to_fp16)[name = tensor("op_102_cast_fp16")]; tensor var_103_cast_fp16 = real_div(x = var_99_cast_fp16, y = var_102_cast_fp16)[name = tensor("op_103_cast_fp16")]; tensor x_std_1_cast_fp16 = sqrt(x = var_103_cast_fp16)[name = tensor("x_std_1_cast_fp16")]; tensor var_7_to_fp16 = const()[name = tensor("op_7_to_fp16"), val = tensor(0x1.5p-17)]; tensor x_std_cast_fp16 = add(x = x_std_1_cast_fp16, y = var_7_to_fp16)[name = tensor("x_std_cast_fp16")]; tensor var_108_axes_0 = const()[name = tensor("op_108_axes_0"), val = tensor([2])]; tensor var_108_cast_fp16 = expand_dims(axes = var_108_axes_0, x = x_std_cast_fp16)[name = tensor("op_108_cast_fp16")]; tensor x_15_cast_fp16 = real_div(x = var_95_cast_fp16, y = var_108_cast_fp16)[name = tensor("x_15_cast_fp16")]; tensor mask_3 = greater_equal(x = var_80, y = var_83)[name = tensor("mask_3")]; tensor var_117_axes_0 = const()[name = tensor("op_117_axes_0"), val = tensor([1])]; tensor var_117 = expand_dims(axes = var_117_axes_0, x = mask_3)[name = tensor("op_117")]; tensor cast_9_to_fp16 = const()[name = tensor("cast_9_to_fp16"), val = tensor(0x0p+0)]; tensor processed_signal_cast_fp16 = select(a = cast_9_to_fp16, b = x_15_cast_fp16, cond = var_117)[name = tensor("processed_signal_cast_fp16")]; tensor var_138 = const()[name = tensor("op_138"), val = tensor(-1)]; tensor x_17_perm_0 = const()[name = tensor("x_17_perm_0"), val = tensor([0, 2, 1])]; tensor var_201_to_fp16_dtype_0 = const()[name = tensor("op_201_to_fp16_dtype_0"), val = tensor("fp16")]; tensor var_202_promoted_to_fp16 = const()[name = tensor("op_202_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor seq_len_1_cast_fp16_to_int32_to_fp16 = cast(dtype = var_201_to_fp16_dtype_0, x = seq_len_1_cast_fp16_to_int32)[name = tensor("cast_177")]; tensor var_203_cast_fp16 = add(x = seq_len_1_cast_fp16_to_int32_to_fp16, y = var_202_promoted_to_fp16)[name = tensor("op_203_cast_fp16")]; tensor _inversed_205_y_0_to_fp16 = const()[name = tensor("_inversed_205_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_205_cast_fp16 = mul(x = var_203_cast_fp16, y = _inversed_205_y_0_to_fp16)[name = tensor("_inversed_205_cast_fp16")]; tensor var_206_to_fp16 = const()[name = tensor("op_206_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_1_cast_fp16 = add(x = _inversed_205_cast_fp16, y = var_206_to_fp16)[name = tensor("lengths_1_cast_fp16")]; tensor lengths_3_cast_fp16 = floor(x = lengths_1_cast_fp16)[name = tensor("lengths_3_cast_fp16")]; tensor var_210_promoted_to_fp16 = const()[name = tensor("op_210_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor var_211_cast_fp16 = add(x = lengths_3_cast_fp16, y = var_210_promoted_to_fp16)[name = tensor("op_211_cast_fp16")]; tensor _inversed_213_y_0_to_fp16 = const()[name = tensor("_inversed_213_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_213_cast_fp16 = mul(x = var_211_cast_fp16, y = _inversed_213_y_0_to_fp16)[name = tensor("_inversed_213_cast_fp16")]; tensor var_214_to_fp16 = const()[name = tensor("op_214_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_7_cast_fp16 = add(x = _inversed_213_cast_fp16, y = var_214_to_fp16)[name = tensor("lengths_7_cast_fp16")]; tensor lengths_9_cast_fp16 = floor(x = lengths_7_cast_fp16)[name = tensor("lengths_9_cast_fp16")]; tensor var_218_promoted_to_fp16 = const()[name = tensor("op_218_promoted_to_fp16"), val = tensor(-0x1p+0)]; tensor var_219_cast_fp16 = add(x = lengths_9_cast_fp16, y = var_218_promoted_to_fp16)[name = tensor("op_219_cast_fp16")]; tensor _inversed_221_y_0_to_fp16 = const()[name = tensor("_inversed_221_y_0_to_fp16"), val = tensor(0x1p-1)]; tensor _inversed_221_cast_fp16 = mul(x = var_219_cast_fp16, y = _inversed_221_y_0_to_fp16)[name = tensor("_inversed_221_cast_fp16")]; tensor var_222_to_fp16 = const()[name = tensor("op_222_to_fp16"), val = tensor(0x1p+0)]; tensor lengths_13_cast_fp16 = add(x = _inversed_221_cast_fp16, y = var_222_to_fp16)[name = tensor("lengths_13_cast_fp16")]; tensor lengths_cast_fp16 = floor(x = lengths_13_cast_fp16)[name = tensor("lengths_cast_fp16")]; tensor input_9_axes_0 = const()[name = tensor("input_9_axes_0"), val = tensor([1])]; tensor x_17_cast_fp16 = transpose(perm = x_17_perm_0, x = processed_signal_cast_fp16)[name = tensor("transpose_207")]; tensor input_9_cast_fp16 = expand_dims(axes = input_9_axes_0, x = x_17_cast_fp16)[name = tensor("input_9_cast_fp16")]; tensor input_11_pad_type_0 = const()[name = tensor("input_11_pad_type_0"), val = tensor("custom")]; tensor input_11_pad_0 = const()[name = tensor("input_11_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_11_strides_0 = const()[name = tensor("input_11_strides_0"), val = tensor([2, 2])]; tensor input_11_dilations_0 = const()[name = tensor("input_11_dilations_0"), val = tensor([1, 1])]; tensor input_11_groups_0 = const()[name = tensor("input_11_groups_0"), val = tensor(1)]; tensor encoder_module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(808000)))]; tensor encoder_module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(812672)))]; tensor input_11_cast_fp16 = conv(bias = encoder_module_pre_encode_conv_0_bias_to_fp16, dilations = input_11_dilations_0, groups = input_11_groups_0, pad = input_11_pad_0, pad_type = input_11_pad_type_0, strides = input_11_strides_0, weight = encoder_module_pre_encode_conv_0_weight_to_fp16, x = input_9_cast_fp16)[name = tensor("input_11_cast_fp16")]; tensor input_13_cast_fp16 = relu(x = input_11_cast_fp16)[name = tensor("input_13_cast_fp16")]; tensor input_15_pad_type_0 = const()[name = tensor("input_15_pad_type_0"), val = tensor("custom")]; tensor input_15_pad_0 = const()[name = tensor("input_15_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_15_strides_0 = const()[name = tensor("input_15_strides_0"), val = tensor([2, 2])]; tensor input_15_groups_0 = const()[name = tensor("input_15_groups_0"), val = tensor(256)]; tensor input_15_dilations_0 = const()[name = tensor("input_15_dilations_0"), val = tensor([1, 1])]; tensor encoder_module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(813248)))]; tensor encoder_module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(817920)))]; tensor input_15_cast_fp16 = conv(bias = encoder_module_pre_encode_conv_2_bias_to_fp16, dilations = input_15_dilations_0, groups = input_15_groups_0, pad = input_15_pad_0, pad_type = input_15_pad_type_0, strides = input_15_strides_0, weight = encoder_module_pre_encode_conv_2_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; tensor input_17_pad_type_0 = const()[name = tensor("input_17_pad_type_0"), val = tensor("valid")]; tensor input_17_strides_0 = const()[name = tensor("input_17_strides_0"), val = tensor([1, 1])]; tensor input_17_pad_0 = const()[name = tensor("input_17_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_17_dilations_0 = const()[name = tensor("input_17_dilations_0"), val = tensor([1, 1])]; tensor input_17_groups_0 = const()[name = tensor("input_17_groups_0"), val = tensor(1)]; tensor encoder_module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(818496)))]; tensor encoder_module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(949632)))]; tensor input_17_cast_fp16 = conv(bias = encoder_module_pre_encode_conv_3_bias_to_fp16, dilations = input_17_dilations_0, groups = input_17_groups_0, pad = input_17_pad_0, pad_type = input_17_pad_type_0, strides = input_17_strides_0, weight = encoder_module_pre_encode_conv_3_weight_to_fp16, x = input_15_cast_fp16)[name = tensor("input_17_cast_fp16")]; tensor input_19_cast_fp16 = relu(x = input_17_cast_fp16)[name = tensor("input_19_cast_fp16")]; tensor input_21_pad_type_0 = const()[name = tensor("input_21_pad_type_0"), val = tensor("custom")]; tensor input_21_pad_0 = const()[name = tensor("input_21_pad_0"), val = tensor([1, 1, 1, 1])]; tensor input_21_strides_0 = const()[name = tensor("input_21_strides_0"), val = tensor([2, 2])]; tensor input_21_groups_0 = const()[name = tensor("input_21_groups_0"), val = tensor(256)]; tensor input_21_dilations_0 = const()[name = tensor("input_21_dilations_0"), val = tensor([1, 1])]; tensor encoder_module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(950208)))]; tensor encoder_module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(954880)))]; tensor input_21_cast_fp16 = conv(bias = encoder_module_pre_encode_conv_5_bias_to_fp16, dilations = input_21_dilations_0, groups = input_21_groups_0, pad = input_21_pad_0, pad_type = input_21_pad_type_0, strides = input_21_strides_0, weight = encoder_module_pre_encode_conv_5_weight_to_fp16, x = input_19_cast_fp16)[name = tensor("input_21_cast_fp16")]; tensor input_23_pad_type_0 = const()[name = tensor("input_23_pad_type_0"), val = tensor("valid")]; tensor input_23_strides_0 = const()[name = tensor("input_23_strides_0"), val = tensor([1, 1])]; tensor input_23_pad_0 = const()[name = tensor("input_23_pad_0"), val = tensor([0, 0, 0, 0])]; tensor input_23_dilations_0 = const()[name = tensor("input_23_dilations_0"), val = tensor([1, 1])]; tensor input_23_groups_0 = const()[name = tensor("input_23_groups_0"), val = tensor(1)]; tensor encoder_module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(955456)))]; tensor encoder_module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("encoder_module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1086592)))]; tensor input_23_cast_fp16 = conv(bias = encoder_module_pre_encode_conv_6_bias_to_fp16, dilations = input_23_dilations_0, groups = input_23_groups_0, pad = input_23_pad_0, pad_type = input_23_pad_type_0, strides = input_23_strides_0, weight = encoder_module_pre_encode_conv_6_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("input_23_cast_fp16")]; tensor x_19_cast_fp16 = relu(x = input_23_cast_fp16)[name = tensor("x_19_cast_fp16")]; tensor var_272_perm_0 = const()[name = tensor("op_272_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_273 = const()[name = tensor("op_273"), val = tensor([1, 188, -1])]; tensor var_272_cast_fp16 = transpose(perm = var_272_perm_0, x = x_19_cast_fp16)[name = tensor("transpose_206")]; tensor input_25_cast_fp16 = reshape(shape = var_273, x = var_272_cast_fp16)[name = tensor("input_25_cast_fp16")]; tensor encoder_module_pre_encode_out_weight_to_fp16 = const()[name = tensor("encoder_module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1087168)))]; tensor encoder_module_pre_encode_out_bias_to_fp16 = const()[name = tensor("encoder_module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3708672)))]; tensor linear_0_cast_fp16 = linear(bias = encoder_module_pre_encode_out_bias_to_fp16, weight = encoder_module_pre_encode_out_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_0_cast_fp16")]; tensor padding_length_dtype_0 = const()[name = tensor("padding_length_dtype_0"), val = tensor("int32")]; tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; tensor var_311_axes_0 = const()[name = tensor("op_311_axes_0"), val = tensor([-1])]; tensor encoder_length = cast(dtype = padding_length_dtype_0, x = lengths_cast_fp16)[name = tensor("cast_176")]; tensor var_311 = expand_dims(axes = var_311_axes_0, x = encoder_length)[name = tensor("op_311")]; tensor pad_mask_1 = less(x = expand_dims_3, y = var_311)[name = tensor("pad_mask_1")]; tensor var_313_axes_0 = const()[name = tensor("op_313_axes_0"), val = tensor([1])]; tensor var_313 = expand_dims(axes = var_313_axes_0, x = pad_mask_1)[name = tensor("op_313")]; tensor var_314 = const()[name = tensor("op_314"), val = tensor([1, 188, 1])]; tensor pad_mask_for_att_mask_1 = tile(reps = var_314, x = var_313)[name = tensor("pad_mask_for_att_mask_1")]; tensor var_316_perm_0 = const()[name = tensor("op_316_perm_0"), val = tensor([0, 2, 1])]; tensor var_316 = transpose(perm = var_316_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_205")]; tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_316)[name = tensor("pad_mask_for_att_mask")]; tensor const_22 = const()[name = tensor("const_22"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_22)[name = tensor("att_mask")]; tensor mask_5 = logical_not(x = att_mask)[name = tensor("mask_5")]; tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; tensor input_29_axes_0 = const()[name = tensor("input_29_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3709760)))]; tensor encoder_module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3710848)))]; tensor var_156_to_fp16 = const()[name = tensor("op_156_to_fp16"), val = tensor(0x1.5p-17)]; tensor input_29_cast_fp16 = layer_norm(axes = input_29_axes_0, beta = encoder_module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_29_cast_fp16")]; tensor encoder_module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(3711936)))]; tensor encoder_module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5809152)))]; tensor linear_1_cast_fp16 = linear(bias = encoder_module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_29_cast_fp16)[name = tensor("linear_1_cast_fp16")]; tensor input_33_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_33_cast_fp16")]; tensor encoder_module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5813312)))]; tensor encoder_module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7910528)))]; tensor linear_2_cast_fp16 = linear(bias = encoder_module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_33_cast_fp16)[name = tensor("linear_2_cast_fp16")]; tensor var_349_to_fp16 = const()[name = tensor("op_349_to_fp16"), val = tensor(0x1p-1)]; tensor var_350_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_349_to_fp16)[name = tensor("op_350_cast_fp16")]; tensor input_39_cast_fp16 = add(x = linear_0_cast_fp16, y = var_350_cast_fp16)[name = tensor("input_39_cast_fp16")]; tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7911616)))]; tensor encoder_module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7912704)))]; tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = encoder_module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_0_norm_self_att_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("query_1_cast_fp16")]; tensor encoder_module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7913792)))]; tensor encoder_module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8438144)))]; tensor linear_3_cast_fp16 = linear(bias = encoder_module_layers_0_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; tensor var_367 = const()[name = tensor("op_367"), val = tensor([1, -1, 8, 64])]; tensor q_1_cast_fp16 = reshape(shape = var_367, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; tensor encoder_module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8439232)))]; tensor encoder_module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8963584)))]; tensor linear_4_cast_fp16 = linear(bias = encoder_module_layers_0_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; tensor var_372 = const()[name = tensor("op_372"), val = tensor([1, -1, 8, 64])]; tensor k_1_cast_fp16 = reshape(shape = var_372, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; tensor encoder_module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8964672)))]; tensor encoder_module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9489024)))]; tensor linear_5_cast_fp16 = linear(bias = encoder_module_layers_0_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; tensor var_377 = const()[name = tensor("op_377"), val = tensor([1, -1, 8, 64])]; tensor v_1_cast_fp16 = reshape(shape = var_377, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9490112)))]; tensor var_389_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_389_cast_fp16")]; tensor encoder_module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9491200)))]; tensor var_391_cast_fp16 = add(x = q_1_cast_fp16, y = encoder_module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_391_cast_fp16")]; tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_23_transpose_x_0 = const()[name = tensor("x_23_transpose_x_0"), val = tensor(false)]; tensor x_23_transpose_y_0 = const()[name = tensor("x_23_transpose_y_0"), val = tensor(false)]; tensor var_393_to_fp16 = const()[name = tensor("op_393_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9492288)))]; tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_391_cast_fp16)[name = tensor("transpose_203")]; tensor x_23_cast_fp16 = matmul(transpose_x = x_23_transpose_x_0, transpose_y = x_23_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_393_to_fp16)[name = tensor("x_23_cast_fp16")]; tensor x_25_pad_0 = const()[name = tensor("x_25_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_25_mode_0 = const()[name = tensor("x_25_mode_0"), val = tensor("constant")]; tensor const_29_to_fp16 = const()[name = tensor("const_29_to_fp16"), val = tensor(0x0p+0)]; tensor x_25_cast_fp16 = pad(constant_val = const_29_to_fp16, mode = x_25_mode_0, pad = x_25_pad_0, x = x_23_cast_fp16)[name = tensor("x_25_cast_fp16")]; tensor var_401 = const()[name = tensor("op_401"), val = tensor([1, 8, -1, 188])]; tensor x_27_cast_fp16 = reshape(shape = var_401, x = x_25_cast_fp16)[name = tensor("x_27_cast_fp16")]; tensor var_405_begin_0 = const()[name = tensor("op_405_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_405_end_0 = const()[name = tensor("op_405_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_405_end_mask_0 = const()[name = tensor("op_405_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_405_cast_fp16 = slice_by_index(begin = var_405_begin_0, end = var_405_end_0, end_mask = var_405_end_mask_0, x = x_27_cast_fp16)[name = tensor("op_405_cast_fp16")]; tensor var_406 = const()[name = tensor("op_406"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_1_cast_fp16 = reshape(shape = var_406, x = var_405_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_201")]; tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_389_cast_fp16)[name = tensor("transpose_202")]; tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; tensor var_415_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_415_cast_fp16")]; tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_1_cast_fp16 = mul(x = var_415_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; tensor mask_7_axes_0 = const()[name = tensor("mask_7_axes_0"), val = tensor([1])]; tensor mask_7 = expand_dims(axes = mask_7_axes_0, x = mask_5)[name = tensor("mask_7")]; tensor var_153_to_fp16 = const()[name = tensor("op_153_to_fp16"), val = tensor(-0x1.388p+13)]; tensor scores_3_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_7)[name = tensor("scores_3_cast_fp16")]; tensor var_421_cast_fp16 = softmax(axis = var_138, x = scores_3_cast_fp16)[name = tensor("op_421_cast_fp16")]; tensor var_154_to_fp16 = const()[name = tensor("op_154_to_fp16"), val = tensor(0x0p+0)]; tensor input_41_cast_fp16 = select(a = var_154_to_fp16, b = var_421_cast_fp16, cond = mask_7)[name = tensor("input_41_cast_fp16")]; tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_204")]; tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = input_41_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_29_cast_fp16")]; tensor var_425_perm_0 = const()[name = tensor("op_425_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_426 = const()[name = tensor("op_426"), val = tensor([1, -1, 512])]; tensor var_425_cast_fp16 = transpose(perm = var_425_perm_0, x = x_29_cast_fp16)[name = tensor("transpose_200")]; tensor input_43_cast_fp16 = reshape(shape = var_426, x = var_425_cast_fp16)[name = tensor("input_43_cast_fp16")]; tensor encoder_module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9876352)))]; tensor encoder_module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10400704)))]; tensor linear_7_cast_fp16 = linear(bias = encoder_module_layers_0_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_43_cast_fp16)[name = tensor("linear_7_cast_fp16")]; tensor input_47_cast_fp16 = add(x = input_39_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_47_cast_fp16")]; tensor x_33_axes_0 = const()[name = tensor("x_33_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10401792)))]; tensor encoder_module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10402880)))]; tensor x_33_cast_fp16 = layer_norm(axes = x_33_axes_0, beta = encoder_module_layers_0_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_0_norm_conv_weight_to_fp16, x = input_47_cast_fp16)[name = tensor("x_33_cast_fp16")]; tensor input_49_perm_0 = const()[name = tensor("input_49_perm_0"), val = tensor([0, 2, 1])]; tensor input_51_pad_type_0 = const()[name = tensor("input_51_pad_type_0"), val = tensor("valid")]; tensor input_51_strides_0 = const()[name = tensor("input_51_strides_0"), val = tensor([1])]; tensor input_51_pad_0 = const()[name = tensor("input_51_pad_0"), val = tensor([0, 0])]; tensor input_51_dilations_0 = const()[name = tensor("input_51_dilations_0"), val = tensor([1])]; tensor input_51_groups_0 = const()[name = tensor("input_51_groups_0"), val = tensor(1)]; tensor encoder_module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10403968)))]; tensor encoder_module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11452608)))]; tensor input_49_cast_fp16 = transpose(perm = input_49_perm_0, x = x_33_cast_fp16)[name = tensor("transpose_199")]; tensor input_51_cast_fp16 = conv(bias = encoder_module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_51_dilations_0, groups = input_51_groups_0, pad = input_51_pad_0, pad_type = input_51_pad_type_0, strides = input_51_strides_0, weight = encoder_module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_49_cast_fp16)[name = tensor("input_51_cast_fp16")]; tensor x_35_split_num_splits_0 = const()[name = tensor("x_35_split_num_splits_0"), val = tensor(2)]; tensor x_35_split_axis_0 = const()[name = tensor("x_35_split_axis_0"), val = tensor(1)]; tensor x_35_split_cast_fp16_0, tensor x_35_split_cast_fp16_1 = split(axis = x_35_split_axis_0, num_splits = x_35_split_num_splits_0, x = input_51_cast_fp16)[name = tensor("x_35_split_cast_fp16")]; tensor x_35_split_1_sigmoid_cast_fp16 = sigmoid(x = x_35_split_cast_fp16_1)[name = tensor("x_35_split_1_sigmoid_cast_fp16")]; tensor x_35_cast_fp16 = mul(x = x_35_split_cast_fp16_0, y = x_35_split_1_sigmoid_cast_fp16)[name = tensor("x_35_cast_fp16")]; tensor var_450_axes_0 = const()[name = tensor("op_450_axes_0"), val = tensor([1])]; tensor var_450 = expand_dims(axes = var_450_axes_0, x = pad_mask)[name = tensor("op_450")]; tensor input_53_cast_fp16 = select(a = var_154_to_fp16, b = x_35_cast_fp16, cond = var_450)[name = tensor("input_53_cast_fp16")]; tensor input_55_pad_0 = const()[name = tensor("input_55_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_55_mode_0 = const()[name = tensor("input_55_mode_0"), val = tensor("constant")]; tensor const_32_to_fp16 = const()[name = tensor("const_32_to_fp16"), val = tensor(0x0p+0)]; tensor input_55_cast_fp16 = pad(constant_val = const_32_to_fp16, mode = input_55_mode_0, pad = input_55_pad_0, x = input_53_cast_fp16)[name = tensor("input_55_cast_fp16")]; tensor input_57_pad_type_0 = const()[name = tensor("input_57_pad_type_0"), val = tensor("valid")]; tensor input_57_groups_0 = const()[name = tensor("input_57_groups_0"), val = tensor(512)]; tensor input_57_strides_0 = const()[name = tensor("input_57_strides_0"), val = tensor([1])]; tensor input_57_pad_0 = const()[name = tensor("input_57_pad_0"), val = tensor([0, 0])]; tensor input_57_dilations_0 = const()[name = tensor("input_57_dilations_0"), val = tensor([1])]; tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11454720)))]; tensor const_194_to_fp16 = const()[name = tensor("const_194_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11464000)))]; tensor input_59_cast_fp16 = conv(bias = const_194_to_fp16, dilations = input_57_dilations_0, groups = input_57_groups_0, pad = input_57_pad_0, pad_type = input_57_pad_type_0, strides = input_57_strides_0, weight = const_193_to_fp16, x = input_55_cast_fp16)[name = tensor("input_59_cast_fp16")]; tensor input_61_cast_fp16 = silu(x = input_59_cast_fp16)[name = tensor("input_61_cast_fp16")]; tensor x_37_pad_type_0 = const()[name = tensor("x_37_pad_type_0"), val = tensor("valid")]; tensor x_37_strides_0 = const()[name = tensor("x_37_strides_0"), val = tensor([1])]; tensor x_37_pad_0 = const()[name = tensor("x_37_pad_0"), val = tensor([0, 0])]; tensor x_37_dilations_0 = const()[name = tensor("x_37_dilations_0"), val = tensor([1])]; tensor x_37_groups_0 = const()[name = tensor("x_37_groups_0"), val = tensor(1)]; tensor encoder_module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11465088)))]; tensor encoder_module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11989440)))]; tensor x_37_cast_fp16 = conv(bias = encoder_module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_37_dilations_0, groups = x_37_groups_0, pad = x_37_pad_0, pad_type = x_37_pad_type_0, strides = x_37_strides_0, weight = encoder_module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_61_cast_fp16)[name = tensor("x_37_cast_fp16")]; tensor input_63_perm_0 = const()[name = tensor("input_63_perm_0"), val = tensor([0, 2, 1])]; tensor input_63_cast_fp16 = transpose(perm = input_63_perm_0, x = x_37_cast_fp16)[name = tensor("transpose_198")]; tensor input_65_cast_fp16 = add(x = input_47_cast_fp16, y = input_63_cast_fp16)[name = tensor("input_65_cast_fp16")]; tensor input_67_axes_0 = const()[name = tensor("input_67_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11990528)))]; tensor encoder_module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11991616)))]; tensor input_67_cast_fp16 = layer_norm(axes = input_67_axes_0, beta = encoder_module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_65_cast_fp16)[name = tensor("input_67_cast_fp16")]; tensor encoder_module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11992704)))]; tensor encoder_module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14089920)))]; tensor linear_8_cast_fp16 = linear(bias = encoder_module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_67_cast_fp16)[name = tensor("linear_8_cast_fp16")]; tensor input_71_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_71_cast_fp16")]; tensor encoder_module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(14094080)))]; tensor encoder_module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16191296)))]; tensor linear_9_cast_fp16 = linear(bias = encoder_module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("linear_9_cast_fp16")]; tensor var_492_to_fp16 = const()[name = tensor("op_492_to_fp16"), val = tensor(0x1p-1)]; tensor var_493_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_492_to_fp16)[name = tensor("op_493_cast_fp16")]; tensor input_77_cast_fp16 = add(x = input_65_cast_fp16, y = var_493_cast_fp16)[name = tensor("input_77_cast_fp16")]; tensor input_79_axes_0 = const()[name = tensor("input_79_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16192384)))]; tensor encoder_module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16193472)))]; tensor input_79_cast_fp16 = layer_norm(axes = input_79_axes_0, beta = encoder_module_layers_0_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_0_norm_out_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("input_79_cast_fp16")]; tensor input_81_axes_0 = const()[name = tensor("input_81_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16194560)))]; tensor encoder_module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16195648)))]; tensor input_81_cast_fp16 = layer_norm(axes = input_81_axes_0, beta = encoder_module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_79_cast_fp16)[name = tensor("input_81_cast_fp16")]; tensor encoder_module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(16196736)))]; tensor encoder_module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18293952)))]; tensor linear_10_cast_fp16 = linear(bias = encoder_module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_81_cast_fp16)[name = tensor("linear_10_cast_fp16")]; tensor input_85_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_85_cast_fp16")]; tensor encoder_module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(18298112)))]; tensor encoder_module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20395328)))]; tensor linear_11_cast_fp16 = linear(bias = encoder_module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_85_cast_fp16)[name = tensor("linear_11_cast_fp16")]; tensor var_523_to_fp16 = const()[name = tensor("op_523_to_fp16"), val = tensor(0x1p-1)]; tensor var_524_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_523_to_fp16)[name = tensor("op_524_cast_fp16")]; tensor input_91_cast_fp16 = add(x = input_79_cast_fp16, y = var_524_cast_fp16)[name = tensor("input_91_cast_fp16")]; tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20396416)))]; tensor encoder_module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20397504)))]; tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = encoder_module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_1_norm_self_att_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("query_3_cast_fp16")]; tensor encoder_module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20398592)))]; tensor encoder_module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20922944)))]; tensor linear_12_cast_fp16 = linear(bias = encoder_module_layers_1_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; tensor var_541 = const()[name = tensor("op_541"), val = tensor([1, -1, 8, 64])]; tensor q_7_cast_fp16 = reshape(shape = var_541, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; tensor encoder_module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20924032)))]; tensor encoder_module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21448384)))]; tensor linear_13_cast_fp16 = linear(bias = encoder_module_layers_1_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; tensor var_546 = const()[name = tensor("op_546"), val = tensor([1, -1, 8, 64])]; tensor k_5_cast_fp16 = reshape(shape = var_546, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; tensor encoder_module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21449472)))]; tensor encoder_module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21973824)))]; tensor linear_14_cast_fp16 = linear(bias = encoder_module_layers_1_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; tensor var_551 = const()[name = tensor("op_551"), val = tensor([1, -1, 8, 64])]; tensor v_3_cast_fp16 = reshape(shape = var_551, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21974912)))]; tensor var_563_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_563_cast_fp16")]; tensor encoder_module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21976000)))]; tensor var_565_cast_fp16 = add(x = q_7_cast_fp16, y = encoder_module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_565_cast_fp16")]; tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_45_transpose_x_0 = const()[name = tensor("x_45_transpose_x_0"), val = tensor(false)]; tensor x_45_transpose_y_0 = const()[name = tensor("x_45_transpose_y_0"), val = tensor(false)]; tensor var_567_to_fp16 = const()[name = tensor("op_567_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21977088)))]; tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_565_cast_fp16)[name = tensor("transpose_196")]; tensor x_45_cast_fp16 = matmul(transpose_x = x_45_transpose_x_0, transpose_y = x_45_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_567_to_fp16)[name = tensor("x_45_cast_fp16")]; tensor x_47_pad_0 = const()[name = tensor("x_47_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_47_mode_0 = const()[name = tensor("x_47_mode_0"), val = tensor("constant")]; tensor const_39_to_fp16 = const()[name = tensor("const_39_to_fp16"), val = tensor(0x0p+0)]; tensor x_47_cast_fp16 = pad(constant_val = const_39_to_fp16, mode = x_47_mode_0, pad = x_47_pad_0, x = x_45_cast_fp16)[name = tensor("x_47_cast_fp16")]; tensor var_575 = const()[name = tensor("op_575"), val = tensor([1, 8, -1, 188])]; tensor x_49_cast_fp16 = reshape(shape = var_575, x = x_47_cast_fp16)[name = tensor("x_49_cast_fp16")]; tensor var_579_begin_0 = const()[name = tensor("op_579_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_579_end_0 = const()[name = tensor("op_579_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_579_end_mask_0 = const()[name = tensor("op_579_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_579_cast_fp16 = slice_by_index(begin = var_579_begin_0, end = var_579_end_0, end_mask = var_579_end_mask_0, x = x_49_cast_fp16)[name = tensor("op_579_cast_fp16")]; tensor var_580 = const()[name = tensor("op_580"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_5_cast_fp16 = reshape(shape = var_580, x = var_579_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_194")]; tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_563_cast_fp16)[name = tensor("transpose_195")]; tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; tensor var_589_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_589_cast_fp16")]; tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_5_cast_fp16 = mul(x = var_589_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; tensor scores_7_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_7)[name = tensor("scores_7_cast_fp16")]; tensor var_595_cast_fp16 = softmax(axis = var_138, x = scores_7_cast_fp16)[name = tensor("op_595_cast_fp16")]; tensor input_93_cast_fp16 = select(a = var_154_to_fp16, b = var_595_cast_fp16, cond = mask_7)[name = tensor("input_93_cast_fp16")]; tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_197")]; tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = input_93_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_51_cast_fp16")]; tensor var_599_perm_0 = const()[name = tensor("op_599_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_600 = const()[name = tensor("op_600"), val = tensor([1, -1, 512])]; tensor var_599_cast_fp16 = transpose(perm = var_599_perm_0, x = x_51_cast_fp16)[name = tensor("transpose_193")]; tensor input_95_cast_fp16 = reshape(shape = var_600, x = var_599_cast_fp16)[name = tensor("input_95_cast_fp16")]; tensor encoder_module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22361152)))]; tensor encoder_module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22885504)))]; tensor linear_16_cast_fp16 = linear(bias = encoder_module_layers_1_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_95_cast_fp16)[name = tensor("linear_16_cast_fp16")]; tensor input_99_cast_fp16 = add(x = input_91_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_99_cast_fp16")]; tensor x_55_axes_0 = const()[name = tensor("x_55_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22886592)))]; tensor encoder_module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22887680)))]; tensor x_55_cast_fp16 = layer_norm(axes = x_55_axes_0, beta = encoder_module_layers_1_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_1_norm_conv_weight_to_fp16, x = input_99_cast_fp16)[name = tensor("x_55_cast_fp16")]; tensor input_101_perm_0 = const()[name = tensor("input_101_perm_0"), val = tensor([0, 2, 1])]; tensor input_103_pad_type_0 = const()[name = tensor("input_103_pad_type_0"), val = tensor("valid")]; tensor input_103_strides_0 = const()[name = tensor("input_103_strides_0"), val = tensor([1])]; tensor input_103_pad_0 = const()[name = tensor("input_103_pad_0"), val = tensor([0, 0])]; tensor input_103_dilations_0 = const()[name = tensor("input_103_dilations_0"), val = tensor([1])]; tensor input_103_groups_0 = const()[name = tensor("input_103_groups_0"), val = tensor(1)]; tensor encoder_module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22888768)))]; tensor encoder_module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23937408)))]; tensor input_101_cast_fp16 = transpose(perm = input_101_perm_0, x = x_55_cast_fp16)[name = tensor("transpose_192")]; tensor input_103_cast_fp16 = conv(bias = encoder_module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_103_dilations_0, groups = input_103_groups_0, pad = input_103_pad_0, pad_type = input_103_pad_type_0, strides = input_103_strides_0, weight = encoder_module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_101_cast_fp16)[name = tensor("input_103_cast_fp16")]; tensor x_57_split_num_splits_0 = const()[name = tensor("x_57_split_num_splits_0"), val = tensor(2)]; tensor x_57_split_axis_0 = const()[name = tensor("x_57_split_axis_0"), val = tensor(1)]; tensor x_57_split_cast_fp16_0, tensor x_57_split_cast_fp16_1 = split(axis = x_57_split_axis_0, num_splits = x_57_split_num_splits_0, x = input_103_cast_fp16)[name = tensor("x_57_split_cast_fp16")]; tensor x_57_split_1_sigmoid_cast_fp16 = sigmoid(x = x_57_split_cast_fp16_1)[name = tensor("x_57_split_1_sigmoid_cast_fp16")]; tensor x_57_cast_fp16 = mul(x = x_57_split_cast_fp16_0, y = x_57_split_1_sigmoid_cast_fp16)[name = tensor("x_57_cast_fp16")]; tensor input_105_cast_fp16 = select(a = var_154_to_fp16, b = x_57_cast_fp16, cond = var_450)[name = tensor("input_105_cast_fp16")]; tensor input_107_pad_0 = const()[name = tensor("input_107_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_107_mode_0 = const()[name = tensor("input_107_mode_0"), val = tensor("constant")]; tensor const_42_to_fp16 = const()[name = tensor("const_42_to_fp16"), val = tensor(0x0p+0)]; tensor input_107_cast_fp16 = pad(constant_val = const_42_to_fp16, mode = input_107_mode_0, pad = input_107_pad_0, x = input_105_cast_fp16)[name = tensor("input_107_cast_fp16")]; tensor input_109_pad_type_0 = const()[name = tensor("input_109_pad_type_0"), val = tensor("valid")]; tensor input_109_groups_0 = const()[name = tensor("input_109_groups_0"), val = tensor(512)]; tensor input_109_strides_0 = const()[name = tensor("input_109_strides_0"), val = tensor([1])]; tensor input_109_pad_0 = const()[name = tensor("input_109_pad_0"), val = tensor([0, 0])]; tensor input_109_dilations_0 = const()[name = tensor("input_109_dilations_0"), val = tensor([1])]; tensor const_195_to_fp16 = const()[name = tensor("const_195_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23939520)))]; tensor const_196_to_fp16 = const()[name = tensor("const_196_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23948800)))]; tensor input_111_cast_fp16 = conv(bias = const_196_to_fp16, dilations = input_109_dilations_0, groups = input_109_groups_0, pad = input_109_pad_0, pad_type = input_109_pad_type_0, strides = input_109_strides_0, weight = const_195_to_fp16, x = input_107_cast_fp16)[name = tensor("input_111_cast_fp16")]; tensor input_113_cast_fp16 = silu(x = input_111_cast_fp16)[name = tensor("input_113_cast_fp16")]; tensor x_59_pad_type_0 = const()[name = tensor("x_59_pad_type_0"), val = tensor("valid")]; tensor x_59_strides_0 = const()[name = tensor("x_59_strides_0"), val = tensor([1])]; tensor x_59_pad_0 = const()[name = tensor("x_59_pad_0"), val = tensor([0, 0])]; tensor x_59_dilations_0 = const()[name = tensor("x_59_dilations_0"), val = tensor([1])]; tensor x_59_groups_0 = const()[name = tensor("x_59_groups_0"), val = tensor(1)]; tensor encoder_module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23949888)))]; tensor encoder_module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24474240)))]; tensor x_59_cast_fp16 = conv(bias = encoder_module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_59_dilations_0, groups = x_59_groups_0, pad = x_59_pad_0, pad_type = x_59_pad_type_0, strides = x_59_strides_0, weight = encoder_module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_113_cast_fp16)[name = tensor("x_59_cast_fp16")]; tensor input_115_perm_0 = const()[name = tensor("input_115_perm_0"), val = tensor([0, 2, 1])]; tensor input_115_cast_fp16 = transpose(perm = input_115_perm_0, x = x_59_cast_fp16)[name = tensor("transpose_191")]; tensor input_117_cast_fp16 = add(x = input_99_cast_fp16, y = input_115_cast_fp16)[name = tensor("input_117_cast_fp16")]; tensor input_119_axes_0 = const()[name = tensor("input_119_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24475328)))]; tensor encoder_module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24476416)))]; tensor input_119_cast_fp16 = layer_norm(axes = input_119_axes_0, beta = encoder_module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_117_cast_fp16)[name = tensor("input_119_cast_fp16")]; tensor encoder_module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(24477504)))]; tensor encoder_module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26574720)))]; tensor linear_17_cast_fp16 = linear(bias = encoder_module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_119_cast_fp16)[name = tensor("linear_17_cast_fp16")]; tensor input_123_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_123_cast_fp16")]; tensor encoder_module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(26578880)))]; tensor encoder_module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28676096)))]; tensor linear_18_cast_fp16 = linear(bias = encoder_module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("linear_18_cast_fp16")]; tensor var_666_to_fp16 = const()[name = tensor("op_666_to_fp16"), val = tensor(0x1p-1)]; tensor var_667_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_666_to_fp16)[name = tensor("op_667_cast_fp16")]; tensor input_129_cast_fp16 = add(x = input_117_cast_fp16, y = var_667_cast_fp16)[name = tensor("input_129_cast_fp16")]; tensor input_131_axes_0 = const()[name = tensor("input_131_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28677184)))]; tensor encoder_module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28678272)))]; tensor input_131_cast_fp16 = layer_norm(axes = input_131_axes_0, beta = encoder_module_layers_1_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_1_norm_out_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("input_131_cast_fp16")]; tensor input_133_axes_0 = const()[name = tensor("input_133_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28679360)))]; tensor encoder_module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28680448)))]; tensor input_133_cast_fp16 = layer_norm(axes = input_133_axes_0, beta = encoder_module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_131_cast_fp16)[name = tensor("input_133_cast_fp16")]; tensor encoder_module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(28681536)))]; tensor encoder_module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30778752)))]; tensor linear_19_cast_fp16 = linear(bias = encoder_module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_133_cast_fp16)[name = tensor("linear_19_cast_fp16")]; tensor input_137_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_137_cast_fp16")]; tensor encoder_module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(30782912)))]; tensor encoder_module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32880128)))]; tensor linear_20_cast_fp16 = linear(bias = encoder_module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_137_cast_fp16)[name = tensor("linear_20_cast_fp16")]; tensor var_697_to_fp16 = const()[name = tensor("op_697_to_fp16"), val = tensor(0x1p-1)]; tensor var_698_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_697_to_fp16)[name = tensor("op_698_cast_fp16")]; tensor input_143_cast_fp16 = add(x = input_131_cast_fp16, y = var_698_cast_fp16)[name = tensor("input_143_cast_fp16")]; tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32881216)))]; tensor encoder_module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32882304)))]; tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = encoder_module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_2_norm_self_att_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("query_5_cast_fp16")]; tensor encoder_module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32883392)))]; tensor encoder_module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33407744)))]; tensor linear_21_cast_fp16 = linear(bias = encoder_module_layers_2_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; tensor var_715 = const()[name = tensor("op_715"), val = tensor([1, -1, 8, 64])]; tensor q_13_cast_fp16 = reshape(shape = var_715, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; tensor encoder_module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33408832)))]; tensor encoder_module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33933184)))]; tensor linear_22_cast_fp16 = linear(bias = encoder_module_layers_2_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; tensor var_720 = const()[name = tensor("op_720"), val = tensor([1, -1, 8, 64])]; tensor k_9_cast_fp16 = reshape(shape = var_720, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; tensor encoder_module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33934272)))]; tensor encoder_module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34458624)))]; tensor linear_23_cast_fp16 = linear(bias = encoder_module_layers_2_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; tensor var_725 = const()[name = tensor("op_725"), val = tensor([1, -1, 8, 64])]; tensor v_5_cast_fp16 = reshape(shape = var_725, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34459712)))]; tensor var_737_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_737_cast_fp16")]; tensor encoder_module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34460800)))]; tensor var_739_cast_fp16 = add(x = q_13_cast_fp16, y = encoder_module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_739_cast_fp16")]; tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_67_transpose_x_0 = const()[name = tensor("x_67_transpose_x_0"), val = tensor(false)]; tensor x_67_transpose_y_0 = const()[name = tensor("x_67_transpose_y_0"), val = tensor(false)]; tensor var_741_to_fp16 = const()[name = tensor("op_741_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34461888)))]; tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_739_cast_fp16)[name = tensor("transpose_189")]; tensor x_67_cast_fp16 = matmul(transpose_x = x_67_transpose_x_0, transpose_y = x_67_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_741_to_fp16)[name = tensor("x_67_cast_fp16")]; tensor x_69_pad_0 = const()[name = tensor("x_69_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_69_mode_0 = const()[name = tensor("x_69_mode_0"), val = tensor("constant")]; tensor const_49_to_fp16 = const()[name = tensor("const_49_to_fp16"), val = tensor(0x0p+0)]; tensor x_69_cast_fp16 = pad(constant_val = const_49_to_fp16, mode = x_69_mode_0, pad = x_69_pad_0, x = x_67_cast_fp16)[name = tensor("x_69_cast_fp16")]; tensor var_749 = const()[name = tensor("op_749"), val = tensor([1, 8, -1, 188])]; tensor x_71_cast_fp16 = reshape(shape = var_749, x = x_69_cast_fp16)[name = tensor("x_71_cast_fp16")]; tensor var_753_begin_0 = const()[name = tensor("op_753_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_753_end_0 = const()[name = tensor("op_753_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_753_end_mask_0 = const()[name = tensor("op_753_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_753_cast_fp16 = slice_by_index(begin = var_753_begin_0, end = var_753_end_0, end_mask = var_753_end_mask_0, x = x_71_cast_fp16)[name = tensor("op_753_cast_fp16")]; tensor var_754 = const()[name = tensor("op_754"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_9_cast_fp16 = reshape(shape = var_754, x = var_753_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_187")]; tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_737_cast_fp16)[name = tensor("transpose_188")]; tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; tensor var_763_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_763_cast_fp16")]; tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_9_cast_fp16 = mul(x = var_763_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; tensor scores_11_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_7)[name = tensor("scores_11_cast_fp16")]; tensor var_769_cast_fp16 = softmax(axis = var_138, x = scores_11_cast_fp16)[name = tensor("op_769_cast_fp16")]; tensor input_145_cast_fp16 = select(a = var_154_to_fp16, b = var_769_cast_fp16, cond = mask_7)[name = tensor("input_145_cast_fp16")]; tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_190")]; tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = input_145_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_73_cast_fp16")]; tensor var_773_perm_0 = const()[name = tensor("op_773_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_774 = const()[name = tensor("op_774"), val = tensor([1, -1, 512])]; tensor var_773_cast_fp16 = transpose(perm = var_773_perm_0, x = x_73_cast_fp16)[name = tensor("transpose_186")]; tensor input_147_cast_fp16 = reshape(shape = var_774, x = var_773_cast_fp16)[name = tensor("input_147_cast_fp16")]; tensor encoder_module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34845952)))]; tensor encoder_module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35370304)))]; tensor linear_25_cast_fp16 = linear(bias = encoder_module_layers_2_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_147_cast_fp16)[name = tensor("linear_25_cast_fp16")]; tensor input_151_cast_fp16 = add(x = input_143_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_151_cast_fp16")]; tensor x_77_axes_0 = const()[name = tensor("x_77_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35371392)))]; tensor encoder_module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35372480)))]; tensor x_77_cast_fp16 = layer_norm(axes = x_77_axes_0, beta = encoder_module_layers_2_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_2_norm_conv_weight_to_fp16, x = input_151_cast_fp16)[name = tensor("x_77_cast_fp16")]; tensor input_153_perm_0 = const()[name = tensor("input_153_perm_0"), val = tensor([0, 2, 1])]; tensor input_155_pad_type_0 = const()[name = tensor("input_155_pad_type_0"), val = tensor("valid")]; tensor input_155_strides_0 = const()[name = tensor("input_155_strides_0"), val = tensor([1])]; tensor input_155_pad_0 = const()[name = tensor("input_155_pad_0"), val = tensor([0, 0])]; tensor input_155_dilations_0 = const()[name = tensor("input_155_dilations_0"), val = tensor([1])]; tensor input_155_groups_0 = const()[name = tensor("input_155_groups_0"), val = tensor(1)]; tensor encoder_module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35373568)))]; tensor encoder_module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36422208)))]; tensor input_153_cast_fp16 = transpose(perm = input_153_perm_0, x = x_77_cast_fp16)[name = tensor("transpose_185")]; tensor input_155_cast_fp16 = conv(bias = encoder_module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_155_dilations_0, groups = input_155_groups_0, pad = input_155_pad_0, pad_type = input_155_pad_type_0, strides = input_155_strides_0, weight = encoder_module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_153_cast_fp16)[name = tensor("input_155_cast_fp16")]; tensor x_79_split_num_splits_0 = const()[name = tensor("x_79_split_num_splits_0"), val = tensor(2)]; tensor x_79_split_axis_0 = const()[name = tensor("x_79_split_axis_0"), val = tensor(1)]; tensor x_79_split_cast_fp16_0, tensor x_79_split_cast_fp16_1 = split(axis = x_79_split_axis_0, num_splits = x_79_split_num_splits_0, x = input_155_cast_fp16)[name = tensor("x_79_split_cast_fp16")]; tensor x_79_split_1_sigmoid_cast_fp16 = sigmoid(x = x_79_split_cast_fp16_1)[name = tensor("x_79_split_1_sigmoid_cast_fp16")]; tensor x_79_cast_fp16 = mul(x = x_79_split_cast_fp16_0, y = x_79_split_1_sigmoid_cast_fp16)[name = tensor("x_79_cast_fp16")]; tensor input_157_cast_fp16 = select(a = var_154_to_fp16, b = x_79_cast_fp16, cond = var_450)[name = tensor("input_157_cast_fp16")]; tensor input_159_pad_0 = const()[name = tensor("input_159_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_159_mode_0 = const()[name = tensor("input_159_mode_0"), val = tensor("constant")]; tensor const_52_to_fp16 = const()[name = tensor("const_52_to_fp16"), val = tensor(0x0p+0)]; tensor input_159_cast_fp16 = pad(constant_val = const_52_to_fp16, mode = input_159_mode_0, pad = input_159_pad_0, x = input_157_cast_fp16)[name = tensor("input_159_cast_fp16")]; tensor input_161_pad_type_0 = const()[name = tensor("input_161_pad_type_0"), val = tensor("valid")]; tensor input_161_groups_0 = const()[name = tensor("input_161_groups_0"), val = tensor(512)]; tensor input_161_strides_0 = const()[name = tensor("input_161_strides_0"), val = tensor([1])]; tensor input_161_pad_0 = const()[name = tensor("input_161_pad_0"), val = tensor([0, 0])]; tensor input_161_dilations_0 = const()[name = tensor("input_161_dilations_0"), val = tensor([1])]; tensor const_197_to_fp16 = const()[name = tensor("const_197_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36424320)))]; tensor const_198_to_fp16 = const()[name = tensor("const_198_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36433600)))]; tensor input_163_cast_fp16 = conv(bias = const_198_to_fp16, dilations = input_161_dilations_0, groups = input_161_groups_0, pad = input_161_pad_0, pad_type = input_161_pad_type_0, strides = input_161_strides_0, weight = const_197_to_fp16, x = input_159_cast_fp16)[name = tensor("input_163_cast_fp16")]; tensor input_165_cast_fp16 = silu(x = input_163_cast_fp16)[name = tensor("input_165_cast_fp16")]; tensor x_81_pad_type_0 = const()[name = tensor("x_81_pad_type_0"), val = tensor("valid")]; tensor x_81_strides_0 = const()[name = tensor("x_81_strides_0"), val = tensor([1])]; tensor x_81_pad_0 = const()[name = tensor("x_81_pad_0"), val = tensor([0, 0])]; tensor x_81_dilations_0 = const()[name = tensor("x_81_dilations_0"), val = tensor([1])]; tensor x_81_groups_0 = const()[name = tensor("x_81_groups_0"), val = tensor(1)]; tensor encoder_module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36434688)))]; tensor encoder_module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36959040)))]; tensor x_81_cast_fp16 = conv(bias = encoder_module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_81_dilations_0, groups = x_81_groups_0, pad = x_81_pad_0, pad_type = x_81_pad_type_0, strides = x_81_strides_0, weight = encoder_module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_165_cast_fp16)[name = tensor("x_81_cast_fp16")]; tensor input_167_perm_0 = const()[name = tensor("input_167_perm_0"), val = tensor([0, 2, 1])]; tensor input_167_cast_fp16 = transpose(perm = input_167_perm_0, x = x_81_cast_fp16)[name = tensor("transpose_184")]; tensor input_169_cast_fp16 = add(x = input_151_cast_fp16, y = input_167_cast_fp16)[name = tensor("input_169_cast_fp16")]; tensor input_171_axes_0 = const()[name = tensor("input_171_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36960128)))]; tensor encoder_module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36961216)))]; tensor input_171_cast_fp16 = layer_norm(axes = input_171_axes_0, beta = encoder_module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_169_cast_fp16)[name = tensor("input_171_cast_fp16")]; tensor encoder_module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36962304)))]; tensor encoder_module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39059520)))]; tensor linear_26_cast_fp16 = linear(bias = encoder_module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_171_cast_fp16)[name = tensor("linear_26_cast_fp16")]; tensor input_175_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_175_cast_fp16")]; tensor encoder_module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(39063680)))]; tensor encoder_module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41160896)))]; tensor linear_27_cast_fp16 = linear(bias = encoder_module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("linear_27_cast_fp16")]; tensor var_840_to_fp16 = const()[name = tensor("op_840_to_fp16"), val = tensor(0x1p-1)]; tensor var_841_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_840_to_fp16)[name = tensor("op_841_cast_fp16")]; tensor input_181_cast_fp16 = add(x = input_169_cast_fp16, y = var_841_cast_fp16)[name = tensor("input_181_cast_fp16")]; tensor input_183_axes_0 = const()[name = tensor("input_183_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41161984)))]; tensor encoder_module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41163072)))]; tensor input_183_cast_fp16 = layer_norm(axes = input_183_axes_0, beta = encoder_module_layers_2_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_2_norm_out_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("input_183_cast_fp16")]; tensor input_185_axes_0 = const()[name = tensor("input_185_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41164160)))]; tensor encoder_module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41165248)))]; tensor input_185_cast_fp16 = layer_norm(axes = input_185_axes_0, beta = encoder_module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_183_cast_fp16)[name = tensor("input_185_cast_fp16")]; tensor encoder_module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(41166336)))]; tensor encoder_module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43263552)))]; tensor linear_28_cast_fp16 = linear(bias = encoder_module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_185_cast_fp16)[name = tensor("linear_28_cast_fp16")]; tensor input_189_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_189_cast_fp16")]; tensor encoder_module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(43267712)))]; tensor encoder_module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45364928)))]; tensor linear_29_cast_fp16 = linear(bias = encoder_module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_189_cast_fp16)[name = tensor("linear_29_cast_fp16")]; tensor var_871_to_fp16 = const()[name = tensor("op_871_to_fp16"), val = tensor(0x1p-1)]; tensor var_872_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_871_to_fp16)[name = tensor("op_872_cast_fp16")]; tensor input_195_cast_fp16 = add(x = input_183_cast_fp16, y = var_872_cast_fp16)[name = tensor("input_195_cast_fp16")]; tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45366016)))]; tensor encoder_module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45367104)))]; tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = encoder_module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_3_norm_self_att_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("query_7_cast_fp16")]; tensor encoder_module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45368192)))]; tensor encoder_module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45892544)))]; tensor linear_30_cast_fp16 = linear(bias = encoder_module_layers_3_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; tensor var_889 = const()[name = tensor("op_889"), val = tensor([1, -1, 8, 64])]; tensor q_19_cast_fp16 = reshape(shape = var_889, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; tensor encoder_module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45893632)))]; tensor encoder_module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46417984)))]; tensor linear_31_cast_fp16 = linear(bias = encoder_module_layers_3_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; tensor var_894 = const()[name = tensor("op_894"), val = tensor([1, -1, 8, 64])]; tensor k_13_cast_fp16 = reshape(shape = var_894, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; tensor encoder_module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46419072)))]; tensor encoder_module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46943424)))]; tensor linear_32_cast_fp16 = linear(bias = encoder_module_layers_3_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; tensor var_899 = const()[name = tensor("op_899"), val = tensor([1, -1, 8, 64])]; tensor v_7_cast_fp16 = reshape(shape = var_899, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46944512)))]; tensor var_911_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_911_cast_fp16")]; tensor encoder_module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46945600)))]; tensor var_913_cast_fp16 = add(x = q_19_cast_fp16, y = encoder_module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_913_cast_fp16")]; tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_89_transpose_x_0 = const()[name = tensor("x_89_transpose_x_0"), val = tensor(false)]; tensor x_89_transpose_y_0 = const()[name = tensor("x_89_transpose_y_0"), val = tensor(false)]; tensor var_915_to_fp16 = const()[name = tensor("op_915_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46946688)))]; tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_913_cast_fp16)[name = tensor("transpose_182")]; tensor x_89_cast_fp16 = matmul(transpose_x = x_89_transpose_x_0, transpose_y = x_89_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_915_to_fp16)[name = tensor("x_89_cast_fp16")]; tensor x_91_pad_0 = const()[name = tensor("x_91_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_91_mode_0 = const()[name = tensor("x_91_mode_0"), val = tensor("constant")]; tensor const_59_to_fp16 = const()[name = tensor("const_59_to_fp16"), val = tensor(0x0p+0)]; tensor x_91_cast_fp16 = pad(constant_val = const_59_to_fp16, mode = x_91_mode_0, pad = x_91_pad_0, x = x_89_cast_fp16)[name = tensor("x_91_cast_fp16")]; tensor var_923 = const()[name = tensor("op_923"), val = tensor([1, 8, -1, 188])]; tensor x_93_cast_fp16 = reshape(shape = var_923, x = x_91_cast_fp16)[name = tensor("x_93_cast_fp16")]; tensor var_927_begin_0 = const()[name = tensor("op_927_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_927_end_0 = const()[name = tensor("op_927_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_927_end_mask_0 = const()[name = tensor("op_927_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_927_cast_fp16 = slice_by_index(begin = var_927_begin_0, end = var_927_end_0, end_mask = var_927_end_mask_0, x = x_93_cast_fp16)[name = tensor("op_927_cast_fp16")]; tensor var_928 = const()[name = tensor("op_928"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_13_cast_fp16 = reshape(shape = var_928, x = var_927_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_180")]; tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_911_cast_fp16)[name = tensor("transpose_181")]; tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; tensor var_937_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_937_cast_fp16")]; tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_13_cast_fp16 = mul(x = var_937_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; tensor scores_15_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_7)[name = tensor("scores_15_cast_fp16")]; tensor var_943_cast_fp16 = softmax(axis = var_138, x = scores_15_cast_fp16)[name = tensor("op_943_cast_fp16")]; tensor input_197_cast_fp16 = select(a = var_154_to_fp16, b = var_943_cast_fp16, cond = mask_7)[name = tensor("input_197_cast_fp16")]; tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_183")]; tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = input_197_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_95_cast_fp16")]; tensor var_947_perm_0 = const()[name = tensor("op_947_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_948 = const()[name = tensor("op_948"), val = tensor([1, -1, 512])]; tensor var_947_cast_fp16 = transpose(perm = var_947_perm_0, x = x_95_cast_fp16)[name = tensor("transpose_179")]; tensor input_199_cast_fp16 = reshape(shape = var_948, x = var_947_cast_fp16)[name = tensor("input_199_cast_fp16")]; tensor encoder_module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47330752)))]; tensor encoder_module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47855104)))]; tensor linear_34_cast_fp16 = linear(bias = encoder_module_layers_3_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_199_cast_fp16)[name = tensor("linear_34_cast_fp16")]; tensor input_203_cast_fp16 = add(x = input_195_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_203_cast_fp16")]; tensor x_99_axes_0 = const()[name = tensor("x_99_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47856192)))]; tensor encoder_module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47857280)))]; tensor x_99_cast_fp16 = layer_norm(axes = x_99_axes_0, beta = encoder_module_layers_3_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_3_norm_conv_weight_to_fp16, x = input_203_cast_fp16)[name = tensor("x_99_cast_fp16")]; tensor input_205_perm_0 = const()[name = tensor("input_205_perm_0"), val = tensor([0, 2, 1])]; tensor input_207_pad_type_0 = const()[name = tensor("input_207_pad_type_0"), val = tensor("valid")]; tensor input_207_strides_0 = const()[name = tensor("input_207_strides_0"), val = tensor([1])]; tensor input_207_pad_0 = const()[name = tensor("input_207_pad_0"), val = tensor([0, 0])]; tensor input_207_dilations_0 = const()[name = tensor("input_207_dilations_0"), val = tensor([1])]; tensor input_207_groups_0 = const()[name = tensor("input_207_groups_0"), val = tensor(1)]; tensor encoder_module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47858368)))]; tensor encoder_module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48907008)))]; tensor input_205_cast_fp16 = transpose(perm = input_205_perm_0, x = x_99_cast_fp16)[name = tensor("transpose_178")]; tensor input_207_cast_fp16 = conv(bias = encoder_module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_207_dilations_0, groups = input_207_groups_0, pad = input_207_pad_0, pad_type = input_207_pad_type_0, strides = input_207_strides_0, weight = encoder_module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_205_cast_fp16)[name = tensor("input_207_cast_fp16")]; tensor x_101_split_num_splits_0 = const()[name = tensor("x_101_split_num_splits_0"), val = tensor(2)]; tensor x_101_split_axis_0 = const()[name = tensor("x_101_split_axis_0"), val = tensor(1)]; tensor x_101_split_cast_fp16_0, tensor x_101_split_cast_fp16_1 = split(axis = x_101_split_axis_0, num_splits = x_101_split_num_splits_0, x = input_207_cast_fp16)[name = tensor("x_101_split_cast_fp16")]; tensor x_101_split_1_sigmoid_cast_fp16 = sigmoid(x = x_101_split_cast_fp16_1)[name = tensor("x_101_split_1_sigmoid_cast_fp16")]; tensor x_101_cast_fp16 = mul(x = x_101_split_cast_fp16_0, y = x_101_split_1_sigmoid_cast_fp16)[name = tensor("x_101_cast_fp16")]; tensor input_209_cast_fp16 = select(a = var_154_to_fp16, b = x_101_cast_fp16, cond = var_450)[name = tensor("input_209_cast_fp16")]; tensor input_211_pad_0 = const()[name = tensor("input_211_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_211_mode_0 = const()[name = tensor("input_211_mode_0"), val = tensor("constant")]; tensor const_62_to_fp16 = const()[name = tensor("const_62_to_fp16"), val = tensor(0x0p+0)]; tensor input_211_cast_fp16 = pad(constant_val = const_62_to_fp16, mode = input_211_mode_0, pad = input_211_pad_0, x = input_209_cast_fp16)[name = tensor("input_211_cast_fp16")]; tensor input_213_pad_type_0 = const()[name = tensor("input_213_pad_type_0"), val = tensor("valid")]; tensor input_213_groups_0 = const()[name = tensor("input_213_groups_0"), val = tensor(512)]; tensor input_213_strides_0 = const()[name = tensor("input_213_strides_0"), val = tensor([1])]; tensor input_213_pad_0 = const()[name = tensor("input_213_pad_0"), val = tensor([0, 0])]; tensor input_213_dilations_0 = const()[name = tensor("input_213_dilations_0"), val = tensor([1])]; tensor const_199_to_fp16 = const()[name = tensor("const_199_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48909120)))]; tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48918400)))]; tensor input_215_cast_fp16 = conv(bias = const_200_to_fp16, dilations = input_213_dilations_0, groups = input_213_groups_0, pad = input_213_pad_0, pad_type = input_213_pad_type_0, strides = input_213_strides_0, weight = const_199_to_fp16, x = input_211_cast_fp16)[name = tensor("input_215_cast_fp16")]; tensor input_217_cast_fp16 = silu(x = input_215_cast_fp16)[name = tensor("input_217_cast_fp16")]; tensor x_103_pad_type_0 = const()[name = tensor("x_103_pad_type_0"), val = tensor("valid")]; tensor x_103_strides_0 = const()[name = tensor("x_103_strides_0"), val = tensor([1])]; tensor x_103_pad_0 = const()[name = tensor("x_103_pad_0"), val = tensor([0, 0])]; tensor x_103_dilations_0 = const()[name = tensor("x_103_dilations_0"), val = tensor([1])]; tensor x_103_groups_0 = const()[name = tensor("x_103_groups_0"), val = tensor(1)]; tensor encoder_module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48919488)))]; tensor encoder_module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49443840)))]; tensor x_103_cast_fp16 = conv(bias = encoder_module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_103_dilations_0, groups = x_103_groups_0, pad = x_103_pad_0, pad_type = x_103_pad_type_0, strides = x_103_strides_0, weight = encoder_module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_217_cast_fp16)[name = tensor("x_103_cast_fp16")]; tensor input_219_perm_0 = const()[name = tensor("input_219_perm_0"), val = tensor([0, 2, 1])]; tensor input_219_cast_fp16 = transpose(perm = input_219_perm_0, x = x_103_cast_fp16)[name = tensor("transpose_177")]; tensor input_221_cast_fp16 = add(x = input_203_cast_fp16, y = input_219_cast_fp16)[name = tensor("input_221_cast_fp16")]; tensor input_223_axes_0 = const()[name = tensor("input_223_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49444928)))]; tensor encoder_module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49446016)))]; tensor input_223_cast_fp16 = layer_norm(axes = input_223_axes_0, beta = encoder_module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_221_cast_fp16)[name = tensor("input_223_cast_fp16")]; tensor encoder_module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(49447104)))]; tensor encoder_module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51544320)))]; tensor linear_35_cast_fp16 = linear(bias = encoder_module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_223_cast_fp16)[name = tensor("linear_35_cast_fp16")]; tensor input_227_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_227_cast_fp16")]; tensor encoder_module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(51548480)))]; tensor encoder_module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53645696)))]; tensor linear_36_cast_fp16 = linear(bias = encoder_module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("linear_36_cast_fp16")]; tensor var_1014_to_fp16 = const()[name = tensor("op_1014_to_fp16"), val = tensor(0x1p-1)]; tensor var_1015_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1014_to_fp16)[name = tensor("op_1015_cast_fp16")]; tensor input_233_cast_fp16 = add(x = input_221_cast_fp16, y = var_1015_cast_fp16)[name = tensor("input_233_cast_fp16")]; tensor input_235_axes_0 = const()[name = tensor("input_235_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53646784)))]; tensor encoder_module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53647872)))]; tensor input_235_cast_fp16 = layer_norm(axes = input_235_axes_0, beta = encoder_module_layers_3_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_3_norm_out_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("input_235_cast_fp16")]; tensor input_237_axes_0 = const()[name = tensor("input_237_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53648960)))]; tensor encoder_module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53650048)))]; tensor input_237_cast_fp16 = layer_norm(axes = input_237_axes_0, beta = encoder_module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_235_cast_fp16)[name = tensor("input_237_cast_fp16")]; tensor encoder_module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(53651136)))]; tensor encoder_module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55748352)))]; tensor linear_37_cast_fp16 = linear(bias = encoder_module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_237_cast_fp16)[name = tensor("linear_37_cast_fp16")]; tensor input_241_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_241_cast_fp16")]; tensor encoder_module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(55752512)))]; tensor encoder_module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57849728)))]; tensor linear_38_cast_fp16 = linear(bias = encoder_module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_241_cast_fp16)[name = tensor("linear_38_cast_fp16")]; tensor var_1045_to_fp16 = const()[name = tensor("op_1045_to_fp16"), val = tensor(0x1p-1)]; tensor var_1046_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1045_to_fp16)[name = tensor("op_1046_cast_fp16")]; tensor input_247_cast_fp16 = add(x = input_235_cast_fp16, y = var_1046_cast_fp16)[name = tensor("input_247_cast_fp16")]; tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57850816)))]; tensor encoder_module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57851904)))]; tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = encoder_module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_4_norm_self_att_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("query_9_cast_fp16")]; tensor encoder_module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57852992)))]; tensor encoder_module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58377344)))]; tensor linear_39_cast_fp16 = linear(bias = encoder_module_layers_4_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; tensor var_1063 = const()[name = tensor("op_1063"), val = tensor([1, -1, 8, 64])]; tensor q_25_cast_fp16 = reshape(shape = var_1063, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; tensor encoder_module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58378432)))]; tensor encoder_module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58902784)))]; tensor linear_40_cast_fp16 = linear(bias = encoder_module_layers_4_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; tensor var_1068 = const()[name = tensor("op_1068"), val = tensor([1, -1, 8, 64])]; tensor k_17_cast_fp16 = reshape(shape = var_1068, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; tensor encoder_module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58903872)))]; tensor encoder_module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59428224)))]; tensor linear_41_cast_fp16 = linear(bias = encoder_module_layers_4_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; tensor var_1073 = const()[name = tensor("op_1073"), val = tensor([1, -1, 8, 64])]; tensor v_9_cast_fp16 = reshape(shape = var_1073, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59429312)))]; tensor var_1085_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1085_cast_fp16")]; tensor encoder_module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59430400)))]; tensor var_1087_cast_fp16 = add(x = q_25_cast_fp16, y = encoder_module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1087_cast_fp16")]; tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_111_transpose_x_0 = const()[name = tensor("x_111_transpose_x_0"), val = tensor(false)]; tensor x_111_transpose_y_0 = const()[name = tensor("x_111_transpose_y_0"), val = tensor(false)]; tensor var_1089_to_fp16 = const()[name = tensor("op_1089_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59431488)))]; tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1087_cast_fp16)[name = tensor("transpose_175")]; tensor x_111_cast_fp16 = matmul(transpose_x = x_111_transpose_x_0, transpose_y = x_111_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1089_to_fp16)[name = tensor("x_111_cast_fp16")]; tensor x_113_pad_0 = const()[name = tensor("x_113_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_113_mode_0 = const()[name = tensor("x_113_mode_0"), val = tensor("constant")]; tensor const_69_to_fp16 = const()[name = tensor("const_69_to_fp16"), val = tensor(0x0p+0)]; tensor x_113_cast_fp16 = pad(constant_val = const_69_to_fp16, mode = x_113_mode_0, pad = x_113_pad_0, x = x_111_cast_fp16)[name = tensor("x_113_cast_fp16")]; tensor var_1097 = const()[name = tensor("op_1097"), val = tensor([1, 8, -1, 188])]; tensor x_115_cast_fp16 = reshape(shape = var_1097, x = x_113_cast_fp16)[name = tensor("x_115_cast_fp16")]; tensor var_1101_begin_0 = const()[name = tensor("op_1101_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1101_end_0 = const()[name = tensor("op_1101_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1101_end_mask_0 = const()[name = tensor("op_1101_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1101_cast_fp16 = slice_by_index(begin = var_1101_begin_0, end = var_1101_end_0, end_mask = var_1101_end_mask_0, x = x_115_cast_fp16)[name = tensor("op_1101_cast_fp16")]; tensor var_1102 = const()[name = tensor("op_1102"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1102, x = var_1101_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_173")]; tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1085_cast_fp16)[name = tensor("transpose_174")]; tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; tensor var_1111_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1111_cast_fp16")]; tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_17_cast_fp16 = mul(x = var_1111_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; tensor scores_19_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_7)[name = tensor("scores_19_cast_fp16")]; tensor var_1117_cast_fp16 = softmax(axis = var_138, x = scores_19_cast_fp16)[name = tensor("op_1117_cast_fp16")]; tensor input_249_cast_fp16 = select(a = var_154_to_fp16, b = var_1117_cast_fp16, cond = mask_7)[name = tensor("input_249_cast_fp16")]; tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_176")]; tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = input_249_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_117_cast_fp16")]; tensor var_1121_perm_0 = const()[name = tensor("op_1121_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1122 = const()[name = tensor("op_1122"), val = tensor([1, -1, 512])]; tensor var_1121_cast_fp16 = transpose(perm = var_1121_perm_0, x = x_117_cast_fp16)[name = tensor("transpose_172")]; tensor input_251_cast_fp16 = reshape(shape = var_1122, x = var_1121_cast_fp16)[name = tensor("input_251_cast_fp16")]; tensor encoder_module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59815552)))]; tensor encoder_module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60339904)))]; tensor linear_43_cast_fp16 = linear(bias = encoder_module_layers_4_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_251_cast_fp16)[name = tensor("linear_43_cast_fp16")]; tensor input_255_cast_fp16 = add(x = input_247_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_255_cast_fp16")]; tensor x_121_axes_0 = const()[name = tensor("x_121_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60340992)))]; tensor encoder_module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60342080)))]; tensor x_121_cast_fp16 = layer_norm(axes = x_121_axes_0, beta = encoder_module_layers_4_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_4_norm_conv_weight_to_fp16, x = input_255_cast_fp16)[name = tensor("x_121_cast_fp16")]; tensor input_257_perm_0 = const()[name = tensor("input_257_perm_0"), val = tensor([0, 2, 1])]; tensor input_259_pad_type_0 = const()[name = tensor("input_259_pad_type_0"), val = tensor("valid")]; tensor input_259_strides_0 = const()[name = tensor("input_259_strides_0"), val = tensor([1])]; tensor input_259_pad_0 = const()[name = tensor("input_259_pad_0"), val = tensor([0, 0])]; tensor input_259_dilations_0 = const()[name = tensor("input_259_dilations_0"), val = tensor([1])]; tensor input_259_groups_0 = const()[name = tensor("input_259_groups_0"), val = tensor(1)]; tensor encoder_module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60343168)))]; tensor encoder_module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61391808)))]; tensor input_257_cast_fp16 = transpose(perm = input_257_perm_0, x = x_121_cast_fp16)[name = tensor("transpose_171")]; tensor input_259_cast_fp16 = conv(bias = encoder_module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_259_dilations_0, groups = input_259_groups_0, pad = input_259_pad_0, pad_type = input_259_pad_type_0, strides = input_259_strides_0, weight = encoder_module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_257_cast_fp16)[name = tensor("input_259_cast_fp16")]; tensor x_123_split_num_splits_0 = const()[name = tensor("x_123_split_num_splits_0"), val = tensor(2)]; tensor x_123_split_axis_0 = const()[name = tensor("x_123_split_axis_0"), val = tensor(1)]; tensor x_123_split_cast_fp16_0, tensor x_123_split_cast_fp16_1 = split(axis = x_123_split_axis_0, num_splits = x_123_split_num_splits_0, x = input_259_cast_fp16)[name = tensor("x_123_split_cast_fp16")]; tensor x_123_split_1_sigmoid_cast_fp16 = sigmoid(x = x_123_split_cast_fp16_1)[name = tensor("x_123_split_1_sigmoid_cast_fp16")]; tensor x_123_cast_fp16 = mul(x = x_123_split_cast_fp16_0, y = x_123_split_1_sigmoid_cast_fp16)[name = tensor("x_123_cast_fp16")]; tensor input_261_cast_fp16 = select(a = var_154_to_fp16, b = x_123_cast_fp16, cond = var_450)[name = tensor("input_261_cast_fp16")]; tensor input_263_pad_0 = const()[name = tensor("input_263_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_263_mode_0 = const()[name = tensor("input_263_mode_0"), val = tensor("constant")]; tensor const_72_to_fp16 = const()[name = tensor("const_72_to_fp16"), val = tensor(0x0p+0)]; tensor input_263_cast_fp16 = pad(constant_val = const_72_to_fp16, mode = input_263_mode_0, pad = input_263_pad_0, x = input_261_cast_fp16)[name = tensor("input_263_cast_fp16")]; tensor input_265_pad_type_0 = const()[name = tensor("input_265_pad_type_0"), val = tensor("valid")]; tensor input_265_groups_0 = const()[name = tensor("input_265_groups_0"), val = tensor(512)]; tensor input_265_strides_0 = const()[name = tensor("input_265_strides_0"), val = tensor([1])]; tensor input_265_pad_0 = const()[name = tensor("input_265_pad_0"), val = tensor([0, 0])]; tensor input_265_dilations_0 = const()[name = tensor("input_265_dilations_0"), val = tensor([1])]; tensor const_201_to_fp16 = const()[name = tensor("const_201_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61393920)))]; tensor const_202_to_fp16 = const()[name = tensor("const_202_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61403200)))]; tensor input_267_cast_fp16 = conv(bias = const_202_to_fp16, dilations = input_265_dilations_0, groups = input_265_groups_0, pad = input_265_pad_0, pad_type = input_265_pad_type_0, strides = input_265_strides_0, weight = const_201_to_fp16, x = input_263_cast_fp16)[name = tensor("input_267_cast_fp16")]; tensor input_269_cast_fp16 = silu(x = input_267_cast_fp16)[name = tensor("input_269_cast_fp16")]; tensor x_125_pad_type_0 = const()[name = tensor("x_125_pad_type_0"), val = tensor("valid")]; tensor x_125_strides_0 = const()[name = tensor("x_125_strides_0"), val = tensor([1])]; tensor x_125_pad_0 = const()[name = tensor("x_125_pad_0"), val = tensor([0, 0])]; tensor x_125_dilations_0 = const()[name = tensor("x_125_dilations_0"), val = tensor([1])]; tensor x_125_groups_0 = const()[name = tensor("x_125_groups_0"), val = tensor(1)]; tensor encoder_module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61404288)))]; tensor encoder_module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61928640)))]; tensor x_125_cast_fp16 = conv(bias = encoder_module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_125_dilations_0, groups = x_125_groups_0, pad = x_125_pad_0, pad_type = x_125_pad_type_0, strides = x_125_strides_0, weight = encoder_module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_269_cast_fp16)[name = tensor("x_125_cast_fp16")]; tensor input_271_perm_0 = const()[name = tensor("input_271_perm_0"), val = tensor([0, 2, 1])]; tensor input_271_cast_fp16 = transpose(perm = input_271_perm_0, x = x_125_cast_fp16)[name = tensor("transpose_170")]; tensor input_273_cast_fp16 = add(x = input_255_cast_fp16, y = input_271_cast_fp16)[name = tensor("input_273_cast_fp16")]; tensor input_275_axes_0 = const()[name = tensor("input_275_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61929728)))]; tensor encoder_module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61930816)))]; tensor input_275_cast_fp16 = layer_norm(axes = input_275_axes_0, beta = encoder_module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_273_cast_fp16)[name = tensor("input_275_cast_fp16")]; tensor encoder_module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61931904)))]; tensor encoder_module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64029120)))]; tensor linear_44_cast_fp16 = linear(bias = encoder_module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_275_cast_fp16)[name = tensor("linear_44_cast_fp16")]; tensor input_279_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_279_cast_fp16")]; tensor encoder_module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64033280)))]; tensor encoder_module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66130496)))]; tensor linear_45_cast_fp16 = linear(bias = encoder_module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("linear_45_cast_fp16")]; tensor var_1188_to_fp16 = const()[name = tensor("op_1188_to_fp16"), val = tensor(0x1p-1)]; tensor var_1189_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1188_to_fp16)[name = tensor("op_1189_cast_fp16")]; tensor input_285_cast_fp16 = add(x = input_273_cast_fp16, y = var_1189_cast_fp16)[name = tensor("input_285_cast_fp16")]; tensor input_287_axes_0 = const()[name = tensor("input_287_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66131584)))]; tensor encoder_module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66132672)))]; tensor input_287_cast_fp16 = layer_norm(axes = input_287_axes_0, beta = encoder_module_layers_4_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_4_norm_out_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("input_287_cast_fp16")]; tensor input_289_axes_0 = const()[name = tensor("input_289_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66133760)))]; tensor encoder_module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66134848)))]; tensor input_289_cast_fp16 = layer_norm(axes = input_289_axes_0, beta = encoder_module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_287_cast_fp16)[name = tensor("input_289_cast_fp16")]; tensor encoder_module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(66135936)))]; tensor encoder_module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68233152)))]; tensor linear_46_cast_fp16 = linear(bias = encoder_module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_289_cast_fp16)[name = tensor("linear_46_cast_fp16")]; tensor input_293_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_293_cast_fp16")]; tensor encoder_module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(68237312)))]; tensor encoder_module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70334528)))]; tensor linear_47_cast_fp16 = linear(bias = encoder_module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_293_cast_fp16)[name = tensor("linear_47_cast_fp16")]; tensor var_1219_to_fp16 = const()[name = tensor("op_1219_to_fp16"), val = tensor(0x1p-1)]; tensor var_1220_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1219_to_fp16)[name = tensor("op_1220_cast_fp16")]; tensor input_299_cast_fp16 = add(x = input_287_cast_fp16, y = var_1220_cast_fp16)[name = tensor("input_299_cast_fp16")]; tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70335616)))]; tensor encoder_module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70336704)))]; tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = encoder_module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_5_norm_self_att_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("query_11_cast_fp16")]; tensor encoder_module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70337792)))]; tensor encoder_module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70862144)))]; tensor linear_48_cast_fp16 = linear(bias = encoder_module_layers_5_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; tensor var_1237 = const()[name = tensor("op_1237"), val = tensor([1, -1, 8, 64])]; tensor q_31_cast_fp16 = reshape(shape = var_1237, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; tensor encoder_module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70863232)))]; tensor encoder_module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71387584)))]; tensor linear_49_cast_fp16 = linear(bias = encoder_module_layers_5_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; tensor var_1242 = const()[name = tensor("op_1242"), val = tensor([1, -1, 8, 64])]; tensor k_21_cast_fp16 = reshape(shape = var_1242, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; tensor encoder_module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71388672)))]; tensor encoder_module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71913024)))]; tensor linear_50_cast_fp16 = linear(bias = encoder_module_layers_5_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; tensor var_1247 = const()[name = tensor("op_1247"), val = tensor([1, -1, 8, 64])]; tensor v_11_cast_fp16 = reshape(shape = var_1247, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71914112)))]; tensor var_1259_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1259_cast_fp16")]; tensor encoder_module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71915200)))]; tensor var_1261_cast_fp16 = add(x = q_31_cast_fp16, y = encoder_module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1261_cast_fp16")]; tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_133_transpose_x_0 = const()[name = tensor("x_133_transpose_x_0"), val = tensor(false)]; tensor x_133_transpose_y_0 = const()[name = tensor("x_133_transpose_y_0"), val = tensor(false)]; tensor var_1263_to_fp16 = const()[name = tensor("op_1263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71916288)))]; tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1261_cast_fp16)[name = tensor("transpose_168")]; tensor x_133_cast_fp16 = matmul(transpose_x = x_133_transpose_x_0, transpose_y = x_133_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1263_to_fp16)[name = tensor("x_133_cast_fp16")]; tensor x_135_pad_0 = const()[name = tensor("x_135_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_135_mode_0 = const()[name = tensor("x_135_mode_0"), val = tensor("constant")]; tensor const_79_to_fp16 = const()[name = tensor("const_79_to_fp16"), val = tensor(0x0p+0)]; tensor x_135_cast_fp16 = pad(constant_val = const_79_to_fp16, mode = x_135_mode_0, pad = x_135_pad_0, x = x_133_cast_fp16)[name = tensor("x_135_cast_fp16")]; tensor var_1271 = const()[name = tensor("op_1271"), val = tensor([1, 8, -1, 188])]; tensor x_137_cast_fp16 = reshape(shape = var_1271, x = x_135_cast_fp16)[name = tensor("x_137_cast_fp16")]; tensor var_1275_begin_0 = const()[name = tensor("op_1275_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1275_end_0 = const()[name = tensor("op_1275_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1275_end_mask_0 = const()[name = tensor("op_1275_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1275_cast_fp16 = slice_by_index(begin = var_1275_begin_0, end = var_1275_end_0, end_mask = var_1275_end_mask_0, x = x_137_cast_fp16)[name = tensor("op_1275_cast_fp16")]; tensor var_1276 = const()[name = tensor("op_1276"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1276, x = var_1275_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_166")]; tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1259_cast_fp16)[name = tensor("transpose_167")]; tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; tensor var_1285_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1285_cast_fp16")]; tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_21_cast_fp16 = mul(x = var_1285_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; tensor scores_23_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_7)[name = tensor("scores_23_cast_fp16")]; tensor var_1291_cast_fp16 = softmax(axis = var_138, x = scores_23_cast_fp16)[name = tensor("op_1291_cast_fp16")]; tensor input_301_cast_fp16 = select(a = var_154_to_fp16, b = var_1291_cast_fp16, cond = mask_7)[name = tensor("input_301_cast_fp16")]; tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_169")]; tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = input_301_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_139_cast_fp16")]; tensor var_1295_perm_0 = const()[name = tensor("op_1295_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1296 = const()[name = tensor("op_1296"), val = tensor([1, -1, 512])]; tensor var_1295_cast_fp16 = transpose(perm = var_1295_perm_0, x = x_139_cast_fp16)[name = tensor("transpose_165")]; tensor input_303_cast_fp16 = reshape(shape = var_1296, x = var_1295_cast_fp16)[name = tensor("input_303_cast_fp16")]; tensor encoder_module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72300352)))]; tensor encoder_module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72824704)))]; tensor linear_52_cast_fp16 = linear(bias = encoder_module_layers_5_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_303_cast_fp16)[name = tensor("linear_52_cast_fp16")]; tensor input_307_cast_fp16 = add(x = input_299_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_307_cast_fp16")]; tensor x_143_axes_0 = const()[name = tensor("x_143_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72825792)))]; tensor encoder_module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72826880)))]; tensor x_143_cast_fp16 = layer_norm(axes = x_143_axes_0, beta = encoder_module_layers_5_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_5_norm_conv_weight_to_fp16, x = input_307_cast_fp16)[name = tensor("x_143_cast_fp16")]; tensor input_309_perm_0 = const()[name = tensor("input_309_perm_0"), val = tensor([0, 2, 1])]; tensor input_311_pad_type_0 = const()[name = tensor("input_311_pad_type_0"), val = tensor("valid")]; tensor input_311_strides_0 = const()[name = tensor("input_311_strides_0"), val = tensor([1])]; tensor input_311_pad_0 = const()[name = tensor("input_311_pad_0"), val = tensor([0, 0])]; tensor input_311_dilations_0 = const()[name = tensor("input_311_dilations_0"), val = tensor([1])]; tensor input_311_groups_0 = const()[name = tensor("input_311_groups_0"), val = tensor(1)]; tensor encoder_module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72827968)))]; tensor encoder_module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73876608)))]; tensor input_309_cast_fp16 = transpose(perm = input_309_perm_0, x = x_143_cast_fp16)[name = tensor("transpose_164")]; tensor input_311_cast_fp16 = conv(bias = encoder_module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_311_dilations_0, groups = input_311_groups_0, pad = input_311_pad_0, pad_type = input_311_pad_type_0, strides = input_311_strides_0, weight = encoder_module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_309_cast_fp16)[name = tensor("input_311_cast_fp16")]; tensor x_145_split_num_splits_0 = const()[name = tensor("x_145_split_num_splits_0"), val = tensor(2)]; tensor x_145_split_axis_0 = const()[name = tensor("x_145_split_axis_0"), val = tensor(1)]; tensor x_145_split_cast_fp16_0, tensor x_145_split_cast_fp16_1 = split(axis = x_145_split_axis_0, num_splits = x_145_split_num_splits_0, x = input_311_cast_fp16)[name = tensor("x_145_split_cast_fp16")]; tensor x_145_split_1_sigmoid_cast_fp16 = sigmoid(x = x_145_split_cast_fp16_1)[name = tensor("x_145_split_1_sigmoid_cast_fp16")]; tensor x_145_cast_fp16 = mul(x = x_145_split_cast_fp16_0, y = x_145_split_1_sigmoid_cast_fp16)[name = tensor("x_145_cast_fp16")]; tensor input_313_cast_fp16 = select(a = var_154_to_fp16, b = x_145_cast_fp16, cond = var_450)[name = tensor("input_313_cast_fp16")]; tensor input_315_pad_0 = const()[name = tensor("input_315_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_315_mode_0 = const()[name = tensor("input_315_mode_0"), val = tensor("constant")]; tensor const_82_to_fp16 = const()[name = tensor("const_82_to_fp16"), val = tensor(0x0p+0)]; tensor input_315_cast_fp16 = pad(constant_val = const_82_to_fp16, mode = input_315_mode_0, pad = input_315_pad_0, x = input_313_cast_fp16)[name = tensor("input_315_cast_fp16")]; tensor input_317_pad_type_0 = const()[name = tensor("input_317_pad_type_0"), val = tensor("valid")]; tensor input_317_groups_0 = const()[name = tensor("input_317_groups_0"), val = tensor(512)]; tensor input_317_strides_0 = const()[name = tensor("input_317_strides_0"), val = tensor([1])]; tensor input_317_pad_0 = const()[name = tensor("input_317_pad_0"), val = tensor([0, 0])]; tensor input_317_dilations_0 = const()[name = tensor("input_317_dilations_0"), val = tensor([1])]; tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73878720)))]; tensor const_204_to_fp16 = const()[name = tensor("const_204_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73888000)))]; tensor input_319_cast_fp16 = conv(bias = const_204_to_fp16, dilations = input_317_dilations_0, groups = input_317_groups_0, pad = input_317_pad_0, pad_type = input_317_pad_type_0, strides = input_317_strides_0, weight = const_203_to_fp16, x = input_315_cast_fp16)[name = tensor("input_319_cast_fp16")]; tensor input_321_cast_fp16 = silu(x = input_319_cast_fp16)[name = tensor("input_321_cast_fp16")]; tensor x_147_pad_type_0 = const()[name = tensor("x_147_pad_type_0"), val = tensor("valid")]; tensor x_147_strides_0 = const()[name = tensor("x_147_strides_0"), val = tensor([1])]; tensor x_147_pad_0 = const()[name = tensor("x_147_pad_0"), val = tensor([0, 0])]; tensor x_147_dilations_0 = const()[name = tensor("x_147_dilations_0"), val = tensor([1])]; tensor x_147_groups_0 = const()[name = tensor("x_147_groups_0"), val = tensor(1)]; tensor encoder_module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73889088)))]; tensor encoder_module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74413440)))]; tensor x_147_cast_fp16 = conv(bias = encoder_module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_147_dilations_0, groups = x_147_groups_0, pad = x_147_pad_0, pad_type = x_147_pad_type_0, strides = x_147_strides_0, weight = encoder_module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_321_cast_fp16)[name = tensor("x_147_cast_fp16")]; tensor input_323_perm_0 = const()[name = tensor("input_323_perm_0"), val = tensor([0, 2, 1])]; tensor input_323_cast_fp16 = transpose(perm = input_323_perm_0, x = x_147_cast_fp16)[name = tensor("transpose_163")]; tensor input_325_cast_fp16 = add(x = input_307_cast_fp16, y = input_323_cast_fp16)[name = tensor("input_325_cast_fp16")]; tensor input_327_axes_0 = const()[name = tensor("input_327_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74414528)))]; tensor encoder_module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74415616)))]; tensor input_327_cast_fp16 = layer_norm(axes = input_327_axes_0, beta = encoder_module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_325_cast_fp16)[name = tensor("input_327_cast_fp16")]; tensor encoder_module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(74416704)))]; tensor encoder_module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76513920)))]; tensor linear_53_cast_fp16 = linear(bias = encoder_module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_327_cast_fp16)[name = tensor("linear_53_cast_fp16")]; tensor input_331_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_331_cast_fp16")]; tensor encoder_module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(76518080)))]; tensor encoder_module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78615296)))]; tensor linear_54_cast_fp16 = linear(bias = encoder_module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("linear_54_cast_fp16")]; tensor var_1362_to_fp16 = const()[name = tensor("op_1362_to_fp16"), val = tensor(0x1p-1)]; tensor var_1363_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1362_to_fp16)[name = tensor("op_1363_cast_fp16")]; tensor input_337_cast_fp16 = add(x = input_325_cast_fp16, y = var_1363_cast_fp16)[name = tensor("input_337_cast_fp16")]; tensor input_339_axes_0 = const()[name = tensor("input_339_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78616384)))]; tensor encoder_module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78617472)))]; tensor input_339_cast_fp16 = layer_norm(axes = input_339_axes_0, beta = encoder_module_layers_5_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_5_norm_out_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("input_339_cast_fp16")]; tensor input_341_axes_0 = const()[name = tensor("input_341_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78618560)))]; tensor encoder_module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78619648)))]; tensor input_341_cast_fp16 = layer_norm(axes = input_341_axes_0, beta = encoder_module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_339_cast_fp16)[name = tensor("input_341_cast_fp16")]; tensor encoder_module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(78620736)))]; tensor encoder_module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80717952)))]; tensor linear_55_cast_fp16 = linear(bias = encoder_module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_341_cast_fp16)[name = tensor("linear_55_cast_fp16")]; tensor input_345_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_345_cast_fp16")]; tensor encoder_module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(80722112)))]; tensor encoder_module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82819328)))]; tensor linear_56_cast_fp16 = linear(bias = encoder_module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_345_cast_fp16)[name = tensor("linear_56_cast_fp16")]; tensor var_1393_to_fp16 = const()[name = tensor("op_1393_to_fp16"), val = tensor(0x1p-1)]; tensor var_1394_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1393_to_fp16)[name = tensor("op_1394_cast_fp16")]; tensor input_351_cast_fp16 = add(x = input_339_cast_fp16, y = var_1394_cast_fp16)[name = tensor("input_351_cast_fp16")]; tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82820416)))]; tensor encoder_module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82821504)))]; tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = encoder_module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_6_norm_self_att_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("query_13_cast_fp16")]; tensor encoder_module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82822592)))]; tensor encoder_module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83346944)))]; tensor linear_57_cast_fp16 = linear(bias = encoder_module_layers_6_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; tensor var_1411 = const()[name = tensor("op_1411"), val = tensor([1, -1, 8, 64])]; tensor q_37_cast_fp16 = reshape(shape = var_1411, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; tensor encoder_module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83348032)))]; tensor encoder_module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83872384)))]; tensor linear_58_cast_fp16 = linear(bias = encoder_module_layers_6_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; tensor var_1416 = const()[name = tensor("op_1416"), val = tensor([1, -1, 8, 64])]; tensor k_25_cast_fp16 = reshape(shape = var_1416, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; tensor encoder_module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83873472)))]; tensor encoder_module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84397824)))]; tensor linear_59_cast_fp16 = linear(bias = encoder_module_layers_6_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; tensor var_1421 = const()[name = tensor("op_1421"), val = tensor([1, -1, 8, 64])]; tensor v_13_cast_fp16 = reshape(shape = var_1421, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84398912)))]; tensor var_1433_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1433_cast_fp16")]; tensor encoder_module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84400000)))]; tensor var_1435_cast_fp16 = add(x = q_37_cast_fp16, y = encoder_module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1435_cast_fp16")]; tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_155_transpose_x_0 = const()[name = tensor("x_155_transpose_x_0"), val = tensor(false)]; tensor x_155_transpose_y_0 = const()[name = tensor("x_155_transpose_y_0"), val = tensor(false)]; tensor var_1437_to_fp16 = const()[name = tensor("op_1437_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84401088)))]; tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1435_cast_fp16)[name = tensor("transpose_161")]; tensor x_155_cast_fp16 = matmul(transpose_x = x_155_transpose_x_0, transpose_y = x_155_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1437_to_fp16)[name = tensor("x_155_cast_fp16")]; tensor x_157_pad_0 = const()[name = tensor("x_157_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_157_mode_0 = const()[name = tensor("x_157_mode_0"), val = tensor("constant")]; tensor const_89_to_fp16 = const()[name = tensor("const_89_to_fp16"), val = tensor(0x0p+0)]; tensor x_157_cast_fp16 = pad(constant_val = const_89_to_fp16, mode = x_157_mode_0, pad = x_157_pad_0, x = x_155_cast_fp16)[name = tensor("x_157_cast_fp16")]; tensor var_1445 = const()[name = tensor("op_1445"), val = tensor([1, 8, -1, 188])]; tensor x_159_cast_fp16 = reshape(shape = var_1445, x = x_157_cast_fp16)[name = tensor("x_159_cast_fp16")]; tensor var_1449_begin_0 = const()[name = tensor("op_1449_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1449_end_0 = const()[name = tensor("op_1449_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1449_end_mask_0 = const()[name = tensor("op_1449_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1449_cast_fp16 = slice_by_index(begin = var_1449_begin_0, end = var_1449_end_0, end_mask = var_1449_end_mask_0, x = x_159_cast_fp16)[name = tensor("op_1449_cast_fp16")]; tensor var_1450 = const()[name = tensor("op_1450"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1450, x = var_1449_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_159")]; tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1433_cast_fp16)[name = tensor("transpose_160")]; tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; tensor var_1459_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1459_cast_fp16")]; tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_25_cast_fp16 = mul(x = var_1459_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; tensor scores_27_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_7)[name = tensor("scores_27_cast_fp16")]; tensor var_1465_cast_fp16 = softmax(axis = var_138, x = scores_27_cast_fp16)[name = tensor("op_1465_cast_fp16")]; tensor input_353_cast_fp16 = select(a = var_154_to_fp16, b = var_1465_cast_fp16, cond = mask_7)[name = tensor("input_353_cast_fp16")]; tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_162")]; tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = input_353_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_161_cast_fp16")]; tensor var_1469_perm_0 = const()[name = tensor("op_1469_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1470 = const()[name = tensor("op_1470"), val = tensor([1, -1, 512])]; tensor var_1469_cast_fp16 = transpose(perm = var_1469_perm_0, x = x_161_cast_fp16)[name = tensor("transpose_158")]; tensor input_355_cast_fp16 = reshape(shape = var_1470, x = var_1469_cast_fp16)[name = tensor("input_355_cast_fp16")]; tensor encoder_module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84785152)))]; tensor encoder_module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85309504)))]; tensor linear_61_cast_fp16 = linear(bias = encoder_module_layers_6_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_355_cast_fp16)[name = tensor("linear_61_cast_fp16")]; tensor input_359_cast_fp16 = add(x = input_351_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_359_cast_fp16")]; tensor x_165_axes_0 = const()[name = tensor("x_165_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85310592)))]; tensor encoder_module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85311680)))]; tensor x_165_cast_fp16 = layer_norm(axes = x_165_axes_0, beta = encoder_module_layers_6_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_6_norm_conv_weight_to_fp16, x = input_359_cast_fp16)[name = tensor("x_165_cast_fp16")]; tensor input_361_perm_0 = const()[name = tensor("input_361_perm_0"), val = tensor([0, 2, 1])]; tensor input_363_pad_type_0 = const()[name = tensor("input_363_pad_type_0"), val = tensor("valid")]; tensor input_363_strides_0 = const()[name = tensor("input_363_strides_0"), val = tensor([1])]; tensor input_363_pad_0 = const()[name = tensor("input_363_pad_0"), val = tensor([0, 0])]; tensor input_363_dilations_0 = const()[name = tensor("input_363_dilations_0"), val = tensor([1])]; tensor input_363_groups_0 = const()[name = tensor("input_363_groups_0"), val = tensor(1)]; tensor encoder_module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85312768)))]; tensor encoder_module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86361408)))]; tensor input_361_cast_fp16 = transpose(perm = input_361_perm_0, x = x_165_cast_fp16)[name = tensor("transpose_157")]; tensor input_363_cast_fp16 = conv(bias = encoder_module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_363_dilations_0, groups = input_363_groups_0, pad = input_363_pad_0, pad_type = input_363_pad_type_0, strides = input_363_strides_0, weight = encoder_module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_361_cast_fp16)[name = tensor("input_363_cast_fp16")]; tensor x_167_split_num_splits_0 = const()[name = tensor("x_167_split_num_splits_0"), val = tensor(2)]; tensor x_167_split_axis_0 = const()[name = tensor("x_167_split_axis_0"), val = tensor(1)]; tensor x_167_split_cast_fp16_0, tensor x_167_split_cast_fp16_1 = split(axis = x_167_split_axis_0, num_splits = x_167_split_num_splits_0, x = input_363_cast_fp16)[name = tensor("x_167_split_cast_fp16")]; tensor x_167_split_1_sigmoid_cast_fp16 = sigmoid(x = x_167_split_cast_fp16_1)[name = tensor("x_167_split_1_sigmoid_cast_fp16")]; tensor x_167_cast_fp16 = mul(x = x_167_split_cast_fp16_0, y = x_167_split_1_sigmoid_cast_fp16)[name = tensor("x_167_cast_fp16")]; tensor input_365_cast_fp16 = select(a = var_154_to_fp16, b = x_167_cast_fp16, cond = var_450)[name = tensor("input_365_cast_fp16")]; tensor input_367_pad_0 = const()[name = tensor("input_367_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_367_mode_0 = const()[name = tensor("input_367_mode_0"), val = tensor("constant")]; tensor const_92_to_fp16 = const()[name = tensor("const_92_to_fp16"), val = tensor(0x0p+0)]; tensor input_367_cast_fp16 = pad(constant_val = const_92_to_fp16, mode = input_367_mode_0, pad = input_367_pad_0, x = input_365_cast_fp16)[name = tensor("input_367_cast_fp16")]; tensor input_369_pad_type_0 = const()[name = tensor("input_369_pad_type_0"), val = tensor("valid")]; tensor input_369_groups_0 = const()[name = tensor("input_369_groups_0"), val = tensor(512)]; tensor input_369_strides_0 = const()[name = tensor("input_369_strides_0"), val = tensor([1])]; tensor input_369_pad_0 = const()[name = tensor("input_369_pad_0"), val = tensor([0, 0])]; tensor input_369_dilations_0 = const()[name = tensor("input_369_dilations_0"), val = tensor([1])]; tensor const_205_to_fp16 = const()[name = tensor("const_205_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86363520)))]; tensor const_206_to_fp16 = const()[name = tensor("const_206_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86372800)))]; tensor input_371_cast_fp16 = conv(bias = const_206_to_fp16, dilations = input_369_dilations_0, groups = input_369_groups_0, pad = input_369_pad_0, pad_type = input_369_pad_type_0, strides = input_369_strides_0, weight = const_205_to_fp16, x = input_367_cast_fp16)[name = tensor("input_371_cast_fp16")]; tensor input_373_cast_fp16 = silu(x = input_371_cast_fp16)[name = tensor("input_373_cast_fp16")]; tensor x_169_pad_type_0 = const()[name = tensor("x_169_pad_type_0"), val = tensor("valid")]; tensor x_169_strides_0 = const()[name = tensor("x_169_strides_0"), val = tensor([1])]; tensor x_169_pad_0 = const()[name = tensor("x_169_pad_0"), val = tensor([0, 0])]; tensor x_169_dilations_0 = const()[name = tensor("x_169_dilations_0"), val = tensor([1])]; tensor x_169_groups_0 = const()[name = tensor("x_169_groups_0"), val = tensor(1)]; tensor encoder_module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86373888)))]; tensor encoder_module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86898240)))]; tensor x_169_cast_fp16 = conv(bias = encoder_module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_169_dilations_0, groups = x_169_groups_0, pad = x_169_pad_0, pad_type = x_169_pad_type_0, strides = x_169_strides_0, weight = encoder_module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_373_cast_fp16)[name = tensor("x_169_cast_fp16")]; tensor input_375_perm_0 = const()[name = tensor("input_375_perm_0"), val = tensor([0, 2, 1])]; tensor input_375_cast_fp16 = transpose(perm = input_375_perm_0, x = x_169_cast_fp16)[name = tensor("transpose_156")]; tensor input_377_cast_fp16 = add(x = input_359_cast_fp16, y = input_375_cast_fp16)[name = tensor("input_377_cast_fp16")]; tensor input_379_axes_0 = const()[name = tensor("input_379_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86899328)))]; tensor encoder_module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86900416)))]; tensor input_379_cast_fp16 = layer_norm(axes = input_379_axes_0, beta = encoder_module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_377_cast_fp16)[name = tensor("input_379_cast_fp16")]; tensor encoder_module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86901504)))]; tensor encoder_module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88998720)))]; tensor linear_62_cast_fp16 = linear(bias = encoder_module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_379_cast_fp16)[name = tensor("linear_62_cast_fp16")]; tensor input_383_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_383_cast_fp16")]; tensor encoder_module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(89002880)))]; tensor encoder_module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91100096)))]; tensor linear_63_cast_fp16 = linear(bias = encoder_module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("linear_63_cast_fp16")]; tensor var_1536_to_fp16 = const()[name = tensor("op_1536_to_fp16"), val = tensor(0x1p-1)]; tensor var_1537_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1536_to_fp16)[name = tensor("op_1537_cast_fp16")]; tensor input_389_cast_fp16 = add(x = input_377_cast_fp16, y = var_1537_cast_fp16)[name = tensor("input_389_cast_fp16")]; tensor input_391_axes_0 = const()[name = tensor("input_391_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91101184)))]; tensor encoder_module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91102272)))]; tensor input_391_cast_fp16 = layer_norm(axes = input_391_axes_0, beta = encoder_module_layers_6_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_6_norm_out_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("input_391_cast_fp16")]; tensor input_393_axes_0 = const()[name = tensor("input_393_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91103360)))]; tensor encoder_module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91104448)))]; tensor input_393_cast_fp16 = layer_norm(axes = input_393_axes_0, beta = encoder_module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_391_cast_fp16)[name = tensor("input_393_cast_fp16")]; tensor encoder_module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(91105536)))]; tensor encoder_module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93202752)))]; tensor linear_64_cast_fp16 = linear(bias = encoder_module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_393_cast_fp16)[name = tensor("linear_64_cast_fp16")]; tensor input_397_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_397_cast_fp16")]; tensor encoder_module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(93206912)))]; tensor encoder_module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95304128)))]; tensor linear_65_cast_fp16 = linear(bias = encoder_module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_397_cast_fp16)[name = tensor("linear_65_cast_fp16")]; tensor var_1567_to_fp16 = const()[name = tensor("op_1567_to_fp16"), val = tensor(0x1p-1)]; tensor var_1568_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1567_to_fp16)[name = tensor("op_1568_cast_fp16")]; tensor input_403_cast_fp16 = add(x = input_391_cast_fp16, y = var_1568_cast_fp16)[name = tensor("input_403_cast_fp16")]; tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95305216)))]; tensor encoder_module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95306304)))]; tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = encoder_module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_7_norm_self_att_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("query_15_cast_fp16")]; tensor encoder_module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95307392)))]; tensor encoder_module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95831744)))]; tensor linear_66_cast_fp16 = linear(bias = encoder_module_layers_7_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; tensor var_1585 = const()[name = tensor("op_1585"), val = tensor([1, -1, 8, 64])]; tensor q_43_cast_fp16 = reshape(shape = var_1585, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; tensor encoder_module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95832832)))]; tensor encoder_module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96357184)))]; tensor linear_67_cast_fp16 = linear(bias = encoder_module_layers_7_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; tensor var_1590 = const()[name = tensor("op_1590"), val = tensor([1, -1, 8, 64])]; tensor k_29_cast_fp16 = reshape(shape = var_1590, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; tensor encoder_module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96358272)))]; tensor encoder_module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96882624)))]; tensor linear_68_cast_fp16 = linear(bias = encoder_module_layers_7_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; tensor var_1595 = const()[name = tensor("op_1595"), val = tensor([1, -1, 8, 64])]; tensor v_15_cast_fp16 = reshape(shape = var_1595, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96883712)))]; tensor var_1607_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1607_cast_fp16")]; tensor encoder_module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96884800)))]; tensor var_1609_cast_fp16 = add(x = q_43_cast_fp16, y = encoder_module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1609_cast_fp16")]; tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_177_transpose_x_0 = const()[name = tensor("x_177_transpose_x_0"), val = tensor(false)]; tensor x_177_transpose_y_0 = const()[name = tensor("x_177_transpose_y_0"), val = tensor(false)]; tensor var_1611_to_fp16 = const()[name = tensor("op_1611_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96885888)))]; tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1609_cast_fp16)[name = tensor("transpose_154")]; tensor x_177_cast_fp16 = matmul(transpose_x = x_177_transpose_x_0, transpose_y = x_177_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1611_to_fp16)[name = tensor("x_177_cast_fp16")]; tensor x_179_pad_0 = const()[name = tensor("x_179_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_179_mode_0 = const()[name = tensor("x_179_mode_0"), val = tensor("constant")]; tensor const_99_to_fp16 = const()[name = tensor("const_99_to_fp16"), val = tensor(0x0p+0)]; tensor x_179_cast_fp16 = pad(constant_val = const_99_to_fp16, mode = x_179_mode_0, pad = x_179_pad_0, x = x_177_cast_fp16)[name = tensor("x_179_cast_fp16")]; tensor var_1619 = const()[name = tensor("op_1619"), val = tensor([1, 8, -1, 188])]; tensor x_181_cast_fp16 = reshape(shape = var_1619, x = x_179_cast_fp16)[name = tensor("x_181_cast_fp16")]; tensor var_1623_begin_0 = const()[name = tensor("op_1623_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1623_end_0 = const()[name = tensor("op_1623_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1623_end_mask_0 = const()[name = tensor("op_1623_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1623_cast_fp16 = slice_by_index(begin = var_1623_begin_0, end = var_1623_end_0, end_mask = var_1623_end_mask_0, x = x_181_cast_fp16)[name = tensor("op_1623_cast_fp16")]; tensor var_1624 = const()[name = tensor("op_1624"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1624, x = var_1623_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_152")]; tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1607_cast_fp16)[name = tensor("transpose_153")]; tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; tensor var_1633_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1633_cast_fp16")]; tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_29_cast_fp16 = mul(x = var_1633_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; tensor scores_31_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_7)[name = tensor("scores_31_cast_fp16")]; tensor var_1639_cast_fp16 = softmax(axis = var_138, x = scores_31_cast_fp16)[name = tensor("op_1639_cast_fp16")]; tensor input_405_cast_fp16 = select(a = var_154_to_fp16, b = var_1639_cast_fp16, cond = mask_7)[name = tensor("input_405_cast_fp16")]; tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_155")]; tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = input_405_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_183_cast_fp16")]; tensor var_1643_perm_0 = const()[name = tensor("op_1643_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1644 = const()[name = tensor("op_1644"), val = tensor([1, -1, 512])]; tensor var_1643_cast_fp16 = transpose(perm = var_1643_perm_0, x = x_183_cast_fp16)[name = tensor("transpose_151")]; tensor input_407_cast_fp16 = reshape(shape = var_1644, x = var_1643_cast_fp16)[name = tensor("input_407_cast_fp16")]; tensor encoder_module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97269952)))]; tensor encoder_module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97794304)))]; tensor linear_70_cast_fp16 = linear(bias = encoder_module_layers_7_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_407_cast_fp16)[name = tensor("linear_70_cast_fp16")]; tensor input_411_cast_fp16 = add(x = input_403_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_411_cast_fp16")]; tensor x_187_axes_0 = const()[name = tensor("x_187_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97795392)))]; tensor encoder_module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97796480)))]; tensor x_187_cast_fp16 = layer_norm(axes = x_187_axes_0, beta = encoder_module_layers_7_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_7_norm_conv_weight_to_fp16, x = input_411_cast_fp16)[name = tensor("x_187_cast_fp16")]; tensor input_413_perm_0 = const()[name = tensor("input_413_perm_0"), val = tensor([0, 2, 1])]; tensor input_415_pad_type_0 = const()[name = tensor("input_415_pad_type_0"), val = tensor("valid")]; tensor input_415_strides_0 = const()[name = tensor("input_415_strides_0"), val = tensor([1])]; tensor input_415_pad_0 = const()[name = tensor("input_415_pad_0"), val = tensor([0, 0])]; tensor input_415_dilations_0 = const()[name = tensor("input_415_dilations_0"), val = tensor([1])]; tensor input_415_groups_0 = const()[name = tensor("input_415_groups_0"), val = tensor(1)]; tensor encoder_module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(97797568)))]; tensor encoder_module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98846208)))]; tensor input_413_cast_fp16 = transpose(perm = input_413_perm_0, x = x_187_cast_fp16)[name = tensor("transpose_150")]; tensor input_415_cast_fp16 = conv(bias = encoder_module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_415_dilations_0, groups = input_415_groups_0, pad = input_415_pad_0, pad_type = input_415_pad_type_0, strides = input_415_strides_0, weight = encoder_module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_413_cast_fp16)[name = tensor("input_415_cast_fp16")]; tensor x_189_split_num_splits_0 = const()[name = tensor("x_189_split_num_splits_0"), val = tensor(2)]; tensor x_189_split_axis_0 = const()[name = tensor("x_189_split_axis_0"), val = tensor(1)]; tensor x_189_split_cast_fp16_0, tensor x_189_split_cast_fp16_1 = split(axis = x_189_split_axis_0, num_splits = x_189_split_num_splits_0, x = input_415_cast_fp16)[name = tensor("x_189_split_cast_fp16")]; tensor x_189_split_1_sigmoid_cast_fp16 = sigmoid(x = x_189_split_cast_fp16_1)[name = tensor("x_189_split_1_sigmoid_cast_fp16")]; tensor x_189_cast_fp16 = mul(x = x_189_split_cast_fp16_0, y = x_189_split_1_sigmoid_cast_fp16)[name = tensor("x_189_cast_fp16")]; tensor input_417_cast_fp16 = select(a = var_154_to_fp16, b = x_189_cast_fp16, cond = var_450)[name = tensor("input_417_cast_fp16")]; tensor input_419_pad_0 = const()[name = tensor("input_419_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_419_mode_0 = const()[name = tensor("input_419_mode_0"), val = tensor("constant")]; tensor const_102_to_fp16 = const()[name = tensor("const_102_to_fp16"), val = tensor(0x0p+0)]; tensor input_419_cast_fp16 = pad(constant_val = const_102_to_fp16, mode = input_419_mode_0, pad = input_419_pad_0, x = input_417_cast_fp16)[name = tensor("input_419_cast_fp16")]; tensor input_421_pad_type_0 = const()[name = tensor("input_421_pad_type_0"), val = tensor("valid")]; tensor input_421_groups_0 = const()[name = tensor("input_421_groups_0"), val = tensor(512)]; tensor input_421_strides_0 = const()[name = tensor("input_421_strides_0"), val = tensor([1])]; tensor input_421_pad_0 = const()[name = tensor("input_421_pad_0"), val = tensor([0, 0])]; tensor input_421_dilations_0 = const()[name = tensor("input_421_dilations_0"), val = tensor([1])]; tensor const_207_to_fp16 = const()[name = tensor("const_207_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98848320)))]; tensor const_208_to_fp16 = const()[name = tensor("const_208_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98857600)))]; tensor input_423_cast_fp16 = conv(bias = const_208_to_fp16, dilations = input_421_dilations_0, groups = input_421_groups_0, pad = input_421_pad_0, pad_type = input_421_pad_type_0, strides = input_421_strides_0, weight = const_207_to_fp16, x = input_419_cast_fp16)[name = tensor("input_423_cast_fp16")]; tensor input_425_cast_fp16 = silu(x = input_423_cast_fp16)[name = tensor("input_425_cast_fp16")]; tensor x_191_pad_type_0 = const()[name = tensor("x_191_pad_type_0"), val = tensor("valid")]; tensor x_191_strides_0 = const()[name = tensor("x_191_strides_0"), val = tensor([1])]; tensor x_191_pad_0 = const()[name = tensor("x_191_pad_0"), val = tensor([0, 0])]; tensor x_191_dilations_0 = const()[name = tensor("x_191_dilations_0"), val = tensor([1])]; tensor x_191_groups_0 = const()[name = tensor("x_191_groups_0"), val = tensor(1)]; tensor encoder_module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98858688)))]; tensor encoder_module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99383040)))]; tensor x_191_cast_fp16 = conv(bias = encoder_module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_191_dilations_0, groups = x_191_groups_0, pad = x_191_pad_0, pad_type = x_191_pad_type_0, strides = x_191_strides_0, weight = encoder_module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_425_cast_fp16)[name = tensor("x_191_cast_fp16")]; tensor input_427_perm_0 = const()[name = tensor("input_427_perm_0"), val = tensor([0, 2, 1])]; tensor input_427_cast_fp16 = transpose(perm = input_427_perm_0, x = x_191_cast_fp16)[name = tensor("transpose_149")]; tensor input_429_cast_fp16 = add(x = input_411_cast_fp16, y = input_427_cast_fp16)[name = tensor("input_429_cast_fp16")]; tensor input_431_axes_0 = const()[name = tensor("input_431_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99384128)))]; tensor encoder_module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99385216)))]; tensor input_431_cast_fp16 = layer_norm(axes = input_431_axes_0, beta = encoder_module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_429_cast_fp16)[name = tensor("input_431_cast_fp16")]; tensor encoder_module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(99386304)))]; tensor encoder_module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101483520)))]; tensor linear_71_cast_fp16 = linear(bias = encoder_module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_431_cast_fp16)[name = tensor("linear_71_cast_fp16")]; tensor input_435_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_435_cast_fp16")]; tensor encoder_module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(101487680)))]; tensor encoder_module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103584896)))]; tensor linear_72_cast_fp16 = linear(bias = encoder_module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("linear_72_cast_fp16")]; tensor var_1710_to_fp16 = const()[name = tensor("op_1710_to_fp16"), val = tensor(0x1p-1)]; tensor var_1711_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1710_to_fp16)[name = tensor("op_1711_cast_fp16")]; tensor input_441_cast_fp16 = add(x = input_429_cast_fp16, y = var_1711_cast_fp16)[name = tensor("input_441_cast_fp16")]; tensor input_443_axes_0 = const()[name = tensor("input_443_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103585984)))]; tensor encoder_module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103587072)))]; tensor input_443_cast_fp16 = layer_norm(axes = input_443_axes_0, beta = encoder_module_layers_7_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_7_norm_out_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("input_443_cast_fp16")]; tensor input_445_axes_0 = const()[name = tensor("input_445_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103588160)))]; tensor encoder_module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103589248)))]; tensor input_445_cast_fp16 = layer_norm(axes = input_445_axes_0, beta = encoder_module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_443_cast_fp16)[name = tensor("input_445_cast_fp16")]; tensor encoder_module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(103590336)))]; tensor encoder_module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105687552)))]; tensor linear_73_cast_fp16 = linear(bias = encoder_module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_445_cast_fp16)[name = tensor("linear_73_cast_fp16")]; tensor input_449_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_449_cast_fp16")]; tensor encoder_module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(105691712)))]; tensor encoder_module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107788928)))]; tensor linear_74_cast_fp16 = linear(bias = encoder_module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_449_cast_fp16)[name = tensor("linear_74_cast_fp16")]; tensor var_1741_to_fp16 = const()[name = tensor("op_1741_to_fp16"), val = tensor(0x1p-1)]; tensor var_1742_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1741_to_fp16)[name = tensor("op_1742_cast_fp16")]; tensor input_455_cast_fp16 = add(x = input_443_cast_fp16, y = var_1742_cast_fp16)[name = tensor("input_455_cast_fp16")]; tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107790016)))]; tensor encoder_module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107791104)))]; tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = encoder_module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_8_norm_self_att_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("query_17_cast_fp16")]; tensor encoder_module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107792192)))]; tensor encoder_module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108316544)))]; tensor linear_75_cast_fp16 = linear(bias = encoder_module_layers_8_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; tensor var_1759 = const()[name = tensor("op_1759"), val = tensor([1, -1, 8, 64])]; tensor q_49_cast_fp16 = reshape(shape = var_1759, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; tensor encoder_module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108317632)))]; tensor encoder_module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108841984)))]; tensor linear_76_cast_fp16 = linear(bias = encoder_module_layers_8_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; tensor var_1764 = const()[name = tensor("op_1764"), val = tensor([1, -1, 8, 64])]; tensor k_33_cast_fp16 = reshape(shape = var_1764, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; tensor encoder_module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108843072)))]; tensor encoder_module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109367424)))]; tensor linear_77_cast_fp16 = linear(bias = encoder_module_layers_8_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; tensor var_1769 = const()[name = tensor("op_1769"), val = tensor([1, -1, 8, 64])]; tensor v_17_cast_fp16 = reshape(shape = var_1769, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109368512)))]; tensor var_1781_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1781_cast_fp16")]; tensor encoder_module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109369600)))]; tensor var_1783_cast_fp16 = add(x = q_49_cast_fp16, y = encoder_module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1783_cast_fp16")]; tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_199_transpose_x_0 = const()[name = tensor("x_199_transpose_x_0"), val = tensor(false)]; tensor x_199_transpose_y_0 = const()[name = tensor("x_199_transpose_y_0"), val = tensor(false)]; tensor var_1785_to_fp16 = const()[name = tensor("op_1785_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109370688)))]; tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1783_cast_fp16)[name = tensor("transpose_147")]; tensor x_199_cast_fp16 = matmul(transpose_x = x_199_transpose_x_0, transpose_y = x_199_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1785_to_fp16)[name = tensor("x_199_cast_fp16")]; tensor x_201_pad_0 = const()[name = tensor("x_201_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_201_mode_0 = const()[name = tensor("x_201_mode_0"), val = tensor("constant")]; tensor const_109_to_fp16 = const()[name = tensor("const_109_to_fp16"), val = tensor(0x0p+0)]; tensor x_201_cast_fp16 = pad(constant_val = const_109_to_fp16, mode = x_201_mode_0, pad = x_201_pad_0, x = x_199_cast_fp16)[name = tensor("x_201_cast_fp16")]; tensor var_1793 = const()[name = tensor("op_1793"), val = tensor([1, 8, -1, 188])]; tensor x_203_cast_fp16 = reshape(shape = var_1793, x = x_201_cast_fp16)[name = tensor("x_203_cast_fp16")]; tensor var_1797_begin_0 = const()[name = tensor("op_1797_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1797_end_0 = const()[name = tensor("op_1797_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1797_end_mask_0 = const()[name = tensor("op_1797_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1797_cast_fp16 = slice_by_index(begin = var_1797_begin_0, end = var_1797_end_0, end_mask = var_1797_end_mask_0, x = x_203_cast_fp16)[name = tensor("op_1797_cast_fp16")]; tensor var_1798 = const()[name = tensor("op_1798"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1798, x = var_1797_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_145")]; tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1781_cast_fp16)[name = tensor("transpose_146")]; tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; tensor var_1807_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1807_cast_fp16")]; tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_33_cast_fp16 = mul(x = var_1807_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; tensor scores_35_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_7)[name = tensor("scores_35_cast_fp16")]; tensor var_1813_cast_fp16 = softmax(axis = var_138, x = scores_35_cast_fp16)[name = tensor("op_1813_cast_fp16")]; tensor input_457_cast_fp16 = select(a = var_154_to_fp16, b = var_1813_cast_fp16, cond = mask_7)[name = tensor("input_457_cast_fp16")]; tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_148")]; tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = input_457_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_205_cast_fp16")]; tensor var_1817_perm_0 = const()[name = tensor("op_1817_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1818 = const()[name = tensor("op_1818"), val = tensor([1, -1, 512])]; tensor var_1817_cast_fp16 = transpose(perm = var_1817_perm_0, x = x_205_cast_fp16)[name = tensor("transpose_144")]; tensor input_459_cast_fp16 = reshape(shape = var_1818, x = var_1817_cast_fp16)[name = tensor("input_459_cast_fp16")]; tensor encoder_module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109754752)))]; tensor encoder_module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110279104)))]; tensor linear_79_cast_fp16 = linear(bias = encoder_module_layers_8_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_459_cast_fp16)[name = tensor("linear_79_cast_fp16")]; tensor input_463_cast_fp16 = add(x = input_455_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_463_cast_fp16")]; tensor x_209_axes_0 = const()[name = tensor("x_209_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110280192)))]; tensor encoder_module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110281280)))]; tensor x_209_cast_fp16 = layer_norm(axes = x_209_axes_0, beta = encoder_module_layers_8_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_8_norm_conv_weight_to_fp16, x = input_463_cast_fp16)[name = tensor("x_209_cast_fp16")]; tensor input_465_perm_0 = const()[name = tensor("input_465_perm_0"), val = tensor([0, 2, 1])]; tensor input_467_pad_type_0 = const()[name = tensor("input_467_pad_type_0"), val = tensor("valid")]; tensor input_467_strides_0 = const()[name = tensor("input_467_strides_0"), val = tensor([1])]; tensor input_467_pad_0 = const()[name = tensor("input_467_pad_0"), val = tensor([0, 0])]; tensor input_467_dilations_0 = const()[name = tensor("input_467_dilations_0"), val = tensor([1])]; tensor input_467_groups_0 = const()[name = tensor("input_467_groups_0"), val = tensor(1)]; tensor encoder_module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110282368)))]; tensor encoder_module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111331008)))]; tensor input_465_cast_fp16 = transpose(perm = input_465_perm_0, x = x_209_cast_fp16)[name = tensor("transpose_143")]; tensor input_467_cast_fp16 = conv(bias = encoder_module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_467_dilations_0, groups = input_467_groups_0, pad = input_467_pad_0, pad_type = input_467_pad_type_0, strides = input_467_strides_0, weight = encoder_module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_465_cast_fp16)[name = tensor("input_467_cast_fp16")]; tensor x_211_split_num_splits_0 = const()[name = tensor("x_211_split_num_splits_0"), val = tensor(2)]; tensor x_211_split_axis_0 = const()[name = tensor("x_211_split_axis_0"), val = tensor(1)]; tensor x_211_split_cast_fp16_0, tensor x_211_split_cast_fp16_1 = split(axis = x_211_split_axis_0, num_splits = x_211_split_num_splits_0, x = input_467_cast_fp16)[name = tensor("x_211_split_cast_fp16")]; tensor x_211_split_1_sigmoid_cast_fp16 = sigmoid(x = x_211_split_cast_fp16_1)[name = tensor("x_211_split_1_sigmoid_cast_fp16")]; tensor x_211_cast_fp16 = mul(x = x_211_split_cast_fp16_0, y = x_211_split_1_sigmoid_cast_fp16)[name = tensor("x_211_cast_fp16")]; tensor input_469_cast_fp16 = select(a = var_154_to_fp16, b = x_211_cast_fp16, cond = var_450)[name = tensor("input_469_cast_fp16")]; tensor input_471_pad_0 = const()[name = tensor("input_471_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_471_mode_0 = const()[name = tensor("input_471_mode_0"), val = tensor("constant")]; tensor const_112_to_fp16 = const()[name = tensor("const_112_to_fp16"), val = tensor(0x0p+0)]; tensor input_471_cast_fp16 = pad(constant_val = const_112_to_fp16, mode = input_471_mode_0, pad = input_471_pad_0, x = input_469_cast_fp16)[name = tensor("input_471_cast_fp16")]; tensor input_473_pad_type_0 = const()[name = tensor("input_473_pad_type_0"), val = tensor("valid")]; tensor input_473_groups_0 = const()[name = tensor("input_473_groups_0"), val = tensor(512)]; tensor input_473_strides_0 = const()[name = tensor("input_473_strides_0"), val = tensor([1])]; tensor input_473_pad_0 = const()[name = tensor("input_473_pad_0"), val = tensor([0, 0])]; tensor input_473_dilations_0 = const()[name = tensor("input_473_dilations_0"), val = tensor([1])]; tensor const_209_to_fp16 = const()[name = tensor("const_209_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111333120)))]; tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111342400)))]; tensor input_475_cast_fp16 = conv(bias = const_210_to_fp16, dilations = input_473_dilations_0, groups = input_473_groups_0, pad = input_473_pad_0, pad_type = input_473_pad_type_0, strides = input_473_strides_0, weight = const_209_to_fp16, x = input_471_cast_fp16)[name = tensor("input_475_cast_fp16")]; tensor input_477_cast_fp16 = silu(x = input_475_cast_fp16)[name = tensor("input_477_cast_fp16")]; tensor x_213_pad_type_0 = const()[name = tensor("x_213_pad_type_0"), val = tensor("valid")]; tensor x_213_strides_0 = const()[name = tensor("x_213_strides_0"), val = tensor([1])]; tensor x_213_pad_0 = const()[name = tensor("x_213_pad_0"), val = tensor([0, 0])]; tensor x_213_dilations_0 = const()[name = tensor("x_213_dilations_0"), val = tensor([1])]; tensor x_213_groups_0 = const()[name = tensor("x_213_groups_0"), val = tensor(1)]; tensor encoder_module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111343488)))]; tensor encoder_module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111867840)))]; tensor x_213_cast_fp16 = conv(bias = encoder_module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_213_dilations_0, groups = x_213_groups_0, pad = x_213_pad_0, pad_type = x_213_pad_type_0, strides = x_213_strides_0, weight = encoder_module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_477_cast_fp16)[name = tensor("x_213_cast_fp16")]; tensor input_479_perm_0 = const()[name = tensor("input_479_perm_0"), val = tensor([0, 2, 1])]; tensor input_479_cast_fp16 = transpose(perm = input_479_perm_0, x = x_213_cast_fp16)[name = tensor("transpose_142")]; tensor input_481_cast_fp16 = add(x = input_463_cast_fp16, y = input_479_cast_fp16)[name = tensor("input_481_cast_fp16")]; tensor input_483_axes_0 = const()[name = tensor("input_483_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111868928)))]; tensor encoder_module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111870016)))]; tensor input_483_cast_fp16 = layer_norm(axes = input_483_axes_0, beta = encoder_module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_481_cast_fp16)[name = tensor("input_483_cast_fp16")]; tensor encoder_module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111871104)))]; tensor encoder_module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113968320)))]; tensor linear_80_cast_fp16 = linear(bias = encoder_module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_483_cast_fp16)[name = tensor("linear_80_cast_fp16")]; tensor input_487_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_487_cast_fp16")]; tensor encoder_module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113972480)))]; tensor encoder_module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116069696)))]; tensor linear_81_cast_fp16 = linear(bias = encoder_module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("linear_81_cast_fp16")]; tensor var_1884_to_fp16 = const()[name = tensor("op_1884_to_fp16"), val = tensor(0x1p-1)]; tensor var_1885_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1884_to_fp16)[name = tensor("op_1885_cast_fp16")]; tensor input_493_cast_fp16 = add(x = input_481_cast_fp16, y = var_1885_cast_fp16)[name = tensor("input_493_cast_fp16")]; tensor input_495_axes_0 = const()[name = tensor("input_495_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116070784)))]; tensor encoder_module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116071872)))]; tensor input_495_cast_fp16 = layer_norm(axes = input_495_axes_0, beta = encoder_module_layers_8_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_8_norm_out_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("input_495_cast_fp16")]; tensor input_497_axes_0 = const()[name = tensor("input_497_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116072960)))]; tensor encoder_module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116074048)))]; tensor input_497_cast_fp16 = layer_norm(axes = input_497_axes_0, beta = encoder_module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_495_cast_fp16)[name = tensor("input_497_cast_fp16")]; tensor encoder_module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(116075136)))]; tensor encoder_module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118172352)))]; tensor linear_82_cast_fp16 = linear(bias = encoder_module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_497_cast_fp16)[name = tensor("linear_82_cast_fp16")]; tensor input_501_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_501_cast_fp16")]; tensor encoder_module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(118176512)))]; tensor encoder_module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120273728)))]; tensor linear_83_cast_fp16 = linear(bias = encoder_module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_501_cast_fp16)[name = tensor("linear_83_cast_fp16")]; tensor var_1915_to_fp16 = const()[name = tensor("op_1915_to_fp16"), val = tensor(0x1p-1)]; tensor var_1916_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1915_to_fp16)[name = tensor("op_1916_cast_fp16")]; tensor input_507_cast_fp16 = add(x = input_495_cast_fp16, y = var_1916_cast_fp16)[name = tensor("input_507_cast_fp16")]; tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120274816)))]; tensor encoder_module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120275904)))]; tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = encoder_module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_9_norm_self_att_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("query_19_cast_fp16")]; tensor encoder_module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120276992)))]; tensor encoder_module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120801344)))]; tensor linear_84_cast_fp16 = linear(bias = encoder_module_layers_9_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; tensor var_1933 = const()[name = tensor("op_1933"), val = tensor([1, -1, 8, 64])]; tensor q_55_cast_fp16 = reshape(shape = var_1933, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; tensor encoder_module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120802432)))]; tensor encoder_module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121326784)))]; tensor linear_85_cast_fp16 = linear(bias = encoder_module_layers_9_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; tensor var_1938 = const()[name = tensor("op_1938"), val = tensor([1, -1, 8, 64])]; tensor k_37_cast_fp16 = reshape(shape = var_1938, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; tensor encoder_module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121327872)))]; tensor encoder_module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121852224)))]; tensor linear_86_cast_fp16 = linear(bias = encoder_module_layers_9_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; tensor var_1943 = const()[name = tensor("op_1943"), val = tensor([1, -1, 8, 64])]; tensor v_19_cast_fp16 = reshape(shape = var_1943, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121853312)))]; tensor var_1955_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1955_cast_fp16")]; tensor encoder_module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121854400)))]; tensor var_1957_cast_fp16 = add(x = q_55_cast_fp16, y = encoder_module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1957_cast_fp16")]; tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_221_transpose_x_0 = const()[name = tensor("x_221_transpose_x_0"), val = tensor(false)]; tensor x_221_transpose_y_0 = const()[name = tensor("x_221_transpose_y_0"), val = tensor(false)]; tensor var_1959_to_fp16 = const()[name = tensor("op_1959_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121855488)))]; tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1957_cast_fp16)[name = tensor("transpose_140")]; tensor x_221_cast_fp16 = matmul(transpose_x = x_221_transpose_x_0, transpose_y = x_221_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1959_to_fp16)[name = tensor("x_221_cast_fp16")]; tensor x_223_pad_0 = const()[name = tensor("x_223_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_223_mode_0 = const()[name = tensor("x_223_mode_0"), val = tensor("constant")]; tensor const_119_to_fp16 = const()[name = tensor("const_119_to_fp16"), val = tensor(0x0p+0)]; tensor x_223_cast_fp16 = pad(constant_val = const_119_to_fp16, mode = x_223_mode_0, pad = x_223_pad_0, x = x_221_cast_fp16)[name = tensor("x_223_cast_fp16")]; tensor var_1967 = const()[name = tensor("op_1967"), val = tensor([1, 8, -1, 188])]; tensor x_225_cast_fp16 = reshape(shape = var_1967, x = x_223_cast_fp16)[name = tensor("x_225_cast_fp16")]; tensor var_1971_begin_0 = const()[name = tensor("op_1971_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_1971_end_0 = const()[name = tensor("op_1971_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_1971_end_mask_0 = const()[name = tensor("op_1971_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_1971_cast_fp16 = slice_by_index(begin = var_1971_begin_0, end = var_1971_end_0, end_mask = var_1971_end_mask_0, x = x_225_cast_fp16)[name = tensor("op_1971_cast_fp16")]; tensor var_1972 = const()[name = tensor("op_1972"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1972, x = var_1971_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_138")]; tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_1955_cast_fp16)[name = tensor("transpose_139")]; tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; tensor var_1981_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1981_cast_fp16")]; tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_37_cast_fp16 = mul(x = var_1981_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; tensor scores_39_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_7)[name = tensor("scores_39_cast_fp16")]; tensor var_1987_cast_fp16 = softmax(axis = var_138, x = scores_39_cast_fp16)[name = tensor("op_1987_cast_fp16")]; tensor input_509_cast_fp16 = select(a = var_154_to_fp16, b = var_1987_cast_fp16, cond = mask_7)[name = tensor("input_509_cast_fp16")]; tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_141")]; tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = input_509_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_227_cast_fp16")]; tensor var_1991_perm_0 = const()[name = tensor("op_1991_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_1992 = const()[name = tensor("op_1992"), val = tensor([1, -1, 512])]; tensor var_1991_cast_fp16 = transpose(perm = var_1991_perm_0, x = x_227_cast_fp16)[name = tensor("transpose_137")]; tensor input_511_cast_fp16 = reshape(shape = var_1992, x = var_1991_cast_fp16)[name = tensor("input_511_cast_fp16")]; tensor encoder_module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122239552)))]; tensor encoder_module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122763904)))]; tensor linear_88_cast_fp16 = linear(bias = encoder_module_layers_9_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_511_cast_fp16)[name = tensor("linear_88_cast_fp16")]; tensor input_515_cast_fp16 = add(x = input_507_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_515_cast_fp16")]; tensor x_231_axes_0 = const()[name = tensor("x_231_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122764992)))]; tensor encoder_module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122766080)))]; tensor x_231_cast_fp16 = layer_norm(axes = x_231_axes_0, beta = encoder_module_layers_9_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_9_norm_conv_weight_to_fp16, x = input_515_cast_fp16)[name = tensor("x_231_cast_fp16")]; tensor input_517_perm_0 = const()[name = tensor("input_517_perm_0"), val = tensor([0, 2, 1])]; tensor input_519_pad_type_0 = const()[name = tensor("input_519_pad_type_0"), val = tensor("valid")]; tensor input_519_strides_0 = const()[name = tensor("input_519_strides_0"), val = tensor([1])]; tensor input_519_pad_0 = const()[name = tensor("input_519_pad_0"), val = tensor([0, 0])]; tensor input_519_dilations_0 = const()[name = tensor("input_519_dilations_0"), val = tensor([1])]; tensor input_519_groups_0 = const()[name = tensor("input_519_groups_0"), val = tensor(1)]; tensor encoder_module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(122767168)))]; tensor encoder_module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123815808)))]; tensor input_517_cast_fp16 = transpose(perm = input_517_perm_0, x = x_231_cast_fp16)[name = tensor("transpose_136")]; tensor input_519_cast_fp16 = conv(bias = encoder_module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_519_dilations_0, groups = input_519_groups_0, pad = input_519_pad_0, pad_type = input_519_pad_type_0, strides = input_519_strides_0, weight = encoder_module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_517_cast_fp16)[name = tensor("input_519_cast_fp16")]; tensor x_233_split_num_splits_0 = const()[name = tensor("x_233_split_num_splits_0"), val = tensor(2)]; tensor x_233_split_axis_0 = const()[name = tensor("x_233_split_axis_0"), val = tensor(1)]; tensor x_233_split_cast_fp16_0, tensor x_233_split_cast_fp16_1 = split(axis = x_233_split_axis_0, num_splits = x_233_split_num_splits_0, x = input_519_cast_fp16)[name = tensor("x_233_split_cast_fp16")]; tensor x_233_split_1_sigmoid_cast_fp16 = sigmoid(x = x_233_split_cast_fp16_1)[name = tensor("x_233_split_1_sigmoid_cast_fp16")]; tensor x_233_cast_fp16 = mul(x = x_233_split_cast_fp16_0, y = x_233_split_1_sigmoid_cast_fp16)[name = tensor("x_233_cast_fp16")]; tensor input_521_cast_fp16 = select(a = var_154_to_fp16, b = x_233_cast_fp16, cond = var_450)[name = tensor("input_521_cast_fp16")]; tensor input_523_pad_0 = const()[name = tensor("input_523_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_523_mode_0 = const()[name = tensor("input_523_mode_0"), val = tensor("constant")]; tensor const_122_to_fp16 = const()[name = tensor("const_122_to_fp16"), val = tensor(0x0p+0)]; tensor input_523_cast_fp16 = pad(constant_val = const_122_to_fp16, mode = input_523_mode_0, pad = input_523_pad_0, x = input_521_cast_fp16)[name = tensor("input_523_cast_fp16")]; tensor input_525_pad_type_0 = const()[name = tensor("input_525_pad_type_0"), val = tensor("valid")]; tensor input_525_groups_0 = const()[name = tensor("input_525_groups_0"), val = tensor(512)]; tensor input_525_strides_0 = const()[name = tensor("input_525_strides_0"), val = tensor([1])]; tensor input_525_pad_0 = const()[name = tensor("input_525_pad_0"), val = tensor([0, 0])]; tensor input_525_dilations_0 = const()[name = tensor("input_525_dilations_0"), val = tensor([1])]; tensor const_211_to_fp16 = const()[name = tensor("const_211_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123817920)))]; tensor const_212_to_fp16 = const()[name = tensor("const_212_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123827200)))]; tensor input_527_cast_fp16 = conv(bias = const_212_to_fp16, dilations = input_525_dilations_0, groups = input_525_groups_0, pad = input_525_pad_0, pad_type = input_525_pad_type_0, strides = input_525_strides_0, weight = const_211_to_fp16, x = input_523_cast_fp16)[name = tensor("input_527_cast_fp16")]; tensor input_529_cast_fp16 = silu(x = input_527_cast_fp16)[name = tensor("input_529_cast_fp16")]; tensor x_235_pad_type_0 = const()[name = tensor("x_235_pad_type_0"), val = tensor("valid")]; tensor x_235_strides_0 = const()[name = tensor("x_235_strides_0"), val = tensor([1])]; tensor x_235_pad_0 = const()[name = tensor("x_235_pad_0"), val = tensor([0, 0])]; tensor x_235_dilations_0 = const()[name = tensor("x_235_dilations_0"), val = tensor([1])]; tensor x_235_groups_0 = const()[name = tensor("x_235_groups_0"), val = tensor(1)]; tensor encoder_module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123828288)))]; tensor encoder_module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(124352640)))]; tensor x_235_cast_fp16 = conv(bias = encoder_module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_235_dilations_0, groups = x_235_groups_0, pad = x_235_pad_0, pad_type = x_235_pad_type_0, strides = x_235_strides_0, weight = encoder_module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_529_cast_fp16)[name = tensor("x_235_cast_fp16")]; tensor input_531_perm_0 = const()[name = tensor("input_531_perm_0"), val = tensor([0, 2, 1])]; tensor input_531_cast_fp16 = transpose(perm = input_531_perm_0, x = x_235_cast_fp16)[name = tensor("transpose_135")]; tensor input_533_cast_fp16 = add(x = input_515_cast_fp16, y = input_531_cast_fp16)[name = tensor("input_533_cast_fp16")]; tensor input_535_axes_0 = const()[name = tensor("input_535_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(124353728)))]; tensor encoder_module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(124354816)))]; tensor input_535_cast_fp16 = layer_norm(axes = input_535_axes_0, beta = encoder_module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_533_cast_fp16)[name = tensor("input_535_cast_fp16")]; tensor encoder_module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(124355904)))]; tensor encoder_module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126453120)))]; tensor linear_89_cast_fp16 = linear(bias = encoder_module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_535_cast_fp16)[name = tensor("linear_89_cast_fp16")]; tensor input_539_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_539_cast_fp16")]; tensor encoder_module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(126457280)))]; tensor encoder_module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128554496)))]; tensor linear_90_cast_fp16 = linear(bias = encoder_module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("linear_90_cast_fp16")]; tensor var_2058_to_fp16 = const()[name = tensor("op_2058_to_fp16"), val = tensor(0x1p-1)]; tensor var_2059_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2058_to_fp16)[name = tensor("op_2059_cast_fp16")]; tensor input_545_cast_fp16 = add(x = input_533_cast_fp16, y = var_2059_cast_fp16)[name = tensor("input_545_cast_fp16")]; tensor input_547_axes_0 = const()[name = tensor("input_547_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128555584)))]; tensor encoder_module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128556672)))]; tensor input_547_cast_fp16 = layer_norm(axes = input_547_axes_0, beta = encoder_module_layers_9_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_9_norm_out_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("input_547_cast_fp16")]; tensor input_549_axes_0 = const()[name = tensor("input_549_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128557760)))]; tensor encoder_module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128558848)))]; tensor input_549_cast_fp16 = layer_norm(axes = input_549_axes_0, beta = encoder_module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_547_cast_fp16)[name = tensor("input_549_cast_fp16")]; tensor encoder_module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(128559936)))]; tensor encoder_module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130657152)))]; tensor linear_91_cast_fp16 = linear(bias = encoder_module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_549_cast_fp16)[name = tensor("linear_91_cast_fp16")]; tensor input_553_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_553_cast_fp16")]; tensor encoder_module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(130661312)))]; tensor encoder_module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132758528)))]; tensor linear_92_cast_fp16 = linear(bias = encoder_module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_553_cast_fp16)[name = tensor("linear_92_cast_fp16")]; tensor var_2089_to_fp16 = const()[name = tensor("op_2089_to_fp16"), val = tensor(0x1p-1)]; tensor var_2090_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2089_to_fp16)[name = tensor("op_2090_cast_fp16")]; tensor input_559_cast_fp16 = add(x = input_547_cast_fp16, y = var_2090_cast_fp16)[name = tensor("input_559_cast_fp16")]; tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132759616)))]; tensor encoder_module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132760704)))]; tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = encoder_module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_10_norm_self_att_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("query_21_cast_fp16")]; tensor encoder_module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132761792)))]; tensor encoder_module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133286144)))]; tensor linear_93_cast_fp16 = linear(bias = encoder_module_layers_10_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; tensor var_2107 = const()[name = tensor("op_2107"), val = tensor([1, -1, 8, 64])]; tensor q_61_cast_fp16 = reshape(shape = var_2107, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; tensor encoder_module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133287232)))]; tensor encoder_module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133811584)))]; tensor linear_94_cast_fp16 = linear(bias = encoder_module_layers_10_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; tensor var_2112 = const()[name = tensor("op_2112"), val = tensor([1, -1, 8, 64])]; tensor k_41_cast_fp16 = reshape(shape = var_2112, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; tensor encoder_module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133812672)))]; tensor encoder_module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134337024)))]; tensor linear_95_cast_fp16 = linear(bias = encoder_module_layers_10_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; tensor var_2117 = const()[name = tensor("op_2117"), val = tensor([1, -1, 8, 64])]; tensor v_21_cast_fp16 = reshape(shape = var_2117, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134338112)))]; tensor var_2129_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2129_cast_fp16")]; tensor encoder_module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134339200)))]; tensor var_2131_cast_fp16 = add(x = q_61_cast_fp16, y = encoder_module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2131_cast_fp16")]; tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_243_transpose_x_0 = const()[name = tensor("x_243_transpose_x_0"), val = tensor(false)]; tensor x_243_transpose_y_0 = const()[name = tensor("x_243_transpose_y_0"), val = tensor(false)]; tensor var_2133_to_fp16 = const()[name = tensor("op_2133_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134340288)))]; tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2131_cast_fp16)[name = tensor("transpose_133")]; tensor x_243_cast_fp16 = matmul(transpose_x = x_243_transpose_x_0, transpose_y = x_243_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2133_to_fp16)[name = tensor("x_243_cast_fp16")]; tensor x_245_pad_0 = const()[name = tensor("x_245_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_245_mode_0 = const()[name = tensor("x_245_mode_0"), val = tensor("constant")]; tensor const_129_to_fp16 = const()[name = tensor("const_129_to_fp16"), val = tensor(0x0p+0)]; tensor x_245_cast_fp16 = pad(constant_val = const_129_to_fp16, mode = x_245_mode_0, pad = x_245_pad_0, x = x_243_cast_fp16)[name = tensor("x_245_cast_fp16")]; tensor var_2141 = const()[name = tensor("op_2141"), val = tensor([1, 8, -1, 188])]; tensor x_247_cast_fp16 = reshape(shape = var_2141, x = x_245_cast_fp16)[name = tensor("x_247_cast_fp16")]; tensor var_2145_begin_0 = const()[name = tensor("op_2145_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2145_end_0 = const()[name = tensor("op_2145_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2145_end_mask_0 = const()[name = tensor("op_2145_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2145_cast_fp16 = slice_by_index(begin = var_2145_begin_0, end = var_2145_end_0, end_mask = var_2145_end_mask_0, x = x_247_cast_fp16)[name = tensor("op_2145_cast_fp16")]; tensor var_2146 = const()[name = tensor("op_2146"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2146, x = var_2145_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_131")]; tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2129_cast_fp16)[name = tensor("transpose_132")]; tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; tensor var_2155_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2155_cast_fp16")]; tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_41_cast_fp16 = mul(x = var_2155_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; tensor scores_43_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_7)[name = tensor("scores_43_cast_fp16")]; tensor var_2161_cast_fp16 = softmax(axis = var_138, x = scores_43_cast_fp16)[name = tensor("op_2161_cast_fp16")]; tensor input_561_cast_fp16 = select(a = var_154_to_fp16, b = var_2161_cast_fp16, cond = mask_7)[name = tensor("input_561_cast_fp16")]; tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_134")]; tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = input_561_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_249_cast_fp16")]; tensor var_2165_perm_0 = const()[name = tensor("op_2165_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2166 = const()[name = tensor("op_2166"), val = tensor([1, -1, 512])]; tensor var_2165_cast_fp16 = transpose(perm = var_2165_perm_0, x = x_249_cast_fp16)[name = tensor("transpose_130")]; tensor input_563_cast_fp16 = reshape(shape = var_2166, x = var_2165_cast_fp16)[name = tensor("input_563_cast_fp16")]; tensor encoder_module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134724352)))]; tensor encoder_module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135248704)))]; tensor linear_97_cast_fp16 = linear(bias = encoder_module_layers_10_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_563_cast_fp16)[name = tensor("linear_97_cast_fp16")]; tensor input_567_cast_fp16 = add(x = input_559_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_567_cast_fp16")]; tensor x_253_axes_0 = const()[name = tensor("x_253_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135249792)))]; tensor encoder_module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135250880)))]; tensor x_253_cast_fp16 = layer_norm(axes = x_253_axes_0, beta = encoder_module_layers_10_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_10_norm_conv_weight_to_fp16, x = input_567_cast_fp16)[name = tensor("x_253_cast_fp16")]; tensor input_569_perm_0 = const()[name = tensor("input_569_perm_0"), val = tensor([0, 2, 1])]; tensor input_571_pad_type_0 = const()[name = tensor("input_571_pad_type_0"), val = tensor("valid")]; tensor input_571_strides_0 = const()[name = tensor("input_571_strides_0"), val = tensor([1])]; tensor input_571_pad_0 = const()[name = tensor("input_571_pad_0"), val = tensor([0, 0])]; tensor input_571_dilations_0 = const()[name = tensor("input_571_dilations_0"), val = tensor([1])]; tensor input_571_groups_0 = const()[name = tensor("input_571_groups_0"), val = tensor(1)]; tensor encoder_module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135251968)))]; tensor encoder_module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136300608)))]; tensor input_569_cast_fp16 = transpose(perm = input_569_perm_0, x = x_253_cast_fp16)[name = tensor("transpose_129")]; tensor input_571_cast_fp16 = conv(bias = encoder_module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_571_dilations_0, groups = input_571_groups_0, pad = input_571_pad_0, pad_type = input_571_pad_type_0, strides = input_571_strides_0, weight = encoder_module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_569_cast_fp16)[name = tensor("input_571_cast_fp16")]; tensor x_255_split_num_splits_0 = const()[name = tensor("x_255_split_num_splits_0"), val = tensor(2)]; tensor x_255_split_axis_0 = const()[name = tensor("x_255_split_axis_0"), val = tensor(1)]; tensor x_255_split_cast_fp16_0, tensor x_255_split_cast_fp16_1 = split(axis = x_255_split_axis_0, num_splits = x_255_split_num_splits_0, x = input_571_cast_fp16)[name = tensor("x_255_split_cast_fp16")]; tensor x_255_split_1_sigmoid_cast_fp16 = sigmoid(x = x_255_split_cast_fp16_1)[name = tensor("x_255_split_1_sigmoid_cast_fp16")]; tensor x_255_cast_fp16 = mul(x = x_255_split_cast_fp16_0, y = x_255_split_1_sigmoid_cast_fp16)[name = tensor("x_255_cast_fp16")]; tensor input_573_cast_fp16 = select(a = var_154_to_fp16, b = x_255_cast_fp16, cond = var_450)[name = tensor("input_573_cast_fp16")]; tensor input_575_pad_0 = const()[name = tensor("input_575_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_575_mode_0 = const()[name = tensor("input_575_mode_0"), val = tensor("constant")]; tensor const_132_to_fp16 = const()[name = tensor("const_132_to_fp16"), val = tensor(0x0p+0)]; tensor input_575_cast_fp16 = pad(constant_val = const_132_to_fp16, mode = input_575_mode_0, pad = input_575_pad_0, x = input_573_cast_fp16)[name = tensor("input_575_cast_fp16")]; tensor input_577_pad_type_0 = const()[name = tensor("input_577_pad_type_0"), val = tensor("valid")]; tensor input_577_groups_0 = const()[name = tensor("input_577_groups_0"), val = tensor(512)]; tensor input_577_strides_0 = const()[name = tensor("input_577_strides_0"), val = tensor([1])]; tensor input_577_pad_0 = const()[name = tensor("input_577_pad_0"), val = tensor([0, 0])]; tensor input_577_dilations_0 = const()[name = tensor("input_577_dilations_0"), val = tensor([1])]; tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136302720)))]; tensor const_214_to_fp16 = const()[name = tensor("const_214_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136312000)))]; tensor input_579_cast_fp16 = conv(bias = const_214_to_fp16, dilations = input_577_dilations_0, groups = input_577_groups_0, pad = input_577_pad_0, pad_type = input_577_pad_type_0, strides = input_577_strides_0, weight = const_213_to_fp16, x = input_575_cast_fp16)[name = tensor("input_579_cast_fp16")]; tensor input_581_cast_fp16 = silu(x = input_579_cast_fp16)[name = tensor("input_581_cast_fp16")]; tensor x_257_pad_type_0 = const()[name = tensor("x_257_pad_type_0"), val = tensor("valid")]; tensor x_257_strides_0 = const()[name = tensor("x_257_strides_0"), val = tensor([1])]; tensor x_257_pad_0 = const()[name = tensor("x_257_pad_0"), val = tensor([0, 0])]; tensor x_257_dilations_0 = const()[name = tensor("x_257_dilations_0"), val = tensor([1])]; tensor x_257_groups_0 = const()[name = tensor("x_257_groups_0"), val = tensor(1)]; tensor encoder_module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136313088)))]; tensor encoder_module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136837440)))]; tensor x_257_cast_fp16 = conv(bias = encoder_module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_257_dilations_0, groups = x_257_groups_0, pad = x_257_pad_0, pad_type = x_257_pad_type_0, strides = x_257_strides_0, weight = encoder_module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_581_cast_fp16)[name = tensor("x_257_cast_fp16")]; tensor input_583_perm_0 = const()[name = tensor("input_583_perm_0"), val = tensor([0, 2, 1])]; tensor input_583_cast_fp16 = transpose(perm = input_583_perm_0, x = x_257_cast_fp16)[name = tensor("transpose_128")]; tensor input_585_cast_fp16 = add(x = input_567_cast_fp16, y = input_583_cast_fp16)[name = tensor("input_585_cast_fp16")]; tensor input_587_axes_0 = const()[name = tensor("input_587_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136838528)))]; tensor encoder_module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136839616)))]; tensor input_587_cast_fp16 = layer_norm(axes = input_587_axes_0, beta = encoder_module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_585_cast_fp16)[name = tensor("input_587_cast_fp16")]; tensor encoder_module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136840704)))]; tensor encoder_module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138937920)))]; tensor linear_98_cast_fp16 = linear(bias = encoder_module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_587_cast_fp16)[name = tensor("linear_98_cast_fp16")]; tensor input_591_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_591_cast_fp16")]; tensor encoder_module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138942080)))]; tensor encoder_module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141039296)))]; tensor linear_99_cast_fp16 = linear(bias = encoder_module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("linear_99_cast_fp16")]; tensor var_2232_to_fp16 = const()[name = tensor("op_2232_to_fp16"), val = tensor(0x1p-1)]; tensor var_2233_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2232_to_fp16)[name = tensor("op_2233_cast_fp16")]; tensor input_597_cast_fp16 = add(x = input_585_cast_fp16, y = var_2233_cast_fp16)[name = tensor("input_597_cast_fp16")]; tensor input_599_axes_0 = const()[name = tensor("input_599_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141040384)))]; tensor encoder_module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141041472)))]; tensor input_599_cast_fp16 = layer_norm(axes = input_599_axes_0, beta = encoder_module_layers_10_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_10_norm_out_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("input_599_cast_fp16")]; tensor input_601_axes_0 = const()[name = tensor("input_601_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141042560)))]; tensor encoder_module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141043648)))]; tensor input_601_cast_fp16 = layer_norm(axes = input_601_axes_0, beta = encoder_module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_599_cast_fp16)[name = tensor("input_601_cast_fp16")]; tensor encoder_module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141044736)))]; tensor encoder_module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143141952)))]; tensor linear_100_cast_fp16 = linear(bias = encoder_module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_601_cast_fp16)[name = tensor("linear_100_cast_fp16")]; tensor input_605_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_605_cast_fp16")]; tensor encoder_module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(143146112)))]; tensor encoder_module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145243328)))]; tensor linear_101_cast_fp16 = linear(bias = encoder_module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_605_cast_fp16)[name = tensor("linear_101_cast_fp16")]; tensor var_2263_to_fp16 = const()[name = tensor("op_2263_to_fp16"), val = tensor(0x1p-1)]; tensor var_2264_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2263_to_fp16)[name = tensor("op_2264_cast_fp16")]; tensor input_611_cast_fp16 = add(x = input_599_cast_fp16, y = var_2264_cast_fp16)[name = tensor("input_611_cast_fp16")]; tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145244416)))]; tensor encoder_module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145245504)))]; tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = encoder_module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_11_norm_self_att_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("query_23_cast_fp16")]; tensor encoder_module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145246592)))]; tensor encoder_module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145770944)))]; tensor linear_102_cast_fp16 = linear(bias = encoder_module_layers_11_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; tensor var_2281 = const()[name = tensor("op_2281"), val = tensor([1, -1, 8, 64])]; tensor q_67_cast_fp16 = reshape(shape = var_2281, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; tensor encoder_module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145772032)))]; tensor encoder_module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146296384)))]; tensor linear_103_cast_fp16 = linear(bias = encoder_module_layers_11_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; tensor var_2286 = const()[name = tensor("op_2286"), val = tensor([1, -1, 8, 64])]; tensor k_45_cast_fp16 = reshape(shape = var_2286, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; tensor encoder_module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146297472)))]; tensor encoder_module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146821824)))]; tensor linear_104_cast_fp16 = linear(bias = encoder_module_layers_11_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; tensor var_2291 = const()[name = tensor("op_2291"), val = tensor([1, -1, 8, 64])]; tensor v_23_cast_fp16 = reshape(shape = var_2291, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146822912)))]; tensor var_2303_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2303_cast_fp16")]; tensor encoder_module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146824000)))]; tensor var_2305_cast_fp16 = add(x = q_67_cast_fp16, y = encoder_module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2305_cast_fp16")]; tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_265_transpose_x_0 = const()[name = tensor("x_265_transpose_x_0"), val = tensor(false)]; tensor x_265_transpose_y_0 = const()[name = tensor("x_265_transpose_y_0"), val = tensor(false)]; tensor var_2307_to_fp16 = const()[name = tensor("op_2307_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146825088)))]; tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2305_cast_fp16)[name = tensor("transpose_126")]; tensor x_265_cast_fp16 = matmul(transpose_x = x_265_transpose_x_0, transpose_y = x_265_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2307_to_fp16)[name = tensor("x_265_cast_fp16")]; tensor x_267_pad_0 = const()[name = tensor("x_267_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_267_mode_0 = const()[name = tensor("x_267_mode_0"), val = tensor("constant")]; tensor const_139_to_fp16 = const()[name = tensor("const_139_to_fp16"), val = tensor(0x0p+0)]; tensor x_267_cast_fp16 = pad(constant_val = const_139_to_fp16, mode = x_267_mode_0, pad = x_267_pad_0, x = x_265_cast_fp16)[name = tensor("x_267_cast_fp16")]; tensor var_2315 = const()[name = tensor("op_2315"), val = tensor([1, 8, -1, 188])]; tensor x_269_cast_fp16 = reshape(shape = var_2315, x = x_267_cast_fp16)[name = tensor("x_269_cast_fp16")]; tensor var_2319_begin_0 = const()[name = tensor("op_2319_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2319_end_0 = const()[name = tensor("op_2319_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2319_end_mask_0 = const()[name = tensor("op_2319_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2319_cast_fp16 = slice_by_index(begin = var_2319_begin_0, end = var_2319_end_0, end_mask = var_2319_end_mask_0, x = x_269_cast_fp16)[name = tensor("op_2319_cast_fp16")]; tensor var_2320 = const()[name = tensor("op_2320"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2320, x = var_2319_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_124")]; tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2303_cast_fp16)[name = tensor("transpose_125")]; tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; tensor var_2329_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2329_cast_fp16")]; tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_45_cast_fp16 = mul(x = var_2329_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; tensor scores_47_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_7)[name = tensor("scores_47_cast_fp16")]; tensor var_2335_cast_fp16 = softmax(axis = var_138, x = scores_47_cast_fp16)[name = tensor("op_2335_cast_fp16")]; tensor input_613_cast_fp16 = select(a = var_154_to_fp16, b = var_2335_cast_fp16, cond = mask_7)[name = tensor("input_613_cast_fp16")]; tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_127")]; tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = input_613_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_271_cast_fp16")]; tensor var_2339_perm_0 = const()[name = tensor("op_2339_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2340 = const()[name = tensor("op_2340"), val = tensor([1, -1, 512])]; tensor var_2339_cast_fp16 = transpose(perm = var_2339_perm_0, x = x_271_cast_fp16)[name = tensor("transpose_123")]; tensor input_615_cast_fp16 = reshape(shape = var_2340, x = var_2339_cast_fp16)[name = tensor("input_615_cast_fp16")]; tensor encoder_module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147209152)))]; tensor encoder_module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147733504)))]; tensor linear_106_cast_fp16 = linear(bias = encoder_module_layers_11_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_615_cast_fp16)[name = tensor("linear_106_cast_fp16")]; tensor input_619_cast_fp16 = add(x = input_611_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_619_cast_fp16")]; tensor x_275_axes_0 = const()[name = tensor("x_275_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147734592)))]; tensor encoder_module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147735680)))]; tensor x_275_cast_fp16 = layer_norm(axes = x_275_axes_0, beta = encoder_module_layers_11_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_11_norm_conv_weight_to_fp16, x = input_619_cast_fp16)[name = tensor("x_275_cast_fp16")]; tensor input_621_perm_0 = const()[name = tensor("input_621_perm_0"), val = tensor([0, 2, 1])]; tensor input_623_pad_type_0 = const()[name = tensor("input_623_pad_type_0"), val = tensor("valid")]; tensor input_623_strides_0 = const()[name = tensor("input_623_strides_0"), val = tensor([1])]; tensor input_623_pad_0 = const()[name = tensor("input_623_pad_0"), val = tensor([0, 0])]; tensor input_623_dilations_0 = const()[name = tensor("input_623_dilations_0"), val = tensor([1])]; tensor input_623_groups_0 = const()[name = tensor("input_623_groups_0"), val = tensor(1)]; tensor encoder_module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147736768)))]; tensor encoder_module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148785408)))]; tensor input_621_cast_fp16 = transpose(perm = input_621_perm_0, x = x_275_cast_fp16)[name = tensor("transpose_122")]; tensor input_623_cast_fp16 = conv(bias = encoder_module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_623_dilations_0, groups = input_623_groups_0, pad = input_623_pad_0, pad_type = input_623_pad_type_0, strides = input_623_strides_0, weight = encoder_module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_621_cast_fp16)[name = tensor("input_623_cast_fp16")]; tensor x_277_split_num_splits_0 = const()[name = tensor("x_277_split_num_splits_0"), val = tensor(2)]; tensor x_277_split_axis_0 = const()[name = tensor("x_277_split_axis_0"), val = tensor(1)]; tensor x_277_split_cast_fp16_0, tensor x_277_split_cast_fp16_1 = split(axis = x_277_split_axis_0, num_splits = x_277_split_num_splits_0, x = input_623_cast_fp16)[name = tensor("x_277_split_cast_fp16")]; tensor x_277_split_1_sigmoid_cast_fp16 = sigmoid(x = x_277_split_cast_fp16_1)[name = tensor("x_277_split_1_sigmoid_cast_fp16")]; tensor x_277_cast_fp16 = mul(x = x_277_split_cast_fp16_0, y = x_277_split_1_sigmoid_cast_fp16)[name = tensor("x_277_cast_fp16")]; tensor input_625_cast_fp16 = select(a = var_154_to_fp16, b = x_277_cast_fp16, cond = var_450)[name = tensor("input_625_cast_fp16")]; tensor input_627_pad_0 = const()[name = tensor("input_627_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_627_mode_0 = const()[name = tensor("input_627_mode_0"), val = tensor("constant")]; tensor const_142_to_fp16 = const()[name = tensor("const_142_to_fp16"), val = tensor(0x0p+0)]; tensor input_627_cast_fp16 = pad(constant_val = const_142_to_fp16, mode = input_627_mode_0, pad = input_627_pad_0, x = input_625_cast_fp16)[name = tensor("input_627_cast_fp16")]; tensor input_629_pad_type_0 = const()[name = tensor("input_629_pad_type_0"), val = tensor("valid")]; tensor input_629_groups_0 = const()[name = tensor("input_629_groups_0"), val = tensor(512)]; tensor input_629_strides_0 = const()[name = tensor("input_629_strides_0"), val = tensor([1])]; tensor input_629_pad_0 = const()[name = tensor("input_629_pad_0"), val = tensor([0, 0])]; tensor input_629_dilations_0 = const()[name = tensor("input_629_dilations_0"), val = tensor([1])]; tensor const_215_to_fp16 = const()[name = tensor("const_215_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148787520)))]; tensor const_216_to_fp16 = const()[name = tensor("const_216_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148796800)))]; tensor input_631_cast_fp16 = conv(bias = const_216_to_fp16, dilations = input_629_dilations_0, groups = input_629_groups_0, pad = input_629_pad_0, pad_type = input_629_pad_type_0, strides = input_629_strides_0, weight = const_215_to_fp16, x = input_627_cast_fp16)[name = tensor("input_631_cast_fp16")]; tensor input_633_cast_fp16 = silu(x = input_631_cast_fp16)[name = tensor("input_633_cast_fp16")]; tensor x_279_pad_type_0 = const()[name = tensor("x_279_pad_type_0"), val = tensor("valid")]; tensor x_279_strides_0 = const()[name = tensor("x_279_strides_0"), val = tensor([1])]; tensor x_279_pad_0 = const()[name = tensor("x_279_pad_0"), val = tensor([0, 0])]; tensor x_279_dilations_0 = const()[name = tensor("x_279_dilations_0"), val = tensor([1])]; tensor x_279_groups_0 = const()[name = tensor("x_279_groups_0"), val = tensor(1)]; tensor encoder_module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148797888)))]; tensor encoder_module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149322240)))]; tensor x_279_cast_fp16 = conv(bias = encoder_module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_279_dilations_0, groups = x_279_groups_0, pad = x_279_pad_0, pad_type = x_279_pad_type_0, strides = x_279_strides_0, weight = encoder_module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_633_cast_fp16)[name = tensor("x_279_cast_fp16")]; tensor input_635_perm_0 = const()[name = tensor("input_635_perm_0"), val = tensor([0, 2, 1])]; tensor input_635_cast_fp16 = transpose(perm = input_635_perm_0, x = x_279_cast_fp16)[name = tensor("transpose_121")]; tensor input_637_cast_fp16 = add(x = input_619_cast_fp16, y = input_635_cast_fp16)[name = tensor("input_637_cast_fp16")]; tensor input_639_axes_0 = const()[name = tensor("input_639_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149323328)))]; tensor encoder_module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149324416)))]; tensor input_639_cast_fp16 = layer_norm(axes = input_639_axes_0, beta = encoder_module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_637_cast_fp16)[name = tensor("input_639_cast_fp16")]; tensor encoder_module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(149325504)))]; tensor encoder_module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151422720)))]; tensor linear_107_cast_fp16 = linear(bias = encoder_module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_639_cast_fp16)[name = tensor("linear_107_cast_fp16")]; tensor input_643_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_643_cast_fp16")]; tensor encoder_module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(151426880)))]; tensor encoder_module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153524096)))]; tensor linear_108_cast_fp16 = linear(bias = encoder_module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("linear_108_cast_fp16")]; tensor var_2406_to_fp16 = const()[name = tensor("op_2406_to_fp16"), val = tensor(0x1p-1)]; tensor var_2407_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2406_to_fp16)[name = tensor("op_2407_cast_fp16")]; tensor input_649_cast_fp16 = add(x = input_637_cast_fp16, y = var_2407_cast_fp16)[name = tensor("input_649_cast_fp16")]; tensor input_651_axes_0 = const()[name = tensor("input_651_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153525184)))]; tensor encoder_module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153526272)))]; tensor input_651_cast_fp16 = layer_norm(axes = input_651_axes_0, beta = encoder_module_layers_11_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_11_norm_out_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("input_651_cast_fp16")]; tensor input_653_axes_0 = const()[name = tensor("input_653_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153527360)))]; tensor encoder_module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153528448)))]; tensor input_653_cast_fp16 = layer_norm(axes = input_653_axes_0, beta = encoder_module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_651_cast_fp16)[name = tensor("input_653_cast_fp16")]; tensor encoder_module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(153529536)))]; tensor encoder_module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155626752)))]; tensor linear_109_cast_fp16 = linear(bias = encoder_module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_653_cast_fp16)[name = tensor("linear_109_cast_fp16")]; tensor input_657_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_657_cast_fp16")]; tensor encoder_module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(155630912)))]; tensor encoder_module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157728128)))]; tensor linear_110_cast_fp16 = linear(bias = encoder_module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_657_cast_fp16)[name = tensor("linear_110_cast_fp16")]; tensor var_2437_to_fp16 = const()[name = tensor("op_2437_to_fp16"), val = tensor(0x1p-1)]; tensor var_2438_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2437_to_fp16)[name = tensor("op_2438_cast_fp16")]; tensor input_663_cast_fp16 = add(x = input_651_cast_fp16, y = var_2438_cast_fp16)[name = tensor("input_663_cast_fp16")]; tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157729216)))]; tensor encoder_module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157730304)))]; tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = encoder_module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_12_norm_self_att_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("query_25_cast_fp16")]; tensor encoder_module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157731392)))]; tensor encoder_module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158255744)))]; tensor linear_111_cast_fp16 = linear(bias = encoder_module_layers_12_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; tensor var_2455 = const()[name = tensor("op_2455"), val = tensor([1, -1, 8, 64])]; tensor q_73_cast_fp16 = reshape(shape = var_2455, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; tensor encoder_module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158256832)))]; tensor encoder_module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158781184)))]; tensor linear_112_cast_fp16 = linear(bias = encoder_module_layers_12_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; tensor var_2460 = const()[name = tensor("op_2460"), val = tensor([1, -1, 8, 64])]; tensor k_49_cast_fp16 = reshape(shape = var_2460, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; tensor encoder_module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158782272)))]; tensor encoder_module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159306624)))]; tensor linear_113_cast_fp16 = linear(bias = encoder_module_layers_12_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; tensor var_2465 = const()[name = tensor("op_2465"), val = tensor([1, -1, 8, 64])]; tensor v_25_cast_fp16 = reshape(shape = var_2465, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159307712)))]; tensor var_2477_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2477_cast_fp16")]; tensor encoder_module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159308800)))]; tensor var_2479_cast_fp16 = add(x = q_73_cast_fp16, y = encoder_module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2479_cast_fp16")]; tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_287_transpose_x_0 = const()[name = tensor("x_287_transpose_x_0"), val = tensor(false)]; tensor x_287_transpose_y_0 = const()[name = tensor("x_287_transpose_y_0"), val = tensor(false)]; tensor var_2481_to_fp16 = const()[name = tensor("op_2481_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159309888)))]; tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2479_cast_fp16)[name = tensor("transpose_119")]; tensor x_287_cast_fp16 = matmul(transpose_x = x_287_transpose_x_0, transpose_y = x_287_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2481_to_fp16)[name = tensor("x_287_cast_fp16")]; tensor x_289_pad_0 = const()[name = tensor("x_289_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_289_mode_0 = const()[name = tensor("x_289_mode_0"), val = tensor("constant")]; tensor const_149_to_fp16 = const()[name = tensor("const_149_to_fp16"), val = tensor(0x0p+0)]; tensor x_289_cast_fp16 = pad(constant_val = const_149_to_fp16, mode = x_289_mode_0, pad = x_289_pad_0, x = x_287_cast_fp16)[name = tensor("x_289_cast_fp16")]; tensor var_2489 = const()[name = tensor("op_2489"), val = tensor([1, 8, -1, 188])]; tensor x_291_cast_fp16 = reshape(shape = var_2489, x = x_289_cast_fp16)[name = tensor("x_291_cast_fp16")]; tensor var_2493_begin_0 = const()[name = tensor("op_2493_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2493_end_0 = const()[name = tensor("op_2493_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2493_end_mask_0 = const()[name = tensor("op_2493_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2493_cast_fp16 = slice_by_index(begin = var_2493_begin_0, end = var_2493_end_0, end_mask = var_2493_end_mask_0, x = x_291_cast_fp16)[name = tensor("op_2493_cast_fp16")]; tensor var_2494 = const()[name = tensor("op_2494"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2494, x = var_2493_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_117")]; tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2477_cast_fp16)[name = tensor("transpose_118")]; tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; tensor var_2503_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2503_cast_fp16")]; tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_49_cast_fp16 = mul(x = var_2503_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; tensor scores_51_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_7)[name = tensor("scores_51_cast_fp16")]; tensor var_2509_cast_fp16 = softmax(axis = var_138, x = scores_51_cast_fp16)[name = tensor("op_2509_cast_fp16")]; tensor input_665_cast_fp16 = select(a = var_154_to_fp16, b = var_2509_cast_fp16, cond = mask_7)[name = tensor("input_665_cast_fp16")]; tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_120")]; tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = input_665_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_293_cast_fp16")]; tensor var_2513_perm_0 = const()[name = tensor("op_2513_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2514 = const()[name = tensor("op_2514"), val = tensor([1, -1, 512])]; tensor var_2513_cast_fp16 = transpose(perm = var_2513_perm_0, x = x_293_cast_fp16)[name = tensor("transpose_116")]; tensor input_667_cast_fp16 = reshape(shape = var_2514, x = var_2513_cast_fp16)[name = tensor("input_667_cast_fp16")]; tensor encoder_module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159693952)))]; tensor encoder_module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160218304)))]; tensor linear_115_cast_fp16 = linear(bias = encoder_module_layers_12_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_667_cast_fp16)[name = tensor("linear_115_cast_fp16")]; tensor input_671_cast_fp16 = add(x = input_663_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_671_cast_fp16")]; tensor x_297_axes_0 = const()[name = tensor("x_297_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160219392)))]; tensor encoder_module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160220480)))]; tensor x_297_cast_fp16 = layer_norm(axes = x_297_axes_0, beta = encoder_module_layers_12_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_12_norm_conv_weight_to_fp16, x = input_671_cast_fp16)[name = tensor("x_297_cast_fp16")]; tensor input_673_perm_0 = const()[name = tensor("input_673_perm_0"), val = tensor([0, 2, 1])]; tensor input_675_pad_type_0 = const()[name = tensor("input_675_pad_type_0"), val = tensor("valid")]; tensor input_675_strides_0 = const()[name = tensor("input_675_strides_0"), val = tensor([1])]; tensor input_675_pad_0 = const()[name = tensor("input_675_pad_0"), val = tensor([0, 0])]; tensor input_675_dilations_0 = const()[name = tensor("input_675_dilations_0"), val = tensor([1])]; tensor input_675_groups_0 = const()[name = tensor("input_675_groups_0"), val = tensor(1)]; tensor encoder_module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160221568)))]; tensor encoder_module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161270208)))]; tensor input_673_cast_fp16 = transpose(perm = input_673_perm_0, x = x_297_cast_fp16)[name = tensor("transpose_115")]; tensor input_675_cast_fp16 = conv(bias = encoder_module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_675_dilations_0, groups = input_675_groups_0, pad = input_675_pad_0, pad_type = input_675_pad_type_0, strides = input_675_strides_0, weight = encoder_module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_673_cast_fp16)[name = tensor("input_675_cast_fp16")]; tensor x_299_split_num_splits_0 = const()[name = tensor("x_299_split_num_splits_0"), val = tensor(2)]; tensor x_299_split_axis_0 = const()[name = tensor("x_299_split_axis_0"), val = tensor(1)]; tensor x_299_split_cast_fp16_0, tensor x_299_split_cast_fp16_1 = split(axis = x_299_split_axis_0, num_splits = x_299_split_num_splits_0, x = input_675_cast_fp16)[name = tensor("x_299_split_cast_fp16")]; tensor x_299_split_1_sigmoid_cast_fp16 = sigmoid(x = x_299_split_cast_fp16_1)[name = tensor("x_299_split_1_sigmoid_cast_fp16")]; tensor x_299_cast_fp16 = mul(x = x_299_split_cast_fp16_0, y = x_299_split_1_sigmoid_cast_fp16)[name = tensor("x_299_cast_fp16")]; tensor input_677_cast_fp16 = select(a = var_154_to_fp16, b = x_299_cast_fp16, cond = var_450)[name = tensor("input_677_cast_fp16")]; tensor input_679_pad_0 = const()[name = tensor("input_679_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_679_mode_0 = const()[name = tensor("input_679_mode_0"), val = tensor("constant")]; tensor const_152_to_fp16 = const()[name = tensor("const_152_to_fp16"), val = tensor(0x0p+0)]; tensor input_679_cast_fp16 = pad(constant_val = const_152_to_fp16, mode = input_679_mode_0, pad = input_679_pad_0, x = input_677_cast_fp16)[name = tensor("input_679_cast_fp16")]; tensor input_681_pad_type_0 = const()[name = tensor("input_681_pad_type_0"), val = tensor("valid")]; tensor input_681_groups_0 = const()[name = tensor("input_681_groups_0"), val = tensor(512)]; tensor input_681_strides_0 = const()[name = tensor("input_681_strides_0"), val = tensor([1])]; tensor input_681_pad_0 = const()[name = tensor("input_681_pad_0"), val = tensor([0, 0])]; tensor input_681_dilations_0 = const()[name = tensor("input_681_dilations_0"), val = tensor([1])]; tensor const_217_to_fp16 = const()[name = tensor("const_217_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161272320)))]; tensor const_218_to_fp16 = const()[name = tensor("const_218_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161281600)))]; tensor input_683_cast_fp16 = conv(bias = const_218_to_fp16, dilations = input_681_dilations_0, groups = input_681_groups_0, pad = input_681_pad_0, pad_type = input_681_pad_type_0, strides = input_681_strides_0, weight = const_217_to_fp16, x = input_679_cast_fp16)[name = tensor("input_683_cast_fp16")]; tensor input_685_cast_fp16 = silu(x = input_683_cast_fp16)[name = tensor("input_685_cast_fp16")]; tensor x_301_pad_type_0 = const()[name = tensor("x_301_pad_type_0"), val = tensor("valid")]; tensor x_301_strides_0 = const()[name = tensor("x_301_strides_0"), val = tensor([1])]; tensor x_301_pad_0 = const()[name = tensor("x_301_pad_0"), val = tensor([0, 0])]; tensor x_301_dilations_0 = const()[name = tensor("x_301_dilations_0"), val = tensor([1])]; tensor x_301_groups_0 = const()[name = tensor("x_301_groups_0"), val = tensor(1)]; tensor encoder_module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161282688)))]; tensor encoder_module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161807040)))]; tensor x_301_cast_fp16 = conv(bias = encoder_module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_301_dilations_0, groups = x_301_groups_0, pad = x_301_pad_0, pad_type = x_301_pad_type_0, strides = x_301_strides_0, weight = encoder_module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_685_cast_fp16)[name = tensor("x_301_cast_fp16")]; tensor input_687_perm_0 = const()[name = tensor("input_687_perm_0"), val = tensor([0, 2, 1])]; tensor input_687_cast_fp16 = transpose(perm = input_687_perm_0, x = x_301_cast_fp16)[name = tensor("transpose_114")]; tensor input_689_cast_fp16 = add(x = input_671_cast_fp16, y = input_687_cast_fp16)[name = tensor("input_689_cast_fp16")]; tensor input_691_axes_0 = const()[name = tensor("input_691_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161808128)))]; tensor encoder_module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161809216)))]; tensor input_691_cast_fp16 = layer_norm(axes = input_691_axes_0, beta = encoder_module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_689_cast_fp16)[name = tensor("input_691_cast_fp16")]; tensor encoder_module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161810304)))]; tensor encoder_module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163907520)))]; tensor linear_116_cast_fp16 = linear(bias = encoder_module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_691_cast_fp16)[name = tensor("linear_116_cast_fp16")]; tensor input_695_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_695_cast_fp16")]; tensor encoder_module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163911680)))]; tensor encoder_module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166008896)))]; tensor linear_117_cast_fp16 = linear(bias = encoder_module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("linear_117_cast_fp16")]; tensor var_2580_to_fp16 = const()[name = tensor("op_2580_to_fp16"), val = tensor(0x1p-1)]; tensor var_2581_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2580_to_fp16)[name = tensor("op_2581_cast_fp16")]; tensor input_701_cast_fp16 = add(x = input_689_cast_fp16, y = var_2581_cast_fp16)[name = tensor("input_701_cast_fp16")]; tensor input_703_axes_0 = const()[name = tensor("input_703_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166009984)))]; tensor encoder_module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166011072)))]; tensor input_703_cast_fp16 = layer_norm(axes = input_703_axes_0, beta = encoder_module_layers_12_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_12_norm_out_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("input_703_cast_fp16")]; tensor input_705_axes_0 = const()[name = tensor("input_705_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166012160)))]; tensor encoder_module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166013248)))]; tensor input_705_cast_fp16 = layer_norm(axes = input_705_axes_0, beta = encoder_module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_703_cast_fp16)[name = tensor("input_705_cast_fp16")]; tensor encoder_module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(166014336)))]; tensor encoder_module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168111552)))]; tensor linear_118_cast_fp16 = linear(bias = encoder_module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_705_cast_fp16)[name = tensor("linear_118_cast_fp16")]; tensor input_709_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_709_cast_fp16")]; tensor encoder_module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(168115712)))]; tensor encoder_module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170212928)))]; tensor linear_119_cast_fp16 = linear(bias = encoder_module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_709_cast_fp16)[name = tensor("linear_119_cast_fp16")]; tensor var_2611_to_fp16 = const()[name = tensor("op_2611_to_fp16"), val = tensor(0x1p-1)]; tensor var_2612_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2611_to_fp16)[name = tensor("op_2612_cast_fp16")]; tensor input_715_cast_fp16 = add(x = input_703_cast_fp16, y = var_2612_cast_fp16)[name = tensor("input_715_cast_fp16")]; tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170214016)))]; tensor encoder_module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170215104)))]; tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = encoder_module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_13_norm_self_att_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("query_27_cast_fp16")]; tensor encoder_module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170216192)))]; tensor encoder_module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170740544)))]; tensor linear_120_cast_fp16 = linear(bias = encoder_module_layers_13_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; tensor var_2629 = const()[name = tensor("op_2629"), val = tensor([1, -1, 8, 64])]; tensor q_79_cast_fp16 = reshape(shape = var_2629, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; tensor encoder_module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170741632)))]; tensor encoder_module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171265984)))]; tensor linear_121_cast_fp16 = linear(bias = encoder_module_layers_13_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; tensor var_2634 = const()[name = tensor("op_2634"), val = tensor([1, -1, 8, 64])]; tensor k_53_cast_fp16 = reshape(shape = var_2634, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; tensor encoder_module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171267072)))]; tensor encoder_module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171791424)))]; tensor linear_122_cast_fp16 = linear(bias = encoder_module_layers_13_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; tensor var_2639 = const()[name = tensor("op_2639"), val = tensor([1, -1, 8, 64])]; tensor v_27_cast_fp16 = reshape(shape = var_2639, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171792512)))]; tensor var_2651_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2651_cast_fp16")]; tensor encoder_module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171793600)))]; tensor var_2653_cast_fp16 = add(x = q_79_cast_fp16, y = encoder_module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2653_cast_fp16")]; tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_309_transpose_x_0 = const()[name = tensor("x_309_transpose_x_0"), val = tensor(false)]; tensor x_309_transpose_y_0 = const()[name = tensor("x_309_transpose_y_0"), val = tensor(false)]; tensor var_2655_to_fp16 = const()[name = tensor("op_2655_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171794688)))]; tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2653_cast_fp16)[name = tensor("transpose_112")]; tensor x_309_cast_fp16 = matmul(transpose_x = x_309_transpose_x_0, transpose_y = x_309_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2655_to_fp16)[name = tensor("x_309_cast_fp16")]; tensor x_311_pad_0 = const()[name = tensor("x_311_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_311_mode_0 = const()[name = tensor("x_311_mode_0"), val = tensor("constant")]; tensor const_159_to_fp16 = const()[name = tensor("const_159_to_fp16"), val = tensor(0x0p+0)]; tensor x_311_cast_fp16 = pad(constant_val = const_159_to_fp16, mode = x_311_mode_0, pad = x_311_pad_0, x = x_309_cast_fp16)[name = tensor("x_311_cast_fp16")]; tensor var_2663 = const()[name = tensor("op_2663"), val = tensor([1, 8, -1, 188])]; tensor x_313_cast_fp16 = reshape(shape = var_2663, x = x_311_cast_fp16)[name = tensor("x_313_cast_fp16")]; tensor var_2667_begin_0 = const()[name = tensor("op_2667_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2667_end_0 = const()[name = tensor("op_2667_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2667_end_mask_0 = const()[name = tensor("op_2667_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2667_cast_fp16 = slice_by_index(begin = var_2667_begin_0, end = var_2667_end_0, end_mask = var_2667_end_mask_0, x = x_313_cast_fp16)[name = tensor("op_2667_cast_fp16")]; tensor var_2668 = const()[name = tensor("op_2668"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2668, x = var_2667_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_110")]; tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2651_cast_fp16)[name = tensor("transpose_111")]; tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; tensor var_2677_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2677_cast_fp16")]; tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_53_cast_fp16 = mul(x = var_2677_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; tensor scores_55_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_7)[name = tensor("scores_55_cast_fp16")]; tensor var_2683_cast_fp16 = softmax(axis = var_138, x = scores_55_cast_fp16)[name = tensor("op_2683_cast_fp16")]; tensor input_717_cast_fp16 = select(a = var_154_to_fp16, b = var_2683_cast_fp16, cond = mask_7)[name = tensor("input_717_cast_fp16")]; tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_113")]; tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = input_717_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_315_cast_fp16")]; tensor var_2687_perm_0 = const()[name = tensor("op_2687_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2688 = const()[name = tensor("op_2688"), val = tensor([1, -1, 512])]; tensor var_2687_cast_fp16 = transpose(perm = var_2687_perm_0, x = x_315_cast_fp16)[name = tensor("transpose_109")]; tensor input_719_cast_fp16 = reshape(shape = var_2688, x = var_2687_cast_fp16)[name = tensor("input_719_cast_fp16")]; tensor encoder_module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172178752)))]; tensor encoder_module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172703104)))]; tensor linear_124_cast_fp16 = linear(bias = encoder_module_layers_13_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_719_cast_fp16)[name = tensor("linear_124_cast_fp16")]; tensor input_723_cast_fp16 = add(x = input_715_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_723_cast_fp16")]; tensor x_319_axes_0 = const()[name = tensor("x_319_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172704192)))]; tensor encoder_module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172705280)))]; tensor x_319_cast_fp16 = layer_norm(axes = x_319_axes_0, beta = encoder_module_layers_13_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_13_norm_conv_weight_to_fp16, x = input_723_cast_fp16)[name = tensor("x_319_cast_fp16")]; tensor input_725_perm_0 = const()[name = tensor("input_725_perm_0"), val = tensor([0, 2, 1])]; tensor input_727_pad_type_0 = const()[name = tensor("input_727_pad_type_0"), val = tensor("valid")]; tensor input_727_strides_0 = const()[name = tensor("input_727_strides_0"), val = tensor([1])]; tensor input_727_pad_0 = const()[name = tensor("input_727_pad_0"), val = tensor([0, 0])]; tensor input_727_dilations_0 = const()[name = tensor("input_727_dilations_0"), val = tensor([1])]; tensor input_727_groups_0 = const()[name = tensor("input_727_groups_0"), val = tensor(1)]; tensor encoder_module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172706368)))]; tensor encoder_module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173755008)))]; tensor input_725_cast_fp16 = transpose(perm = input_725_perm_0, x = x_319_cast_fp16)[name = tensor("transpose_108")]; tensor input_727_cast_fp16 = conv(bias = encoder_module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_727_dilations_0, groups = input_727_groups_0, pad = input_727_pad_0, pad_type = input_727_pad_type_0, strides = input_727_strides_0, weight = encoder_module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_725_cast_fp16)[name = tensor("input_727_cast_fp16")]; tensor x_321_split_num_splits_0 = const()[name = tensor("x_321_split_num_splits_0"), val = tensor(2)]; tensor x_321_split_axis_0 = const()[name = tensor("x_321_split_axis_0"), val = tensor(1)]; tensor x_321_split_cast_fp16_0, tensor x_321_split_cast_fp16_1 = split(axis = x_321_split_axis_0, num_splits = x_321_split_num_splits_0, x = input_727_cast_fp16)[name = tensor("x_321_split_cast_fp16")]; tensor x_321_split_1_sigmoid_cast_fp16 = sigmoid(x = x_321_split_cast_fp16_1)[name = tensor("x_321_split_1_sigmoid_cast_fp16")]; tensor x_321_cast_fp16 = mul(x = x_321_split_cast_fp16_0, y = x_321_split_1_sigmoid_cast_fp16)[name = tensor("x_321_cast_fp16")]; tensor input_729_cast_fp16 = select(a = var_154_to_fp16, b = x_321_cast_fp16, cond = var_450)[name = tensor("input_729_cast_fp16")]; tensor input_731_pad_0 = const()[name = tensor("input_731_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_731_mode_0 = const()[name = tensor("input_731_mode_0"), val = tensor("constant")]; tensor const_162_to_fp16 = const()[name = tensor("const_162_to_fp16"), val = tensor(0x0p+0)]; tensor input_731_cast_fp16 = pad(constant_val = const_162_to_fp16, mode = input_731_mode_0, pad = input_731_pad_0, x = input_729_cast_fp16)[name = tensor("input_731_cast_fp16")]; tensor input_733_pad_type_0 = const()[name = tensor("input_733_pad_type_0"), val = tensor("valid")]; tensor input_733_groups_0 = const()[name = tensor("input_733_groups_0"), val = tensor(512)]; tensor input_733_strides_0 = const()[name = tensor("input_733_strides_0"), val = tensor([1])]; tensor input_733_pad_0 = const()[name = tensor("input_733_pad_0"), val = tensor([0, 0])]; tensor input_733_dilations_0 = const()[name = tensor("input_733_dilations_0"), val = tensor([1])]; tensor const_219_to_fp16 = const()[name = tensor("const_219_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173757120)))]; tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173766400)))]; tensor input_735_cast_fp16 = conv(bias = const_220_to_fp16, dilations = input_733_dilations_0, groups = input_733_groups_0, pad = input_733_pad_0, pad_type = input_733_pad_type_0, strides = input_733_strides_0, weight = const_219_to_fp16, x = input_731_cast_fp16)[name = tensor("input_735_cast_fp16")]; tensor input_737_cast_fp16 = silu(x = input_735_cast_fp16)[name = tensor("input_737_cast_fp16")]; tensor x_323_pad_type_0 = const()[name = tensor("x_323_pad_type_0"), val = tensor("valid")]; tensor x_323_strides_0 = const()[name = tensor("x_323_strides_0"), val = tensor([1])]; tensor x_323_pad_0 = const()[name = tensor("x_323_pad_0"), val = tensor([0, 0])]; tensor x_323_dilations_0 = const()[name = tensor("x_323_dilations_0"), val = tensor([1])]; tensor x_323_groups_0 = const()[name = tensor("x_323_groups_0"), val = tensor(1)]; tensor encoder_module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173767488)))]; tensor encoder_module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174291840)))]; tensor x_323_cast_fp16 = conv(bias = encoder_module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_323_dilations_0, groups = x_323_groups_0, pad = x_323_pad_0, pad_type = x_323_pad_type_0, strides = x_323_strides_0, weight = encoder_module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_737_cast_fp16)[name = tensor("x_323_cast_fp16")]; tensor input_739_perm_0 = const()[name = tensor("input_739_perm_0"), val = tensor([0, 2, 1])]; tensor input_739_cast_fp16 = transpose(perm = input_739_perm_0, x = x_323_cast_fp16)[name = tensor("transpose_107")]; tensor input_741_cast_fp16 = add(x = input_723_cast_fp16, y = input_739_cast_fp16)[name = tensor("input_741_cast_fp16")]; tensor input_743_axes_0 = const()[name = tensor("input_743_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174292928)))]; tensor encoder_module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174294016)))]; tensor input_743_cast_fp16 = layer_norm(axes = input_743_axes_0, beta = encoder_module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_741_cast_fp16)[name = tensor("input_743_cast_fp16")]; tensor encoder_module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(174295104)))]; tensor encoder_module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176392320)))]; tensor linear_125_cast_fp16 = linear(bias = encoder_module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_743_cast_fp16)[name = tensor("linear_125_cast_fp16")]; tensor input_747_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_747_cast_fp16")]; tensor encoder_module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(176396480)))]; tensor encoder_module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178493696)))]; tensor linear_126_cast_fp16 = linear(bias = encoder_module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("linear_126_cast_fp16")]; tensor var_2754_to_fp16 = const()[name = tensor("op_2754_to_fp16"), val = tensor(0x1p-1)]; tensor var_2755_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2754_to_fp16)[name = tensor("op_2755_cast_fp16")]; tensor input_753_cast_fp16 = add(x = input_741_cast_fp16, y = var_2755_cast_fp16)[name = tensor("input_753_cast_fp16")]; tensor input_755_axes_0 = const()[name = tensor("input_755_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178494784)))]; tensor encoder_module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178495872)))]; tensor input_755_cast_fp16 = layer_norm(axes = input_755_axes_0, beta = encoder_module_layers_13_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_13_norm_out_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("input_755_cast_fp16")]; tensor input_757_axes_0 = const()[name = tensor("input_757_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178496960)))]; tensor encoder_module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178498048)))]; tensor input_757_cast_fp16 = layer_norm(axes = input_757_axes_0, beta = encoder_module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_755_cast_fp16)[name = tensor("input_757_cast_fp16")]; tensor encoder_module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(178499136)))]; tensor encoder_module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180596352)))]; tensor linear_127_cast_fp16 = linear(bias = encoder_module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_757_cast_fp16)[name = tensor("linear_127_cast_fp16")]; tensor input_761_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_761_cast_fp16")]; tensor encoder_module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(180600512)))]; tensor encoder_module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182697728)))]; tensor linear_128_cast_fp16 = linear(bias = encoder_module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_761_cast_fp16)[name = tensor("linear_128_cast_fp16")]; tensor var_2785_to_fp16 = const()[name = tensor("op_2785_to_fp16"), val = tensor(0x1p-1)]; tensor var_2786_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2785_to_fp16)[name = tensor("op_2786_cast_fp16")]; tensor input_767_cast_fp16 = add(x = input_755_cast_fp16, y = var_2786_cast_fp16)[name = tensor("input_767_cast_fp16")]; tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182698816)))]; tensor encoder_module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182699904)))]; tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = encoder_module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_14_norm_self_att_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("query_29_cast_fp16")]; tensor encoder_module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182700992)))]; tensor encoder_module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183225344)))]; tensor linear_129_cast_fp16 = linear(bias = encoder_module_layers_14_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; tensor var_2803 = const()[name = tensor("op_2803"), val = tensor([1, -1, 8, 64])]; tensor q_85_cast_fp16 = reshape(shape = var_2803, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; tensor encoder_module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183226432)))]; tensor encoder_module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183750784)))]; tensor linear_130_cast_fp16 = linear(bias = encoder_module_layers_14_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; tensor var_2808 = const()[name = tensor("op_2808"), val = tensor([1, -1, 8, 64])]; tensor k_57_cast_fp16 = reshape(shape = var_2808, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; tensor encoder_module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183751872)))]; tensor encoder_module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184276224)))]; tensor linear_131_cast_fp16 = linear(bias = encoder_module_layers_14_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; tensor var_2813 = const()[name = tensor("op_2813"), val = tensor([1, -1, 8, 64])]; tensor v_29_cast_fp16 = reshape(shape = var_2813, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184277312)))]; tensor var_2825_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2825_cast_fp16")]; tensor encoder_module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184278400)))]; tensor var_2827_cast_fp16 = add(x = q_85_cast_fp16, y = encoder_module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2827_cast_fp16")]; tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_331_transpose_x_0 = const()[name = tensor("x_331_transpose_x_0"), val = tensor(false)]; tensor x_331_transpose_y_0 = const()[name = tensor("x_331_transpose_y_0"), val = tensor(false)]; tensor var_2829_to_fp16 = const()[name = tensor("op_2829_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184279488)))]; tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2827_cast_fp16)[name = tensor("transpose_105")]; tensor x_331_cast_fp16 = matmul(transpose_x = x_331_transpose_x_0, transpose_y = x_331_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2829_to_fp16)[name = tensor("x_331_cast_fp16")]; tensor x_333_pad_0 = const()[name = tensor("x_333_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_333_mode_0 = const()[name = tensor("x_333_mode_0"), val = tensor("constant")]; tensor const_169_to_fp16 = const()[name = tensor("const_169_to_fp16"), val = tensor(0x0p+0)]; tensor x_333_cast_fp16 = pad(constant_val = const_169_to_fp16, mode = x_333_mode_0, pad = x_333_pad_0, x = x_331_cast_fp16)[name = tensor("x_333_cast_fp16")]; tensor var_2837 = const()[name = tensor("op_2837"), val = tensor([1, 8, -1, 188])]; tensor x_335_cast_fp16 = reshape(shape = var_2837, x = x_333_cast_fp16)[name = tensor("x_335_cast_fp16")]; tensor var_2841_begin_0 = const()[name = tensor("op_2841_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_2841_end_0 = const()[name = tensor("op_2841_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_2841_end_mask_0 = const()[name = tensor("op_2841_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_2841_cast_fp16 = slice_by_index(begin = var_2841_begin_0, end = var_2841_end_0, end_mask = var_2841_end_mask_0, x = x_335_cast_fp16)[name = tensor("op_2841_cast_fp16")]; tensor var_2842 = const()[name = tensor("op_2842"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2842, x = var_2841_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_103")]; tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_2825_cast_fp16)[name = tensor("transpose_104")]; tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; tensor var_2851_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2851_cast_fp16")]; tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_57_cast_fp16 = mul(x = var_2851_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; tensor scores_59_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_7)[name = tensor("scores_59_cast_fp16")]; tensor var_2857_cast_fp16 = softmax(axis = var_138, x = scores_59_cast_fp16)[name = tensor("op_2857_cast_fp16")]; tensor input_769_cast_fp16 = select(a = var_154_to_fp16, b = var_2857_cast_fp16, cond = mask_7)[name = tensor("input_769_cast_fp16")]; tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_106")]; tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = input_769_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_337_cast_fp16")]; tensor var_2861_perm_0 = const()[name = tensor("op_2861_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_2862 = const()[name = tensor("op_2862"), val = tensor([1, -1, 512])]; tensor var_2861_cast_fp16 = transpose(perm = var_2861_perm_0, x = x_337_cast_fp16)[name = tensor("transpose_102")]; tensor input_771_cast_fp16 = reshape(shape = var_2862, x = var_2861_cast_fp16)[name = tensor("input_771_cast_fp16")]; tensor encoder_module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184663552)))]; tensor encoder_module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185187904)))]; tensor linear_133_cast_fp16 = linear(bias = encoder_module_layers_14_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_771_cast_fp16)[name = tensor("linear_133_cast_fp16")]; tensor input_775_cast_fp16 = add(x = input_767_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_775_cast_fp16")]; tensor x_341_axes_0 = const()[name = tensor("x_341_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185188992)))]; tensor encoder_module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185190080)))]; tensor x_341_cast_fp16 = layer_norm(axes = x_341_axes_0, beta = encoder_module_layers_14_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_14_norm_conv_weight_to_fp16, x = input_775_cast_fp16)[name = tensor("x_341_cast_fp16")]; tensor input_777_perm_0 = const()[name = tensor("input_777_perm_0"), val = tensor([0, 2, 1])]; tensor input_779_pad_type_0 = const()[name = tensor("input_779_pad_type_0"), val = tensor("valid")]; tensor input_779_strides_0 = const()[name = tensor("input_779_strides_0"), val = tensor([1])]; tensor input_779_pad_0 = const()[name = tensor("input_779_pad_0"), val = tensor([0, 0])]; tensor input_779_dilations_0 = const()[name = tensor("input_779_dilations_0"), val = tensor([1])]; tensor input_779_groups_0 = const()[name = tensor("input_779_groups_0"), val = tensor(1)]; tensor encoder_module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185191168)))]; tensor encoder_module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186239808)))]; tensor input_777_cast_fp16 = transpose(perm = input_777_perm_0, x = x_341_cast_fp16)[name = tensor("transpose_101")]; tensor input_779_cast_fp16 = conv(bias = encoder_module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_779_dilations_0, groups = input_779_groups_0, pad = input_779_pad_0, pad_type = input_779_pad_type_0, strides = input_779_strides_0, weight = encoder_module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_777_cast_fp16)[name = tensor("input_779_cast_fp16")]; tensor x_343_split_num_splits_0 = const()[name = tensor("x_343_split_num_splits_0"), val = tensor(2)]; tensor x_343_split_axis_0 = const()[name = tensor("x_343_split_axis_0"), val = tensor(1)]; tensor x_343_split_cast_fp16_0, tensor x_343_split_cast_fp16_1 = split(axis = x_343_split_axis_0, num_splits = x_343_split_num_splits_0, x = input_779_cast_fp16)[name = tensor("x_343_split_cast_fp16")]; tensor x_343_split_1_sigmoid_cast_fp16 = sigmoid(x = x_343_split_cast_fp16_1)[name = tensor("x_343_split_1_sigmoid_cast_fp16")]; tensor x_343_cast_fp16 = mul(x = x_343_split_cast_fp16_0, y = x_343_split_1_sigmoid_cast_fp16)[name = tensor("x_343_cast_fp16")]; tensor input_781_cast_fp16 = select(a = var_154_to_fp16, b = x_343_cast_fp16, cond = var_450)[name = tensor("input_781_cast_fp16")]; tensor input_783_pad_0 = const()[name = tensor("input_783_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_783_mode_0 = const()[name = tensor("input_783_mode_0"), val = tensor("constant")]; tensor const_172_to_fp16 = const()[name = tensor("const_172_to_fp16"), val = tensor(0x0p+0)]; tensor input_783_cast_fp16 = pad(constant_val = const_172_to_fp16, mode = input_783_mode_0, pad = input_783_pad_0, x = input_781_cast_fp16)[name = tensor("input_783_cast_fp16")]; tensor input_785_pad_type_0 = const()[name = tensor("input_785_pad_type_0"), val = tensor("valid")]; tensor input_785_groups_0 = const()[name = tensor("input_785_groups_0"), val = tensor(512)]; tensor input_785_strides_0 = const()[name = tensor("input_785_strides_0"), val = tensor([1])]; tensor input_785_pad_0 = const()[name = tensor("input_785_pad_0"), val = tensor([0, 0])]; tensor input_785_dilations_0 = const()[name = tensor("input_785_dilations_0"), val = tensor([1])]; tensor const_221_to_fp16 = const()[name = tensor("const_221_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186241920)))]; tensor const_222_to_fp16 = const()[name = tensor("const_222_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186251200)))]; tensor input_787_cast_fp16 = conv(bias = const_222_to_fp16, dilations = input_785_dilations_0, groups = input_785_groups_0, pad = input_785_pad_0, pad_type = input_785_pad_type_0, strides = input_785_strides_0, weight = const_221_to_fp16, x = input_783_cast_fp16)[name = tensor("input_787_cast_fp16")]; tensor input_789_cast_fp16 = silu(x = input_787_cast_fp16)[name = tensor("input_789_cast_fp16")]; tensor x_345_pad_type_0 = const()[name = tensor("x_345_pad_type_0"), val = tensor("valid")]; tensor x_345_strides_0 = const()[name = tensor("x_345_strides_0"), val = tensor([1])]; tensor x_345_pad_0 = const()[name = tensor("x_345_pad_0"), val = tensor([0, 0])]; tensor x_345_dilations_0 = const()[name = tensor("x_345_dilations_0"), val = tensor([1])]; tensor x_345_groups_0 = const()[name = tensor("x_345_groups_0"), val = tensor(1)]; tensor encoder_module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186252288)))]; tensor encoder_module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186776640)))]; tensor x_345_cast_fp16 = conv(bias = encoder_module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_345_dilations_0, groups = x_345_groups_0, pad = x_345_pad_0, pad_type = x_345_pad_type_0, strides = x_345_strides_0, weight = encoder_module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_789_cast_fp16)[name = tensor("x_345_cast_fp16")]; tensor input_791_perm_0 = const()[name = tensor("input_791_perm_0"), val = tensor([0, 2, 1])]; tensor input_791_cast_fp16 = transpose(perm = input_791_perm_0, x = x_345_cast_fp16)[name = tensor("transpose_100")]; tensor input_793_cast_fp16 = add(x = input_775_cast_fp16, y = input_791_cast_fp16)[name = tensor("input_793_cast_fp16")]; tensor input_795_axes_0 = const()[name = tensor("input_795_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186777728)))]; tensor encoder_module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186778816)))]; tensor input_795_cast_fp16 = layer_norm(axes = input_795_axes_0, beta = encoder_module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_793_cast_fp16)[name = tensor("input_795_cast_fp16")]; tensor encoder_module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(186779904)))]; tensor encoder_module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188877120)))]; tensor linear_134_cast_fp16 = linear(bias = encoder_module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_795_cast_fp16)[name = tensor("linear_134_cast_fp16")]; tensor input_799_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_799_cast_fp16")]; tensor encoder_module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188881280)))]; tensor encoder_module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190978496)))]; tensor linear_135_cast_fp16 = linear(bias = encoder_module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("linear_135_cast_fp16")]; tensor var_2928_to_fp16 = const()[name = tensor("op_2928_to_fp16"), val = tensor(0x1p-1)]; tensor var_2929_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2928_to_fp16)[name = tensor("op_2929_cast_fp16")]; tensor input_805_cast_fp16 = add(x = input_793_cast_fp16, y = var_2929_cast_fp16)[name = tensor("input_805_cast_fp16")]; tensor input_807_axes_0 = const()[name = tensor("input_807_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190979584)))]; tensor encoder_module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190980672)))]; tensor input_807_cast_fp16 = layer_norm(axes = input_807_axes_0, beta = encoder_module_layers_14_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_14_norm_out_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("input_807_cast_fp16")]; tensor input_809_axes_0 = const()[name = tensor("input_809_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190981760)))]; tensor encoder_module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190982848)))]; tensor input_809_cast_fp16 = layer_norm(axes = input_809_axes_0, beta = encoder_module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_807_cast_fp16)[name = tensor("input_809_cast_fp16")]; tensor encoder_module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190983936)))]; tensor encoder_module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193081152)))]; tensor linear_136_cast_fp16 = linear(bias = encoder_module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_809_cast_fp16)[name = tensor("linear_136_cast_fp16")]; tensor input_813_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_813_cast_fp16")]; tensor encoder_module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(193085312)))]; tensor encoder_module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195182528)))]; tensor linear_137_cast_fp16 = linear(bias = encoder_module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_813_cast_fp16)[name = tensor("linear_137_cast_fp16")]; tensor var_2959_to_fp16 = const()[name = tensor("op_2959_to_fp16"), val = tensor(0x1p-1)]; tensor var_2960_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2959_to_fp16)[name = tensor("op_2960_cast_fp16")]; tensor input_819_cast_fp16 = add(x = input_807_cast_fp16, y = var_2960_cast_fp16)[name = tensor("input_819_cast_fp16")]; tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195183616)))]; tensor encoder_module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195184704)))]; tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = encoder_module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_15_norm_self_att_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("query_31_cast_fp16")]; tensor encoder_module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195185792)))]; tensor encoder_module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195710144)))]; tensor linear_138_cast_fp16 = linear(bias = encoder_module_layers_15_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; tensor var_2977 = const()[name = tensor("op_2977"), val = tensor([1, -1, 8, 64])]; tensor q_91_cast_fp16 = reshape(shape = var_2977, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; tensor encoder_module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195711232)))]; tensor encoder_module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196235584)))]; tensor linear_139_cast_fp16 = linear(bias = encoder_module_layers_15_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; tensor var_2982 = const()[name = tensor("op_2982"), val = tensor([1, -1, 8, 64])]; tensor k_61_cast_fp16 = reshape(shape = var_2982, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; tensor encoder_module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196236672)))]; tensor encoder_module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196761024)))]; tensor linear_140_cast_fp16 = linear(bias = encoder_module_layers_15_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; tensor var_2987 = const()[name = tensor("op_2987"), val = tensor([1, -1, 8, 64])]; tensor v_31_cast_fp16 = reshape(shape = var_2987, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; tensor value_35_perm_0 = const()[name = tensor("value_35_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196762112)))]; tensor var_2999_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2999_cast_fp16")]; tensor encoder_module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196763200)))]; tensor var_3001_cast_fp16 = add(x = q_91_cast_fp16, y = encoder_module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3001_cast_fp16")]; tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_353_transpose_x_0 = const()[name = tensor("x_353_transpose_x_0"), val = tensor(false)]; tensor x_353_transpose_y_0 = const()[name = tensor("x_353_transpose_y_0"), val = tensor(false)]; tensor var_3003_to_fp16 = const()[name = tensor("op_3003_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196764288)))]; tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3001_cast_fp16)[name = tensor("transpose_98")]; tensor x_353_cast_fp16 = matmul(transpose_x = x_353_transpose_x_0, transpose_y = x_353_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3003_to_fp16)[name = tensor("x_353_cast_fp16")]; tensor x_355_pad_0 = const()[name = tensor("x_355_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_355_mode_0 = const()[name = tensor("x_355_mode_0"), val = tensor("constant")]; tensor const_179_to_fp16 = const()[name = tensor("const_179_to_fp16"), val = tensor(0x0p+0)]; tensor x_355_cast_fp16 = pad(constant_val = const_179_to_fp16, mode = x_355_mode_0, pad = x_355_pad_0, x = x_353_cast_fp16)[name = tensor("x_355_cast_fp16")]; tensor var_3011 = const()[name = tensor("op_3011"), val = tensor([1, 8, -1, 188])]; tensor x_357_cast_fp16 = reshape(shape = var_3011, x = x_355_cast_fp16)[name = tensor("x_357_cast_fp16")]; tensor var_3015_begin_0 = const()[name = tensor("op_3015_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3015_end_0 = const()[name = tensor("op_3015_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3015_end_mask_0 = const()[name = tensor("op_3015_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3015_cast_fp16 = slice_by_index(begin = var_3015_begin_0, end = var_3015_end_0, end_mask = var_3015_end_mask_0, x = x_357_cast_fp16)[name = tensor("op_3015_cast_fp16")]; tensor var_3016 = const()[name = tensor("op_3016"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3016, x = var_3015_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_96")]; tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_2999_cast_fp16)[name = tensor("transpose_97")]; tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; tensor var_3025_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3025_cast_fp16")]; tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_61_cast_fp16 = mul(x = var_3025_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; tensor scores_63_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_7)[name = tensor("scores_63_cast_fp16")]; tensor var_3031_cast_fp16 = softmax(axis = var_138, x = scores_63_cast_fp16)[name = tensor("op_3031_cast_fp16")]; tensor input_821_cast_fp16 = select(a = var_154_to_fp16, b = var_3031_cast_fp16, cond = mask_7)[name = tensor("input_821_cast_fp16")]; tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; tensor value_35_cast_fp16 = transpose(perm = value_35_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_99")]; tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = input_821_cast_fp16, y = value_35_cast_fp16)[name = tensor("x_359_cast_fp16")]; tensor var_3035_perm_0 = const()[name = tensor("op_3035_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3036 = const()[name = tensor("op_3036"), val = tensor([1, -1, 512])]; tensor var_3035_cast_fp16 = transpose(perm = var_3035_perm_0, x = x_359_cast_fp16)[name = tensor("transpose_95")]; tensor input_823_cast_fp16 = reshape(shape = var_3036, x = var_3035_cast_fp16)[name = tensor("input_823_cast_fp16")]; tensor encoder_module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197148352)))]; tensor encoder_module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197672704)))]; tensor linear_142_cast_fp16 = linear(bias = encoder_module_layers_15_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_823_cast_fp16)[name = tensor("linear_142_cast_fp16")]; tensor input_827_cast_fp16 = add(x = input_819_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_827_cast_fp16")]; tensor x_363_axes_0 = const()[name = tensor("x_363_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197673792)))]; tensor encoder_module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197674880)))]; tensor x_363_cast_fp16 = layer_norm(axes = x_363_axes_0, beta = encoder_module_layers_15_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_15_norm_conv_weight_to_fp16, x = input_827_cast_fp16)[name = tensor("x_363_cast_fp16")]; tensor input_829_perm_0 = const()[name = tensor("input_829_perm_0"), val = tensor([0, 2, 1])]; tensor input_831_pad_type_0 = const()[name = tensor("input_831_pad_type_0"), val = tensor("valid")]; tensor input_831_strides_0 = const()[name = tensor("input_831_strides_0"), val = tensor([1])]; tensor input_831_pad_0 = const()[name = tensor("input_831_pad_0"), val = tensor([0, 0])]; tensor input_831_dilations_0 = const()[name = tensor("input_831_dilations_0"), val = tensor([1])]; tensor input_831_groups_0 = const()[name = tensor("input_831_groups_0"), val = tensor(1)]; tensor encoder_module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197675968)))]; tensor encoder_module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198724608)))]; tensor input_829_cast_fp16 = transpose(perm = input_829_perm_0, x = x_363_cast_fp16)[name = tensor("transpose_94")]; tensor input_831_cast_fp16 = conv(bias = encoder_module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_831_dilations_0, groups = input_831_groups_0, pad = input_831_pad_0, pad_type = input_831_pad_type_0, strides = input_831_strides_0, weight = encoder_module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_829_cast_fp16)[name = tensor("input_831_cast_fp16")]; tensor x_365_split_num_splits_0 = const()[name = tensor("x_365_split_num_splits_0"), val = tensor(2)]; tensor x_365_split_axis_0 = const()[name = tensor("x_365_split_axis_0"), val = tensor(1)]; tensor x_365_split_cast_fp16_0, tensor x_365_split_cast_fp16_1 = split(axis = x_365_split_axis_0, num_splits = x_365_split_num_splits_0, x = input_831_cast_fp16)[name = tensor("x_365_split_cast_fp16")]; tensor x_365_split_1_sigmoid_cast_fp16 = sigmoid(x = x_365_split_cast_fp16_1)[name = tensor("x_365_split_1_sigmoid_cast_fp16")]; tensor x_365_cast_fp16 = mul(x = x_365_split_cast_fp16_0, y = x_365_split_1_sigmoid_cast_fp16)[name = tensor("x_365_cast_fp16")]; tensor input_833_cast_fp16 = select(a = var_154_to_fp16, b = x_365_cast_fp16, cond = var_450)[name = tensor("input_833_cast_fp16")]; tensor input_835_pad_0 = const()[name = tensor("input_835_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_835_mode_0 = const()[name = tensor("input_835_mode_0"), val = tensor("constant")]; tensor const_182_to_fp16 = const()[name = tensor("const_182_to_fp16"), val = tensor(0x0p+0)]; tensor input_835_cast_fp16 = pad(constant_val = const_182_to_fp16, mode = input_835_mode_0, pad = input_835_pad_0, x = input_833_cast_fp16)[name = tensor("input_835_cast_fp16")]; tensor input_837_pad_type_0 = const()[name = tensor("input_837_pad_type_0"), val = tensor("valid")]; tensor input_837_groups_0 = const()[name = tensor("input_837_groups_0"), val = tensor(512)]; tensor input_837_strides_0 = const()[name = tensor("input_837_strides_0"), val = tensor([1])]; tensor input_837_pad_0 = const()[name = tensor("input_837_pad_0"), val = tensor([0, 0])]; tensor input_837_dilations_0 = const()[name = tensor("input_837_dilations_0"), val = tensor([1])]; tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198726720)))]; tensor const_224_to_fp16 = const()[name = tensor("const_224_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198736000)))]; tensor input_839_cast_fp16 = conv(bias = const_224_to_fp16, dilations = input_837_dilations_0, groups = input_837_groups_0, pad = input_837_pad_0, pad_type = input_837_pad_type_0, strides = input_837_strides_0, weight = const_223_to_fp16, x = input_835_cast_fp16)[name = tensor("input_839_cast_fp16")]; tensor input_841_cast_fp16 = silu(x = input_839_cast_fp16)[name = tensor("input_841_cast_fp16")]; tensor x_367_pad_type_0 = const()[name = tensor("x_367_pad_type_0"), val = tensor("valid")]; tensor x_367_strides_0 = const()[name = tensor("x_367_strides_0"), val = tensor([1])]; tensor x_367_pad_0 = const()[name = tensor("x_367_pad_0"), val = tensor([0, 0])]; tensor x_367_dilations_0 = const()[name = tensor("x_367_dilations_0"), val = tensor([1])]; tensor x_367_groups_0 = const()[name = tensor("x_367_groups_0"), val = tensor(1)]; tensor encoder_module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198737088)))]; tensor encoder_module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199261440)))]; tensor x_367_cast_fp16 = conv(bias = encoder_module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_367_dilations_0, groups = x_367_groups_0, pad = x_367_pad_0, pad_type = x_367_pad_type_0, strides = x_367_strides_0, weight = encoder_module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_841_cast_fp16)[name = tensor("x_367_cast_fp16")]; tensor input_843_perm_0 = const()[name = tensor("input_843_perm_0"), val = tensor([0, 2, 1])]; tensor input_843_cast_fp16 = transpose(perm = input_843_perm_0, x = x_367_cast_fp16)[name = tensor("transpose_93")]; tensor input_845_cast_fp16 = add(x = input_827_cast_fp16, y = input_843_cast_fp16)[name = tensor("input_845_cast_fp16")]; tensor input_847_axes_0 = const()[name = tensor("input_847_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199262528)))]; tensor encoder_module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199263616)))]; tensor input_847_cast_fp16 = layer_norm(axes = input_847_axes_0, beta = encoder_module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_845_cast_fp16)[name = tensor("input_847_cast_fp16")]; tensor encoder_module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(199264704)))]; tensor encoder_module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201361920)))]; tensor linear_143_cast_fp16 = linear(bias = encoder_module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_847_cast_fp16)[name = tensor("linear_143_cast_fp16")]; tensor input_851_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_851_cast_fp16")]; tensor encoder_module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(201366080)))]; tensor encoder_module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203463296)))]; tensor linear_144_cast_fp16 = linear(bias = encoder_module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("linear_144_cast_fp16")]; tensor var_3102_to_fp16 = const()[name = tensor("op_3102_to_fp16"), val = tensor(0x1p-1)]; tensor var_3103_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3102_to_fp16)[name = tensor("op_3103_cast_fp16")]; tensor input_857_cast_fp16 = add(x = input_845_cast_fp16, y = var_3103_cast_fp16)[name = tensor("input_857_cast_fp16")]; tensor input_859_axes_0 = const()[name = tensor("input_859_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203464384)))]; tensor encoder_module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203465472)))]; tensor input_859_cast_fp16 = layer_norm(axes = input_859_axes_0, beta = encoder_module_layers_15_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_15_norm_out_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("input_859_cast_fp16")]; tensor input_861_axes_0 = const()[name = tensor("input_861_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203466560)))]; tensor encoder_module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203467648)))]; tensor input_861_cast_fp16 = layer_norm(axes = input_861_axes_0, beta = encoder_module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_859_cast_fp16)[name = tensor("input_861_cast_fp16")]; tensor encoder_module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(203468736)))]; tensor encoder_module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205565952)))]; tensor linear_145_cast_fp16 = linear(bias = encoder_module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = encoder_module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_861_cast_fp16)[name = tensor("linear_145_cast_fp16")]; tensor input_865_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_865_cast_fp16")]; tensor encoder_module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(205570112)))]; tensor encoder_module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207667328)))]; tensor linear_146_cast_fp16 = linear(bias = encoder_module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = encoder_module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_865_cast_fp16)[name = tensor("linear_146_cast_fp16")]; tensor var_3133_to_fp16 = const()[name = tensor("op_3133_to_fp16"), val = tensor(0x1p-1)]; tensor var_3134_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3133_to_fp16)[name = tensor("op_3134_cast_fp16")]; tensor input_871_cast_fp16 = add(x = input_859_cast_fp16, y = var_3134_cast_fp16)[name = tensor("input_871_cast_fp16")]; tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207668416)))]; tensor encoder_module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207669504)))]; tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = encoder_module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_16_norm_self_att_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("query_cast_fp16")]; tensor encoder_module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207670592)))]; tensor encoder_module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208194944)))]; tensor linear_147_cast_fp16 = linear(bias = encoder_module_layers_16_self_attn_linear_q_bias_to_fp16, weight = encoder_module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_147_cast_fp16")]; tensor var_3151 = const()[name = tensor("op_3151"), val = tensor([1, -1, 8, 64])]; tensor q_97_cast_fp16 = reshape(shape = var_3151, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; tensor encoder_module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208196032)))]; tensor encoder_module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208720384)))]; tensor linear_148_cast_fp16 = linear(bias = encoder_module_layers_16_self_attn_linear_k_bias_to_fp16, weight = encoder_module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_148_cast_fp16")]; tensor var_3156 = const()[name = tensor("op_3156"), val = tensor([1, -1, 8, 64])]; tensor k_65_cast_fp16 = reshape(shape = var_3156, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; tensor encoder_module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208721472)))]; tensor encoder_module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209245824)))]; tensor linear_149_cast_fp16 = linear(bias = encoder_module_layers_16_self_attn_linear_v_bias_to_fp16, weight = encoder_module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_149_cast_fp16")]; tensor var_3161 = const()[name = tensor("op_3161"), val = tensor([1, -1, 8, 64])]; tensor v_cast_fp16 = reshape(shape = var_3161, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; tensor encoder_module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209246912)))]; tensor var_3173_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3173_cast_fp16")]; tensor encoder_module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209248000)))]; tensor var_3175_cast_fp16 = add(x = q_97_cast_fp16, y = encoder_module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3175_cast_fp16")]; tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; tensor x_375_transpose_x_0 = const()[name = tensor("x_375_transpose_x_0"), val = tensor(false)]; tensor x_375_transpose_y_0 = const()[name = tensor("x_375_transpose_y_0"), val = tensor(false)]; tensor var_3177_to_fp16 = const()[name = tensor("op_3177_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209249088)))]; tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3175_cast_fp16)[name = tensor("transpose_91")]; tensor x_375_cast_fp16 = matmul(transpose_x = x_375_transpose_x_0, transpose_y = x_375_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3177_to_fp16)[name = tensor("x_375_cast_fp16")]; tensor x_377_pad_0 = const()[name = tensor("x_377_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; tensor x_377_mode_0 = const()[name = tensor("x_377_mode_0"), val = tensor("constant")]; tensor const_189_to_fp16 = const()[name = tensor("const_189_to_fp16"), val = tensor(0x0p+0)]; tensor x_377_cast_fp16 = pad(constant_val = const_189_to_fp16, mode = x_377_mode_0, pad = x_377_pad_0, x = x_375_cast_fp16)[name = tensor("x_377_cast_fp16")]; tensor var_3185 = const()[name = tensor("op_3185"), val = tensor([1, 8, -1, 188])]; tensor x_379_cast_fp16 = reshape(shape = var_3185, x = x_377_cast_fp16)[name = tensor("x_379_cast_fp16")]; tensor var_3189_begin_0 = const()[name = tensor("op_3189_begin_0"), val = tensor([0, 0, 1, 0])]; tensor var_3189_end_0 = const()[name = tensor("op_3189_end_0"), val = tensor([1, 8, 376, 188])]; tensor var_3189_end_mask_0 = const()[name = tensor("op_3189_end_mask_0"), val = tensor([true, true, true, true])]; tensor var_3189_cast_fp16 = slice_by_index(begin = var_3189_begin_0, end = var_3189_end_0, end_mask = var_3189_end_mask_0, x = x_379_cast_fp16)[name = tensor("op_3189_cast_fp16")]; tensor var_3190 = const()[name = tensor("op_3190"), val = tensor([1, 8, 188, 375])]; tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3190, x = var_3189_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_89")]; tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3173_cast_fp16)[name = tensor("transpose_90")]; tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; tensor var_3199_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3199_cast_fp16")]; tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; tensor _inversed_scores_65_cast_fp16 = mul(x = var_3199_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; tensor scores_cast_fp16 = select(a = var_153_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_7)[name = tensor("scores_cast_fp16")]; tensor var_3205_cast_fp16 = softmax(axis = var_138, x = scores_cast_fp16)[name = tensor("op_3205_cast_fp16")]; tensor input_873_cast_fp16 = select(a = var_154_to_fp16, b = var_3205_cast_fp16, cond = mask_7)[name = tensor("input_873_cast_fp16")]; tensor x_381_transpose_x_0 = const()[name = tensor("x_381_transpose_x_0"), val = tensor(false)]; tensor x_381_transpose_y_0 = const()[name = tensor("x_381_transpose_y_0"), val = tensor(false)]; tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_92")]; tensor x_381_cast_fp16 = matmul(transpose_x = x_381_transpose_x_0, transpose_y = x_381_transpose_y_0, x = input_873_cast_fp16, y = value_cast_fp16)[name = tensor("x_381_cast_fp16")]; tensor var_3209_perm_0 = const()[name = tensor("op_3209_perm_0"), val = tensor([0, 2, 1, 3])]; tensor var_3210 = const()[name = tensor("op_3210"), val = tensor([1, -1, 512])]; tensor var_3209_cast_fp16 = transpose(perm = var_3209_perm_0, x = x_381_cast_fp16)[name = tensor("transpose_88")]; tensor input_875_cast_fp16 = reshape(shape = var_3210, x = var_3209_cast_fp16)[name = tensor("input_875_cast_fp16")]; tensor encoder_module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209633152)))]; tensor encoder_module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210157504)))]; tensor linear_151_cast_fp16 = linear(bias = encoder_module_layers_16_self_attn_linear_out_bias_to_fp16, weight = encoder_module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_875_cast_fp16)[name = tensor("linear_151_cast_fp16")]; tensor input_879_cast_fp16 = add(x = input_871_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_879_cast_fp16")]; tensor x_385_axes_0 = const()[name = tensor("x_385_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210158592)))]; tensor encoder_module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210159680)))]; tensor x_385_cast_fp16 = layer_norm(axes = x_385_axes_0, beta = encoder_module_layers_16_norm_conv_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_16_norm_conv_weight_to_fp16, x = input_879_cast_fp16)[name = tensor("x_385_cast_fp16")]; tensor input_881_perm_0 = const()[name = tensor("input_881_perm_0"), val = tensor([0, 2, 1])]; tensor input_883_pad_type_0 = const()[name = tensor("input_883_pad_type_0"), val = tensor("valid")]; tensor input_883_strides_0 = const()[name = tensor("input_883_strides_0"), val = tensor([1])]; tensor input_883_pad_0 = const()[name = tensor("input_883_pad_0"), val = tensor([0, 0])]; tensor input_883_dilations_0 = const()[name = tensor("input_883_dilations_0"), val = tensor([1])]; tensor input_883_groups_0 = const()[name = tensor("input_883_groups_0"), val = tensor(1)]; tensor encoder_module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210160768)))]; tensor encoder_module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211209408)))]; tensor input_881_cast_fp16 = transpose(perm = input_881_perm_0, x = x_385_cast_fp16)[name = tensor("transpose_87")]; tensor input_883_cast_fp16 = conv(bias = encoder_module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_883_dilations_0, groups = input_883_groups_0, pad = input_883_pad_0, pad_type = input_883_pad_type_0, strides = input_883_strides_0, weight = encoder_module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_881_cast_fp16)[name = tensor("input_883_cast_fp16")]; tensor x_387_split_num_splits_0 = const()[name = tensor("x_387_split_num_splits_0"), val = tensor(2)]; tensor x_387_split_axis_0 = const()[name = tensor("x_387_split_axis_0"), val = tensor(1)]; tensor x_387_split_cast_fp16_0, tensor x_387_split_cast_fp16_1 = split(axis = x_387_split_axis_0, num_splits = x_387_split_num_splits_0, x = input_883_cast_fp16)[name = tensor("x_387_split_cast_fp16")]; tensor x_387_split_1_sigmoid_cast_fp16 = sigmoid(x = x_387_split_cast_fp16_1)[name = tensor("x_387_split_1_sigmoid_cast_fp16")]; tensor x_387_cast_fp16 = mul(x = x_387_split_cast_fp16_0, y = x_387_split_1_sigmoid_cast_fp16)[name = tensor("x_387_cast_fp16")]; tensor input_885_cast_fp16 = select(a = var_154_to_fp16, b = x_387_cast_fp16, cond = var_450)[name = tensor("input_885_cast_fp16")]; tensor input_887_pad_0 = const()[name = tensor("input_887_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; tensor input_887_mode_0 = const()[name = tensor("input_887_mode_0"), val = tensor("constant")]; tensor const_192_to_fp16 = const()[name = tensor("const_192_to_fp16"), val = tensor(0x0p+0)]; tensor input_887_cast_fp16 = pad(constant_val = const_192_to_fp16, mode = input_887_mode_0, pad = input_887_pad_0, x = input_885_cast_fp16)[name = tensor("input_887_cast_fp16")]; tensor input_889_pad_type_0 = const()[name = tensor("input_889_pad_type_0"), val = tensor("valid")]; tensor input_889_groups_0 = const()[name = tensor("input_889_groups_0"), val = tensor(512)]; tensor input_889_strides_0 = const()[name = tensor("input_889_strides_0"), val = tensor([1])]; tensor input_889_pad_0 = const()[name = tensor("input_889_pad_0"), val = tensor([0, 0])]; tensor input_889_dilations_0 = const()[name = tensor("input_889_dilations_0"), val = tensor([1])]; tensor const_225_to_fp16 = const()[name = tensor("const_225_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211211520)))]; tensor const_226_to_fp16 = const()[name = tensor("const_226_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211220800)))]; tensor input_891_cast_fp16 = conv(bias = const_226_to_fp16, dilations = input_889_dilations_0, groups = input_889_groups_0, pad = input_889_pad_0, pad_type = input_889_pad_type_0, strides = input_889_strides_0, weight = const_225_to_fp16, x = input_887_cast_fp16)[name = tensor("input_891_cast_fp16")]; tensor input_893_cast_fp16 = silu(x = input_891_cast_fp16)[name = tensor("input_893_cast_fp16")]; tensor x_389_pad_type_0 = const()[name = tensor("x_389_pad_type_0"), val = tensor("valid")]; tensor x_389_strides_0 = const()[name = tensor("x_389_strides_0"), val = tensor([1])]; tensor x_389_pad_0 = const()[name = tensor("x_389_pad_0"), val = tensor([0, 0])]; tensor x_389_dilations_0 = const()[name = tensor("x_389_dilations_0"), val = tensor([1])]; tensor x_389_groups_0 = const()[name = tensor("x_389_groups_0"), val = tensor(1)]; tensor encoder_module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211221888)))]; tensor encoder_module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211746240)))]; tensor x_389_cast_fp16 = conv(bias = encoder_module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_389_dilations_0, groups = x_389_groups_0, pad = x_389_pad_0, pad_type = x_389_pad_type_0, strides = x_389_strides_0, weight = encoder_module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_893_cast_fp16)[name = tensor("x_389_cast_fp16")]; tensor input_895_perm_0 = const()[name = tensor("input_895_perm_0"), val = tensor([0, 2, 1])]; tensor input_895_cast_fp16 = transpose(perm = input_895_perm_0, x = x_389_cast_fp16)[name = tensor("transpose_86")]; tensor input_897_cast_fp16 = add(x = input_879_cast_fp16, y = input_895_cast_fp16)[name = tensor("input_897_cast_fp16")]; tensor input_899_axes_0 = const()[name = tensor("input_899_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211747328)))]; tensor encoder_module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211748416)))]; tensor input_899_cast_fp16 = layer_norm(axes = input_899_axes_0, beta = encoder_module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_897_cast_fp16)[name = tensor("input_899_cast_fp16")]; tensor encoder_module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(211749504)))]; tensor encoder_module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213846720)))]; tensor linear_152_cast_fp16 = linear(bias = encoder_module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = encoder_module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_899_cast_fp16)[name = tensor("linear_152_cast_fp16")]; tensor input_903_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_903_cast_fp16")]; tensor encoder_module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213850880)))]; tensor encoder_module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215948096)))]; tensor linear_153_cast_fp16 = linear(bias = encoder_module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = encoder_module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_903_cast_fp16)[name = tensor("linear_153_cast_fp16")]; tensor var_3276_to_fp16 = const()[name = tensor("op_3276_to_fp16"), val = tensor(0x1p-1)]; tensor var_3277_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3276_to_fp16)[name = tensor("op_3277_cast_fp16")]; tensor input_cast_fp16 = add(x = input_897_cast_fp16, y = var_3277_cast_fp16)[name = tensor("input_cast_fp16")]; tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; tensor encoder_module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215949184)))]; tensor encoder_module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("encoder_module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215950272)))]; tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = encoder_module_layers_16_norm_out_bias_to_fp16, epsilon = var_156_to_fp16, gamma = encoder_module_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; tensor obj_3_perm_0 = const()[name = tensor("obj_3_perm_0"), val = tensor([0, 2, 1])]; tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor obj_3_cast_fp16 = transpose(perm = obj_3_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; tensor encoder = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_175")]; } -> (encoder, encoder_length); }