diff --git "a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil" "b/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil" deleted file mode 100644--- "a/convert/parakeet-tdt-ctc-110m/coreml/compiled_models/Encoder.mlmodelc/model.mil" +++ /dev/null @@ -1,2690 +0,0 @@ -program(1.0) -[buildInfo = dict, tensor>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.9.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] -{ - func main(tensor mel_features, tensor mel_length) { - tensor var_23 = const()[name = tensor("op_23"), val = tensor(-1)]; - tensor x_1_perm_0 = const()[name = tensor("x_1_perm_0"), val = tensor([0, 2, 1])]; - tensor mel_features_to_fp16_dtype_0 = const()[name = tensor("mel_features_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor tensor_1_axes_0 = const()[name = tensor("tensor_1_axes_0"), val = tensor([1])]; - tensor mel_features_to_fp16 = cast(dtype = mel_features_to_fp16_dtype_0, x = mel_features)[name = tensor("cast_182")]; - tensor x_1_cast_fp16 = transpose(perm = x_1_perm_0, x = mel_features_to_fp16)[name = tensor("transpose_207")]; - tensor tensor_1_cast_fp16 = expand_dims(axes = tensor_1_axes_0, x = x_1_cast_fp16)[name = tensor("tensor_1_cast_fp16")]; - tensor expand_dims_0 = const()[name = tensor("expand_dims_0"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750, 751, 752, 753, 754, 755, 756, 757, 758, 759, 760, 761, 762, 763, 764, 765, 766, 767, 768, 769, 770, 771, 772, 773, 774, 775, 776, 777, 778, 779, 780, 781, 782, 783, 784, 785, 786, 787, 788, 789, 790, 791, 792, 793, 794, 795, 796, 797, 798, 799, 800, 801, 802, 803, 804, 805, 806, 807, 808, 809, 810, 811, 812, 813, 814, 815, 816, 817, 818, 819, 820, 821, 822, 823, 824, 825, 826, 827, 828, 829, 830, 831, 832, 833, 834, 835, 836, 837, 838, 839, 840, 841, 842, 843, 844, 845, 846, 847, 848, 849, 850, 851, 852, 853, 854, 855, 856, 857, 858, 859, 860, 861, 862, 863, 864, 865, 866, 867, 868, 869, 870, 871, 872, 873, 874, 875, 876, 877, 878, 879, 880, 881, 882, 883, 884, 885, 886, 887, 888, 889, 890, 891, 892, 893, 894, 895, 896, 897, 898, 899, 900, 901, 902, 903, 904, 905, 906, 907, 908, 909, 910, 911, 912, 913, 914, 915, 916, 917, 918, 919, 920, 921, 922, 923, 924, 925, 926, 927, 928, 929, 930, 931, 932, 933, 934, 935, 936, 937, 938, 939, 940, 941, 942, 943, 944, 945, 946, 947, 948, 949, 950, 951, 952, 953, 954, 955, 956, 957, 958, 959, 960, 961, 962, 963, 964, 965, 966, 967, 968, 969, 970, 971, 972, 973, 974, 975, 976, 977, 978, 979, 980, 981, 982, 983, 984, 985, 986, 987, 988, 989, 990, 991, 992, 993, 994, 995, 996, 997, 998, 999, 1000, 1001, 1002, 1003, 1004, 1005, 1006, 1007, 1008, 1009, 1010, 1011, 1012, 1013, 1014, 1015, 1016, 1017, 1018, 1019, 1020, 1021, 1022, 1023, 1024, 1025, 1026, 1027, 1028, 1029, 1030, 1031, 1032, 1033, 1034, 1035, 1036, 1037, 1038, 1039, 1040, 1041, 1042, 1043, 1044, 1045, 1046, 1047, 1048, 1049, 1050, 1051, 1052, 1053, 1054, 1055, 1056, 1057, 1058, 1059, 1060, 1061, 1062, 1063, 1064, 1065, 1066, 1067, 1068, 1069, 1070, 1071, 1072, 1073, 1074, 1075, 1076, 1077, 1078, 1079, 1080, 1081, 1082, 1083, 1084, 1085, 1086, 1087, 1088, 1089, 1090, 1091, 1092, 1093, 1094, 1095, 1096, 1097, 1098, 1099, 1100, 1101, 1102, 1103, 1104, 1105, 1106, 1107, 1108, 1109, 1110, 1111, 1112, 1113, 1114, 1115, 1116, 1117, 1118, 1119, 1120, 1121, 1122, 1123, 1124, 1125, 1126, 1127, 1128, 1129, 1130, 1131, 1132, 1133, 1134, 1135, 1136, 1137, 1138, 1139, 1140, 1141, 1142, 1143, 1144, 1145, 1146, 1147, 1148, 1149, 1150, 1151, 1152, 1153, 1154, 1155, 1156, 1157, 1158, 1159, 1160, 1161, 1162, 1163, 1164, 1165, 1166, 1167, 1168, 1169, 1170, 1171, 1172, 1173, 1174, 1175, 1176, 1177, 1178, 1179, 1180, 1181, 1182, 1183, 1184, 1185, 1186, 1187, 1188, 1189, 1190, 1191, 1192, 1193, 1194, 1195, 1196, 1197, 1198, 1199, 1200, 1201, 1202, 1203, 1204, 1205, 1206, 1207, 1208, 1209, 1210, 1211, 1212, 1213, 1214, 1215, 1216, 1217, 1218, 1219, 1220, 1221, 1222, 1223, 1224, 1225, 1226, 1227, 1228, 1229, 1230, 1231, 1232, 1233, 1234, 1235, 1236, 1237, 1238, 1239, 1240, 1241, 1242, 1243, 1244, 1245, 1246, 1247, 1248, 1249, 1250, 1251, 1252, 1253, 1254, 1255, 1256, 1257, 1258, 1259, 1260, 1261, 1262, 1263, 1264, 1265, 1266, 1267, 1268, 1269, 1270, 1271, 1272, 1273, 1274, 1275, 1276, 1277, 1278, 1279, 1280, 1281, 1282, 1283, 1284, 1285, 1286, 1287, 1288, 1289, 1290, 1291, 1292, 1293, 1294, 1295, 1296, 1297, 1298, 1299, 1300, 1301, 1302, 1303, 1304, 1305, 1306, 1307, 1308, 1309, 1310, 1311, 1312, 1313, 1314, 1315, 1316, 1317, 1318, 1319, 1320, 1321, 1322, 1323, 1324, 1325, 1326, 1327, 1328, 1329, 1330, 1331, 1332, 1333, 1334, 1335, 1336, 1337, 1338, 1339, 1340, 1341, 1342, 1343, 1344, 1345, 1346, 1347, 1348, 1349, 1350, 1351, 1352, 1353, 1354, 1355, 1356, 1357, 1358, 1359, 1360, 1361, 1362, 1363, 1364, 1365, 1366, 1367, 1368, 1369, 1370, 1371, 1372, 1373, 1374, 1375, 1376, 1377, 1378, 1379, 1380, 1381, 1382, 1383, 1384, 1385, 1386, 1387, 1388, 1389, 1390, 1391, 1392, 1393, 1394, 1395, 1396, 1397, 1398, 1399, 1400, 1401, 1402, 1403, 1404, 1405, 1406, 1407, 1408, 1409, 1410, 1411, 1412, 1413, 1414, 1415, 1416, 1417, 1418, 1419, 1420, 1421, 1422, 1423, 1424, 1425, 1426, 1427, 1428, 1429, 1430, 1431, 1432, 1433, 1434, 1435, 1436, 1437, 1438, 1439, 1440, 1441, 1442, 1443, 1444, 1445, 1446, 1447, 1448, 1449, 1450, 1451, 1452, 1453, 1454, 1455, 1456, 1457, 1458, 1459, 1460, 1461, 1462, 1463, 1464, 1465, 1466, 1467, 1468, 1469, 1470, 1471, 1472, 1473, 1474, 1475, 1476, 1477, 1478, 1479, 1480, 1481, 1482, 1483, 1484, 1485, 1486, 1487, 1488, 1489, 1490, 1491, 1492, 1493, 1494, 1495, 1496, 1497, 1498, 1499, 1500]])]; - tensor var_87_axes_0 = const()[name = tensor("op_87_axes_0"), val = tensor([1])]; - tensor var_87 = expand_dims(axes = var_87_axes_0, x = mel_length)[name = tensor("op_87")]; - tensor time_mask_1 = less(x = expand_dims_0, y = var_87)[name = tensor("time_mask_1")]; - tensor var_89_axes_0 = const()[name = tensor("op_89_axes_0"), val = tensor([-1])]; - tensor var_89 = expand_dims(axes = var_89_axes_0, x = time_mask_1)[name = tensor("op_89")]; - tensor var_91_reps_0 = const()[name = tensor("op_91_reps_0"), val = tensor([1, 1, 80])]; - tensor var_91 = tile(reps = var_91_reps_0, x = var_89)[name = tensor("op_91")]; - tensor var_97_axes_0 = const()[name = tensor("op_97_axes_0"), val = tensor([1])]; - tensor cast_3_to_fp16_dtype_0 = const()[name = tensor("cast_3_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_91_to_fp16 = cast(dtype = cast_3_to_fp16_dtype_0, x = var_91)[name = tensor("cast_181")]; - tensor var_97_cast_fp16 = expand_dims(axes = var_97_axes_0, x = var_91_to_fp16)[name = tensor("op_97_cast_fp16")]; - tensor input_1_cast_fp16 = mul(x = tensor_1_cast_fp16, y = var_97_cast_fp16)[name = tensor("input_1_cast_fp16")]; - tensor tensor_3_pad_type_0 = const()[name = tensor("tensor_3_pad_type_0"), val = tensor("custom")]; - tensor tensor_3_pad_0 = const()[name = tensor("tensor_3_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_3_strides_0 = const()[name = tensor("tensor_3_strides_0"), val = tensor([2, 2])]; - tensor tensor_3_dilations_0 = const()[name = tensor("tensor_3_dilations_0"), val = tensor([1, 1])]; - tensor tensor_3_groups_0 = const()[name = tensor("tensor_3_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_0_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; - tensor module_pre_encode_conv_0_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_0_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4736)))]; - tensor tensor_3_cast_fp16 = conv(bias = module_pre_encode_conv_0_bias_to_fp16, dilations = tensor_3_dilations_0, groups = tensor_3_groups_0, pad = tensor_3_pad_0, pad_type = tensor_3_pad_type_0, strides = tensor_3_strides_0, weight = module_pre_encode_conv_0_weight_to_fp16, x = input_1_cast_fp16)[name = tensor("tensor_3_cast_fp16")]; - tensor cast_1_to_fp16_dtype_0 = const()[name = tensor("cast_1_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_108_promoted_to_fp16 = const()[name = tensor("op_108_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor mel_length_to_fp16 = cast(dtype = cast_1_to_fp16_dtype_0, x = mel_length)[name = tensor("cast_180")]; - tensor var_109_cast_fp16 = add(x = mel_length_to_fp16, y = var_108_promoted_to_fp16)[name = tensor("op_109_cast_fp16")]; - tensor var_110_promoted_to_fp16 = const()[name = tensor("op_110_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_111_cast_fp16 = add(x = var_109_cast_fp16, y = var_110_promoted_to_fp16)[name = tensor("op_111_cast_fp16")]; - tensor var_112_promoted_to_fp16 = const()[name = tensor("op_112_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_113_cast_fp16 = sub(x = var_111_cast_fp16, y = var_112_promoted_to_fp16)[name = tensor("op_113_cast_fp16")]; - tensor var_21_promoted_to_fp16 = const()[name = tensor("op_21_promoted_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_0_cast_fp16 = floor_div(x = var_113_cast_fp16, y = var_21_promoted_to_fp16)[name = tensor("floor_div_0_cast_fp16")]; - tensor var_115_promoted_to_fp16 = const()[name = tensor("op_115_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_3_cast_fp16 = add(x = floor_div_0_cast_fp16, y = var_115_promoted_to_fp16)[name = tensor("current_lengths_3_cast_fp16")]; - tensor cast_4_dtype_0 = const()[name = tensor("cast_4_dtype_0"), val = tensor("int32")]; - tensor expand_dims_1 = const()[name = tensor("expand_dims_1"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375, 376, 377, 378, 379, 380, 381, 382, 383, 384, 385, 386, 387, 388, 389, 390, 391, 392, 393, 394, 395, 396, 397, 398, 399, 400, 401, 402, 403, 404, 405, 406, 407, 408, 409, 410, 411, 412, 413, 414, 415, 416, 417, 418, 419, 420, 421, 422, 423, 424, 425, 426, 427, 428, 429, 430, 431, 432, 433, 434, 435, 436, 437, 438, 439, 440, 441, 442, 443, 444, 445, 446, 447, 448, 449, 450, 451, 452, 453, 454, 455, 456, 457, 458, 459, 460, 461, 462, 463, 464, 465, 466, 467, 468, 469, 470, 471, 472, 473, 474, 475, 476, 477, 478, 479, 480, 481, 482, 483, 484, 485, 486, 487, 488, 489, 490, 491, 492, 493, 494, 495, 496, 497, 498, 499, 500, 501, 502, 503, 504, 505, 506, 507, 508, 509, 510, 511, 512, 513, 514, 515, 516, 517, 518, 519, 520, 521, 522, 523, 524, 525, 526, 527, 528, 529, 530, 531, 532, 533, 534, 535, 536, 537, 538, 539, 540, 541, 542, 543, 544, 545, 546, 547, 548, 549, 550, 551, 552, 553, 554, 555, 556, 557, 558, 559, 560, 561, 562, 563, 564, 565, 566, 567, 568, 569, 570, 571, 572, 573, 574, 575, 576, 577, 578, 579, 580, 581, 582, 583, 584, 585, 586, 587, 588, 589, 590, 591, 592, 593, 594, 595, 596, 597, 598, 599, 600, 601, 602, 603, 604, 605, 606, 607, 608, 609, 610, 611, 612, 613, 614, 615, 616, 617, 618, 619, 620, 621, 622, 623, 624, 625, 626, 627, 628, 629, 630, 631, 632, 633, 634, 635, 636, 637, 638, 639, 640, 641, 642, 643, 644, 645, 646, 647, 648, 649, 650, 651, 652, 653, 654, 655, 656, 657, 658, 659, 660, 661, 662, 663, 664, 665, 666, 667, 668, 669, 670, 671, 672, 673, 674, 675, 676, 677, 678, 679, 680, 681, 682, 683, 684, 685, 686, 687, 688, 689, 690, 691, 692, 693, 694, 695, 696, 697, 698, 699, 700, 701, 702, 703, 704, 705, 706, 707, 708, 709, 710, 711, 712, 713, 714, 715, 716, 717, 718, 719, 720, 721, 722, 723, 724, 725, 726, 727, 728, 729, 730, 731, 732, 733, 734, 735, 736, 737, 738, 739, 740, 741, 742, 743, 744, 745, 746, 747, 748, 749, 750]])]; - tensor var_124_axes_0 = const()[name = tensor("op_124_axes_0"), val = tensor([1])]; - tensor current_lengths_3_cast_fp16_to_int32 = cast(dtype = cast_4_dtype_0, x = current_lengths_3_cast_fp16)[name = tensor("cast_179")]; - tensor var_124 = expand_dims(axes = var_124_axes_0, x = current_lengths_3_cast_fp16_to_int32)[name = tensor("op_124")]; - tensor time_mask_3 = less(x = expand_dims_1, y = var_124)[name = tensor("time_mask_3")]; - tensor var_126_axes_0 = const()[name = tensor("op_126_axes_0"), val = tensor([-1])]; - tensor var_126 = expand_dims(axes = var_126_axes_0, x = time_mask_3)[name = tensor("op_126")]; - tensor var_128_reps_0 = const()[name = tensor("op_128_reps_0"), val = tensor([1, 1, 40])]; - tensor var_128 = tile(reps = var_128_reps_0, x = var_126)[name = tensor("op_128")]; - tensor var_134_axes_0 = const()[name = tensor("op_134_axes_0"), val = tensor([1])]; - tensor cast_5_to_fp16_dtype_0 = const()[name = tensor("cast_5_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_128_to_fp16 = cast(dtype = cast_5_to_fp16_dtype_0, x = var_128)[name = tensor("cast_178")]; - tensor var_134_cast_fp16 = expand_dims(axes = var_134_axes_0, x = var_128_to_fp16)[name = tensor("op_134_cast_fp16")]; - tensor expanded_mask_3_reps_0 = const()[name = tensor("expanded_mask_3_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_3_cast_fp16 = tile(reps = expanded_mask_3_reps_0, x = var_134_cast_fp16)[name = tensor("expanded_mask_3_cast_fp16")]; - tensor input_3_cast_fp16 = mul(x = tensor_3_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_3_cast_fp16")]; - tensor tensor_5_cast_fp16 = relu(x = input_3_cast_fp16)[name = tensor("tensor_5_cast_fp16")]; - tensor input_5_cast_fp16 = mul(x = tensor_5_cast_fp16, y = expanded_mask_3_cast_fp16)[name = tensor("input_5_cast_fp16")]; - tensor tensor_7_pad_type_0 = const()[name = tensor("tensor_7_pad_type_0"), val = tensor("custom")]; - tensor tensor_7_pad_0 = const()[name = tensor("tensor_7_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_7_strides_0 = const()[name = tensor("tensor_7_strides_0"), val = tensor([2, 2])]; - tensor tensor_7_groups_0 = const()[name = tensor("tensor_7_groups_0"), val = tensor(256)]; - tensor tensor_7_dilations_0 = const()[name = tensor("tensor_7_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_2_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5312)))]; - tensor module_pre_encode_conv_2_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9984)))]; - tensor tensor_7_cast_fp16 = conv(bias = module_pre_encode_conv_2_bias_to_fp16, dilations = tensor_7_dilations_0, groups = tensor_7_groups_0, pad = tensor_7_pad_0, pad_type = tensor_7_pad_type_0, strides = tensor_7_strides_0, weight = module_pre_encode_conv_2_weight_to_fp16, x = input_5_cast_fp16)[name = tensor("tensor_7_cast_fp16")]; - tensor var_154_promoted_to_fp16 = const()[name = tensor("op_154_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_155_cast_fp16 = add(x = current_lengths_3_cast_fp16, y = var_154_promoted_to_fp16)[name = tensor("op_155_cast_fp16")]; - tensor var_156_promoted_to_fp16 = const()[name = tensor("op_156_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_157_cast_fp16 = add(x = var_155_cast_fp16, y = var_156_promoted_to_fp16)[name = tensor("op_157_cast_fp16")]; - tensor var_158_promoted_to_fp16 = const()[name = tensor("op_158_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_159_cast_fp16 = sub(x = var_157_cast_fp16, y = var_158_promoted_to_fp16)[name = tensor("op_159_cast_fp16")]; - tensor var_21_promoted_1_to_fp16 = const()[name = tensor("op_21_promoted_1_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_1_cast_fp16 = floor_div(x = var_159_cast_fp16, y = var_21_promoted_1_to_fp16)[name = tensor("floor_div_1_cast_fp16")]; - tensor var_161_promoted_to_fp16 = const()[name = tensor("op_161_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_5_cast_fp16 = add(x = floor_div_1_cast_fp16, y = var_161_promoted_to_fp16)[name = tensor("current_lengths_5_cast_fp16")]; - tensor cast_6_dtype_0 = const()[name = tensor("cast_6_dtype_0"), val = tensor("int32")]; - tensor expand_dims_2 = const()[name = tensor("expand_dims_2"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187, 188, 189, 190, 191, 192, 193, 194, 195, 196, 197, 198, 199, 200, 201, 202, 203, 204, 205, 206, 207, 208, 209, 210, 211, 212, 213, 214, 215, 216, 217, 218, 219, 220, 221, 222, 223, 224, 225, 226, 227, 228, 229, 230, 231, 232, 233, 234, 235, 236, 237, 238, 239, 240, 241, 242, 243, 244, 245, 246, 247, 248, 249, 250, 251, 252, 253, 254, 255, 256, 257, 258, 259, 260, 261, 262, 263, 264, 265, 266, 267, 268, 269, 270, 271, 272, 273, 274, 275, 276, 277, 278, 279, 280, 281, 282, 283, 284, 285, 286, 287, 288, 289, 290, 291, 292, 293, 294, 295, 296, 297, 298, 299, 300, 301, 302, 303, 304, 305, 306, 307, 308, 309, 310, 311, 312, 313, 314, 315, 316, 317, 318, 319, 320, 321, 322, 323, 324, 325, 326, 327, 328, 329, 330, 331, 332, 333, 334, 335, 336, 337, 338, 339, 340, 341, 342, 343, 344, 345, 346, 347, 348, 349, 350, 351, 352, 353, 354, 355, 356, 357, 358, 359, 360, 361, 362, 363, 364, 365, 366, 367, 368, 369, 370, 371, 372, 373, 374, 375]])]; - tensor var_170_axes_0 = const()[name = tensor("op_170_axes_0"), val = tensor([1])]; - tensor current_lengths_5_cast_fp16_to_int32 = cast(dtype = cast_6_dtype_0, x = current_lengths_5_cast_fp16)[name = tensor("cast_177")]; - tensor var_170 = expand_dims(axes = var_170_axes_0, x = current_lengths_5_cast_fp16_to_int32)[name = tensor("op_170")]; - tensor time_mask_5 = less(x = expand_dims_2, y = var_170)[name = tensor("time_mask_5")]; - tensor var_172_axes_0 = const()[name = tensor("op_172_axes_0"), val = tensor([-1])]; - tensor var_172 = expand_dims(axes = var_172_axes_0, x = time_mask_5)[name = tensor("op_172")]; - tensor var_174_reps_0 = const()[name = tensor("op_174_reps_0"), val = tensor([1, 1, 20])]; - tensor var_174 = tile(reps = var_174_reps_0, x = var_172)[name = tensor("op_174")]; - tensor var_180_axes_0 = const()[name = tensor("op_180_axes_0"), val = tensor([1])]; - tensor cast_7_to_fp16_dtype_0 = const()[name = tensor("cast_7_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_174_to_fp16 = cast(dtype = cast_7_to_fp16_dtype_0, x = var_174)[name = tensor("cast_176")]; - tensor var_180_cast_fp16 = expand_dims(axes = var_180_axes_0, x = var_174_to_fp16)[name = tensor("op_180_cast_fp16")]; - tensor expanded_mask_7_reps_0 = const()[name = tensor("expanded_mask_7_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_7_cast_fp16 = tile(reps = expanded_mask_7_reps_0, x = var_180_cast_fp16)[name = tensor("expanded_mask_7_cast_fp16")]; - tensor input_7_cast_fp16 = mul(x = tensor_7_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_7_cast_fp16")]; - tensor tensor_9_pad_type_0 = const()[name = tensor("tensor_9_pad_type_0"), val = tensor("valid")]; - tensor tensor_9_strides_0 = const()[name = tensor("tensor_9_strides_0"), val = tensor([1, 1])]; - tensor tensor_9_pad_0 = const()[name = tensor("tensor_9_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_9_dilations_0 = const()[name = tensor("tensor_9_dilations_0"), val = tensor([1, 1])]; - tensor tensor_9_groups_0 = const()[name = tensor("tensor_9_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_3_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10560)))]; - tensor module_pre_encode_conv_3_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_3_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(141696)))]; - tensor tensor_9_cast_fp16 = conv(bias = module_pre_encode_conv_3_bias_to_fp16, dilations = tensor_9_dilations_0, groups = tensor_9_groups_0, pad = tensor_9_pad_0, pad_type = tensor_9_pad_type_0, strides = tensor_9_strides_0, weight = module_pre_encode_conv_3_weight_to_fp16, x = input_7_cast_fp16)[name = tensor("tensor_9_cast_fp16")]; - tensor input_9_cast_fp16 = mul(x = tensor_9_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_9_cast_fp16")]; - tensor tensor_11_cast_fp16 = relu(x = input_9_cast_fp16)[name = tensor("tensor_11_cast_fp16")]; - tensor input_11_cast_fp16 = mul(x = tensor_11_cast_fp16, y = expanded_mask_7_cast_fp16)[name = tensor("input_11_cast_fp16")]; - tensor tensor_13_pad_type_0 = const()[name = tensor("tensor_13_pad_type_0"), val = tensor("custom")]; - tensor tensor_13_pad_0 = const()[name = tensor("tensor_13_pad_0"), val = tensor([1, 1, 1, 1])]; - tensor tensor_13_strides_0 = const()[name = tensor("tensor_13_strides_0"), val = tensor([2, 2])]; - tensor tensor_13_groups_0 = const()[name = tensor("tensor_13_groups_0"), val = tensor(256)]; - tensor tensor_13_dilations_0 = const()[name = tensor("tensor_13_dilations_0"), val = tensor([1, 1])]; - tensor module_pre_encode_conv_5_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142272)))]; - tensor module_pre_encode_conv_5_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_5_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146944)))]; - tensor tensor_13_cast_fp16 = conv(bias = module_pre_encode_conv_5_bias_to_fp16, dilations = tensor_13_dilations_0, groups = tensor_13_groups_0, pad = tensor_13_pad_0, pad_type = tensor_13_pad_type_0, strides = tensor_13_strides_0, weight = module_pre_encode_conv_5_weight_to_fp16, x = input_11_cast_fp16)[name = tensor("tensor_13_cast_fp16")]; - tensor var_215_promoted_to_fp16 = const()[name = tensor("op_215_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_216_cast_fp16 = add(x = current_lengths_5_cast_fp16, y = var_215_promoted_to_fp16)[name = tensor("op_216_cast_fp16")]; - tensor var_217_promoted_to_fp16 = const()[name = tensor("op_217_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor var_218_cast_fp16 = add(x = var_216_cast_fp16, y = var_217_promoted_to_fp16)[name = tensor("op_218_cast_fp16")]; - tensor var_219_promoted_to_fp16 = const()[name = tensor("op_219_promoted_to_fp16"), val = tensor(0x1.8p+1)]; - tensor var_220_cast_fp16 = sub(x = var_218_cast_fp16, y = var_219_promoted_to_fp16)[name = tensor("op_220_cast_fp16")]; - tensor var_21_promoted_2_to_fp16 = const()[name = tensor("op_21_promoted_2_to_fp16"), val = tensor(0x1p+1)]; - tensor floor_div_2_cast_fp16 = floor_div(x = var_220_cast_fp16, y = var_21_promoted_2_to_fp16)[name = tensor("floor_div_2_cast_fp16")]; - tensor var_222_promoted_to_fp16 = const()[name = tensor("op_222_promoted_to_fp16"), val = tensor(0x1p+0)]; - tensor current_lengths_cast_fp16 = add(x = floor_div_2_cast_fp16, y = var_222_promoted_to_fp16)[name = tensor("current_lengths_cast_fp16")]; - tensor cast_8_dtype_0 = const()[name = tensor("cast_8_dtype_0"), val = tensor("int32")]; - tensor expand_dims_3 = const()[name = tensor("expand_dims_3"), val = tensor([[0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26, 27, 28, 29, 30, 31, 32, 33, 34, 35, 36, 37, 38, 39, 40, 41, 42, 43, 44, 45, 46, 47, 48, 49, 50, 51, 52, 53, 54, 55, 56, 57, 58, 59, 60, 61, 62, 63, 64, 65, 66, 67, 68, 69, 70, 71, 72, 73, 74, 75, 76, 77, 78, 79, 80, 81, 82, 83, 84, 85, 86, 87, 88, 89, 90, 91, 92, 93, 94, 95, 96, 97, 98, 99, 100, 101, 102, 103, 104, 105, 106, 107, 108, 109, 110, 111, 112, 113, 114, 115, 116, 117, 118, 119, 120, 121, 122, 123, 124, 125, 126, 127, 128, 129, 130, 131, 132, 133, 134, 135, 136, 137, 138, 139, 140, 141, 142, 143, 144, 145, 146, 147, 148, 149, 150, 151, 152, 153, 154, 155, 156, 157, 158, 159, 160, 161, 162, 163, 164, 165, 166, 167, 168, 169, 170, 171, 172, 173, 174, 175, 176, 177, 178, 179, 180, 181, 182, 183, 184, 185, 186, 187]])]; - tensor var_231_axes_0 = const()[name = tensor("op_231_axes_0"), val = tensor([1])]; - tensor current_lengths_cast_fp16_to_int32 = cast(dtype = cast_8_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_175")]; - tensor var_231 = expand_dims(axes = var_231_axes_0, x = current_lengths_cast_fp16_to_int32)[name = tensor("op_231")]; - tensor time_mask = less(x = expand_dims_3, y = var_231)[name = tensor("time_mask")]; - tensor var_233_axes_0 = const()[name = tensor("op_233_axes_0"), val = tensor([-1])]; - tensor var_233 = expand_dims(axes = var_233_axes_0, x = time_mask)[name = tensor("op_233")]; - tensor var_235_reps_0 = const()[name = tensor("op_235_reps_0"), val = tensor([1, 1, 10])]; - tensor var_235 = tile(reps = var_235_reps_0, x = var_233)[name = tensor("op_235")]; - tensor var_241_axes_0 = const()[name = tensor("op_241_axes_0"), val = tensor([1])]; - tensor cast_9_to_fp16_dtype_0 = const()[name = tensor("cast_9_to_fp16_dtype_0"), val = tensor("fp16")]; - tensor var_235_to_fp16 = cast(dtype = cast_9_to_fp16_dtype_0, x = var_235)[name = tensor("cast_174")]; - tensor var_241_cast_fp16 = expand_dims(axes = var_241_axes_0, x = var_235_to_fp16)[name = tensor("op_241_cast_fp16")]; - tensor expanded_mask_13_reps_0 = const()[name = tensor("expanded_mask_13_reps_0"), val = tensor([1, 256, 1, 1])]; - tensor expanded_mask_13_cast_fp16 = tile(reps = expanded_mask_13_reps_0, x = var_241_cast_fp16)[name = tensor("expanded_mask_13_cast_fp16")]; - tensor input_13_cast_fp16 = mul(x = tensor_13_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_13_cast_fp16")]; - tensor tensor_15_pad_type_0 = const()[name = tensor("tensor_15_pad_type_0"), val = tensor("valid")]; - tensor tensor_15_strides_0 = const()[name = tensor("tensor_15_strides_0"), val = tensor([1, 1])]; - tensor tensor_15_pad_0 = const()[name = tensor("tensor_15_pad_0"), val = tensor([0, 0, 0, 0])]; - tensor tensor_15_dilations_0 = const()[name = tensor("tensor_15_dilations_0"), val = tensor([1, 1])]; - tensor tensor_15_groups_0 = const()[name = tensor("tensor_15_groups_0"), val = tensor(1)]; - tensor module_pre_encode_conv_6_weight_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147520)))]; - tensor module_pre_encode_conv_6_bias_to_fp16 = const()[name = tensor("module_pre_encode_conv_6_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(278656)))]; - tensor tensor_15_cast_fp16 = conv(bias = module_pre_encode_conv_6_bias_to_fp16, dilations = tensor_15_dilations_0, groups = tensor_15_groups_0, pad = tensor_15_pad_0, pad_type = tensor_15_pad_type_0, strides = tensor_15_strides_0, weight = module_pre_encode_conv_6_weight_to_fp16, x = input_13_cast_fp16)[name = tensor("tensor_15_cast_fp16")]; - tensor input_15_cast_fp16 = mul(x = tensor_15_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("input_15_cast_fp16")]; - tensor tensor_cast_fp16 = relu(x = input_15_cast_fp16)[name = tensor("tensor_cast_fp16")]; - tensor x_3_cast_fp16 = mul(x = tensor_cast_fp16, y = expanded_mask_13_cast_fp16)[name = tensor("x_3_cast_fp16")]; - tensor var_275_perm_0 = const()[name = tensor("op_275_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_276 = const()[name = tensor("op_276"), val = tensor([1, 188, -1])]; - tensor var_275_cast_fp16 = transpose(perm = var_275_perm_0, x = x_3_cast_fp16)[name = tensor("transpose_206")]; - tensor input_17_cast_fp16 = reshape(shape = var_276, x = var_275_cast_fp16)[name = tensor("input_17_cast_fp16")]; - tensor module_pre_encode_out_weight_to_fp16 = const()[name = tensor("module_pre_encode_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(279232)))]; - tensor module_pre_encode_out_bias_to_fp16 = const()[name = tensor("module_pre_encode_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2900736)))]; - tensor linear_0_cast_fp16 = linear(bias = module_pre_encode_out_bias_to_fp16, weight = module_pre_encode_out_weight_to_fp16, x = input_17_cast_fp16)[name = tensor("linear_0_cast_fp16")]; - tensor cast_12_dtype_0 = const()[name = tensor("cast_12_dtype_0"), val = tensor("int32")]; - tensor var_314_axes_0 = const()[name = tensor("op_314_axes_0"), val = tensor([-1])]; - tensor encoder_length = cast(dtype = cast_12_dtype_0, x = current_lengths_cast_fp16)[name = tensor("cast_173")]; - tensor var_314 = expand_dims(axes = var_314_axes_0, x = encoder_length)[name = tensor("op_314")]; - tensor pad_mask_1 = less(x = expand_dims_3, y = var_314)[name = tensor("pad_mask_1")]; - tensor var_316_axes_0 = const()[name = tensor("op_316_axes_0"), val = tensor([1])]; - tensor var_316 = expand_dims(axes = var_316_axes_0, x = pad_mask_1)[name = tensor("op_316")]; - tensor var_317 = const()[name = tensor("op_317"), val = tensor([1, 188, 1])]; - tensor pad_mask_for_att_mask_1 = tile(reps = var_317, x = var_316)[name = tensor("pad_mask_for_att_mask_1")]; - tensor var_319_perm_0 = const()[name = tensor("op_319_perm_0"), val = tensor([0, 2, 1])]; - tensor var_319 = transpose(perm = var_319_perm_0, x = pad_mask_for_att_mask_1)[name = tensor("transpose_205")]; - tensor pad_mask_for_att_mask = logical_and(x = pad_mask_for_att_mask_1, y = var_319)[name = tensor("pad_mask_for_att_mask")]; - tensor const_63 = const()[name = tensor("const_63"), val = tensor([[[true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true], [true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true, true]]])]; - tensor att_mask = logical_and(x = pad_mask_for_att_mask, y = const_63)[name = tensor("att_mask")]; - tensor mask_9 = logical_not(x = att_mask)[name = tensor("mask_9")]; - tensor pad_mask = logical_not(x = pad_mask_1)[name = tensor("pad_mask")]; - tensor input_21_axes_0 = const()[name = tensor("input_21_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2901824)))]; - tensor module_layers_0_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2902912)))]; - tensor var_9_to_fp16 = const()[name = tensor("op_9_to_fp16"), val = tensor(0x1.5p-17)]; - tensor input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = module_layers_0_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward1_weight_to_fp16, x = linear_0_cast_fp16)[name = tensor("input_21_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(2904000)))]; - tensor module_layers_0_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5001216)))]; - tensor linear_1_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear1_bias_to_fp16, weight = module_layers_0_feed_forward1_linear1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor("linear_1_cast_fp16")]; - tensor input_25_cast_fp16 = silu(x = linear_1_cast_fp16)[name = tensor("input_25_cast_fp16")]; - tensor module_layers_0_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(5005376)))]; - tensor module_layers_0_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7102592)))]; - tensor linear_2_cast_fp16 = linear(bias = module_layers_0_feed_forward1_linear2_bias_to_fp16, weight = module_layers_0_feed_forward1_linear2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor("linear_2_cast_fp16")]; - tensor var_352_to_fp16 = const()[name = tensor("op_352_to_fp16"), val = tensor(0x1p-1)]; - tensor var_353_cast_fp16 = mul(x = linear_2_cast_fp16, y = var_352_to_fp16)[name = tensor("op_353_cast_fp16")]; - tensor input_31_cast_fp16 = add(x = linear_0_cast_fp16, y = var_353_cast_fp16)[name = tensor("input_31_cast_fp16")]; - tensor query_1_axes_0 = const()[name = tensor("query_1_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7103680)))]; - tensor module_layers_0_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7104768)))]; - tensor query_1_cast_fp16 = layer_norm(axes = query_1_axes_0, beta = module_layers_0_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_self_att_weight_to_fp16, x = input_31_cast_fp16)[name = tensor("query_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7105856)))]; - tensor module_layers_0_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7630208)))]; - tensor linear_3_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_q_bias_to_fp16, weight = module_layers_0_self_attn_linear_q_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_3_cast_fp16")]; - tensor var_370 = const()[name = tensor("op_370"), val = tensor([1, -1, 8, 64])]; - tensor q_1_cast_fp16 = reshape(shape = var_370, x = linear_3_cast_fp16)[name = tensor("q_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7631296)))]; - tensor module_layers_0_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8155648)))]; - tensor linear_4_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_k_bias_to_fp16, weight = module_layers_0_self_attn_linear_k_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_4_cast_fp16")]; - tensor var_375 = const()[name = tensor("op_375"), val = tensor([1, -1, 8, 64])]; - tensor k_1_cast_fp16 = reshape(shape = var_375, x = linear_4_cast_fp16)[name = tensor("k_1_cast_fp16")]; - tensor module_layers_0_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8156736)))]; - tensor module_layers_0_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8681088)))]; - tensor linear_5_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_v_bias_to_fp16, weight = module_layers_0_self_attn_linear_v_weight_to_fp16, x = query_1_cast_fp16)[name = tensor("linear_5_cast_fp16")]; - tensor var_380 = const()[name = tensor("op_380"), val = tensor([1, -1, 8, 64])]; - tensor v_1_cast_fp16 = reshape(shape = var_380, x = linear_5_cast_fp16)[name = tensor("v_1_cast_fp16")]; - tensor value_3_perm_0 = const()[name = tensor("value_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_0_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8682176)))]; - tensor var_392_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_u_to_fp16)[name = tensor("op_392_cast_fp16")]; - tensor module_layers_0_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_0_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8683264)))]; - tensor var_394_cast_fp16 = add(x = q_1_cast_fp16, y = module_layers_0_self_attn_pos_bias_v_to_fp16)[name = tensor("op_394_cast_fp16")]; - tensor q_with_bias_v_1_perm_0 = const()[name = tensor("q_with_bias_v_1_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_7_transpose_x_0 = const()[name = tensor("x_7_transpose_x_0"), val = tensor(false)]; - tensor x_7_transpose_y_0 = const()[name = tensor("x_7_transpose_y_0"), val = tensor(false)]; - tensor var_396_to_fp16 = const()[name = tensor("op_396_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(8684352)))]; - tensor q_with_bias_v_1_cast_fp16 = transpose(perm = q_with_bias_v_1_perm_0, x = var_394_cast_fp16)[name = tensor("transpose_203")]; - tensor x_7_cast_fp16 = matmul(transpose_x = x_7_transpose_x_0, transpose_y = x_7_transpose_y_0, x = q_with_bias_v_1_cast_fp16, y = var_396_to_fp16)[name = tensor("x_7_cast_fp16")]; - tensor x_9_pad_0 = const()[name = tensor("x_9_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_9_mode_0 = const()[name = tensor("x_9_mode_0"), val = tensor("constant")]; - tensor const_70_to_fp16 = const()[name = tensor("const_70_to_fp16"), val = tensor(0x0p+0)]; - tensor x_9_cast_fp16 = pad(constant_val = const_70_to_fp16, mode = x_9_mode_0, pad = x_9_pad_0, x = x_7_cast_fp16)[name = tensor("x_9_cast_fp16")]; - tensor var_404 = const()[name = tensor("op_404"), val = tensor([1, 8, -1, 188])]; - tensor x_11_cast_fp16 = reshape(shape = var_404, x = x_9_cast_fp16)[name = tensor("x_11_cast_fp16")]; - tensor var_408_begin_0 = const()[name = tensor("op_408_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_408_end_0 = const()[name = tensor("op_408_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_408_end_mask_0 = const()[name = tensor("op_408_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_408_cast_fp16 = slice_by_index(begin = var_408_begin_0, end = var_408_end_0, end_mask = var_408_end_mask_0, x = x_11_cast_fp16)[name = tensor("op_408_cast_fp16")]; - tensor var_409 = const()[name = tensor("op_409"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_1_cast_fp16 = reshape(shape = var_409, x = var_408_cast_fp16)[name = tensor("matrix_bd_1_cast_fp16")]; - tensor matrix_ac_1_transpose_x_0 = const()[name = tensor("matrix_ac_1_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_1_transpose_y_0 = const()[name = tensor("matrix_ac_1_transpose_y_0"), val = tensor(false)]; - tensor transpose_51_perm_0 = const()[name = tensor("transpose_51_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_52_perm_0 = const()[name = tensor("transpose_52_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_52 = transpose(perm = transpose_52_perm_0, x = k_1_cast_fp16)[name = tensor("transpose_201")]; - tensor transpose_51 = transpose(perm = transpose_51_perm_0, x = var_392_cast_fp16)[name = tensor("transpose_202")]; - tensor matrix_ac_1_cast_fp16 = matmul(transpose_x = matrix_ac_1_transpose_x_0, transpose_y = matrix_ac_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor("matrix_ac_1_cast_fp16")]; - tensor matrix_bd_3_begin_0 = const()[name = tensor("matrix_bd_3_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_3_end_0 = const()[name = tensor("matrix_bd_3_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_3_end_mask_0 = const()[name = tensor("matrix_bd_3_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_3_cast_fp16 = slice_by_index(begin = matrix_bd_3_begin_0, end = matrix_bd_3_end_0, end_mask = matrix_bd_3_end_mask_0, x = matrix_bd_1_cast_fp16)[name = tensor("matrix_bd_3_cast_fp16")]; - tensor var_418_cast_fp16 = add(x = matrix_ac_1_cast_fp16, y = matrix_bd_3_cast_fp16)[name = tensor("op_418_cast_fp16")]; - tensor _inversed_scores_1_y_0_to_fp16 = const()[name = tensor("_inversed_scores_1_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_1_cast_fp16 = mul(x = var_418_cast_fp16, y = _inversed_scores_1_y_0_to_fp16)[name = tensor("_inversed_scores_1_cast_fp16")]; - tensor mask_11_axes_0 = const()[name = tensor("mask_11_axes_0"), val = tensor([1])]; - tensor mask_11 = expand_dims(axes = mask_11_axes_0, x = mask_9)[name = tensor("mask_11")]; - tensor var_12_to_fp16 = const()[name = tensor("op_12_to_fp16"), val = tensor(-0x1.388p+13)]; - tensor scores_3_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_1_cast_fp16, cond = mask_11)[name = tensor("scores_3_cast_fp16")]; - tensor var_424_cast_fp16 = softmax(axis = var_23, x = scores_3_cast_fp16)[name = tensor("op_424_cast_fp16")]; - tensor var_11_to_fp16 = const()[name = tensor("op_11_to_fp16"), val = tensor(0x0p+0)]; - tensor input_33_cast_fp16 = select(a = var_11_to_fp16, b = var_424_cast_fp16, cond = mask_11)[name = tensor("input_33_cast_fp16")]; - tensor x_13_transpose_x_0 = const()[name = tensor("x_13_transpose_x_0"), val = tensor(false)]; - tensor x_13_transpose_y_0 = const()[name = tensor("x_13_transpose_y_0"), val = tensor(false)]; - tensor value_3_cast_fp16 = transpose(perm = value_3_perm_0, x = v_1_cast_fp16)[name = tensor("transpose_204")]; - tensor x_13_cast_fp16 = matmul(transpose_x = x_13_transpose_x_0, transpose_y = x_13_transpose_y_0, x = input_33_cast_fp16, y = value_3_cast_fp16)[name = tensor("x_13_cast_fp16")]; - tensor var_428_perm_0 = const()[name = tensor("op_428_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_429 = const()[name = tensor("op_429"), val = tensor([1, -1, 512])]; - tensor var_428_cast_fp16 = transpose(perm = var_428_perm_0, x = x_13_cast_fp16)[name = tensor("transpose_200")]; - tensor input_35_cast_fp16 = reshape(shape = var_429, x = var_428_cast_fp16)[name = tensor("input_35_cast_fp16")]; - tensor module_layers_0_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9068416)))]; - tensor module_layers_0_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_0_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9592768)))]; - tensor linear_7_cast_fp16 = linear(bias = module_layers_0_self_attn_linear_out_bias_to_fp16, weight = module_layers_0_self_attn_linear_out_weight_to_fp16, x = input_35_cast_fp16)[name = tensor("linear_7_cast_fp16")]; - tensor input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_7_cast_fp16)[name = tensor("input_39_cast_fp16")]; - tensor x_17_axes_0 = const()[name = tensor("x_17_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9593856)))]; - tensor module_layers_0_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9594944)))]; - tensor x_17_cast_fp16 = layer_norm(axes = x_17_axes_0, beta = module_layers_0_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_conv_weight_to_fp16, x = input_39_cast_fp16)[name = tensor("x_17_cast_fp16")]; - tensor input_41_perm_0 = const()[name = tensor("input_41_perm_0"), val = tensor([0, 2, 1])]; - tensor input_43_pad_type_0 = const()[name = tensor("input_43_pad_type_0"), val = tensor("valid")]; - tensor input_43_strides_0 = const()[name = tensor("input_43_strides_0"), val = tensor([1])]; - tensor input_43_pad_0 = const()[name = tensor("input_43_pad_0"), val = tensor([0, 0])]; - tensor input_43_dilations_0 = const()[name = tensor("input_43_dilations_0"), val = tensor([1])]; - tensor input_43_groups_0 = const()[name = tensor("input_43_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(9596032)))]; - tensor module_layers_0_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10644672)))]; - tensor input_41_cast_fp16 = transpose(perm = input_41_perm_0, x = x_17_cast_fp16)[name = tensor("transpose_199")]; - tensor input_43_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv1_bias_to_fp16, dilations = input_43_dilations_0, groups = input_43_groups_0, pad = input_43_pad_0, pad_type = input_43_pad_type_0, strides = input_43_strides_0, weight = module_layers_0_conv_pointwise_conv1_weight_to_fp16, x = input_41_cast_fp16)[name = tensor("input_43_cast_fp16")]; - tensor x_19_split_num_splits_0 = const()[name = tensor("x_19_split_num_splits_0"), val = tensor(2)]; - tensor x_19_split_axis_0 = const()[name = tensor("x_19_split_axis_0"), val = tensor(1)]; - tensor x_19_split_cast_fp16_0, tensor x_19_split_cast_fp16_1 = split(axis = x_19_split_axis_0, num_splits = x_19_split_num_splits_0, x = input_43_cast_fp16)[name = tensor("x_19_split_cast_fp16")]; - tensor x_19_split_1_sigmoid_cast_fp16 = sigmoid(x = x_19_split_cast_fp16_1)[name = tensor("x_19_split_1_sigmoid_cast_fp16")]; - tensor x_19_cast_fp16 = mul(x = x_19_split_cast_fp16_0, y = x_19_split_1_sigmoid_cast_fp16)[name = tensor("x_19_cast_fp16")]; - tensor var_453_axes_0 = const()[name = tensor("op_453_axes_0"), val = tensor([1])]; - tensor var_453 = expand_dims(axes = var_453_axes_0, x = pad_mask)[name = tensor("op_453")]; - tensor input_45_cast_fp16 = select(a = var_11_to_fp16, b = x_19_cast_fp16, cond = var_453)[name = tensor("input_45_cast_fp16")]; - tensor input_47_pad_0 = const()[name = tensor("input_47_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_47_mode_0 = const()[name = tensor("input_47_mode_0"), val = tensor("constant")]; - tensor const_73_to_fp16 = const()[name = tensor("const_73_to_fp16"), val = tensor(0x0p+0)]; - tensor input_47_cast_fp16 = pad(constant_val = const_73_to_fp16, mode = input_47_mode_0, pad = input_47_pad_0, x = input_45_cast_fp16)[name = tensor("input_47_cast_fp16")]; - tensor input_49_pad_type_0 = const()[name = tensor("input_49_pad_type_0"), val = tensor("valid")]; - tensor input_49_groups_0 = const()[name = tensor("input_49_groups_0"), val = tensor(512)]; - tensor input_49_strides_0 = const()[name = tensor("input_49_strides_0"), val = tensor([1])]; - tensor input_49_pad_0 = const()[name = tensor("input_49_pad_0"), val = tensor([0, 0])]; - tensor input_49_dilations_0 = const()[name = tensor("input_49_dilations_0"), val = tensor([1])]; - tensor const_234_to_fp16 = const()[name = tensor("const_234_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10646784)))]; - tensor const_235_to_fp16 = const()[name = tensor("const_235_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10656064)))]; - tensor input_51_cast_fp16 = conv(bias = const_235_to_fp16, dilations = input_49_dilations_0, groups = input_49_groups_0, pad = input_49_pad_0, pad_type = input_49_pad_type_0, strides = input_49_strides_0, weight = const_234_to_fp16, x = input_47_cast_fp16)[name = tensor("input_51_cast_fp16")]; - tensor input_53_cast_fp16 = silu(x = input_51_cast_fp16)[name = tensor("input_53_cast_fp16")]; - tensor x_21_pad_type_0 = const()[name = tensor("x_21_pad_type_0"), val = tensor("valid")]; - tensor x_21_strides_0 = const()[name = tensor("x_21_strides_0"), val = tensor([1])]; - tensor x_21_pad_0 = const()[name = tensor("x_21_pad_0"), val = tensor([0, 0])]; - tensor x_21_dilations_0 = const()[name = tensor("x_21_dilations_0"), val = tensor([1])]; - tensor x_21_groups_0 = const()[name = tensor("x_21_groups_0"), val = tensor(1)]; - tensor module_layers_0_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(10657152)))]; - tensor module_layers_0_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_0_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11181504)))]; - tensor x_21_cast_fp16 = conv(bias = module_layers_0_conv_pointwise_conv2_bias_to_fp16, dilations = x_21_dilations_0, groups = x_21_groups_0, pad = x_21_pad_0, pad_type = x_21_pad_type_0, strides = x_21_strides_0, weight = module_layers_0_conv_pointwise_conv2_weight_to_fp16, x = input_53_cast_fp16)[name = tensor("x_21_cast_fp16")]; - tensor input_55_perm_0 = const()[name = tensor("input_55_perm_0"), val = tensor([0, 2, 1])]; - tensor input_55_cast_fp16 = transpose(perm = input_55_perm_0, x = x_21_cast_fp16)[name = tensor("transpose_198")]; - tensor input_57_cast_fp16 = add(x = input_39_cast_fp16, y = input_55_cast_fp16)[name = tensor("input_57_cast_fp16")]; - tensor input_59_axes_0 = const()[name = tensor("input_59_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11182592)))]; - tensor module_layers_0_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11183680)))]; - tensor input_59_cast_fp16 = layer_norm(axes = input_59_axes_0, beta = module_layers_0_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_feed_forward2_weight_to_fp16, x = input_57_cast_fp16)[name = tensor("input_59_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(11184768)))]; - tensor module_layers_0_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13281984)))]; - tensor linear_8_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear1_bias_to_fp16, weight = module_layers_0_feed_forward2_linear1_weight_to_fp16, x = input_59_cast_fp16)[name = tensor("linear_8_cast_fp16")]; - tensor input_63_cast_fp16 = silu(x = linear_8_cast_fp16)[name = tensor("input_63_cast_fp16")]; - tensor module_layers_0_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(13286144)))]; - tensor module_layers_0_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_0_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15383360)))]; - tensor linear_9_cast_fp16 = linear(bias = module_layers_0_feed_forward2_linear2_bias_to_fp16, weight = module_layers_0_feed_forward2_linear2_weight_to_fp16, x = input_63_cast_fp16)[name = tensor("linear_9_cast_fp16")]; - tensor var_495_to_fp16 = const()[name = tensor("op_495_to_fp16"), val = tensor(0x1p-1)]; - tensor var_496_cast_fp16 = mul(x = linear_9_cast_fp16, y = var_495_to_fp16)[name = tensor("op_496_cast_fp16")]; - tensor input_69_cast_fp16 = add(x = input_57_cast_fp16, y = var_496_cast_fp16)[name = tensor("input_69_cast_fp16")]; - tensor input_71_axes_0 = const()[name = tensor("input_71_axes_0"), val = tensor([-1])]; - tensor module_layers_0_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_0_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15384448)))]; - tensor module_layers_0_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_0_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15385536)))]; - tensor input_71_cast_fp16 = layer_norm(axes = input_71_axes_0, beta = module_layers_0_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_0_norm_out_weight_to_fp16, x = input_69_cast_fp16)[name = tensor("input_71_cast_fp16")]; - tensor input_73_axes_0 = const()[name = tensor("input_73_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15386624)))]; - tensor module_layers_1_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15387712)))]; - tensor input_73_cast_fp16 = layer_norm(axes = input_73_axes_0, beta = module_layers_1_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward1_weight_to_fp16, x = input_71_cast_fp16)[name = tensor("input_73_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(15388800)))]; - tensor module_layers_1_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17486016)))]; - tensor linear_10_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear1_bias_to_fp16, weight = module_layers_1_feed_forward1_linear1_weight_to_fp16, x = input_73_cast_fp16)[name = tensor("linear_10_cast_fp16")]; - tensor input_77_cast_fp16 = silu(x = linear_10_cast_fp16)[name = tensor("input_77_cast_fp16")]; - tensor module_layers_1_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(17490176)))]; - tensor module_layers_1_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19587392)))]; - tensor linear_11_cast_fp16 = linear(bias = module_layers_1_feed_forward1_linear2_bias_to_fp16, weight = module_layers_1_feed_forward1_linear2_weight_to_fp16, x = input_77_cast_fp16)[name = tensor("linear_11_cast_fp16")]; - tensor var_526_to_fp16 = const()[name = tensor("op_526_to_fp16"), val = tensor(0x1p-1)]; - tensor var_527_cast_fp16 = mul(x = linear_11_cast_fp16, y = var_526_to_fp16)[name = tensor("op_527_cast_fp16")]; - tensor input_83_cast_fp16 = add(x = input_71_cast_fp16, y = var_527_cast_fp16)[name = tensor("input_83_cast_fp16")]; - tensor query_3_axes_0 = const()[name = tensor("query_3_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19588480)))]; - tensor module_layers_1_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19589568)))]; - tensor query_3_cast_fp16 = layer_norm(axes = query_3_axes_0, beta = module_layers_1_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_self_att_weight_to_fp16, x = input_83_cast_fp16)[name = tensor("query_3_cast_fp16")]; - tensor module_layers_1_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(19590656)))]; - tensor module_layers_1_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20115008)))]; - tensor linear_12_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_q_bias_to_fp16, weight = module_layers_1_self_attn_linear_q_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_12_cast_fp16")]; - tensor var_544 = const()[name = tensor("op_544"), val = tensor([1, -1, 8, 64])]; - tensor q_7_cast_fp16 = reshape(shape = var_544, x = linear_12_cast_fp16)[name = tensor("q_7_cast_fp16")]; - tensor module_layers_1_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20116096)))]; - tensor module_layers_1_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20640448)))]; - tensor linear_13_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_k_bias_to_fp16, weight = module_layers_1_self_attn_linear_k_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_13_cast_fp16")]; - tensor var_549 = const()[name = tensor("op_549"), val = tensor([1, -1, 8, 64])]; - tensor k_5_cast_fp16 = reshape(shape = var_549, x = linear_13_cast_fp16)[name = tensor("k_5_cast_fp16")]; - tensor module_layers_1_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(20641536)))]; - tensor module_layers_1_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21165888)))]; - tensor linear_14_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_v_bias_to_fp16, weight = module_layers_1_self_attn_linear_v_weight_to_fp16, x = query_3_cast_fp16)[name = tensor("linear_14_cast_fp16")]; - tensor var_554 = const()[name = tensor("op_554"), val = tensor([1, -1, 8, 64])]; - tensor v_3_cast_fp16 = reshape(shape = var_554, x = linear_14_cast_fp16)[name = tensor("v_3_cast_fp16")]; - tensor value_5_perm_0 = const()[name = tensor("value_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_1_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21166976)))]; - tensor var_566_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_u_to_fp16)[name = tensor("op_566_cast_fp16")]; - tensor module_layers_1_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_1_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21168064)))]; - tensor var_568_cast_fp16 = add(x = q_7_cast_fp16, y = module_layers_1_self_attn_pos_bias_v_to_fp16)[name = tensor("op_568_cast_fp16")]; - tensor q_with_bias_v_3_perm_0 = const()[name = tensor("q_with_bias_v_3_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_29_transpose_x_0 = const()[name = tensor("x_29_transpose_x_0"), val = tensor(false)]; - tensor x_29_transpose_y_0 = const()[name = tensor("x_29_transpose_y_0"), val = tensor(false)]; - tensor var_570_to_fp16 = const()[name = tensor("op_570_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21169152)))]; - tensor q_with_bias_v_3_cast_fp16 = transpose(perm = q_with_bias_v_3_perm_0, x = var_568_cast_fp16)[name = tensor("transpose_196")]; - tensor x_29_cast_fp16 = matmul(transpose_x = x_29_transpose_x_0, transpose_y = x_29_transpose_y_0, x = q_with_bias_v_3_cast_fp16, y = var_570_to_fp16)[name = tensor("x_29_cast_fp16")]; - tensor x_31_pad_0 = const()[name = tensor("x_31_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_31_mode_0 = const()[name = tensor("x_31_mode_0"), val = tensor("constant")]; - tensor const_80_to_fp16 = const()[name = tensor("const_80_to_fp16"), val = tensor(0x0p+0)]; - tensor x_31_cast_fp16 = pad(constant_val = const_80_to_fp16, mode = x_31_mode_0, pad = x_31_pad_0, x = x_29_cast_fp16)[name = tensor("x_31_cast_fp16")]; - tensor var_578 = const()[name = tensor("op_578"), val = tensor([1, 8, -1, 188])]; - tensor x_33_cast_fp16 = reshape(shape = var_578, x = x_31_cast_fp16)[name = tensor("x_33_cast_fp16")]; - tensor var_582_begin_0 = const()[name = tensor("op_582_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_582_end_0 = const()[name = tensor("op_582_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_582_end_mask_0 = const()[name = tensor("op_582_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_582_cast_fp16 = slice_by_index(begin = var_582_begin_0, end = var_582_end_0, end_mask = var_582_end_mask_0, x = x_33_cast_fp16)[name = tensor("op_582_cast_fp16")]; - tensor var_583 = const()[name = tensor("op_583"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_5_cast_fp16 = reshape(shape = var_583, x = var_582_cast_fp16)[name = tensor("matrix_bd_5_cast_fp16")]; - tensor matrix_ac_3_transpose_x_0 = const()[name = tensor("matrix_ac_3_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_3_transpose_y_0 = const()[name = tensor("matrix_ac_3_transpose_y_0"), val = tensor(false)]; - tensor transpose_53_perm_0 = const()[name = tensor("transpose_53_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_54_perm_0 = const()[name = tensor("transpose_54_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_54 = transpose(perm = transpose_54_perm_0, x = k_5_cast_fp16)[name = tensor("transpose_194")]; - tensor transpose_53 = transpose(perm = transpose_53_perm_0, x = var_566_cast_fp16)[name = tensor("transpose_195")]; - tensor matrix_ac_3_cast_fp16 = matmul(transpose_x = matrix_ac_3_transpose_x_0, transpose_y = matrix_ac_3_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor("matrix_ac_3_cast_fp16")]; - tensor matrix_bd_7_begin_0 = const()[name = tensor("matrix_bd_7_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_7_end_0 = const()[name = tensor("matrix_bd_7_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_7_end_mask_0 = const()[name = tensor("matrix_bd_7_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_7_cast_fp16 = slice_by_index(begin = matrix_bd_7_begin_0, end = matrix_bd_7_end_0, end_mask = matrix_bd_7_end_mask_0, x = matrix_bd_5_cast_fp16)[name = tensor("matrix_bd_7_cast_fp16")]; - tensor var_592_cast_fp16 = add(x = matrix_ac_3_cast_fp16, y = matrix_bd_7_cast_fp16)[name = tensor("op_592_cast_fp16")]; - tensor _inversed_scores_5_y_0_to_fp16 = const()[name = tensor("_inversed_scores_5_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_5_cast_fp16 = mul(x = var_592_cast_fp16, y = _inversed_scores_5_y_0_to_fp16)[name = tensor("_inversed_scores_5_cast_fp16")]; - tensor scores_7_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_5_cast_fp16, cond = mask_11)[name = tensor("scores_7_cast_fp16")]; - tensor var_598_cast_fp16 = softmax(axis = var_23, x = scores_7_cast_fp16)[name = tensor("op_598_cast_fp16")]; - tensor input_85_cast_fp16 = select(a = var_11_to_fp16, b = var_598_cast_fp16, cond = mask_11)[name = tensor("input_85_cast_fp16")]; - tensor x_35_transpose_x_0 = const()[name = tensor("x_35_transpose_x_0"), val = tensor(false)]; - tensor x_35_transpose_y_0 = const()[name = tensor("x_35_transpose_y_0"), val = tensor(false)]; - tensor value_5_cast_fp16 = transpose(perm = value_5_perm_0, x = v_3_cast_fp16)[name = tensor("transpose_197")]; - tensor x_35_cast_fp16 = matmul(transpose_x = x_35_transpose_x_0, transpose_y = x_35_transpose_y_0, x = input_85_cast_fp16, y = value_5_cast_fp16)[name = tensor("x_35_cast_fp16")]; - tensor var_602_perm_0 = const()[name = tensor("op_602_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_603 = const()[name = tensor("op_603"), val = tensor([1, -1, 512])]; - tensor var_602_cast_fp16 = transpose(perm = var_602_perm_0, x = x_35_cast_fp16)[name = tensor("transpose_193")]; - tensor input_87_cast_fp16 = reshape(shape = var_603, x = var_602_cast_fp16)[name = tensor("input_87_cast_fp16")]; - tensor module_layers_1_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(21553216)))]; - tensor module_layers_1_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_1_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22077568)))]; - tensor linear_16_cast_fp16 = linear(bias = module_layers_1_self_attn_linear_out_bias_to_fp16, weight = module_layers_1_self_attn_linear_out_weight_to_fp16, x = input_87_cast_fp16)[name = tensor("linear_16_cast_fp16")]; - tensor input_91_cast_fp16 = add(x = input_83_cast_fp16, y = linear_16_cast_fp16)[name = tensor("input_91_cast_fp16")]; - tensor x_39_axes_0 = const()[name = tensor("x_39_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22078656)))]; - tensor module_layers_1_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22079744)))]; - tensor x_39_cast_fp16 = layer_norm(axes = x_39_axes_0, beta = module_layers_1_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_conv_weight_to_fp16, x = input_91_cast_fp16)[name = tensor("x_39_cast_fp16")]; - tensor input_93_perm_0 = const()[name = tensor("input_93_perm_0"), val = tensor([0, 2, 1])]; - tensor input_95_pad_type_0 = const()[name = tensor("input_95_pad_type_0"), val = tensor("valid")]; - tensor input_95_strides_0 = const()[name = tensor("input_95_strides_0"), val = tensor([1])]; - tensor input_95_pad_0 = const()[name = tensor("input_95_pad_0"), val = tensor([0, 0])]; - tensor input_95_dilations_0 = const()[name = tensor("input_95_dilations_0"), val = tensor([1])]; - tensor input_95_groups_0 = const()[name = tensor("input_95_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(22080832)))]; - tensor module_layers_1_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23129472)))]; - tensor input_93_cast_fp16 = transpose(perm = input_93_perm_0, x = x_39_cast_fp16)[name = tensor("transpose_192")]; - tensor input_95_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv1_bias_to_fp16, dilations = input_95_dilations_0, groups = input_95_groups_0, pad = input_95_pad_0, pad_type = input_95_pad_type_0, strides = input_95_strides_0, weight = module_layers_1_conv_pointwise_conv1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor("input_95_cast_fp16")]; - tensor x_41_split_num_splits_0 = const()[name = tensor("x_41_split_num_splits_0"), val = tensor(2)]; - tensor x_41_split_axis_0 = const()[name = tensor("x_41_split_axis_0"), val = tensor(1)]; - tensor x_41_split_cast_fp16_0, tensor x_41_split_cast_fp16_1 = split(axis = x_41_split_axis_0, num_splits = x_41_split_num_splits_0, x = input_95_cast_fp16)[name = tensor("x_41_split_cast_fp16")]; - tensor x_41_split_1_sigmoid_cast_fp16 = sigmoid(x = x_41_split_cast_fp16_1)[name = tensor("x_41_split_1_sigmoid_cast_fp16")]; - tensor x_41_cast_fp16 = mul(x = x_41_split_cast_fp16_0, y = x_41_split_1_sigmoid_cast_fp16)[name = tensor("x_41_cast_fp16")]; - tensor input_97_cast_fp16 = select(a = var_11_to_fp16, b = x_41_cast_fp16, cond = var_453)[name = tensor("input_97_cast_fp16")]; - tensor input_99_pad_0 = const()[name = tensor("input_99_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_99_mode_0 = const()[name = tensor("input_99_mode_0"), val = tensor("constant")]; - tensor const_83_to_fp16 = const()[name = tensor("const_83_to_fp16"), val = tensor(0x0p+0)]; - tensor input_99_cast_fp16 = pad(constant_val = const_83_to_fp16, mode = input_99_mode_0, pad = input_99_pad_0, x = input_97_cast_fp16)[name = tensor("input_99_cast_fp16")]; - tensor input_101_pad_type_0 = const()[name = tensor("input_101_pad_type_0"), val = tensor("valid")]; - tensor input_101_groups_0 = const()[name = tensor("input_101_groups_0"), val = tensor(512)]; - tensor input_101_strides_0 = const()[name = tensor("input_101_strides_0"), val = tensor([1])]; - tensor input_101_pad_0 = const()[name = tensor("input_101_pad_0"), val = tensor([0, 0])]; - tensor input_101_dilations_0 = const()[name = tensor("input_101_dilations_0"), val = tensor([1])]; - tensor const_236_to_fp16 = const()[name = tensor("const_236_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23131584)))]; - tensor const_237_to_fp16 = const()[name = tensor("const_237_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23140864)))]; - tensor input_103_cast_fp16 = conv(bias = const_237_to_fp16, dilations = input_101_dilations_0, groups = input_101_groups_0, pad = input_101_pad_0, pad_type = input_101_pad_type_0, strides = input_101_strides_0, weight = const_236_to_fp16, x = input_99_cast_fp16)[name = tensor("input_103_cast_fp16")]; - tensor input_105_cast_fp16 = silu(x = input_103_cast_fp16)[name = tensor("input_105_cast_fp16")]; - tensor x_43_pad_type_0 = const()[name = tensor("x_43_pad_type_0"), val = tensor("valid")]; - tensor x_43_strides_0 = const()[name = tensor("x_43_strides_0"), val = tensor([1])]; - tensor x_43_pad_0 = const()[name = tensor("x_43_pad_0"), val = tensor([0, 0])]; - tensor x_43_dilations_0 = const()[name = tensor("x_43_dilations_0"), val = tensor([1])]; - tensor x_43_groups_0 = const()[name = tensor("x_43_groups_0"), val = tensor(1)]; - tensor module_layers_1_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23141952)))]; - tensor module_layers_1_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_1_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23666304)))]; - tensor x_43_cast_fp16 = conv(bias = module_layers_1_conv_pointwise_conv2_bias_to_fp16, dilations = x_43_dilations_0, groups = x_43_groups_0, pad = x_43_pad_0, pad_type = x_43_pad_type_0, strides = x_43_strides_0, weight = module_layers_1_conv_pointwise_conv2_weight_to_fp16, x = input_105_cast_fp16)[name = tensor("x_43_cast_fp16")]; - tensor input_107_perm_0 = const()[name = tensor("input_107_perm_0"), val = tensor([0, 2, 1])]; - tensor input_107_cast_fp16 = transpose(perm = input_107_perm_0, x = x_43_cast_fp16)[name = tensor("transpose_191")]; - tensor input_109_cast_fp16 = add(x = input_91_cast_fp16, y = input_107_cast_fp16)[name = tensor("input_109_cast_fp16")]; - tensor input_111_axes_0 = const()[name = tensor("input_111_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23667392)))]; - tensor module_layers_1_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23668480)))]; - tensor input_111_cast_fp16 = layer_norm(axes = input_111_axes_0, beta = module_layers_1_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_feed_forward2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor("input_111_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(23669568)))]; - tensor module_layers_1_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25766784)))]; - tensor linear_17_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear1_bias_to_fp16, weight = module_layers_1_feed_forward2_linear1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor("linear_17_cast_fp16")]; - tensor input_115_cast_fp16 = silu(x = linear_17_cast_fp16)[name = tensor("input_115_cast_fp16")]; - tensor module_layers_1_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(25770944)))]; - tensor module_layers_1_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_1_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27868160)))]; - tensor linear_18_cast_fp16 = linear(bias = module_layers_1_feed_forward2_linear2_bias_to_fp16, weight = module_layers_1_feed_forward2_linear2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor("linear_18_cast_fp16")]; - tensor var_669_to_fp16 = const()[name = tensor("op_669_to_fp16"), val = tensor(0x1p-1)]; - tensor var_670_cast_fp16 = mul(x = linear_18_cast_fp16, y = var_669_to_fp16)[name = tensor("op_670_cast_fp16")]; - tensor input_121_cast_fp16 = add(x = input_109_cast_fp16, y = var_670_cast_fp16)[name = tensor("input_121_cast_fp16")]; - tensor input_123_axes_0 = const()[name = tensor("input_123_axes_0"), val = tensor([-1])]; - tensor module_layers_1_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_1_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27869248)))]; - tensor module_layers_1_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_1_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27870336)))]; - tensor input_123_cast_fp16 = layer_norm(axes = input_123_axes_0, beta = module_layers_1_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_1_norm_out_weight_to_fp16, x = input_121_cast_fp16)[name = tensor("input_123_cast_fp16")]; - tensor input_125_axes_0 = const()[name = tensor("input_125_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27871424)))]; - tensor module_layers_2_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27872512)))]; - tensor input_125_cast_fp16 = layer_norm(axes = input_125_axes_0, beta = module_layers_2_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor("input_125_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(27873600)))]; - tensor module_layers_2_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29970816)))]; - tensor linear_19_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear1_bias_to_fp16, weight = module_layers_2_feed_forward1_linear1_weight_to_fp16, x = input_125_cast_fp16)[name = tensor("linear_19_cast_fp16")]; - tensor input_129_cast_fp16 = silu(x = linear_19_cast_fp16)[name = tensor("input_129_cast_fp16")]; - tensor module_layers_2_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(29974976)))]; - tensor module_layers_2_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32072192)))]; - tensor linear_20_cast_fp16 = linear(bias = module_layers_2_feed_forward1_linear2_bias_to_fp16, weight = module_layers_2_feed_forward1_linear2_weight_to_fp16, x = input_129_cast_fp16)[name = tensor("linear_20_cast_fp16")]; - tensor var_700_to_fp16 = const()[name = tensor("op_700_to_fp16"), val = tensor(0x1p-1)]; - tensor var_701_cast_fp16 = mul(x = linear_20_cast_fp16, y = var_700_to_fp16)[name = tensor("op_701_cast_fp16")]; - tensor input_135_cast_fp16 = add(x = input_123_cast_fp16, y = var_701_cast_fp16)[name = tensor("input_135_cast_fp16")]; - tensor query_5_axes_0 = const()[name = tensor("query_5_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32073280)))]; - tensor module_layers_2_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32074368)))]; - tensor query_5_cast_fp16 = layer_norm(axes = query_5_axes_0, beta = module_layers_2_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_self_att_weight_to_fp16, x = input_135_cast_fp16)[name = tensor("query_5_cast_fp16")]; - tensor module_layers_2_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32075456)))]; - tensor module_layers_2_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32599808)))]; - tensor linear_21_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_q_bias_to_fp16, weight = module_layers_2_self_attn_linear_q_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_21_cast_fp16")]; - tensor var_718 = const()[name = tensor("op_718"), val = tensor([1, -1, 8, 64])]; - tensor q_13_cast_fp16 = reshape(shape = var_718, x = linear_21_cast_fp16)[name = tensor("q_13_cast_fp16")]; - tensor module_layers_2_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(32600896)))]; - tensor module_layers_2_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33125248)))]; - tensor linear_22_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_k_bias_to_fp16, weight = module_layers_2_self_attn_linear_k_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_22_cast_fp16")]; - tensor var_723 = const()[name = tensor("op_723"), val = tensor([1, -1, 8, 64])]; - tensor k_9_cast_fp16 = reshape(shape = var_723, x = linear_22_cast_fp16)[name = tensor("k_9_cast_fp16")]; - tensor module_layers_2_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33126336)))]; - tensor module_layers_2_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33650688)))]; - tensor linear_23_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_v_bias_to_fp16, weight = module_layers_2_self_attn_linear_v_weight_to_fp16, x = query_5_cast_fp16)[name = tensor("linear_23_cast_fp16")]; - tensor var_728 = const()[name = tensor("op_728"), val = tensor([1, -1, 8, 64])]; - tensor v_5_cast_fp16 = reshape(shape = var_728, x = linear_23_cast_fp16)[name = tensor("v_5_cast_fp16")]; - tensor value_7_perm_0 = const()[name = tensor("value_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_2_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33651776)))]; - tensor var_740_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_u_to_fp16)[name = tensor("op_740_cast_fp16")]; - tensor module_layers_2_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_2_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33652864)))]; - tensor var_742_cast_fp16 = add(x = q_13_cast_fp16, y = module_layers_2_self_attn_pos_bias_v_to_fp16)[name = tensor("op_742_cast_fp16")]; - tensor q_with_bias_v_5_perm_0 = const()[name = tensor("q_with_bias_v_5_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_51_transpose_x_0 = const()[name = tensor("x_51_transpose_x_0"), val = tensor(false)]; - tensor x_51_transpose_y_0 = const()[name = tensor("x_51_transpose_y_0"), val = tensor(false)]; - tensor var_744_to_fp16 = const()[name = tensor("op_744_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(33653952)))]; - tensor q_with_bias_v_5_cast_fp16 = transpose(perm = q_with_bias_v_5_perm_0, x = var_742_cast_fp16)[name = tensor("transpose_189")]; - tensor x_51_cast_fp16 = matmul(transpose_x = x_51_transpose_x_0, transpose_y = x_51_transpose_y_0, x = q_with_bias_v_5_cast_fp16, y = var_744_to_fp16)[name = tensor("x_51_cast_fp16")]; - tensor x_53_pad_0 = const()[name = tensor("x_53_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_53_mode_0 = const()[name = tensor("x_53_mode_0"), val = tensor("constant")]; - tensor const_90_to_fp16 = const()[name = tensor("const_90_to_fp16"), val = tensor(0x0p+0)]; - tensor x_53_cast_fp16 = pad(constant_val = const_90_to_fp16, mode = x_53_mode_0, pad = x_53_pad_0, x = x_51_cast_fp16)[name = tensor("x_53_cast_fp16")]; - tensor var_752 = const()[name = tensor("op_752"), val = tensor([1, 8, -1, 188])]; - tensor x_55_cast_fp16 = reshape(shape = var_752, x = x_53_cast_fp16)[name = tensor("x_55_cast_fp16")]; - tensor var_756_begin_0 = const()[name = tensor("op_756_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_756_end_0 = const()[name = tensor("op_756_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_756_end_mask_0 = const()[name = tensor("op_756_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_756_cast_fp16 = slice_by_index(begin = var_756_begin_0, end = var_756_end_0, end_mask = var_756_end_mask_0, x = x_55_cast_fp16)[name = tensor("op_756_cast_fp16")]; - tensor var_757 = const()[name = tensor("op_757"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_9_cast_fp16 = reshape(shape = var_757, x = var_756_cast_fp16)[name = tensor("matrix_bd_9_cast_fp16")]; - tensor matrix_ac_5_transpose_x_0 = const()[name = tensor("matrix_ac_5_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_5_transpose_y_0 = const()[name = tensor("matrix_ac_5_transpose_y_0"), val = tensor(false)]; - tensor transpose_55_perm_0 = const()[name = tensor("transpose_55_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_56_perm_0 = const()[name = tensor("transpose_56_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_56 = transpose(perm = transpose_56_perm_0, x = k_9_cast_fp16)[name = tensor("transpose_187")]; - tensor transpose_55 = transpose(perm = transpose_55_perm_0, x = var_740_cast_fp16)[name = tensor("transpose_188")]; - tensor matrix_ac_5_cast_fp16 = matmul(transpose_x = matrix_ac_5_transpose_x_0, transpose_y = matrix_ac_5_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor("matrix_ac_5_cast_fp16")]; - tensor matrix_bd_11_begin_0 = const()[name = tensor("matrix_bd_11_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_11_end_0 = const()[name = tensor("matrix_bd_11_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_11_end_mask_0 = const()[name = tensor("matrix_bd_11_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_11_cast_fp16 = slice_by_index(begin = matrix_bd_11_begin_0, end = matrix_bd_11_end_0, end_mask = matrix_bd_11_end_mask_0, x = matrix_bd_9_cast_fp16)[name = tensor("matrix_bd_11_cast_fp16")]; - tensor var_766_cast_fp16 = add(x = matrix_ac_5_cast_fp16, y = matrix_bd_11_cast_fp16)[name = tensor("op_766_cast_fp16")]; - tensor _inversed_scores_9_y_0_to_fp16 = const()[name = tensor("_inversed_scores_9_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_9_cast_fp16 = mul(x = var_766_cast_fp16, y = _inversed_scores_9_y_0_to_fp16)[name = tensor("_inversed_scores_9_cast_fp16")]; - tensor scores_11_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_9_cast_fp16, cond = mask_11)[name = tensor("scores_11_cast_fp16")]; - tensor var_772_cast_fp16 = softmax(axis = var_23, x = scores_11_cast_fp16)[name = tensor("op_772_cast_fp16")]; - tensor input_137_cast_fp16 = select(a = var_11_to_fp16, b = var_772_cast_fp16, cond = mask_11)[name = tensor("input_137_cast_fp16")]; - tensor x_57_transpose_x_0 = const()[name = tensor("x_57_transpose_x_0"), val = tensor(false)]; - tensor x_57_transpose_y_0 = const()[name = tensor("x_57_transpose_y_0"), val = tensor(false)]; - tensor value_7_cast_fp16 = transpose(perm = value_7_perm_0, x = v_5_cast_fp16)[name = tensor("transpose_190")]; - tensor x_57_cast_fp16 = matmul(transpose_x = x_57_transpose_x_0, transpose_y = x_57_transpose_y_0, x = input_137_cast_fp16, y = value_7_cast_fp16)[name = tensor("x_57_cast_fp16")]; - tensor var_776_perm_0 = const()[name = tensor("op_776_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_777 = const()[name = tensor("op_777"), val = tensor([1, -1, 512])]; - tensor var_776_cast_fp16 = transpose(perm = var_776_perm_0, x = x_57_cast_fp16)[name = tensor("transpose_186")]; - tensor input_139_cast_fp16 = reshape(shape = var_777, x = var_776_cast_fp16)[name = tensor("input_139_cast_fp16")]; - tensor module_layers_2_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34038016)))]; - tensor module_layers_2_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_2_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34562368)))]; - tensor linear_25_cast_fp16 = linear(bias = module_layers_2_self_attn_linear_out_bias_to_fp16, weight = module_layers_2_self_attn_linear_out_weight_to_fp16, x = input_139_cast_fp16)[name = tensor("linear_25_cast_fp16")]; - tensor input_143_cast_fp16 = add(x = input_135_cast_fp16, y = linear_25_cast_fp16)[name = tensor("input_143_cast_fp16")]; - tensor x_61_axes_0 = const()[name = tensor("x_61_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34563456)))]; - tensor module_layers_2_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34564544)))]; - tensor x_61_cast_fp16 = layer_norm(axes = x_61_axes_0, beta = module_layers_2_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_conv_weight_to_fp16, x = input_143_cast_fp16)[name = tensor("x_61_cast_fp16")]; - tensor input_145_perm_0 = const()[name = tensor("input_145_perm_0"), val = tensor([0, 2, 1])]; - tensor input_147_pad_type_0 = const()[name = tensor("input_147_pad_type_0"), val = tensor("valid")]; - tensor input_147_strides_0 = const()[name = tensor("input_147_strides_0"), val = tensor([1])]; - tensor input_147_pad_0 = const()[name = tensor("input_147_pad_0"), val = tensor([0, 0])]; - tensor input_147_dilations_0 = const()[name = tensor("input_147_dilations_0"), val = tensor([1])]; - tensor input_147_groups_0 = const()[name = tensor("input_147_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(34565632)))]; - tensor module_layers_2_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35614272)))]; - tensor input_145_cast_fp16 = transpose(perm = input_145_perm_0, x = x_61_cast_fp16)[name = tensor("transpose_185")]; - tensor input_147_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv1_bias_to_fp16, dilations = input_147_dilations_0, groups = input_147_groups_0, pad = input_147_pad_0, pad_type = input_147_pad_type_0, strides = input_147_strides_0, weight = module_layers_2_conv_pointwise_conv1_weight_to_fp16, x = input_145_cast_fp16)[name = tensor("input_147_cast_fp16")]; - tensor x_63_split_num_splits_0 = const()[name = tensor("x_63_split_num_splits_0"), val = tensor(2)]; - tensor x_63_split_axis_0 = const()[name = tensor("x_63_split_axis_0"), val = tensor(1)]; - tensor x_63_split_cast_fp16_0, tensor x_63_split_cast_fp16_1 = split(axis = x_63_split_axis_0, num_splits = x_63_split_num_splits_0, x = input_147_cast_fp16)[name = tensor("x_63_split_cast_fp16")]; - tensor x_63_split_1_sigmoid_cast_fp16 = sigmoid(x = x_63_split_cast_fp16_1)[name = tensor("x_63_split_1_sigmoid_cast_fp16")]; - tensor x_63_cast_fp16 = mul(x = x_63_split_cast_fp16_0, y = x_63_split_1_sigmoid_cast_fp16)[name = tensor("x_63_cast_fp16")]; - tensor input_149_cast_fp16 = select(a = var_11_to_fp16, b = x_63_cast_fp16, cond = var_453)[name = tensor("input_149_cast_fp16")]; - tensor input_151_pad_0 = const()[name = tensor("input_151_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_151_mode_0 = const()[name = tensor("input_151_mode_0"), val = tensor("constant")]; - tensor const_93_to_fp16 = const()[name = tensor("const_93_to_fp16"), val = tensor(0x0p+0)]; - tensor input_151_cast_fp16 = pad(constant_val = const_93_to_fp16, mode = input_151_mode_0, pad = input_151_pad_0, x = input_149_cast_fp16)[name = tensor("input_151_cast_fp16")]; - tensor input_153_pad_type_0 = const()[name = tensor("input_153_pad_type_0"), val = tensor("valid")]; - tensor input_153_groups_0 = const()[name = tensor("input_153_groups_0"), val = tensor(512)]; - tensor input_153_strides_0 = const()[name = tensor("input_153_strides_0"), val = tensor([1])]; - tensor input_153_pad_0 = const()[name = tensor("input_153_pad_0"), val = tensor([0, 0])]; - tensor input_153_dilations_0 = const()[name = tensor("input_153_dilations_0"), val = tensor([1])]; - tensor const_238_to_fp16 = const()[name = tensor("const_238_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35616384)))]; - tensor const_239_to_fp16 = const()[name = tensor("const_239_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35625664)))]; - tensor input_155_cast_fp16 = conv(bias = const_239_to_fp16, dilations = input_153_dilations_0, groups = input_153_groups_0, pad = input_153_pad_0, pad_type = input_153_pad_type_0, strides = input_153_strides_0, weight = const_238_to_fp16, x = input_151_cast_fp16)[name = tensor("input_155_cast_fp16")]; - tensor input_157_cast_fp16 = silu(x = input_155_cast_fp16)[name = tensor("input_157_cast_fp16")]; - tensor x_65_pad_type_0 = const()[name = tensor("x_65_pad_type_0"), val = tensor("valid")]; - tensor x_65_strides_0 = const()[name = tensor("x_65_strides_0"), val = tensor([1])]; - tensor x_65_pad_0 = const()[name = tensor("x_65_pad_0"), val = tensor([0, 0])]; - tensor x_65_dilations_0 = const()[name = tensor("x_65_dilations_0"), val = tensor([1])]; - tensor x_65_groups_0 = const()[name = tensor("x_65_groups_0"), val = tensor(1)]; - tensor module_layers_2_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(35626752)))]; - tensor module_layers_2_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_2_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36151104)))]; - tensor x_65_cast_fp16 = conv(bias = module_layers_2_conv_pointwise_conv2_bias_to_fp16, dilations = x_65_dilations_0, groups = x_65_groups_0, pad = x_65_pad_0, pad_type = x_65_pad_type_0, strides = x_65_strides_0, weight = module_layers_2_conv_pointwise_conv2_weight_to_fp16, x = input_157_cast_fp16)[name = tensor("x_65_cast_fp16")]; - tensor input_159_perm_0 = const()[name = tensor("input_159_perm_0"), val = tensor([0, 2, 1])]; - tensor input_159_cast_fp16 = transpose(perm = input_159_perm_0, x = x_65_cast_fp16)[name = tensor("transpose_184")]; - tensor input_161_cast_fp16 = add(x = input_143_cast_fp16, y = input_159_cast_fp16)[name = tensor("input_161_cast_fp16")]; - tensor input_163_axes_0 = const()[name = tensor("input_163_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36152192)))]; - tensor module_layers_2_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36153280)))]; - tensor input_163_cast_fp16 = layer_norm(axes = input_163_axes_0, beta = module_layers_2_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_feed_forward2_weight_to_fp16, x = input_161_cast_fp16)[name = tensor("input_163_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(36154368)))]; - tensor module_layers_2_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38251584)))]; - tensor linear_26_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear1_bias_to_fp16, weight = module_layers_2_feed_forward2_linear1_weight_to_fp16, x = input_163_cast_fp16)[name = tensor("linear_26_cast_fp16")]; - tensor input_167_cast_fp16 = silu(x = linear_26_cast_fp16)[name = tensor("input_167_cast_fp16")]; - tensor module_layers_2_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(38255744)))]; - tensor module_layers_2_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_2_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40352960)))]; - tensor linear_27_cast_fp16 = linear(bias = module_layers_2_feed_forward2_linear2_bias_to_fp16, weight = module_layers_2_feed_forward2_linear2_weight_to_fp16, x = input_167_cast_fp16)[name = tensor("linear_27_cast_fp16")]; - tensor var_843_to_fp16 = const()[name = tensor("op_843_to_fp16"), val = tensor(0x1p-1)]; - tensor var_844_cast_fp16 = mul(x = linear_27_cast_fp16, y = var_843_to_fp16)[name = tensor("op_844_cast_fp16")]; - tensor input_173_cast_fp16 = add(x = input_161_cast_fp16, y = var_844_cast_fp16)[name = tensor("input_173_cast_fp16")]; - tensor input_175_axes_0 = const()[name = tensor("input_175_axes_0"), val = tensor([-1])]; - tensor module_layers_2_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_2_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40354048)))]; - tensor module_layers_2_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_2_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40355136)))]; - tensor input_175_cast_fp16 = layer_norm(axes = input_175_axes_0, beta = module_layers_2_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_2_norm_out_weight_to_fp16, x = input_173_cast_fp16)[name = tensor("input_175_cast_fp16")]; - tensor input_177_axes_0 = const()[name = tensor("input_177_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40356224)))]; - tensor module_layers_3_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40357312)))]; - tensor input_177_cast_fp16 = layer_norm(axes = input_177_axes_0, beta = module_layers_3_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward1_weight_to_fp16, x = input_175_cast_fp16)[name = tensor("input_177_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(40358400)))]; - tensor module_layers_3_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42455616)))]; - tensor linear_28_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear1_bias_to_fp16, weight = module_layers_3_feed_forward1_linear1_weight_to_fp16, x = input_177_cast_fp16)[name = tensor("linear_28_cast_fp16")]; - tensor input_181_cast_fp16 = silu(x = linear_28_cast_fp16)[name = tensor("input_181_cast_fp16")]; - tensor module_layers_3_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(42459776)))]; - tensor module_layers_3_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44556992)))]; - tensor linear_29_cast_fp16 = linear(bias = module_layers_3_feed_forward1_linear2_bias_to_fp16, weight = module_layers_3_feed_forward1_linear2_weight_to_fp16, x = input_181_cast_fp16)[name = tensor("linear_29_cast_fp16")]; - tensor var_874_to_fp16 = const()[name = tensor("op_874_to_fp16"), val = tensor(0x1p-1)]; - tensor var_875_cast_fp16 = mul(x = linear_29_cast_fp16, y = var_874_to_fp16)[name = tensor("op_875_cast_fp16")]; - tensor input_187_cast_fp16 = add(x = input_175_cast_fp16, y = var_875_cast_fp16)[name = tensor("input_187_cast_fp16")]; - tensor query_7_axes_0 = const()[name = tensor("query_7_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44558080)))]; - tensor module_layers_3_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44559168)))]; - tensor query_7_cast_fp16 = layer_norm(axes = query_7_axes_0, beta = module_layers_3_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_self_att_weight_to_fp16, x = input_187_cast_fp16)[name = tensor("query_7_cast_fp16")]; - tensor module_layers_3_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(44560256)))]; - tensor module_layers_3_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45084608)))]; - tensor linear_30_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_q_bias_to_fp16, weight = module_layers_3_self_attn_linear_q_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_30_cast_fp16")]; - tensor var_892 = const()[name = tensor("op_892"), val = tensor([1, -1, 8, 64])]; - tensor q_19_cast_fp16 = reshape(shape = var_892, x = linear_30_cast_fp16)[name = tensor("q_19_cast_fp16")]; - tensor module_layers_3_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45085696)))]; - tensor module_layers_3_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45610048)))]; - tensor linear_31_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_k_bias_to_fp16, weight = module_layers_3_self_attn_linear_k_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_31_cast_fp16")]; - tensor var_897 = const()[name = tensor("op_897"), val = tensor([1, -1, 8, 64])]; - tensor k_13_cast_fp16 = reshape(shape = var_897, x = linear_31_cast_fp16)[name = tensor("k_13_cast_fp16")]; - tensor module_layers_3_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(45611136)))]; - tensor module_layers_3_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46135488)))]; - tensor linear_32_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_v_bias_to_fp16, weight = module_layers_3_self_attn_linear_v_weight_to_fp16, x = query_7_cast_fp16)[name = tensor("linear_32_cast_fp16")]; - tensor var_902 = const()[name = tensor("op_902"), val = tensor([1, -1, 8, 64])]; - tensor v_7_cast_fp16 = reshape(shape = var_902, x = linear_32_cast_fp16)[name = tensor("v_7_cast_fp16")]; - tensor value_9_perm_0 = const()[name = tensor("value_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_3_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46136576)))]; - tensor var_914_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_u_to_fp16)[name = tensor("op_914_cast_fp16")]; - tensor module_layers_3_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_3_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46137664)))]; - tensor var_916_cast_fp16 = add(x = q_19_cast_fp16, y = module_layers_3_self_attn_pos_bias_v_to_fp16)[name = tensor("op_916_cast_fp16")]; - tensor q_with_bias_v_7_perm_0 = const()[name = tensor("q_with_bias_v_7_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_73_transpose_x_0 = const()[name = tensor("x_73_transpose_x_0"), val = tensor(false)]; - tensor x_73_transpose_y_0 = const()[name = tensor("x_73_transpose_y_0"), val = tensor(false)]; - tensor var_918_to_fp16 = const()[name = tensor("op_918_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46138752)))]; - tensor q_with_bias_v_7_cast_fp16 = transpose(perm = q_with_bias_v_7_perm_0, x = var_916_cast_fp16)[name = tensor("transpose_182")]; - tensor x_73_cast_fp16 = matmul(transpose_x = x_73_transpose_x_0, transpose_y = x_73_transpose_y_0, x = q_with_bias_v_7_cast_fp16, y = var_918_to_fp16)[name = tensor("x_73_cast_fp16")]; - tensor x_75_pad_0 = const()[name = tensor("x_75_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_75_mode_0 = const()[name = tensor("x_75_mode_0"), val = tensor("constant")]; - tensor const_100_to_fp16 = const()[name = tensor("const_100_to_fp16"), val = tensor(0x0p+0)]; - tensor x_75_cast_fp16 = pad(constant_val = const_100_to_fp16, mode = x_75_mode_0, pad = x_75_pad_0, x = x_73_cast_fp16)[name = tensor("x_75_cast_fp16")]; - tensor var_926 = const()[name = tensor("op_926"), val = tensor([1, 8, -1, 188])]; - tensor x_77_cast_fp16 = reshape(shape = var_926, x = x_75_cast_fp16)[name = tensor("x_77_cast_fp16")]; - tensor var_930_begin_0 = const()[name = tensor("op_930_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_930_end_0 = const()[name = tensor("op_930_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_930_end_mask_0 = const()[name = tensor("op_930_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_930_cast_fp16 = slice_by_index(begin = var_930_begin_0, end = var_930_end_0, end_mask = var_930_end_mask_0, x = x_77_cast_fp16)[name = tensor("op_930_cast_fp16")]; - tensor var_931 = const()[name = tensor("op_931"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_13_cast_fp16 = reshape(shape = var_931, x = var_930_cast_fp16)[name = tensor("matrix_bd_13_cast_fp16")]; - tensor matrix_ac_7_transpose_x_0 = const()[name = tensor("matrix_ac_7_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_7_transpose_y_0 = const()[name = tensor("matrix_ac_7_transpose_y_0"), val = tensor(false)]; - tensor transpose_57_perm_0 = const()[name = tensor("transpose_57_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_58_perm_0 = const()[name = tensor("transpose_58_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_58 = transpose(perm = transpose_58_perm_0, x = k_13_cast_fp16)[name = tensor("transpose_180")]; - tensor transpose_57 = transpose(perm = transpose_57_perm_0, x = var_914_cast_fp16)[name = tensor("transpose_181")]; - tensor matrix_ac_7_cast_fp16 = matmul(transpose_x = matrix_ac_7_transpose_x_0, transpose_y = matrix_ac_7_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor("matrix_ac_7_cast_fp16")]; - tensor matrix_bd_15_begin_0 = const()[name = tensor("matrix_bd_15_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_15_end_0 = const()[name = tensor("matrix_bd_15_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_15_end_mask_0 = const()[name = tensor("matrix_bd_15_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_15_cast_fp16 = slice_by_index(begin = matrix_bd_15_begin_0, end = matrix_bd_15_end_0, end_mask = matrix_bd_15_end_mask_0, x = matrix_bd_13_cast_fp16)[name = tensor("matrix_bd_15_cast_fp16")]; - tensor var_940_cast_fp16 = add(x = matrix_ac_7_cast_fp16, y = matrix_bd_15_cast_fp16)[name = tensor("op_940_cast_fp16")]; - tensor _inversed_scores_13_y_0_to_fp16 = const()[name = tensor("_inversed_scores_13_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_13_cast_fp16 = mul(x = var_940_cast_fp16, y = _inversed_scores_13_y_0_to_fp16)[name = tensor("_inversed_scores_13_cast_fp16")]; - tensor scores_15_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_13_cast_fp16, cond = mask_11)[name = tensor("scores_15_cast_fp16")]; - tensor var_946_cast_fp16 = softmax(axis = var_23, x = scores_15_cast_fp16)[name = tensor("op_946_cast_fp16")]; - tensor input_189_cast_fp16 = select(a = var_11_to_fp16, b = var_946_cast_fp16, cond = mask_11)[name = tensor("input_189_cast_fp16")]; - tensor x_79_transpose_x_0 = const()[name = tensor("x_79_transpose_x_0"), val = tensor(false)]; - tensor x_79_transpose_y_0 = const()[name = tensor("x_79_transpose_y_0"), val = tensor(false)]; - tensor value_9_cast_fp16 = transpose(perm = value_9_perm_0, x = v_7_cast_fp16)[name = tensor("transpose_183")]; - tensor x_79_cast_fp16 = matmul(transpose_x = x_79_transpose_x_0, transpose_y = x_79_transpose_y_0, x = input_189_cast_fp16, y = value_9_cast_fp16)[name = tensor("x_79_cast_fp16")]; - tensor var_950_perm_0 = const()[name = tensor("op_950_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_951 = const()[name = tensor("op_951"), val = tensor([1, -1, 512])]; - tensor var_950_cast_fp16 = transpose(perm = var_950_perm_0, x = x_79_cast_fp16)[name = tensor("transpose_179")]; - tensor input_191_cast_fp16 = reshape(shape = var_951, x = var_950_cast_fp16)[name = tensor("input_191_cast_fp16")]; - tensor module_layers_3_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(46522816)))]; - tensor module_layers_3_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_3_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47047168)))]; - tensor linear_34_cast_fp16 = linear(bias = module_layers_3_self_attn_linear_out_bias_to_fp16, weight = module_layers_3_self_attn_linear_out_weight_to_fp16, x = input_191_cast_fp16)[name = tensor("linear_34_cast_fp16")]; - tensor input_195_cast_fp16 = add(x = input_187_cast_fp16, y = linear_34_cast_fp16)[name = tensor("input_195_cast_fp16")]; - tensor x_83_axes_0 = const()[name = tensor("x_83_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47048256)))]; - tensor module_layers_3_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47049344)))]; - tensor x_83_cast_fp16 = layer_norm(axes = x_83_axes_0, beta = module_layers_3_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_conv_weight_to_fp16, x = input_195_cast_fp16)[name = tensor("x_83_cast_fp16")]; - tensor input_197_perm_0 = const()[name = tensor("input_197_perm_0"), val = tensor([0, 2, 1])]; - tensor input_199_pad_type_0 = const()[name = tensor("input_199_pad_type_0"), val = tensor("valid")]; - tensor input_199_strides_0 = const()[name = tensor("input_199_strides_0"), val = tensor([1])]; - tensor input_199_pad_0 = const()[name = tensor("input_199_pad_0"), val = tensor([0, 0])]; - tensor input_199_dilations_0 = const()[name = tensor("input_199_dilations_0"), val = tensor([1])]; - tensor input_199_groups_0 = const()[name = tensor("input_199_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(47050432)))]; - tensor module_layers_3_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48099072)))]; - tensor input_197_cast_fp16 = transpose(perm = input_197_perm_0, x = x_83_cast_fp16)[name = tensor("transpose_178")]; - tensor input_199_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv1_bias_to_fp16, dilations = input_199_dilations_0, groups = input_199_groups_0, pad = input_199_pad_0, pad_type = input_199_pad_type_0, strides = input_199_strides_0, weight = module_layers_3_conv_pointwise_conv1_weight_to_fp16, x = input_197_cast_fp16)[name = tensor("input_199_cast_fp16")]; - tensor x_85_split_num_splits_0 = const()[name = tensor("x_85_split_num_splits_0"), val = tensor(2)]; - tensor x_85_split_axis_0 = const()[name = tensor("x_85_split_axis_0"), val = tensor(1)]; - tensor x_85_split_cast_fp16_0, tensor x_85_split_cast_fp16_1 = split(axis = x_85_split_axis_0, num_splits = x_85_split_num_splits_0, x = input_199_cast_fp16)[name = tensor("x_85_split_cast_fp16")]; - tensor x_85_split_1_sigmoid_cast_fp16 = sigmoid(x = x_85_split_cast_fp16_1)[name = tensor("x_85_split_1_sigmoid_cast_fp16")]; - tensor x_85_cast_fp16 = mul(x = x_85_split_cast_fp16_0, y = x_85_split_1_sigmoid_cast_fp16)[name = tensor("x_85_cast_fp16")]; - tensor input_201_cast_fp16 = select(a = var_11_to_fp16, b = x_85_cast_fp16, cond = var_453)[name = tensor("input_201_cast_fp16")]; - tensor input_203_pad_0 = const()[name = tensor("input_203_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_203_mode_0 = const()[name = tensor("input_203_mode_0"), val = tensor("constant")]; - tensor const_103_to_fp16 = const()[name = tensor("const_103_to_fp16"), val = tensor(0x0p+0)]; - tensor input_203_cast_fp16 = pad(constant_val = const_103_to_fp16, mode = input_203_mode_0, pad = input_203_pad_0, x = input_201_cast_fp16)[name = tensor("input_203_cast_fp16")]; - tensor input_205_pad_type_0 = const()[name = tensor("input_205_pad_type_0"), val = tensor("valid")]; - tensor input_205_groups_0 = const()[name = tensor("input_205_groups_0"), val = tensor(512)]; - tensor input_205_strides_0 = const()[name = tensor("input_205_strides_0"), val = tensor([1])]; - tensor input_205_pad_0 = const()[name = tensor("input_205_pad_0"), val = tensor([0, 0])]; - tensor input_205_dilations_0 = const()[name = tensor("input_205_dilations_0"), val = tensor([1])]; - tensor const_240_to_fp16 = const()[name = tensor("const_240_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48101184)))]; - tensor const_241_to_fp16 = const()[name = tensor("const_241_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48110464)))]; - tensor input_207_cast_fp16 = conv(bias = const_241_to_fp16, dilations = input_205_dilations_0, groups = input_205_groups_0, pad = input_205_pad_0, pad_type = input_205_pad_type_0, strides = input_205_strides_0, weight = const_240_to_fp16, x = input_203_cast_fp16)[name = tensor("input_207_cast_fp16")]; - tensor input_209_cast_fp16 = silu(x = input_207_cast_fp16)[name = tensor("input_209_cast_fp16")]; - tensor x_87_pad_type_0 = const()[name = tensor("x_87_pad_type_0"), val = tensor("valid")]; - tensor x_87_strides_0 = const()[name = tensor("x_87_strides_0"), val = tensor([1])]; - tensor x_87_pad_0 = const()[name = tensor("x_87_pad_0"), val = tensor([0, 0])]; - tensor x_87_dilations_0 = const()[name = tensor("x_87_dilations_0"), val = tensor([1])]; - tensor x_87_groups_0 = const()[name = tensor("x_87_groups_0"), val = tensor(1)]; - tensor module_layers_3_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48111552)))]; - tensor module_layers_3_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_3_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48635904)))]; - tensor x_87_cast_fp16 = conv(bias = module_layers_3_conv_pointwise_conv2_bias_to_fp16, dilations = x_87_dilations_0, groups = x_87_groups_0, pad = x_87_pad_0, pad_type = x_87_pad_type_0, strides = x_87_strides_0, weight = module_layers_3_conv_pointwise_conv2_weight_to_fp16, x = input_209_cast_fp16)[name = tensor("x_87_cast_fp16")]; - tensor input_211_perm_0 = const()[name = tensor("input_211_perm_0"), val = tensor([0, 2, 1])]; - tensor input_211_cast_fp16 = transpose(perm = input_211_perm_0, x = x_87_cast_fp16)[name = tensor("transpose_177")]; - tensor input_213_cast_fp16 = add(x = input_195_cast_fp16, y = input_211_cast_fp16)[name = tensor("input_213_cast_fp16")]; - tensor input_215_axes_0 = const()[name = tensor("input_215_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48636992)))]; - tensor module_layers_3_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48638080)))]; - tensor input_215_cast_fp16 = layer_norm(axes = input_215_axes_0, beta = module_layers_3_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_feed_forward2_weight_to_fp16, x = input_213_cast_fp16)[name = tensor("input_215_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(48639168)))]; - tensor module_layers_3_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50736384)))]; - tensor linear_35_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear1_bias_to_fp16, weight = module_layers_3_feed_forward2_linear1_weight_to_fp16, x = input_215_cast_fp16)[name = tensor("linear_35_cast_fp16")]; - tensor input_219_cast_fp16 = silu(x = linear_35_cast_fp16)[name = tensor("input_219_cast_fp16")]; - tensor module_layers_3_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(50740544)))]; - tensor module_layers_3_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_3_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52837760)))]; - tensor linear_36_cast_fp16 = linear(bias = module_layers_3_feed_forward2_linear2_bias_to_fp16, weight = module_layers_3_feed_forward2_linear2_weight_to_fp16, x = input_219_cast_fp16)[name = tensor("linear_36_cast_fp16")]; - tensor var_1017_to_fp16 = const()[name = tensor("op_1017_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1018_cast_fp16 = mul(x = linear_36_cast_fp16, y = var_1017_to_fp16)[name = tensor("op_1018_cast_fp16")]; - tensor input_225_cast_fp16 = add(x = input_213_cast_fp16, y = var_1018_cast_fp16)[name = tensor("input_225_cast_fp16")]; - tensor input_227_axes_0 = const()[name = tensor("input_227_axes_0"), val = tensor([-1])]; - tensor module_layers_3_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_3_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52838848)))]; - tensor module_layers_3_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_3_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52839936)))]; - tensor input_227_cast_fp16 = layer_norm(axes = input_227_axes_0, beta = module_layers_3_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_3_norm_out_weight_to_fp16, x = input_225_cast_fp16)[name = tensor("input_227_cast_fp16")]; - tensor input_229_axes_0 = const()[name = tensor("input_229_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52841024)))]; - tensor module_layers_4_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52842112)))]; - tensor input_229_cast_fp16 = layer_norm(axes = input_229_axes_0, beta = module_layers_4_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward1_weight_to_fp16, x = input_227_cast_fp16)[name = tensor("input_229_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(52843200)))]; - tensor module_layers_4_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54940416)))]; - tensor linear_37_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear1_bias_to_fp16, weight = module_layers_4_feed_forward1_linear1_weight_to_fp16, x = input_229_cast_fp16)[name = tensor("linear_37_cast_fp16")]; - tensor input_233_cast_fp16 = silu(x = linear_37_cast_fp16)[name = tensor("input_233_cast_fp16")]; - tensor module_layers_4_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(54944576)))]; - tensor module_layers_4_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57041792)))]; - tensor linear_38_cast_fp16 = linear(bias = module_layers_4_feed_forward1_linear2_bias_to_fp16, weight = module_layers_4_feed_forward1_linear2_weight_to_fp16, x = input_233_cast_fp16)[name = tensor("linear_38_cast_fp16")]; - tensor var_1048_to_fp16 = const()[name = tensor("op_1048_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1049_cast_fp16 = mul(x = linear_38_cast_fp16, y = var_1048_to_fp16)[name = tensor("op_1049_cast_fp16")]; - tensor input_239_cast_fp16 = add(x = input_227_cast_fp16, y = var_1049_cast_fp16)[name = tensor("input_239_cast_fp16")]; - tensor query_9_axes_0 = const()[name = tensor("query_9_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57042880)))]; - tensor module_layers_4_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57043968)))]; - tensor query_9_cast_fp16 = layer_norm(axes = query_9_axes_0, beta = module_layers_4_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_self_att_weight_to_fp16, x = input_239_cast_fp16)[name = tensor("query_9_cast_fp16")]; - tensor module_layers_4_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57045056)))]; - tensor module_layers_4_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57569408)))]; - tensor linear_39_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_q_bias_to_fp16, weight = module_layers_4_self_attn_linear_q_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_39_cast_fp16")]; - tensor var_1066 = const()[name = tensor("op_1066"), val = tensor([1, -1, 8, 64])]; - tensor q_25_cast_fp16 = reshape(shape = var_1066, x = linear_39_cast_fp16)[name = tensor("q_25_cast_fp16")]; - tensor module_layers_4_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(57570496)))]; - tensor module_layers_4_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58094848)))]; - tensor linear_40_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_k_bias_to_fp16, weight = module_layers_4_self_attn_linear_k_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_40_cast_fp16")]; - tensor var_1071 = const()[name = tensor("op_1071"), val = tensor([1, -1, 8, 64])]; - tensor k_17_cast_fp16 = reshape(shape = var_1071, x = linear_40_cast_fp16)[name = tensor("k_17_cast_fp16")]; - tensor module_layers_4_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58095936)))]; - tensor module_layers_4_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58620288)))]; - tensor linear_41_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_v_bias_to_fp16, weight = module_layers_4_self_attn_linear_v_weight_to_fp16, x = query_9_cast_fp16)[name = tensor("linear_41_cast_fp16")]; - tensor var_1076 = const()[name = tensor("op_1076"), val = tensor([1, -1, 8, 64])]; - tensor v_9_cast_fp16 = reshape(shape = var_1076, x = linear_41_cast_fp16)[name = tensor("v_9_cast_fp16")]; - tensor value_11_perm_0 = const()[name = tensor("value_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_4_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58621376)))]; - tensor var_1088_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1088_cast_fp16")]; - tensor module_layers_4_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_4_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58622464)))]; - tensor var_1090_cast_fp16 = add(x = q_25_cast_fp16, y = module_layers_4_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1090_cast_fp16")]; - tensor q_with_bias_v_9_perm_0 = const()[name = tensor("q_with_bias_v_9_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_95_transpose_x_0 = const()[name = tensor("x_95_transpose_x_0"), val = tensor(false)]; - tensor x_95_transpose_y_0 = const()[name = tensor("x_95_transpose_y_0"), val = tensor(false)]; - tensor var_1092_to_fp16 = const()[name = tensor("op_1092_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(58623552)))]; - tensor q_with_bias_v_9_cast_fp16 = transpose(perm = q_with_bias_v_9_perm_0, x = var_1090_cast_fp16)[name = tensor("transpose_175")]; - tensor x_95_cast_fp16 = matmul(transpose_x = x_95_transpose_x_0, transpose_y = x_95_transpose_y_0, x = q_with_bias_v_9_cast_fp16, y = var_1092_to_fp16)[name = tensor("x_95_cast_fp16")]; - tensor x_97_pad_0 = const()[name = tensor("x_97_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_97_mode_0 = const()[name = tensor("x_97_mode_0"), val = tensor("constant")]; - tensor const_110_to_fp16 = const()[name = tensor("const_110_to_fp16"), val = tensor(0x0p+0)]; - tensor x_97_cast_fp16 = pad(constant_val = const_110_to_fp16, mode = x_97_mode_0, pad = x_97_pad_0, x = x_95_cast_fp16)[name = tensor("x_97_cast_fp16")]; - tensor var_1100 = const()[name = tensor("op_1100"), val = tensor([1, 8, -1, 188])]; - tensor x_99_cast_fp16 = reshape(shape = var_1100, x = x_97_cast_fp16)[name = tensor("x_99_cast_fp16")]; - tensor var_1104_begin_0 = const()[name = tensor("op_1104_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1104_end_0 = const()[name = tensor("op_1104_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1104_end_mask_0 = const()[name = tensor("op_1104_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1104_cast_fp16 = slice_by_index(begin = var_1104_begin_0, end = var_1104_end_0, end_mask = var_1104_end_mask_0, x = x_99_cast_fp16)[name = tensor("op_1104_cast_fp16")]; - tensor var_1105 = const()[name = tensor("op_1105"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_17_cast_fp16 = reshape(shape = var_1105, x = var_1104_cast_fp16)[name = tensor("matrix_bd_17_cast_fp16")]; - tensor matrix_ac_9_transpose_x_0 = const()[name = tensor("matrix_ac_9_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_9_transpose_y_0 = const()[name = tensor("matrix_ac_9_transpose_y_0"), val = tensor(false)]; - tensor transpose_59_perm_0 = const()[name = tensor("transpose_59_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_60_perm_0 = const()[name = tensor("transpose_60_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_60 = transpose(perm = transpose_60_perm_0, x = k_17_cast_fp16)[name = tensor("transpose_173")]; - tensor transpose_59 = transpose(perm = transpose_59_perm_0, x = var_1088_cast_fp16)[name = tensor("transpose_174")]; - tensor matrix_ac_9_cast_fp16 = matmul(transpose_x = matrix_ac_9_transpose_x_0, transpose_y = matrix_ac_9_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor("matrix_ac_9_cast_fp16")]; - tensor matrix_bd_19_begin_0 = const()[name = tensor("matrix_bd_19_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_19_end_0 = const()[name = tensor("matrix_bd_19_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_19_end_mask_0 = const()[name = tensor("matrix_bd_19_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_19_cast_fp16 = slice_by_index(begin = matrix_bd_19_begin_0, end = matrix_bd_19_end_0, end_mask = matrix_bd_19_end_mask_0, x = matrix_bd_17_cast_fp16)[name = tensor("matrix_bd_19_cast_fp16")]; - tensor var_1114_cast_fp16 = add(x = matrix_ac_9_cast_fp16, y = matrix_bd_19_cast_fp16)[name = tensor("op_1114_cast_fp16")]; - tensor _inversed_scores_17_y_0_to_fp16 = const()[name = tensor("_inversed_scores_17_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_17_cast_fp16 = mul(x = var_1114_cast_fp16, y = _inversed_scores_17_y_0_to_fp16)[name = tensor("_inversed_scores_17_cast_fp16")]; - tensor scores_19_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_17_cast_fp16, cond = mask_11)[name = tensor("scores_19_cast_fp16")]; - tensor var_1120_cast_fp16 = softmax(axis = var_23, x = scores_19_cast_fp16)[name = tensor("op_1120_cast_fp16")]; - tensor input_241_cast_fp16 = select(a = var_11_to_fp16, b = var_1120_cast_fp16, cond = mask_11)[name = tensor("input_241_cast_fp16")]; - tensor x_101_transpose_x_0 = const()[name = tensor("x_101_transpose_x_0"), val = tensor(false)]; - tensor x_101_transpose_y_0 = const()[name = tensor("x_101_transpose_y_0"), val = tensor(false)]; - tensor value_11_cast_fp16 = transpose(perm = value_11_perm_0, x = v_9_cast_fp16)[name = tensor("transpose_176")]; - tensor x_101_cast_fp16 = matmul(transpose_x = x_101_transpose_x_0, transpose_y = x_101_transpose_y_0, x = input_241_cast_fp16, y = value_11_cast_fp16)[name = tensor("x_101_cast_fp16")]; - tensor var_1124_perm_0 = const()[name = tensor("op_1124_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1125 = const()[name = tensor("op_1125"), val = tensor([1, -1, 512])]; - tensor var_1124_cast_fp16 = transpose(perm = var_1124_perm_0, x = x_101_cast_fp16)[name = tensor("transpose_172")]; - tensor input_243_cast_fp16 = reshape(shape = var_1125, x = var_1124_cast_fp16)[name = tensor("input_243_cast_fp16")]; - tensor module_layers_4_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59007616)))]; - tensor module_layers_4_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_4_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59531968)))]; - tensor linear_43_cast_fp16 = linear(bias = module_layers_4_self_attn_linear_out_bias_to_fp16, weight = module_layers_4_self_attn_linear_out_weight_to_fp16, x = input_243_cast_fp16)[name = tensor("linear_43_cast_fp16")]; - tensor input_247_cast_fp16 = add(x = input_239_cast_fp16, y = linear_43_cast_fp16)[name = tensor("input_247_cast_fp16")]; - tensor x_105_axes_0 = const()[name = tensor("x_105_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59533056)))]; - tensor module_layers_4_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59534144)))]; - tensor x_105_cast_fp16 = layer_norm(axes = x_105_axes_0, beta = module_layers_4_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_conv_weight_to_fp16, x = input_247_cast_fp16)[name = tensor("x_105_cast_fp16")]; - tensor input_249_perm_0 = const()[name = tensor("input_249_perm_0"), val = tensor([0, 2, 1])]; - tensor input_251_pad_type_0 = const()[name = tensor("input_251_pad_type_0"), val = tensor("valid")]; - tensor input_251_strides_0 = const()[name = tensor("input_251_strides_0"), val = tensor([1])]; - tensor input_251_pad_0 = const()[name = tensor("input_251_pad_0"), val = tensor([0, 0])]; - tensor input_251_dilations_0 = const()[name = tensor("input_251_dilations_0"), val = tensor([1])]; - tensor input_251_groups_0 = const()[name = tensor("input_251_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(59535232)))]; - tensor module_layers_4_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60583872)))]; - tensor input_249_cast_fp16 = transpose(perm = input_249_perm_0, x = x_105_cast_fp16)[name = tensor("transpose_171")]; - tensor input_251_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv1_bias_to_fp16, dilations = input_251_dilations_0, groups = input_251_groups_0, pad = input_251_pad_0, pad_type = input_251_pad_type_0, strides = input_251_strides_0, weight = module_layers_4_conv_pointwise_conv1_weight_to_fp16, x = input_249_cast_fp16)[name = tensor("input_251_cast_fp16")]; - tensor x_107_split_num_splits_0 = const()[name = tensor("x_107_split_num_splits_0"), val = tensor(2)]; - tensor x_107_split_axis_0 = const()[name = tensor("x_107_split_axis_0"), val = tensor(1)]; - tensor x_107_split_cast_fp16_0, tensor x_107_split_cast_fp16_1 = split(axis = x_107_split_axis_0, num_splits = x_107_split_num_splits_0, x = input_251_cast_fp16)[name = tensor("x_107_split_cast_fp16")]; - tensor x_107_split_1_sigmoid_cast_fp16 = sigmoid(x = x_107_split_cast_fp16_1)[name = tensor("x_107_split_1_sigmoid_cast_fp16")]; - tensor x_107_cast_fp16 = mul(x = x_107_split_cast_fp16_0, y = x_107_split_1_sigmoid_cast_fp16)[name = tensor("x_107_cast_fp16")]; - tensor input_253_cast_fp16 = select(a = var_11_to_fp16, b = x_107_cast_fp16, cond = var_453)[name = tensor("input_253_cast_fp16")]; - tensor input_255_pad_0 = const()[name = tensor("input_255_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_255_mode_0 = const()[name = tensor("input_255_mode_0"), val = tensor("constant")]; - tensor const_113_to_fp16 = const()[name = tensor("const_113_to_fp16"), val = tensor(0x0p+0)]; - tensor input_255_cast_fp16 = pad(constant_val = const_113_to_fp16, mode = input_255_mode_0, pad = input_255_pad_0, x = input_253_cast_fp16)[name = tensor("input_255_cast_fp16")]; - tensor input_257_pad_type_0 = const()[name = tensor("input_257_pad_type_0"), val = tensor("valid")]; - tensor input_257_groups_0 = const()[name = tensor("input_257_groups_0"), val = tensor(512)]; - tensor input_257_strides_0 = const()[name = tensor("input_257_strides_0"), val = tensor([1])]; - tensor input_257_pad_0 = const()[name = tensor("input_257_pad_0"), val = tensor([0, 0])]; - tensor input_257_dilations_0 = const()[name = tensor("input_257_dilations_0"), val = tensor([1])]; - tensor const_242_to_fp16 = const()[name = tensor("const_242_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60585984)))]; - tensor const_243_to_fp16 = const()[name = tensor("const_243_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60595264)))]; - tensor input_259_cast_fp16 = conv(bias = const_243_to_fp16, dilations = input_257_dilations_0, groups = input_257_groups_0, pad = input_257_pad_0, pad_type = input_257_pad_type_0, strides = input_257_strides_0, weight = const_242_to_fp16, x = input_255_cast_fp16)[name = tensor("input_259_cast_fp16")]; - tensor input_261_cast_fp16 = silu(x = input_259_cast_fp16)[name = tensor("input_261_cast_fp16")]; - tensor x_109_pad_type_0 = const()[name = tensor("x_109_pad_type_0"), val = tensor("valid")]; - tensor x_109_strides_0 = const()[name = tensor("x_109_strides_0"), val = tensor([1])]; - tensor x_109_pad_0 = const()[name = tensor("x_109_pad_0"), val = tensor([0, 0])]; - tensor x_109_dilations_0 = const()[name = tensor("x_109_dilations_0"), val = tensor([1])]; - tensor x_109_groups_0 = const()[name = tensor("x_109_groups_0"), val = tensor(1)]; - tensor module_layers_4_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(60596352)))]; - tensor module_layers_4_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_4_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61120704)))]; - tensor x_109_cast_fp16 = conv(bias = module_layers_4_conv_pointwise_conv2_bias_to_fp16, dilations = x_109_dilations_0, groups = x_109_groups_0, pad = x_109_pad_0, pad_type = x_109_pad_type_0, strides = x_109_strides_0, weight = module_layers_4_conv_pointwise_conv2_weight_to_fp16, x = input_261_cast_fp16)[name = tensor("x_109_cast_fp16")]; - tensor input_263_perm_0 = const()[name = tensor("input_263_perm_0"), val = tensor([0, 2, 1])]; - tensor input_263_cast_fp16 = transpose(perm = input_263_perm_0, x = x_109_cast_fp16)[name = tensor("transpose_170")]; - tensor input_265_cast_fp16 = add(x = input_247_cast_fp16, y = input_263_cast_fp16)[name = tensor("input_265_cast_fp16")]; - tensor input_267_axes_0 = const()[name = tensor("input_267_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61121792)))]; - tensor module_layers_4_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61122880)))]; - tensor input_267_cast_fp16 = layer_norm(axes = input_267_axes_0, beta = module_layers_4_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_feed_forward2_weight_to_fp16, x = input_265_cast_fp16)[name = tensor("input_267_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(61123968)))]; - tensor module_layers_4_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63221184)))]; - tensor linear_44_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear1_bias_to_fp16, weight = module_layers_4_feed_forward2_linear1_weight_to_fp16, x = input_267_cast_fp16)[name = tensor("linear_44_cast_fp16")]; - tensor input_271_cast_fp16 = silu(x = linear_44_cast_fp16)[name = tensor("input_271_cast_fp16")]; - tensor module_layers_4_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(63225344)))]; - tensor module_layers_4_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_4_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65322560)))]; - tensor linear_45_cast_fp16 = linear(bias = module_layers_4_feed_forward2_linear2_bias_to_fp16, weight = module_layers_4_feed_forward2_linear2_weight_to_fp16, x = input_271_cast_fp16)[name = tensor("linear_45_cast_fp16")]; - tensor var_1191_to_fp16 = const()[name = tensor("op_1191_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1192_cast_fp16 = mul(x = linear_45_cast_fp16, y = var_1191_to_fp16)[name = tensor("op_1192_cast_fp16")]; - tensor input_277_cast_fp16 = add(x = input_265_cast_fp16, y = var_1192_cast_fp16)[name = tensor("input_277_cast_fp16")]; - tensor input_279_axes_0 = const()[name = tensor("input_279_axes_0"), val = tensor([-1])]; - tensor module_layers_4_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_4_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65323648)))]; - tensor module_layers_4_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_4_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65324736)))]; - tensor input_279_cast_fp16 = layer_norm(axes = input_279_axes_0, beta = module_layers_4_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_4_norm_out_weight_to_fp16, x = input_277_cast_fp16)[name = tensor("input_279_cast_fp16")]; - tensor input_281_axes_0 = const()[name = tensor("input_281_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65325824)))]; - tensor module_layers_5_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65326912)))]; - tensor input_281_cast_fp16 = layer_norm(axes = input_281_axes_0, beta = module_layers_5_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward1_weight_to_fp16, x = input_279_cast_fp16)[name = tensor("input_281_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(65328000)))]; - tensor module_layers_5_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67425216)))]; - tensor linear_46_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear1_bias_to_fp16, weight = module_layers_5_feed_forward1_linear1_weight_to_fp16, x = input_281_cast_fp16)[name = tensor("linear_46_cast_fp16")]; - tensor input_285_cast_fp16 = silu(x = linear_46_cast_fp16)[name = tensor("input_285_cast_fp16")]; - tensor module_layers_5_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(67429376)))]; - tensor module_layers_5_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69526592)))]; - tensor linear_47_cast_fp16 = linear(bias = module_layers_5_feed_forward1_linear2_bias_to_fp16, weight = module_layers_5_feed_forward1_linear2_weight_to_fp16, x = input_285_cast_fp16)[name = tensor("linear_47_cast_fp16")]; - tensor var_1222_to_fp16 = const()[name = tensor("op_1222_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1223_cast_fp16 = mul(x = linear_47_cast_fp16, y = var_1222_to_fp16)[name = tensor("op_1223_cast_fp16")]; - tensor input_291_cast_fp16 = add(x = input_279_cast_fp16, y = var_1223_cast_fp16)[name = tensor("input_291_cast_fp16")]; - tensor query_11_axes_0 = const()[name = tensor("query_11_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69527680)))]; - tensor module_layers_5_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69528768)))]; - tensor query_11_cast_fp16 = layer_norm(axes = query_11_axes_0, beta = module_layers_5_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_self_att_weight_to_fp16, x = input_291_cast_fp16)[name = tensor("query_11_cast_fp16")]; - tensor module_layers_5_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(69529856)))]; - tensor module_layers_5_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70054208)))]; - tensor linear_48_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_q_bias_to_fp16, weight = module_layers_5_self_attn_linear_q_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_48_cast_fp16")]; - tensor var_1240 = const()[name = tensor("op_1240"), val = tensor([1, -1, 8, 64])]; - tensor q_31_cast_fp16 = reshape(shape = var_1240, x = linear_48_cast_fp16)[name = tensor("q_31_cast_fp16")]; - tensor module_layers_5_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70055296)))]; - tensor module_layers_5_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70579648)))]; - tensor linear_49_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_k_bias_to_fp16, weight = module_layers_5_self_attn_linear_k_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_49_cast_fp16")]; - tensor var_1245 = const()[name = tensor("op_1245"), val = tensor([1, -1, 8, 64])]; - tensor k_21_cast_fp16 = reshape(shape = var_1245, x = linear_49_cast_fp16)[name = tensor("k_21_cast_fp16")]; - tensor module_layers_5_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(70580736)))]; - tensor module_layers_5_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71105088)))]; - tensor linear_50_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_v_bias_to_fp16, weight = module_layers_5_self_attn_linear_v_weight_to_fp16, x = query_11_cast_fp16)[name = tensor("linear_50_cast_fp16")]; - tensor var_1250 = const()[name = tensor("op_1250"), val = tensor([1, -1, 8, 64])]; - tensor v_11_cast_fp16 = reshape(shape = var_1250, x = linear_50_cast_fp16)[name = tensor("v_11_cast_fp16")]; - tensor value_13_perm_0 = const()[name = tensor("value_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_5_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71106176)))]; - tensor var_1262_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1262_cast_fp16")]; - tensor module_layers_5_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_5_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71107264)))]; - tensor var_1264_cast_fp16 = add(x = q_31_cast_fp16, y = module_layers_5_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1264_cast_fp16")]; - tensor q_with_bias_v_11_perm_0 = const()[name = tensor("q_with_bias_v_11_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_117_transpose_x_0 = const()[name = tensor("x_117_transpose_x_0"), val = tensor(false)]; - tensor x_117_transpose_y_0 = const()[name = tensor("x_117_transpose_y_0"), val = tensor(false)]; - tensor var_1266_to_fp16 = const()[name = tensor("op_1266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71108352)))]; - tensor q_with_bias_v_11_cast_fp16 = transpose(perm = q_with_bias_v_11_perm_0, x = var_1264_cast_fp16)[name = tensor("transpose_168")]; - tensor x_117_cast_fp16 = matmul(transpose_x = x_117_transpose_x_0, transpose_y = x_117_transpose_y_0, x = q_with_bias_v_11_cast_fp16, y = var_1266_to_fp16)[name = tensor("x_117_cast_fp16")]; - tensor x_119_pad_0 = const()[name = tensor("x_119_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_119_mode_0 = const()[name = tensor("x_119_mode_0"), val = tensor("constant")]; - tensor const_120_to_fp16 = const()[name = tensor("const_120_to_fp16"), val = tensor(0x0p+0)]; - tensor x_119_cast_fp16 = pad(constant_val = const_120_to_fp16, mode = x_119_mode_0, pad = x_119_pad_0, x = x_117_cast_fp16)[name = tensor("x_119_cast_fp16")]; - tensor var_1274 = const()[name = tensor("op_1274"), val = tensor([1, 8, -1, 188])]; - tensor x_121_cast_fp16 = reshape(shape = var_1274, x = x_119_cast_fp16)[name = tensor("x_121_cast_fp16")]; - tensor var_1278_begin_0 = const()[name = tensor("op_1278_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1278_end_0 = const()[name = tensor("op_1278_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1278_end_mask_0 = const()[name = tensor("op_1278_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1278_cast_fp16 = slice_by_index(begin = var_1278_begin_0, end = var_1278_end_0, end_mask = var_1278_end_mask_0, x = x_121_cast_fp16)[name = tensor("op_1278_cast_fp16")]; - tensor var_1279 = const()[name = tensor("op_1279"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_21_cast_fp16 = reshape(shape = var_1279, x = var_1278_cast_fp16)[name = tensor("matrix_bd_21_cast_fp16")]; - tensor matrix_ac_11_transpose_x_0 = const()[name = tensor("matrix_ac_11_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_11_transpose_y_0 = const()[name = tensor("matrix_ac_11_transpose_y_0"), val = tensor(false)]; - tensor transpose_61_perm_0 = const()[name = tensor("transpose_61_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_62_perm_0 = const()[name = tensor("transpose_62_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_62 = transpose(perm = transpose_62_perm_0, x = k_21_cast_fp16)[name = tensor("transpose_166")]; - tensor transpose_61 = transpose(perm = transpose_61_perm_0, x = var_1262_cast_fp16)[name = tensor("transpose_167")]; - tensor matrix_ac_11_cast_fp16 = matmul(transpose_x = matrix_ac_11_transpose_x_0, transpose_y = matrix_ac_11_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor("matrix_ac_11_cast_fp16")]; - tensor matrix_bd_23_begin_0 = const()[name = tensor("matrix_bd_23_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_23_end_0 = const()[name = tensor("matrix_bd_23_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_23_end_mask_0 = const()[name = tensor("matrix_bd_23_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_23_cast_fp16 = slice_by_index(begin = matrix_bd_23_begin_0, end = matrix_bd_23_end_0, end_mask = matrix_bd_23_end_mask_0, x = matrix_bd_21_cast_fp16)[name = tensor("matrix_bd_23_cast_fp16")]; - tensor var_1288_cast_fp16 = add(x = matrix_ac_11_cast_fp16, y = matrix_bd_23_cast_fp16)[name = tensor("op_1288_cast_fp16")]; - tensor _inversed_scores_21_y_0_to_fp16 = const()[name = tensor("_inversed_scores_21_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_21_cast_fp16 = mul(x = var_1288_cast_fp16, y = _inversed_scores_21_y_0_to_fp16)[name = tensor("_inversed_scores_21_cast_fp16")]; - tensor scores_23_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_21_cast_fp16, cond = mask_11)[name = tensor("scores_23_cast_fp16")]; - tensor var_1294_cast_fp16 = softmax(axis = var_23, x = scores_23_cast_fp16)[name = tensor("op_1294_cast_fp16")]; - tensor input_293_cast_fp16 = select(a = var_11_to_fp16, b = var_1294_cast_fp16, cond = mask_11)[name = tensor("input_293_cast_fp16")]; - tensor x_123_transpose_x_0 = const()[name = tensor("x_123_transpose_x_0"), val = tensor(false)]; - tensor x_123_transpose_y_0 = const()[name = tensor("x_123_transpose_y_0"), val = tensor(false)]; - tensor value_13_cast_fp16 = transpose(perm = value_13_perm_0, x = v_11_cast_fp16)[name = tensor("transpose_169")]; - tensor x_123_cast_fp16 = matmul(transpose_x = x_123_transpose_x_0, transpose_y = x_123_transpose_y_0, x = input_293_cast_fp16, y = value_13_cast_fp16)[name = tensor("x_123_cast_fp16")]; - tensor var_1298_perm_0 = const()[name = tensor("op_1298_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1299 = const()[name = tensor("op_1299"), val = tensor([1, -1, 512])]; - tensor var_1298_cast_fp16 = transpose(perm = var_1298_perm_0, x = x_123_cast_fp16)[name = tensor("transpose_165")]; - tensor input_295_cast_fp16 = reshape(shape = var_1299, x = var_1298_cast_fp16)[name = tensor("input_295_cast_fp16")]; - tensor module_layers_5_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(71492416)))]; - tensor module_layers_5_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_5_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72016768)))]; - tensor linear_52_cast_fp16 = linear(bias = module_layers_5_self_attn_linear_out_bias_to_fp16, weight = module_layers_5_self_attn_linear_out_weight_to_fp16, x = input_295_cast_fp16)[name = tensor("linear_52_cast_fp16")]; - tensor input_299_cast_fp16 = add(x = input_291_cast_fp16, y = linear_52_cast_fp16)[name = tensor("input_299_cast_fp16")]; - tensor x_127_axes_0 = const()[name = tensor("x_127_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72017856)))]; - tensor module_layers_5_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72018944)))]; - tensor x_127_cast_fp16 = layer_norm(axes = x_127_axes_0, beta = module_layers_5_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_conv_weight_to_fp16, x = input_299_cast_fp16)[name = tensor("x_127_cast_fp16")]; - tensor input_301_perm_0 = const()[name = tensor("input_301_perm_0"), val = tensor([0, 2, 1])]; - tensor input_303_pad_type_0 = const()[name = tensor("input_303_pad_type_0"), val = tensor("valid")]; - tensor input_303_strides_0 = const()[name = tensor("input_303_strides_0"), val = tensor([1])]; - tensor input_303_pad_0 = const()[name = tensor("input_303_pad_0"), val = tensor([0, 0])]; - tensor input_303_dilations_0 = const()[name = tensor("input_303_dilations_0"), val = tensor([1])]; - tensor input_303_groups_0 = const()[name = tensor("input_303_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(72020032)))]; - tensor module_layers_5_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73068672)))]; - tensor input_301_cast_fp16 = transpose(perm = input_301_perm_0, x = x_127_cast_fp16)[name = tensor("transpose_164")]; - tensor input_303_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv1_bias_to_fp16, dilations = input_303_dilations_0, groups = input_303_groups_0, pad = input_303_pad_0, pad_type = input_303_pad_type_0, strides = input_303_strides_0, weight = module_layers_5_conv_pointwise_conv1_weight_to_fp16, x = input_301_cast_fp16)[name = tensor("input_303_cast_fp16")]; - tensor x_129_split_num_splits_0 = const()[name = tensor("x_129_split_num_splits_0"), val = tensor(2)]; - tensor x_129_split_axis_0 = const()[name = tensor("x_129_split_axis_0"), val = tensor(1)]; - tensor x_129_split_cast_fp16_0, tensor x_129_split_cast_fp16_1 = split(axis = x_129_split_axis_0, num_splits = x_129_split_num_splits_0, x = input_303_cast_fp16)[name = tensor("x_129_split_cast_fp16")]; - tensor x_129_split_1_sigmoid_cast_fp16 = sigmoid(x = x_129_split_cast_fp16_1)[name = tensor("x_129_split_1_sigmoid_cast_fp16")]; - tensor x_129_cast_fp16 = mul(x = x_129_split_cast_fp16_0, y = x_129_split_1_sigmoid_cast_fp16)[name = tensor("x_129_cast_fp16")]; - tensor input_305_cast_fp16 = select(a = var_11_to_fp16, b = x_129_cast_fp16, cond = var_453)[name = tensor("input_305_cast_fp16")]; - tensor input_307_pad_0 = const()[name = tensor("input_307_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_307_mode_0 = const()[name = tensor("input_307_mode_0"), val = tensor("constant")]; - tensor const_123_to_fp16 = const()[name = tensor("const_123_to_fp16"), val = tensor(0x0p+0)]; - tensor input_307_cast_fp16 = pad(constant_val = const_123_to_fp16, mode = input_307_mode_0, pad = input_307_pad_0, x = input_305_cast_fp16)[name = tensor("input_307_cast_fp16")]; - tensor input_309_pad_type_0 = const()[name = tensor("input_309_pad_type_0"), val = tensor("valid")]; - tensor input_309_groups_0 = const()[name = tensor("input_309_groups_0"), val = tensor(512)]; - tensor input_309_strides_0 = const()[name = tensor("input_309_strides_0"), val = tensor([1])]; - tensor input_309_pad_0 = const()[name = tensor("input_309_pad_0"), val = tensor([0, 0])]; - tensor input_309_dilations_0 = const()[name = tensor("input_309_dilations_0"), val = tensor([1])]; - tensor const_244_to_fp16 = const()[name = tensor("const_244_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73070784)))]; - tensor const_245_to_fp16 = const()[name = tensor("const_245_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73080064)))]; - tensor input_311_cast_fp16 = conv(bias = const_245_to_fp16, dilations = input_309_dilations_0, groups = input_309_groups_0, pad = input_309_pad_0, pad_type = input_309_pad_type_0, strides = input_309_strides_0, weight = const_244_to_fp16, x = input_307_cast_fp16)[name = tensor("input_311_cast_fp16")]; - tensor input_313_cast_fp16 = silu(x = input_311_cast_fp16)[name = tensor("input_313_cast_fp16")]; - tensor x_131_pad_type_0 = const()[name = tensor("x_131_pad_type_0"), val = tensor("valid")]; - tensor x_131_strides_0 = const()[name = tensor("x_131_strides_0"), val = tensor([1])]; - tensor x_131_pad_0 = const()[name = tensor("x_131_pad_0"), val = tensor([0, 0])]; - tensor x_131_dilations_0 = const()[name = tensor("x_131_dilations_0"), val = tensor([1])]; - tensor x_131_groups_0 = const()[name = tensor("x_131_groups_0"), val = tensor(1)]; - tensor module_layers_5_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73081152)))]; - tensor module_layers_5_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_5_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73605504)))]; - tensor x_131_cast_fp16 = conv(bias = module_layers_5_conv_pointwise_conv2_bias_to_fp16, dilations = x_131_dilations_0, groups = x_131_groups_0, pad = x_131_pad_0, pad_type = x_131_pad_type_0, strides = x_131_strides_0, weight = module_layers_5_conv_pointwise_conv2_weight_to_fp16, x = input_313_cast_fp16)[name = tensor("x_131_cast_fp16")]; - tensor input_315_perm_0 = const()[name = tensor("input_315_perm_0"), val = tensor([0, 2, 1])]; - tensor input_315_cast_fp16 = transpose(perm = input_315_perm_0, x = x_131_cast_fp16)[name = tensor("transpose_163")]; - tensor input_317_cast_fp16 = add(x = input_299_cast_fp16, y = input_315_cast_fp16)[name = tensor("input_317_cast_fp16")]; - tensor input_319_axes_0 = const()[name = tensor("input_319_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73606592)))]; - tensor module_layers_5_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73607680)))]; - tensor input_319_cast_fp16 = layer_norm(axes = input_319_axes_0, beta = module_layers_5_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_feed_forward2_weight_to_fp16, x = input_317_cast_fp16)[name = tensor("input_319_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(73608768)))]; - tensor module_layers_5_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75705984)))]; - tensor linear_53_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear1_bias_to_fp16, weight = module_layers_5_feed_forward2_linear1_weight_to_fp16, x = input_319_cast_fp16)[name = tensor("linear_53_cast_fp16")]; - tensor input_323_cast_fp16 = silu(x = linear_53_cast_fp16)[name = tensor("input_323_cast_fp16")]; - tensor module_layers_5_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(75710144)))]; - tensor module_layers_5_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_5_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77807360)))]; - tensor linear_54_cast_fp16 = linear(bias = module_layers_5_feed_forward2_linear2_bias_to_fp16, weight = module_layers_5_feed_forward2_linear2_weight_to_fp16, x = input_323_cast_fp16)[name = tensor("linear_54_cast_fp16")]; - tensor var_1365_to_fp16 = const()[name = tensor("op_1365_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1366_cast_fp16 = mul(x = linear_54_cast_fp16, y = var_1365_to_fp16)[name = tensor("op_1366_cast_fp16")]; - tensor input_329_cast_fp16 = add(x = input_317_cast_fp16, y = var_1366_cast_fp16)[name = tensor("input_329_cast_fp16")]; - tensor input_331_axes_0 = const()[name = tensor("input_331_axes_0"), val = tensor([-1])]; - tensor module_layers_5_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_5_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77808448)))]; - tensor module_layers_5_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_5_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77809536)))]; - tensor input_331_cast_fp16 = layer_norm(axes = input_331_axes_0, beta = module_layers_5_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_5_norm_out_weight_to_fp16, x = input_329_cast_fp16)[name = tensor("input_331_cast_fp16")]; - tensor input_333_axes_0 = const()[name = tensor("input_333_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77810624)))]; - tensor module_layers_6_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77811712)))]; - tensor input_333_cast_fp16 = layer_norm(axes = input_333_axes_0, beta = module_layers_6_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward1_weight_to_fp16, x = input_331_cast_fp16)[name = tensor("input_333_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(77812800)))]; - tensor module_layers_6_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79910016)))]; - tensor linear_55_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear1_bias_to_fp16, weight = module_layers_6_feed_forward1_linear1_weight_to_fp16, x = input_333_cast_fp16)[name = tensor("linear_55_cast_fp16")]; - tensor input_337_cast_fp16 = silu(x = linear_55_cast_fp16)[name = tensor("input_337_cast_fp16")]; - tensor module_layers_6_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(79914176)))]; - tensor module_layers_6_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82011392)))]; - tensor linear_56_cast_fp16 = linear(bias = module_layers_6_feed_forward1_linear2_bias_to_fp16, weight = module_layers_6_feed_forward1_linear2_weight_to_fp16, x = input_337_cast_fp16)[name = tensor("linear_56_cast_fp16")]; - tensor var_1396_to_fp16 = const()[name = tensor("op_1396_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1397_cast_fp16 = mul(x = linear_56_cast_fp16, y = var_1396_to_fp16)[name = tensor("op_1397_cast_fp16")]; - tensor input_343_cast_fp16 = add(x = input_331_cast_fp16, y = var_1397_cast_fp16)[name = tensor("input_343_cast_fp16")]; - tensor query_13_axes_0 = const()[name = tensor("query_13_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82012480)))]; - tensor module_layers_6_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82013568)))]; - tensor query_13_cast_fp16 = layer_norm(axes = query_13_axes_0, beta = module_layers_6_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_self_att_weight_to_fp16, x = input_343_cast_fp16)[name = tensor("query_13_cast_fp16")]; - tensor module_layers_6_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82014656)))]; - tensor module_layers_6_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82539008)))]; - tensor linear_57_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_q_bias_to_fp16, weight = module_layers_6_self_attn_linear_q_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_57_cast_fp16")]; - tensor var_1414 = const()[name = tensor("op_1414"), val = tensor([1, -1, 8, 64])]; - tensor q_37_cast_fp16 = reshape(shape = var_1414, x = linear_57_cast_fp16)[name = tensor("q_37_cast_fp16")]; - tensor module_layers_6_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(82540096)))]; - tensor module_layers_6_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83064448)))]; - tensor linear_58_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_k_bias_to_fp16, weight = module_layers_6_self_attn_linear_k_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_58_cast_fp16")]; - tensor var_1419 = const()[name = tensor("op_1419"), val = tensor([1, -1, 8, 64])]; - tensor k_25_cast_fp16 = reshape(shape = var_1419, x = linear_58_cast_fp16)[name = tensor("k_25_cast_fp16")]; - tensor module_layers_6_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83065536)))]; - tensor module_layers_6_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83589888)))]; - tensor linear_59_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_v_bias_to_fp16, weight = module_layers_6_self_attn_linear_v_weight_to_fp16, x = query_13_cast_fp16)[name = tensor("linear_59_cast_fp16")]; - tensor var_1424 = const()[name = tensor("op_1424"), val = tensor([1, -1, 8, 64])]; - tensor v_13_cast_fp16 = reshape(shape = var_1424, x = linear_59_cast_fp16)[name = tensor("v_13_cast_fp16")]; - tensor value_15_perm_0 = const()[name = tensor("value_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_6_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83590976)))]; - tensor var_1436_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1436_cast_fp16")]; - tensor module_layers_6_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_6_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83592064)))]; - tensor var_1438_cast_fp16 = add(x = q_37_cast_fp16, y = module_layers_6_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1438_cast_fp16")]; - tensor q_with_bias_v_13_perm_0 = const()[name = tensor("q_with_bias_v_13_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_139_transpose_x_0 = const()[name = tensor("x_139_transpose_x_0"), val = tensor(false)]; - tensor x_139_transpose_y_0 = const()[name = tensor("x_139_transpose_y_0"), val = tensor(false)]; - tensor var_1440_to_fp16 = const()[name = tensor("op_1440_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83593152)))]; - tensor q_with_bias_v_13_cast_fp16 = transpose(perm = q_with_bias_v_13_perm_0, x = var_1438_cast_fp16)[name = tensor("transpose_161")]; - tensor x_139_cast_fp16 = matmul(transpose_x = x_139_transpose_x_0, transpose_y = x_139_transpose_y_0, x = q_with_bias_v_13_cast_fp16, y = var_1440_to_fp16)[name = tensor("x_139_cast_fp16")]; - tensor x_141_pad_0 = const()[name = tensor("x_141_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_141_mode_0 = const()[name = tensor("x_141_mode_0"), val = tensor("constant")]; - tensor const_130_to_fp16 = const()[name = tensor("const_130_to_fp16"), val = tensor(0x0p+0)]; - tensor x_141_cast_fp16 = pad(constant_val = const_130_to_fp16, mode = x_141_mode_0, pad = x_141_pad_0, x = x_139_cast_fp16)[name = tensor("x_141_cast_fp16")]; - tensor var_1448 = const()[name = tensor("op_1448"), val = tensor([1, 8, -1, 188])]; - tensor x_143_cast_fp16 = reshape(shape = var_1448, x = x_141_cast_fp16)[name = tensor("x_143_cast_fp16")]; - tensor var_1452_begin_0 = const()[name = tensor("op_1452_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1452_end_0 = const()[name = tensor("op_1452_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1452_end_mask_0 = const()[name = tensor("op_1452_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1452_cast_fp16 = slice_by_index(begin = var_1452_begin_0, end = var_1452_end_0, end_mask = var_1452_end_mask_0, x = x_143_cast_fp16)[name = tensor("op_1452_cast_fp16")]; - tensor var_1453 = const()[name = tensor("op_1453"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_25_cast_fp16 = reshape(shape = var_1453, x = var_1452_cast_fp16)[name = tensor("matrix_bd_25_cast_fp16")]; - tensor matrix_ac_13_transpose_x_0 = const()[name = tensor("matrix_ac_13_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_13_transpose_y_0 = const()[name = tensor("matrix_ac_13_transpose_y_0"), val = tensor(false)]; - tensor transpose_63_perm_0 = const()[name = tensor("transpose_63_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_64_perm_0 = const()[name = tensor("transpose_64_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_64 = transpose(perm = transpose_64_perm_0, x = k_25_cast_fp16)[name = tensor("transpose_159")]; - tensor transpose_63 = transpose(perm = transpose_63_perm_0, x = var_1436_cast_fp16)[name = tensor("transpose_160")]; - tensor matrix_ac_13_cast_fp16 = matmul(transpose_x = matrix_ac_13_transpose_x_0, transpose_y = matrix_ac_13_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor("matrix_ac_13_cast_fp16")]; - tensor matrix_bd_27_begin_0 = const()[name = tensor("matrix_bd_27_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_27_end_0 = const()[name = tensor("matrix_bd_27_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_27_end_mask_0 = const()[name = tensor("matrix_bd_27_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_27_cast_fp16 = slice_by_index(begin = matrix_bd_27_begin_0, end = matrix_bd_27_end_0, end_mask = matrix_bd_27_end_mask_0, x = matrix_bd_25_cast_fp16)[name = tensor("matrix_bd_27_cast_fp16")]; - tensor var_1462_cast_fp16 = add(x = matrix_ac_13_cast_fp16, y = matrix_bd_27_cast_fp16)[name = tensor("op_1462_cast_fp16")]; - tensor _inversed_scores_25_y_0_to_fp16 = const()[name = tensor("_inversed_scores_25_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_25_cast_fp16 = mul(x = var_1462_cast_fp16, y = _inversed_scores_25_y_0_to_fp16)[name = tensor("_inversed_scores_25_cast_fp16")]; - tensor scores_27_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_25_cast_fp16, cond = mask_11)[name = tensor("scores_27_cast_fp16")]; - tensor var_1468_cast_fp16 = softmax(axis = var_23, x = scores_27_cast_fp16)[name = tensor("op_1468_cast_fp16")]; - tensor input_345_cast_fp16 = select(a = var_11_to_fp16, b = var_1468_cast_fp16, cond = mask_11)[name = tensor("input_345_cast_fp16")]; - tensor x_145_transpose_x_0 = const()[name = tensor("x_145_transpose_x_0"), val = tensor(false)]; - tensor x_145_transpose_y_0 = const()[name = tensor("x_145_transpose_y_0"), val = tensor(false)]; - tensor value_15_cast_fp16 = transpose(perm = value_15_perm_0, x = v_13_cast_fp16)[name = tensor("transpose_162")]; - tensor x_145_cast_fp16 = matmul(transpose_x = x_145_transpose_x_0, transpose_y = x_145_transpose_y_0, x = input_345_cast_fp16, y = value_15_cast_fp16)[name = tensor("x_145_cast_fp16")]; - tensor var_1472_perm_0 = const()[name = tensor("op_1472_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1473 = const()[name = tensor("op_1473"), val = tensor([1, -1, 512])]; - tensor var_1472_cast_fp16 = transpose(perm = var_1472_perm_0, x = x_145_cast_fp16)[name = tensor("transpose_158")]; - tensor input_347_cast_fp16 = reshape(shape = var_1473, x = var_1472_cast_fp16)[name = tensor("input_347_cast_fp16")]; - tensor module_layers_6_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(83977216)))]; - tensor module_layers_6_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_6_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84501568)))]; - tensor linear_61_cast_fp16 = linear(bias = module_layers_6_self_attn_linear_out_bias_to_fp16, weight = module_layers_6_self_attn_linear_out_weight_to_fp16, x = input_347_cast_fp16)[name = tensor("linear_61_cast_fp16")]; - tensor input_351_cast_fp16 = add(x = input_343_cast_fp16, y = linear_61_cast_fp16)[name = tensor("input_351_cast_fp16")]; - tensor x_149_axes_0 = const()[name = tensor("x_149_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84502656)))]; - tensor module_layers_6_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84503744)))]; - tensor x_149_cast_fp16 = layer_norm(axes = x_149_axes_0, beta = module_layers_6_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_conv_weight_to_fp16, x = input_351_cast_fp16)[name = tensor("x_149_cast_fp16")]; - tensor input_353_perm_0 = const()[name = tensor("input_353_perm_0"), val = tensor([0, 2, 1])]; - tensor input_355_pad_type_0 = const()[name = tensor("input_355_pad_type_0"), val = tensor("valid")]; - tensor input_355_strides_0 = const()[name = tensor("input_355_strides_0"), val = tensor([1])]; - tensor input_355_pad_0 = const()[name = tensor("input_355_pad_0"), val = tensor([0, 0])]; - tensor input_355_dilations_0 = const()[name = tensor("input_355_dilations_0"), val = tensor([1])]; - tensor input_355_groups_0 = const()[name = tensor("input_355_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(84504832)))]; - tensor module_layers_6_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85553472)))]; - tensor input_353_cast_fp16 = transpose(perm = input_353_perm_0, x = x_149_cast_fp16)[name = tensor("transpose_157")]; - tensor input_355_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv1_bias_to_fp16, dilations = input_355_dilations_0, groups = input_355_groups_0, pad = input_355_pad_0, pad_type = input_355_pad_type_0, strides = input_355_strides_0, weight = module_layers_6_conv_pointwise_conv1_weight_to_fp16, x = input_353_cast_fp16)[name = tensor("input_355_cast_fp16")]; - tensor x_151_split_num_splits_0 = const()[name = tensor("x_151_split_num_splits_0"), val = tensor(2)]; - tensor x_151_split_axis_0 = const()[name = tensor("x_151_split_axis_0"), val = tensor(1)]; - tensor x_151_split_cast_fp16_0, tensor x_151_split_cast_fp16_1 = split(axis = x_151_split_axis_0, num_splits = x_151_split_num_splits_0, x = input_355_cast_fp16)[name = tensor("x_151_split_cast_fp16")]; - tensor x_151_split_1_sigmoid_cast_fp16 = sigmoid(x = x_151_split_cast_fp16_1)[name = tensor("x_151_split_1_sigmoid_cast_fp16")]; - tensor x_151_cast_fp16 = mul(x = x_151_split_cast_fp16_0, y = x_151_split_1_sigmoid_cast_fp16)[name = tensor("x_151_cast_fp16")]; - tensor input_357_cast_fp16 = select(a = var_11_to_fp16, b = x_151_cast_fp16, cond = var_453)[name = tensor("input_357_cast_fp16")]; - tensor input_359_pad_0 = const()[name = tensor("input_359_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_359_mode_0 = const()[name = tensor("input_359_mode_0"), val = tensor("constant")]; - tensor const_133_to_fp16 = const()[name = tensor("const_133_to_fp16"), val = tensor(0x0p+0)]; - tensor input_359_cast_fp16 = pad(constant_val = const_133_to_fp16, mode = input_359_mode_0, pad = input_359_pad_0, x = input_357_cast_fp16)[name = tensor("input_359_cast_fp16")]; - tensor input_361_pad_type_0 = const()[name = tensor("input_361_pad_type_0"), val = tensor("valid")]; - tensor input_361_groups_0 = const()[name = tensor("input_361_groups_0"), val = tensor(512)]; - tensor input_361_strides_0 = const()[name = tensor("input_361_strides_0"), val = tensor([1])]; - tensor input_361_pad_0 = const()[name = tensor("input_361_pad_0"), val = tensor([0, 0])]; - tensor input_361_dilations_0 = const()[name = tensor("input_361_dilations_0"), val = tensor([1])]; - tensor const_246_to_fp16 = const()[name = tensor("const_246_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85555584)))]; - tensor const_247_to_fp16 = const()[name = tensor("const_247_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85564864)))]; - tensor input_363_cast_fp16 = conv(bias = const_247_to_fp16, dilations = input_361_dilations_0, groups = input_361_groups_0, pad = input_361_pad_0, pad_type = input_361_pad_type_0, strides = input_361_strides_0, weight = const_246_to_fp16, x = input_359_cast_fp16)[name = tensor("input_363_cast_fp16")]; - tensor input_365_cast_fp16 = silu(x = input_363_cast_fp16)[name = tensor("input_365_cast_fp16")]; - tensor x_153_pad_type_0 = const()[name = tensor("x_153_pad_type_0"), val = tensor("valid")]; - tensor x_153_strides_0 = const()[name = tensor("x_153_strides_0"), val = tensor([1])]; - tensor x_153_pad_0 = const()[name = tensor("x_153_pad_0"), val = tensor([0, 0])]; - tensor x_153_dilations_0 = const()[name = tensor("x_153_dilations_0"), val = tensor([1])]; - tensor x_153_groups_0 = const()[name = tensor("x_153_groups_0"), val = tensor(1)]; - tensor module_layers_6_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(85565952)))]; - tensor module_layers_6_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_6_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86090304)))]; - tensor x_153_cast_fp16 = conv(bias = module_layers_6_conv_pointwise_conv2_bias_to_fp16, dilations = x_153_dilations_0, groups = x_153_groups_0, pad = x_153_pad_0, pad_type = x_153_pad_type_0, strides = x_153_strides_0, weight = module_layers_6_conv_pointwise_conv2_weight_to_fp16, x = input_365_cast_fp16)[name = tensor("x_153_cast_fp16")]; - tensor input_367_perm_0 = const()[name = tensor("input_367_perm_0"), val = tensor([0, 2, 1])]; - tensor input_367_cast_fp16 = transpose(perm = input_367_perm_0, x = x_153_cast_fp16)[name = tensor("transpose_156")]; - tensor input_369_cast_fp16 = add(x = input_351_cast_fp16, y = input_367_cast_fp16)[name = tensor("input_369_cast_fp16")]; - tensor input_371_axes_0 = const()[name = tensor("input_371_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86091392)))]; - tensor module_layers_6_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86092480)))]; - tensor input_371_cast_fp16 = layer_norm(axes = input_371_axes_0, beta = module_layers_6_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_feed_forward2_weight_to_fp16, x = input_369_cast_fp16)[name = tensor("input_371_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(86093568)))]; - tensor module_layers_6_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88190784)))]; - tensor linear_62_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear1_bias_to_fp16, weight = module_layers_6_feed_forward2_linear1_weight_to_fp16, x = input_371_cast_fp16)[name = tensor("linear_62_cast_fp16")]; - tensor input_375_cast_fp16 = silu(x = linear_62_cast_fp16)[name = tensor("input_375_cast_fp16")]; - tensor module_layers_6_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(88194944)))]; - tensor module_layers_6_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_6_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90292160)))]; - tensor linear_63_cast_fp16 = linear(bias = module_layers_6_feed_forward2_linear2_bias_to_fp16, weight = module_layers_6_feed_forward2_linear2_weight_to_fp16, x = input_375_cast_fp16)[name = tensor("linear_63_cast_fp16")]; - tensor var_1539_to_fp16 = const()[name = tensor("op_1539_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1540_cast_fp16 = mul(x = linear_63_cast_fp16, y = var_1539_to_fp16)[name = tensor("op_1540_cast_fp16")]; - tensor input_381_cast_fp16 = add(x = input_369_cast_fp16, y = var_1540_cast_fp16)[name = tensor("input_381_cast_fp16")]; - tensor input_383_axes_0 = const()[name = tensor("input_383_axes_0"), val = tensor([-1])]; - tensor module_layers_6_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_6_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90293248)))]; - tensor module_layers_6_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_6_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90294336)))]; - tensor input_383_cast_fp16 = layer_norm(axes = input_383_axes_0, beta = module_layers_6_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_6_norm_out_weight_to_fp16, x = input_381_cast_fp16)[name = tensor("input_383_cast_fp16")]; - tensor input_385_axes_0 = const()[name = tensor("input_385_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90295424)))]; - tensor module_layers_7_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90296512)))]; - tensor input_385_cast_fp16 = layer_norm(axes = input_385_axes_0, beta = module_layers_7_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward1_weight_to_fp16, x = input_383_cast_fp16)[name = tensor("input_385_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(90297600)))]; - tensor module_layers_7_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92394816)))]; - tensor linear_64_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear1_bias_to_fp16, weight = module_layers_7_feed_forward1_linear1_weight_to_fp16, x = input_385_cast_fp16)[name = tensor("linear_64_cast_fp16")]; - tensor input_389_cast_fp16 = silu(x = linear_64_cast_fp16)[name = tensor("input_389_cast_fp16")]; - tensor module_layers_7_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(92398976)))]; - tensor module_layers_7_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94496192)))]; - tensor linear_65_cast_fp16 = linear(bias = module_layers_7_feed_forward1_linear2_bias_to_fp16, weight = module_layers_7_feed_forward1_linear2_weight_to_fp16, x = input_389_cast_fp16)[name = tensor("linear_65_cast_fp16")]; - tensor var_1570_to_fp16 = const()[name = tensor("op_1570_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1571_cast_fp16 = mul(x = linear_65_cast_fp16, y = var_1570_to_fp16)[name = tensor("op_1571_cast_fp16")]; - tensor input_395_cast_fp16 = add(x = input_383_cast_fp16, y = var_1571_cast_fp16)[name = tensor("input_395_cast_fp16")]; - tensor query_15_axes_0 = const()[name = tensor("query_15_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94497280)))]; - tensor module_layers_7_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94498368)))]; - tensor query_15_cast_fp16 = layer_norm(axes = query_15_axes_0, beta = module_layers_7_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_self_att_weight_to_fp16, x = input_395_cast_fp16)[name = tensor("query_15_cast_fp16")]; - tensor module_layers_7_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(94499456)))]; - tensor module_layers_7_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95023808)))]; - tensor linear_66_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_q_bias_to_fp16, weight = module_layers_7_self_attn_linear_q_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_66_cast_fp16")]; - tensor var_1588 = const()[name = tensor("op_1588"), val = tensor([1, -1, 8, 64])]; - tensor q_43_cast_fp16 = reshape(shape = var_1588, x = linear_66_cast_fp16)[name = tensor("q_43_cast_fp16")]; - tensor module_layers_7_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95024896)))]; - tensor module_layers_7_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95549248)))]; - tensor linear_67_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_k_bias_to_fp16, weight = module_layers_7_self_attn_linear_k_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_67_cast_fp16")]; - tensor var_1593 = const()[name = tensor("op_1593"), val = tensor([1, -1, 8, 64])]; - tensor k_29_cast_fp16 = reshape(shape = var_1593, x = linear_67_cast_fp16)[name = tensor("k_29_cast_fp16")]; - tensor module_layers_7_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(95550336)))]; - tensor module_layers_7_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96074688)))]; - tensor linear_68_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_v_bias_to_fp16, weight = module_layers_7_self_attn_linear_v_weight_to_fp16, x = query_15_cast_fp16)[name = tensor("linear_68_cast_fp16")]; - tensor var_1598 = const()[name = tensor("op_1598"), val = tensor([1, -1, 8, 64])]; - tensor v_15_cast_fp16 = reshape(shape = var_1598, x = linear_68_cast_fp16)[name = tensor("v_15_cast_fp16")]; - tensor value_17_perm_0 = const()[name = tensor("value_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_7_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96075776)))]; - tensor var_1610_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1610_cast_fp16")]; - tensor module_layers_7_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_7_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96076864)))]; - tensor var_1612_cast_fp16 = add(x = q_43_cast_fp16, y = module_layers_7_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1612_cast_fp16")]; - tensor q_with_bias_v_15_perm_0 = const()[name = tensor("q_with_bias_v_15_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_161_transpose_x_0 = const()[name = tensor("x_161_transpose_x_0"), val = tensor(false)]; - tensor x_161_transpose_y_0 = const()[name = tensor("x_161_transpose_y_0"), val = tensor(false)]; - tensor var_1614_to_fp16 = const()[name = tensor("op_1614_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96077952)))]; - tensor q_with_bias_v_15_cast_fp16 = transpose(perm = q_with_bias_v_15_perm_0, x = var_1612_cast_fp16)[name = tensor("transpose_154")]; - tensor x_161_cast_fp16 = matmul(transpose_x = x_161_transpose_x_0, transpose_y = x_161_transpose_y_0, x = q_with_bias_v_15_cast_fp16, y = var_1614_to_fp16)[name = tensor("x_161_cast_fp16")]; - tensor x_163_pad_0 = const()[name = tensor("x_163_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_163_mode_0 = const()[name = tensor("x_163_mode_0"), val = tensor("constant")]; - tensor const_140_to_fp16 = const()[name = tensor("const_140_to_fp16"), val = tensor(0x0p+0)]; - tensor x_163_cast_fp16 = pad(constant_val = const_140_to_fp16, mode = x_163_mode_0, pad = x_163_pad_0, x = x_161_cast_fp16)[name = tensor("x_163_cast_fp16")]; - tensor var_1622 = const()[name = tensor("op_1622"), val = tensor([1, 8, -1, 188])]; - tensor x_165_cast_fp16 = reshape(shape = var_1622, x = x_163_cast_fp16)[name = tensor("x_165_cast_fp16")]; - tensor var_1626_begin_0 = const()[name = tensor("op_1626_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1626_end_0 = const()[name = tensor("op_1626_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1626_end_mask_0 = const()[name = tensor("op_1626_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1626_cast_fp16 = slice_by_index(begin = var_1626_begin_0, end = var_1626_end_0, end_mask = var_1626_end_mask_0, x = x_165_cast_fp16)[name = tensor("op_1626_cast_fp16")]; - tensor var_1627 = const()[name = tensor("op_1627"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_29_cast_fp16 = reshape(shape = var_1627, x = var_1626_cast_fp16)[name = tensor("matrix_bd_29_cast_fp16")]; - tensor matrix_ac_15_transpose_x_0 = const()[name = tensor("matrix_ac_15_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_15_transpose_y_0 = const()[name = tensor("matrix_ac_15_transpose_y_0"), val = tensor(false)]; - tensor transpose_65_perm_0 = const()[name = tensor("transpose_65_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_66_perm_0 = const()[name = tensor("transpose_66_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_66 = transpose(perm = transpose_66_perm_0, x = k_29_cast_fp16)[name = tensor("transpose_152")]; - tensor transpose_65 = transpose(perm = transpose_65_perm_0, x = var_1610_cast_fp16)[name = tensor("transpose_153")]; - tensor matrix_ac_15_cast_fp16 = matmul(transpose_x = matrix_ac_15_transpose_x_0, transpose_y = matrix_ac_15_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor("matrix_ac_15_cast_fp16")]; - tensor matrix_bd_31_begin_0 = const()[name = tensor("matrix_bd_31_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_31_end_0 = const()[name = tensor("matrix_bd_31_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_31_end_mask_0 = const()[name = tensor("matrix_bd_31_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_31_cast_fp16 = slice_by_index(begin = matrix_bd_31_begin_0, end = matrix_bd_31_end_0, end_mask = matrix_bd_31_end_mask_0, x = matrix_bd_29_cast_fp16)[name = tensor("matrix_bd_31_cast_fp16")]; - tensor var_1636_cast_fp16 = add(x = matrix_ac_15_cast_fp16, y = matrix_bd_31_cast_fp16)[name = tensor("op_1636_cast_fp16")]; - tensor _inversed_scores_29_y_0_to_fp16 = const()[name = tensor("_inversed_scores_29_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_29_cast_fp16 = mul(x = var_1636_cast_fp16, y = _inversed_scores_29_y_0_to_fp16)[name = tensor("_inversed_scores_29_cast_fp16")]; - tensor scores_31_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_29_cast_fp16, cond = mask_11)[name = tensor("scores_31_cast_fp16")]; - tensor var_1642_cast_fp16 = softmax(axis = var_23, x = scores_31_cast_fp16)[name = tensor("op_1642_cast_fp16")]; - tensor input_397_cast_fp16 = select(a = var_11_to_fp16, b = var_1642_cast_fp16, cond = mask_11)[name = tensor("input_397_cast_fp16")]; - tensor x_167_transpose_x_0 = const()[name = tensor("x_167_transpose_x_0"), val = tensor(false)]; - tensor x_167_transpose_y_0 = const()[name = tensor("x_167_transpose_y_0"), val = tensor(false)]; - tensor value_17_cast_fp16 = transpose(perm = value_17_perm_0, x = v_15_cast_fp16)[name = tensor("transpose_155")]; - tensor x_167_cast_fp16 = matmul(transpose_x = x_167_transpose_x_0, transpose_y = x_167_transpose_y_0, x = input_397_cast_fp16, y = value_17_cast_fp16)[name = tensor("x_167_cast_fp16")]; - tensor var_1646_perm_0 = const()[name = tensor("op_1646_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1647 = const()[name = tensor("op_1647"), val = tensor([1, -1, 512])]; - tensor var_1646_cast_fp16 = transpose(perm = var_1646_perm_0, x = x_167_cast_fp16)[name = tensor("transpose_151")]; - tensor input_399_cast_fp16 = reshape(shape = var_1647, x = var_1646_cast_fp16)[name = tensor("input_399_cast_fp16")]; - tensor module_layers_7_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96462016)))]; - tensor module_layers_7_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_7_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96986368)))]; - tensor linear_70_cast_fp16 = linear(bias = module_layers_7_self_attn_linear_out_bias_to_fp16, weight = module_layers_7_self_attn_linear_out_weight_to_fp16, x = input_399_cast_fp16)[name = tensor("linear_70_cast_fp16")]; - tensor input_403_cast_fp16 = add(x = input_395_cast_fp16, y = linear_70_cast_fp16)[name = tensor("input_403_cast_fp16")]; - tensor x_171_axes_0 = const()[name = tensor("x_171_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96987456)))]; - tensor module_layers_7_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96988544)))]; - tensor x_171_cast_fp16 = layer_norm(axes = x_171_axes_0, beta = module_layers_7_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_conv_weight_to_fp16, x = input_403_cast_fp16)[name = tensor("x_171_cast_fp16")]; - tensor input_405_perm_0 = const()[name = tensor("input_405_perm_0"), val = tensor([0, 2, 1])]; - tensor input_407_pad_type_0 = const()[name = tensor("input_407_pad_type_0"), val = tensor("valid")]; - tensor input_407_strides_0 = const()[name = tensor("input_407_strides_0"), val = tensor([1])]; - tensor input_407_pad_0 = const()[name = tensor("input_407_pad_0"), val = tensor([0, 0])]; - tensor input_407_dilations_0 = const()[name = tensor("input_407_dilations_0"), val = tensor([1])]; - tensor input_407_groups_0 = const()[name = tensor("input_407_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(96989632)))]; - tensor module_layers_7_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98038272)))]; - tensor input_405_cast_fp16 = transpose(perm = input_405_perm_0, x = x_171_cast_fp16)[name = tensor("transpose_150")]; - tensor input_407_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv1_bias_to_fp16, dilations = input_407_dilations_0, groups = input_407_groups_0, pad = input_407_pad_0, pad_type = input_407_pad_type_0, strides = input_407_strides_0, weight = module_layers_7_conv_pointwise_conv1_weight_to_fp16, x = input_405_cast_fp16)[name = tensor("input_407_cast_fp16")]; - tensor x_173_split_num_splits_0 = const()[name = tensor("x_173_split_num_splits_0"), val = tensor(2)]; - tensor x_173_split_axis_0 = const()[name = tensor("x_173_split_axis_0"), val = tensor(1)]; - tensor x_173_split_cast_fp16_0, tensor x_173_split_cast_fp16_1 = split(axis = x_173_split_axis_0, num_splits = x_173_split_num_splits_0, x = input_407_cast_fp16)[name = tensor("x_173_split_cast_fp16")]; - tensor x_173_split_1_sigmoid_cast_fp16 = sigmoid(x = x_173_split_cast_fp16_1)[name = tensor("x_173_split_1_sigmoid_cast_fp16")]; - tensor x_173_cast_fp16 = mul(x = x_173_split_cast_fp16_0, y = x_173_split_1_sigmoid_cast_fp16)[name = tensor("x_173_cast_fp16")]; - tensor input_409_cast_fp16 = select(a = var_11_to_fp16, b = x_173_cast_fp16, cond = var_453)[name = tensor("input_409_cast_fp16")]; - tensor input_411_pad_0 = const()[name = tensor("input_411_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_411_mode_0 = const()[name = tensor("input_411_mode_0"), val = tensor("constant")]; - tensor const_143_to_fp16 = const()[name = tensor("const_143_to_fp16"), val = tensor(0x0p+0)]; - tensor input_411_cast_fp16 = pad(constant_val = const_143_to_fp16, mode = input_411_mode_0, pad = input_411_pad_0, x = input_409_cast_fp16)[name = tensor("input_411_cast_fp16")]; - tensor input_413_pad_type_0 = const()[name = tensor("input_413_pad_type_0"), val = tensor("valid")]; - tensor input_413_groups_0 = const()[name = tensor("input_413_groups_0"), val = tensor(512)]; - tensor input_413_strides_0 = const()[name = tensor("input_413_strides_0"), val = tensor([1])]; - tensor input_413_pad_0 = const()[name = tensor("input_413_pad_0"), val = tensor([0, 0])]; - tensor input_413_dilations_0 = const()[name = tensor("input_413_dilations_0"), val = tensor([1])]; - tensor const_248_to_fp16 = const()[name = tensor("const_248_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98040384)))]; - tensor const_249_to_fp16 = const()[name = tensor("const_249_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98049664)))]; - tensor input_415_cast_fp16 = conv(bias = const_249_to_fp16, dilations = input_413_dilations_0, groups = input_413_groups_0, pad = input_413_pad_0, pad_type = input_413_pad_type_0, strides = input_413_strides_0, weight = const_248_to_fp16, x = input_411_cast_fp16)[name = tensor("input_415_cast_fp16")]; - tensor input_417_cast_fp16 = silu(x = input_415_cast_fp16)[name = tensor("input_417_cast_fp16")]; - tensor x_175_pad_type_0 = const()[name = tensor("x_175_pad_type_0"), val = tensor("valid")]; - tensor x_175_strides_0 = const()[name = tensor("x_175_strides_0"), val = tensor([1])]; - tensor x_175_pad_0 = const()[name = tensor("x_175_pad_0"), val = tensor([0, 0])]; - tensor x_175_dilations_0 = const()[name = tensor("x_175_dilations_0"), val = tensor([1])]; - tensor x_175_groups_0 = const()[name = tensor("x_175_groups_0"), val = tensor(1)]; - tensor module_layers_7_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98050752)))]; - tensor module_layers_7_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_7_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98575104)))]; - tensor x_175_cast_fp16 = conv(bias = module_layers_7_conv_pointwise_conv2_bias_to_fp16, dilations = x_175_dilations_0, groups = x_175_groups_0, pad = x_175_pad_0, pad_type = x_175_pad_type_0, strides = x_175_strides_0, weight = module_layers_7_conv_pointwise_conv2_weight_to_fp16, x = input_417_cast_fp16)[name = tensor("x_175_cast_fp16")]; - tensor input_419_perm_0 = const()[name = tensor("input_419_perm_0"), val = tensor([0, 2, 1])]; - tensor input_419_cast_fp16 = transpose(perm = input_419_perm_0, x = x_175_cast_fp16)[name = tensor("transpose_149")]; - tensor input_421_cast_fp16 = add(x = input_403_cast_fp16, y = input_419_cast_fp16)[name = tensor("input_421_cast_fp16")]; - tensor input_423_axes_0 = const()[name = tensor("input_423_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98576192)))]; - tensor module_layers_7_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98577280)))]; - tensor input_423_cast_fp16 = layer_norm(axes = input_423_axes_0, beta = module_layers_7_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_feed_forward2_weight_to_fp16, x = input_421_cast_fp16)[name = tensor("input_423_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(98578368)))]; - tensor module_layers_7_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100675584)))]; - tensor linear_71_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear1_bias_to_fp16, weight = module_layers_7_feed_forward2_linear1_weight_to_fp16, x = input_423_cast_fp16)[name = tensor("linear_71_cast_fp16")]; - tensor input_427_cast_fp16 = silu(x = linear_71_cast_fp16)[name = tensor("input_427_cast_fp16")]; - tensor module_layers_7_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(100679744)))]; - tensor module_layers_7_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_7_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102776960)))]; - tensor linear_72_cast_fp16 = linear(bias = module_layers_7_feed_forward2_linear2_bias_to_fp16, weight = module_layers_7_feed_forward2_linear2_weight_to_fp16, x = input_427_cast_fp16)[name = tensor("linear_72_cast_fp16")]; - tensor var_1713_to_fp16 = const()[name = tensor("op_1713_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1714_cast_fp16 = mul(x = linear_72_cast_fp16, y = var_1713_to_fp16)[name = tensor("op_1714_cast_fp16")]; - tensor input_433_cast_fp16 = add(x = input_421_cast_fp16, y = var_1714_cast_fp16)[name = tensor("input_433_cast_fp16")]; - tensor input_435_axes_0 = const()[name = tensor("input_435_axes_0"), val = tensor([-1])]; - tensor module_layers_7_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_7_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102778048)))]; - tensor module_layers_7_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_7_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102779136)))]; - tensor input_435_cast_fp16 = layer_norm(axes = input_435_axes_0, beta = module_layers_7_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_7_norm_out_weight_to_fp16, x = input_433_cast_fp16)[name = tensor("input_435_cast_fp16")]; - tensor input_437_axes_0 = const()[name = tensor("input_437_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102780224)))]; - tensor module_layers_8_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102781312)))]; - tensor input_437_cast_fp16 = layer_norm(axes = input_437_axes_0, beta = module_layers_8_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward1_weight_to_fp16, x = input_435_cast_fp16)[name = tensor("input_437_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(102782400)))]; - tensor module_layers_8_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104879616)))]; - tensor linear_73_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear1_bias_to_fp16, weight = module_layers_8_feed_forward1_linear1_weight_to_fp16, x = input_437_cast_fp16)[name = tensor("linear_73_cast_fp16")]; - tensor input_441_cast_fp16 = silu(x = linear_73_cast_fp16)[name = tensor("input_441_cast_fp16")]; - tensor module_layers_8_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(104883776)))]; - tensor module_layers_8_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106980992)))]; - tensor linear_74_cast_fp16 = linear(bias = module_layers_8_feed_forward1_linear2_bias_to_fp16, weight = module_layers_8_feed_forward1_linear2_weight_to_fp16, x = input_441_cast_fp16)[name = tensor("linear_74_cast_fp16")]; - tensor var_1744_to_fp16 = const()[name = tensor("op_1744_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1745_cast_fp16 = mul(x = linear_74_cast_fp16, y = var_1744_to_fp16)[name = tensor("op_1745_cast_fp16")]; - tensor input_447_cast_fp16 = add(x = input_435_cast_fp16, y = var_1745_cast_fp16)[name = tensor("input_447_cast_fp16")]; - tensor query_17_axes_0 = const()[name = tensor("query_17_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106982080)))]; - tensor module_layers_8_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106983168)))]; - tensor query_17_cast_fp16 = layer_norm(axes = query_17_axes_0, beta = module_layers_8_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_self_att_weight_to_fp16, x = input_447_cast_fp16)[name = tensor("query_17_cast_fp16")]; - tensor module_layers_8_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(106984256)))]; - tensor module_layers_8_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107508608)))]; - tensor linear_75_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_q_bias_to_fp16, weight = module_layers_8_self_attn_linear_q_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_75_cast_fp16")]; - tensor var_1762 = const()[name = tensor("op_1762"), val = tensor([1, -1, 8, 64])]; - tensor q_49_cast_fp16 = reshape(shape = var_1762, x = linear_75_cast_fp16)[name = tensor("q_49_cast_fp16")]; - tensor module_layers_8_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(107509696)))]; - tensor module_layers_8_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108034048)))]; - tensor linear_76_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_k_bias_to_fp16, weight = module_layers_8_self_attn_linear_k_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_76_cast_fp16")]; - tensor var_1767 = const()[name = tensor("op_1767"), val = tensor([1, -1, 8, 64])]; - tensor k_33_cast_fp16 = reshape(shape = var_1767, x = linear_76_cast_fp16)[name = tensor("k_33_cast_fp16")]; - tensor module_layers_8_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108035136)))]; - tensor module_layers_8_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108559488)))]; - tensor linear_77_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_v_bias_to_fp16, weight = module_layers_8_self_attn_linear_v_weight_to_fp16, x = query_17_cast_fp16)[name = tensor("linear_77_cast_fp16")]; - tensor var_1772 = const()[name = tensor("op_1772"), val = tensor([1, -1, 8, 64])]; - tensor v_17_cast_fp16 = reshape(shape = var_1772, x = linear_77_cast_fp16)[name = tensor("v_17_cast_fp16")]; - tensor value_19_perm_0 = const()[name = tensor("value_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_8_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108560576)))]; - tensor var_1784_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1784_cast_fp16")]; - tensor module_layers_8_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_8_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108561664)))]; - tensor var_1786_cast_fp16 = add(x = q_49_cast_fp16, y = module_layers_8_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1786_cast_fp16")]; - tensor q_with_bias_v_17_perm_0 = const()[name = tensor("q_with_bias_v_17_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_183_transpose_x_0 = const()[name = tensor("x_183_transpose_x_0"), val = tensor(false)]; - tensor x_183_transpose_y_0 = const()[name = tensor("x_183_transpose_y_0"), val = tensor(false)]; - tensor var_1788_to_fp16 = const()[name = tensor("op_1788_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108562752)))]; - tensor q_with_bias_v_17_cast_fp16 = transpose(perm = q_with_bias_v_17_perm_0, x = var_1786_cast_fp16)[name = tensor("transpose_147")]; - tensor x_183_cast_fp16 = matmul(transpose_x = x_183_transpose_x_0, transpose_y = x_183_transpose_y_0, x = q_with_bias_v_17_cast_fp16, y = var_1788_to_fp16)[name = tensor("x_183_cast_fp16")]; - tensor x_185_pad_0 = const()[name = tensor("x_185_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_185_mode_0 = const()[name = tensor("x_185_mode_0"), val = tensor("constant")]; - tensor const_150_to_fp16 = const()[name = tensor("const_150_to_fp16"), val = tensor(0x0p+0)]; - tensor x_185_cast_fp16 = pad(constant_val = const_150_to_fp16, mode = x_185_mode_0, pad = x_185_pad_0, x = x_183_cast_fp16)[name = tensor("x_185_cast_fp16")]; - tensor var_1796 = const()[name = tensor("op_1796"), val = tensor([1, 8, -1, 188])]; - tensor x_187_cast_fp16 = reshape(shape = var_1796, x = x_185_cast_fp16)[name = tensor("x_187_cast_fp16")]; - tensor var_1800_begin_0 = const()[name = tensor("op_1800_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1800_end_0 = const()[name = tensor("op_1800_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1800_end_mask_0 = const()[name = tensor("op_1800_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1800_cast_fp16 = slice_by_index(begin = var_1800_begin_0, end = var_1800_end_0, end_mask = var_1800_end_mask_0, x = x_187_cast_fp16)[name = tensor("op_1800_cast_fp16")]; - tensor var_1801 = const()[name = tensor("op_1801"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_33_cast_fp16 = reshape(shape = var_1801, x = var_1800_cast_fp16)[name = tensor("matrix_bd_33_cast_fp16")]; - tensor matrix_ac_17_transpose_x_0 = const()[name = tensor("matrix_ac_17_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_17_transpose_y_0 = const()[name = tensor("matrix_ac_17_transpose_y_0"), val = tensor(false)]; - tensor transpose_67_perm_0 = const()[name = tensor("transpose_67_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_68_perm_0 = const()[name = tensor("transpose_68_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_68 = transpose(perm = transpose_68_perm_0, x = k_33_cast_fp16)[name = tensor("transpose_145")]; - tensor transpose_67 = transpose(perm = transpose_67_perm_0, x = var_1784_cast_fp16)[name = tensor("transpose_146")]; - tensor matrix_ac_17_cast_fp16 = matmul(transpose_x = matrix_ac_17_transpose_x_0, transpose_y = matrix_ac_17_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor("matrix_ac_17_cast_fp16")]; - tensor matrix_bd_35_begin_0 = const()[name = tensor("matrix_bd_35_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_35_end_0 = const()[name = tensor("matrix_bd_35_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_35_end_mask_0 = const()[name = tensor("matrix_bd_35_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_35_cast_fp16 = slice_by_index(begin = matrix_bd_35_begin_0, end = matrix_bd_35_end_0, end_mask = matrix_bd_35_end_mask_0, x = matrix_bd_33_cast_fp16)[name = tensor("matrix_bd_35_cast_fp16")]; - tensor var_1810_cast_fp16 = add(x = matrix_ac_17_cast_fp16, y = matrix_bd_35_cast_fp16)[name = tensor("op_1810_cast_fp16")]; - tensor _inversed_scores_33_y_0_to_fp16 = const()[name = tensor("_inversed_scores_33_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_33_cast_fp16 = mul(x = var_1810_cast_fp16, y = _inversed_scores_33_y_0_to_fp16)[name = tensor("_inversed_scores_33_cast_fp16")]; - tensor scores_35_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_33_cast_fp16, cond = mask_11)[name = tensor("scores_35_cast_fp16")]; - tensor var_1816_cast_fp16 = softmax(axis = var_23, x = scores_35_cast_fp16)[name = tensor("op_1816_cast_fp16")]; - tensor input_449_cast_fp16 = select(a = var_11_to_fp16, b = var_1816_cast_fp16, cond = mask_11)[name = tensor("input_449_cast_fp16")]; - tensor x_189_transpose_x_0 = const()[name = tensor("x_189_transpose_x_0"), val = tensor(false)]; - tensor x_189_transpose_y_0 = const()[name = tensor("x_189_transpose_y_0"), val = tensor(false)]; - tensor value_19_cast_fp16 = transpose(perm = value_19_perm_0, x = v_17_cast_fp16)[name = tensor("transpose_148")]; - tensor x_189_cast_fp16 = matmul(transpose_x = x_189_transpose_x_0, transpose_y = x_189_transpose_y_0, x = input_449_cast_fp16, y = value_19_cast_fp16)[name = tensor("x_189_cast_fp16")]; - tensor var_1820_perm_0 = const()[name = tensor("op_1820_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1821 = const()[name = tensor("op_1821"), val = tensor([1, -1, 512])]; - tensor var_1820_cast_fp16 = transpose(perm = var_1820_perm_0, x = x_189_cast_fp16)[name = tensor("transpose_144")]; - tensor input_451_cast_fp16 = reshape(shape = var_1821, x = var_1820_cast_fp16)[name = tensor("input_451_cast_fp16")]; - tensor module_layers_8_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(108946816)))]; - tensor module_layers_8_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_8_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109471168)))]; - tensor linear_79_cast_fp16 = linear(bias = module_layers_8_self_attn_linear_out_bias_to_fp16, weight = module_layers_8_self_attn_linear_out_weight_to_fp16, x = input_451_cast_fp16)[name = tensor("linear_79_cast_fp16")]; - tensor input_455_cast_fp16 = add(x = input_447_cast_fp16, y = linear_79_cast_fp16)[name = tensor("input_455_cast_fp16")]; - tensor x_193_axes_0 = const()[name = tensor("x_193_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109472256)))]; - tensor module_layers_8_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109473344)))]; - tensor x_193_cast_fp16 = layer_norm(axes = x_193_axes_0, beta = module_layers_8_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_conv_weight_to_fp16, x = input_455_cast_fp16)[name = tensor("x_193_cast_fp16")]; - tensor input_457_perm_0 = const()[name = tensor("input_457_perm_0"), val = tensor([0, 2, 1])]; - tensor input_459_pad_type_0 = const()[name = tensor("input_459_pad_type_0"), val = tensor("valid")]; - tensor input_459_strides_0 = const()[name = tensor("input_459_strides_0"), val = tensor([1])]; - tensor input_459_pad_0 = const()[name = tensor("input_459_pad_0"), val = tensor([0, 0])]; - tensor input_459_dilations_0 = const()[name = tensor("input_459_dilations_0"), val = tensor([1])]; - tensor input_459_groups_0 = const()[name = tensor("input_459_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(109474432)))]; - tensor module_layers_8_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110523072)))]; - tensor input_457_cast_fp16 = transpose(perm = input_457_perm_0, x = x_193_cast_fp16)[name = tensor("transpose_143")]; - tensor input_459_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv1_bias_to_fp16, dilations = input_459_dilations_0, groups = input_459_groups_0, pad = input_459_pad_0, pad_type = input_459_pad_type_0, strides = input_459_strides_0, weight = module_layers_8_conv_pointwise_conv1_weight_to_fp16, x = input_457_cast_fp16)[name = tensor("input_459_cast_fp16")]; - tensor x_195_split_num_splits_0 = const()[name = tensor("x_195_split_num_splits_0"), val = tensor(2)]; - tensor x_195_split_axis_0 = const()[name = tensor("x_195_split_axis_0"), val = tensor(1)]; - tensor x_195_split_cast_fp16_0, tensor x_195_split_cast_fp16_1 = split(axis = x_195_split_axis_0, num_splits = x_195_split_num_splits_0, x = input_459_cast_fp16)[name = tensor("x_195_split_cast_fp16")]; - tensor x_195_split_1_sigmoid_cast_fp16 = sigmoid(x = x_195_split_cast_fp16_1)[name = tensor("x_195_split_1_sigmoid_cast_fp16")]; - tensor x_195_cast_fp16 = mul(x = x_195_split_cast_fp16_0, y = x_195_split_1_sigmoid_cast_fp16)[name = tensor("x_195_cast_fp16")]; - tensor input_461_cast_fp16 = select(a = var_11_to_fp16, b = x_195_cast_fp16, cond = var_453)[name = tensor("input_461_cast_fp16")]; - tensor input_463_pad_0 = const()[name = tensor("input_463_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_463_mode_0 = const()[name = tensor("input_463_mode_0"), val = tensor("constant")]; - tensor const_153_to_fp16 = const()[name = tensor("const_153_to_fp16"), val = tensor(0x0p+0)]; - tensor input_463_cast_fp16 = pad(constant_val = const_153_to_fp16, mode = input_463_mode_0, pad = input_463_pad_0, x = input_461_cast_fp16)[name = tensor("input_463_cast_fp16")]; - tensor input_465_pad_type_0 = const()[name = tensor("input_465_pad_type_0"), val = tensor("valid")]; - tensor input_465_groups_0 = const()[name = tensor("input_465_groups_0"), val = tensor(512)]; - tensor input_465_strides_0 = const()[name = tensor("input_465_strides_0"), val = tensor([1])]; - tensor input_465_pad_0 = const()[name = tensor("input_465_pad_0"), val = tensor([0, 0])]; - tensor input_465_dilations_0 = const()[name = tensor("input_465_dilations_0"), val = tensor([1])]; - tensor const_250_to_fp16 = const()[name = tensor("const_250_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110525184)))]; - tensor const_251_to_fp16 = const()[name = tensor("const_251_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110534464)))]; - tensor input_467_cast_fp16 = conv(bias = const_251_to_fp16, dilations = input_465_dilations_0, groups = input_465_groups_0, pad = input_465_pad_0, pad_type = input_465_pad_type_0, strides = input_465_strides_0, weight = const_250_to_fp16, x = input_463_cast_fp16)[name = tensor("input_467_cast_fp16")]; - tensor input_469_cast_fp16 = silu(x = input_467_cast_fp16)[name = tensor("input_469_cast_fp16")]; - tensor x_197_pad_type_0 = const()[name = tensor("x_197_pad_type_0"), val = tensor("valid")]; - tensor x_197_strides_0 = const()[name = tensor("x_197_strides_0"), val = tensor([1])]; - tensor x_197_pad_0 = const()[name = tensor("x_197_pad_0"), val = tensor([0, 0])]; - tensor x_197_dilations_0 = const()[name = tensor("x_197_dilations_0"), val = tensor([1])]; - tensor x_197_groups_0 = const()[name = tensor("x_197_groups_0"), val = tensor(1)]; - tensor module_layers_8_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(110535552)))]; - tensor module_layers_8_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_8_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111059904)))]; - tensor x_197_cast_fp16 = conv(bias = module_layers_8_conv_pointwise_conv2_bias_to_fp16, dilations = x_197_dilations_0, groups = x_197_groups_0, pad = x_197_pad_0, pad_type = x_197_pad_type_0, strides = x_197_strides_0, weight = module_layers_8_conv_pointwise_conv2_weight_to_fp16, x = input_469_cast_fp16)[name = tensor("x_197_cast_fp16")]; - tensor input_471_perm_0 = const()[name = tensor("input_471_perm_0"), val = tensor([0, 2, 1])]; - tensor input_471_cast_fp16 = transpose(perm = input_471_perm_0, x = x_197_cast_fp16)[name = tensor("transpose_142")]; - tensor input_473_cast_fp16 = add(x = input_455_cast_fp16, y = input_471_cast_fp16)[name = tensor("input_473_cast_fp16")]; - tensor input_475_axes_0 = const()[name = tensor("input_475_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111060992)))]; - tensor module_layers_8_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111062080)))]; - tensor input_475_cast_fp16 = layer_norm(axes = input_475_axes_0, beta = module_layers_8_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_feed_forward2_weight_to_fp16, x = input_473_cast_fp16)[name = tensor("input_475_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(111063168)))]; - tensor module_layers_8_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113160384)))]; - tensor linear_80_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear1_bias_to_fp16, weight = module_layers_8_feed_forward2_linear1_weight_to_fp16, x = input_475_cast_fp16)[name = tensor("linear_80_cast_fp16")]; - tensor input_479_cast_fp16 = silu(x = linear_80_cast_fp16)[name = tensor("input_479_cast_fp16")]; - tensor module_layers_8_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(113164544)))]; - tensor module_layers_8_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_8_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115261760)))]; - tensor linear_81_cast_fp16 = linear(bias = module_layers_8_feed_forward2_linear2_bias_to_fp16, weight = module_layers_8_feed_forward2_linear2_weight_to_fp16, x = input_479_cast_fp16)[name = tensor("linear_81_cast_fp16")]; - tensor var_1887_to_fp16 = const()[name = tensor("op_1887_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1888_cast_fp16 = mul(x = linear_81_cast_fp16, y = var_1887_to_fp16)[name = tensor("op_1888_cast_fp16")]; - tensor input_485_cast_fp16 = add(x = input_473_cast_fp16, y = var_1888_cast_fp16)[name = tensor("input_485_cast_fp16")]; - tensor input_487_axes_0 = const()[name = tensor("input_487_axes_0"), val = tensor([-1])]; - tensor module_layers_8_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_8_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115262848)))]; - tensor module_layers_8_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_8_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115263936)))]; - tensor input_487_cast_fp16 = layer_norm(axes = input_487_axes_0, beta = module_layers_8_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_8_norm_out_weight_to_fp16, x = input_485_cast_fp16)[name = tensor("input_487_cast_fp16")]; - tensor input_489_axes_0 = const()[name = tensor("input_489_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115265024)))]; - tensor module_layers_9_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115266112)))]; - tensor input_489_cast_fp16 = layer_norm(axes = input_489_axes_0, beta = module_layers_9_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward1_weight_to_fp16, x = input_487_cast_fp16)[name = tensor("input_489_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(115267200)))]; - tensor module_layers_9_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117364416)))]; - tensor linear_82_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear1_bias_to_fp16, weight = module_layers_9_feed_forward1_linear1_weight_to_fp16, x = input_489_cast_fp16)[name = tensor("linear_82_cast_fp16")]; - tensor input_493_cast_fp16 = silu(x = linear_82_cast_fp16)[name = tensor("input_493_cast_fp16")]; - tensor module_layers_9_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(117368576)))]; - tensor module_layers_9_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119465792)))]; - tensor linear_83_cast_fp16 = linear(bias = module_layers_9_feed_forward1_linear2_bias_to_fp16, weight = module_layers_9_feed_forward1_linear2_weight_to_fp16, x = input_493_cast_fp16)[name = tensor("linear_83_cast_fp16")]; - tensor var_1918_to_fp16 = const()[name = tensor("op_1918_to_fp16"), val = tensor(0x1p-1)]; - tensor var_1919_cast_fp16 = mul(x = linear_83_cast_fp16, y = var_1918_to_fp16)[name = tensor("op_1919_cast_fp16")]; - tensor input_499_cast_fp16 = add(x = input_487_cast_fp16, y = var_1919_cast_fp16)[name = tensor("input_499_cast_fp16")]; - tensor query_19_axes_0 = const()[name = tensor("query_19_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119466880)))]; - tensor module_layers_9_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119467968)))]; - tensor query_19_cast_fp16 = layer_norm(axes = query_19_axes_0, beta = module_layers_9_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_self_att_weight_to_fp16, x = input_499_cast_fp16)[name = tensor("query_19_cast_fp16")]; - tensor module_layers_9_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119469056)))]; - tensor module_layers_9_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119993408)))]; - tensor linear_84_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_q_bias_to_fp16, weight = module_layers_9_self_attn_linear_q_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_84_cast_fp16")]; - tensor var_1936 = const()[name = tensor("op_1936"), val = tensor([1, -1, 8, 64])]; - tensor q_55_cast_fp16 = reshape(shape = var_1936, x = linear_84_cast_fp16)[name = tensor("q_55_cast_fp16")]; - tensor module_layers_9_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(119994496)))]; - tensor module_layers_9_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120518848)))]; - tensor linear_85_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_k_bias_to_fp16, weight = module_layers_9_self_attn_linear_k_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_85_cast_fp16")]; - tensor var_1941 = const()[name = tensor("op_1941"), val = tensor([1, -1, 8, 64])]; - tensor k_37_cast_fp16 = reshape(shape = var_1941, x = linear_85_cast_fp16)[name = tensor("k_37_cast_fp16")]; - tensor module_layers_9_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(120519936)))]; - tensor module_layers_9_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121044288)))]; - tensor linear_86_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_v_bias_to_fp16, weight = module_layers_9_self_attn_linear_v_weight_to_fp16, x = query_19_cast_fp16)[name = tensor("linear_86_cast_fp16")]; - tensor var_1946 = const()[name = tensor("op_1946"), val = tensor([1, -1, 8, 64])]; - tensor v_19_cast_fp16 = reshape(shape = var_1946, x = linear_86_cast_fp16)[name = tensor("v_19_cast_fp16")]; - tensor value_21_perm_0 = const()[name = tensor("value_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_9_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121045376)))]; - tensor var_1958_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_u_to_fp16)[name = tensor("op_1958_cast_fp16")]; - tensor module_layers_9_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_9_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121046464)))]; - tensor var_1960_cast_fp16 = add(x = q_55_cast_fp16, y = module_layers_9_self_attn_pos_bias_v_to_fp16)[name = tensor("op_1960_cast_fp16")]; - tensor q_with_bias_v_19_perm_0 = const()[name = tensor("q_with_bias_v_19_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_205_transpose_x_0 = const()[name = tensor("x_205_transpose_x_0"), val = tensor(false)]; - tensor x_205_transpose_y_0 = const()[name = tensor("x_205_transpose_y_0"), val = tensor(false)]; - tensor var_1962_to_fp16 = const()[name = tensor("op_1962_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121047552)))]; - tensor q_with_bias_v_19_cast_fp16 = transpose(perm = q_with_bias_v_19_perm_0, x = var_1960_cast_fp16)[name = tensor("transpose_140")]; - tensor x_205_cast_fp16 = matmul(transpose_x = x_205_transpose_x_0, transpose_y = x_205_transpose_y_0, x = q_with_bias_v_19_cast_fp16, y = var_1962_to_fp16)[name = tensor("x_205_cast_fp16")]; - tensor x_207_pad_0 = const()[name = tensor("x_207_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_207_mode_0 = const()[name = tensor("x_207_mode_0"), val = tensor("constant")]; - tensor const_160_to_fp16 = const()[name = tensor("const_160_to_fp16"), val = tensor(0x0p+0)]; - tensor x_207_cast_fp16 = pad(constant_val = const_160_to_fp16, mode = x_207_mode_0, pad = x_207_pad_0, x = x_205_cast_fp16)[name = tensor("x_207_cast_fp16")]; - tensor var_1970 = const()[name = tensor("op_1970"), val = tensor([1, 8, -1, 188])]; - tensor x_209_cast_fp16 = reshape(shape = var_1970, x = x_207_cast_fp16)[name = tensor("x_209_cast_fp16")]; - tensor var_1974_begin_0 = const()[name = tensor("op_1974_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_1974_end_0 = const()[name = tensor("op_1974_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_1974_end_mask_0 = const()[name = tensor("op_1974_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_1974_cast_fp16 = slice_by_index(begin = var_1974_begin_0, end = var_1974_end_0, end_mask = var_1974_end_mask_0, x = x_209_cast_fp16)[name = tensor("op_1974_cast_fp16")]; - tensor var_1975 = const()[name = tensor("op_1975"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_37_cast_fp16 = reshape(shape = var_1975, x = var_1974_cast_fp16)[name = tensor("matrix_bd_37_cast_fp16")]; - tensor matrix_ac_19_transpose_x_0 = const()[name = tensor("matrix_ac_19_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_19_transpose_y_0 = const()[name = tensor("matrix_ac_19_transpose_y_0"), val = tensor(false)]; - tensor transpose_69_perm_0 = const()[name = tensor("transpose_69_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_70_perm_0 = const()[name = tensor("transpose_70_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_70 = transpose(perm = transpose_70_perm_0, x = k_37_cast_fp16)[name = tensor("transpose_138")]; - tensor transpose_69 = transpose(perm = transpose_69_perm_0, x = var_1958_cast_fp16)[name = tensor("transpose_139")]; - tensor matrix_ac_19_cast_fp16 = matmul(transpose_x = matrix_ac_19_transpose_x_0, transpose_y = matrix_ac_19_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor("matrix_ac_19_cast_fp16")]; - tensor matrix_bd_39_begin_0 = const()[name = tensor("matrix_bd_39_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_39_end_0 = const()[name = tensor("matrix_bd_39_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_39_end_mask_0 = const()[name = tensor("matrix_bd_39_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_39_cast_fp16 = slice_by_index(begin = matrix_bd_39_begin_0, end = matrix_bd_39_end_0, end_mask = matrix_bd_39_end_mask_0, x = matrix_bd_37_cast_fp16)[name = tensor("matrix_bd_39_cast_fp16")]; - tensor var_1984_cast_fp16 = add(x = matrix_ac_19_cast_fp16, y = matrix_bd_39_cast_fp16)[name = tensor("op_1984_cast_fp16")]; - tensor _inversed_scores_37_y_0_to_fp16 = const()[name = tensor("_inversed_scores_37_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_37_cast_fp16 = mul(x = var_1984_cast_fp16, y = _inversed_scores_37_y_0_to_fp16)[name = tensor("_inversed_scores_37_cast_fp16")]; - tensor scores_39_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_37_cast_fp16, cond = mask_11)[name = tensor("scores_39_cast_fp16")]; - tensor var_1990_cast_fp16 = softmax(axis = var_23, x = scores_39_cast_fp16)[name = tensor("op_1990_cast_fp16")]; - tensor input_501_cast_fp16 = select(a = var_11_to_fp16, b = var_1990_cast_fp16, cond = mask_11)[name = tensor("input_501_cast_fp16")]; - tensor x_211_transpose_x_0 = const()[name = tensor("x_211_transpose_x_0"), val = tensor(false)]; - tensor x_211_transpose_y_0 = const()[name = tensor("x_211_transpose_y_0"), val = tensor(false)]; - tensor value_21_cast_fp16 = transpose(perm = value_21_perm_0, x = v_19_cast_fp16)[name = tensor("transpose_141")]; - tensor x_211_cast_fp16 = matmul(transpose_x = x_211_transpose_x_0, transpose_y = x_211_transpose_y_0, x = input_501_cast_fp16, y = value_21_cast_fp16)[name = tensor("x_211_cast_fp16")]; - tensor var_1994_perm_0 = const()[name = tensor("op_1994_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_1995 = const()[name = tensor("op_1995"), val = tensor([1, -1, 512])]; - tensor var_1994_cast_fp16 = transpose(perm = var_1994_perm_0, x = x_211_cast_fp16)[name = tensor("transpose_137")]; - tensor input_503_cast_fp16 = reshape(shape = var_1995, x = var_1994_cast_fp16)[name = tensor("input_503_cast_fp16")]; - tensor module_layers_9_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121431616)))]; - tensor module_layers_9_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_9_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121955968)))]; - tensor linear_88_cast_fp16 = linear(bias = module_layers_9_self_attn_linear_out_bias_to_fp16, weight = module_layers_9_self_attn_linear_out_weight_to_fp16, x = input_503_cast_fp16)[name = tensor("linear_88_cast_fp16")]; - tensor input_507_cast_fp16 = add(x = input_499_cast_fp16, y = linear_88_cast_fp16)[name = tensor("input_507_cast_fp16")]; - tensor x_215_axes_0 = const()[name = tensor("x_215_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121957056)))]; - tensor module_layers_9_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121958144)))]; - tensor x_215_cast_fp16 = layer_norm(axes = x_215_axes_0, beta = module_layers_9_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_conv_weight_to_fp16, x = input_507_cast_fp16)[name = tensor("x_215_cast_fp16")]; - tensor input_509_perm_0 = const()[name = tensor("input_509_perm_0"), val = tensor([0, 2, 1])]; - tensor input_511_pad_type_0 = const()[name = tensor("input_511_pad_type_0"), val = tensor("valid")]; - tensor input_511_strides_0 = const()[name = tensor("input_511_strides_0"), val = tensor([1])]; - tensor input_511_pad_0 = const()[name = tensor("input_511_pad_0"), val = tensor([0, 0])]; - tensor input_511_dilations_0 = const()[name = tensor("input_511_dilations_0"), val = tensor([1])]; - tensor input_511_groups_0 = const()[name = tensor("input_511_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(121959232)))]; - tensor module_layers_9_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123007872)))]; - tensor input_509_cast_fp16 = transpose(perm = input_509_perm_0, x = x_215_cast_fp16)[name = tensor("transpose_136")]; - tensor input_511_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv1_bias_to_fp16, dilations = input_511_dilations_0, groups = input_511_groups_0, pad = input_511_pad_0, pad_type = input_511_pad_type_0, strides = input_511_strides_0, weight = module_layers_9_conv_pointwise_conv1_weight_to_fp16, x = input_509_cast_fp16)[name = tensor("input_511_cast_fp16")]; - tensor x_217_split_num_splits_0 = const()[name = tensor("x_217_split_num_splits_0"), val = tensor(2)]; - tensor x_217_split_axis_0 = const()[name = tensor("x_217_split_axis_0"), val = tensor(1)]; - tensor x_217_split_cast_fp16_0, tensor x_217_split_cast_fp16_1 = split(axis = x_217_split_axis_0, num_splits = x_217_split_num_splits_0, x = input_511_cast_fp16)[name = tensor("x_217_split_cast_fp16")]; - tensor x_217_split_1_sigmoid_cast_fp16 = sigmoid(x = x_217_split_cast_fp16_1)[name = tensor("x_217_split_1_sigmoid_cast_fp16")]; - tensor x_217_cast_fp16 = mul(x = x_217_split_cast_fp16_0, y = x_217_split_1_sigmoid_cast_fp16)[name = tensor("x_217_cast_fp16")]; - tensor input_513_cast_fp16 = select(a = var_11_to_fp16, b = x_217_cast_fp16, cond = var_453)[name = tensor("input_513_cast_fp16")]; - tensor input_515_pad_0 = const()[name = tensor("input_515_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_515_mode_0 = const()[name = tensor("input_515_mode_0"), val = tensor("constant")]; - tensor const_163_to_fp16 = const()[name = tensor("const_163_to_fp16"), val = tensor(0x0p+0)]; - tensor input_515_cast_fp16 = pad(constant_val = const_163_to_fp16, mode = input_515_mode_0, pad = input_515_pad_0, x = input_513_cast_fp16)[name = tensor("input_515_cast_fp16")]; - tensor input_517_pad_type_0 = const()[name = tensor("input_517_pad_type_0"), val = tensor("valid")]; - tensor input_517_groups_0 = const()[name = tensor("input_517_groups_0"), val = tensor(512)]; - tensor input_517_strides_0 = const()[name = tensor("input_517_strides_0"), val = tensor([1])]; - tensor input_517_pad_0 = const()[name = tensor("input_517_pad_0"), val = tensor([0, 0])]; - tensor input_517_dilations_0 = const()[name = tensor("input_517_dilations_0"), val = tensor([1])]; - tensor const_252_to_fp16 = const()[name = tensor("const_252_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123009984)))]; - tensor const_253_to_fp16 = const()[name = tensor("const_253_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123019264)))]; - tensor input_519_cast_fp16 = conv(bias = const_253_to_fp16, dilations = input_517_dilations_0, groups = input_517_groups_0, pad = input_517_pad_0, pad_type = input_517_pad_type_0, strides = input_517_strides_0, weight = const_252_to_fp16, x = input_515_cast_fp16)[name = tensor("input_519_cast_fp16")]; - tensor input_521_cast_fp16 = silu(x = input_519_cast_fp16)[name = tensor("input_521_cast_fp16")]; - tensor x_219_pad_type_0 = const()[name = tensor("x_219_pad_type_0"), val = tensor("valid")]; - tensor x_219_strides_0 = const()[name = tensor("x_219_strides_0"), val = tensor([1])]; - tensor x_219_pad_0 = const()[name = tensor("x_219_pad_0"), val = tensor([0, 0])]; - tensor x_219_dilations_0 = const()[name = tensor("x_219_dilations_0"), val = tensor([1])]; - tensor x_219_groups_0 = const()[name = tensor("x_219_groups_0"), val = tensor(1)]; - tensor module_layers_9_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123020352)))]; - tensor module_layers_9_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_9_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123544704)))]; - tensor x_219_cast_fp16 = conv(bias = module_layers_9_conv_pointwise_conv2_bias_to_fp16, dilations = x_219_dilations_0, groups = x_219_groups_0, pad = x_219_pad_0, pad_type = x_219_pad_type_0, strides = x_219_strides_0, weight = module_layers_9_conv_pointwise_conv2_weight_to_fp16, x = input_521_cast_fp16)[name = tensor("x_219_cast_fp16")]; - tensor input_523_perm_0 = const()[name = tensor("input_523_perm_0"), val = tensor([0, 2, 1])]; - tensor input_523_cast_fp16 = transpose(perm = input_523_perm_0, x = x_219_cast_fp16)[name = tensor("transpose_135")]; - tensor input_525_cast_fp16 = add(x = input_507_cast_fp16, y = input_523_cast_fp16)[name = tensor("input_525_cast_fp16")]; - tensor input_527_axes_0 = const()[name = tensor("input_527_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123545792)))]; - tensor module_layers_9_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123546880)))]; - tensor input_527_cast_fp16 = layer_norm(axes = input_527_axes_0, beta = module_layers_9_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_feed_forward2_weight_to_fp16, x = input_525_cast_fp16)[name = tensor("input_527_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(123547968)))]; - tensor module_layers_9_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125645184)))]; - tensor linear_89_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear1_bias_to_fp16, weight = module_layers_9_feed_forward2_linear1_weight_to_fp16, x = input_527_cast_fp16)[name = tensor("linear_89_cast_fp16")]; - tensor input_531_cast_fp16 = silu(x = linear_89_cast_fp16)[name = tensor("input_531_cast_fp16")]; - tensor module_layers_9_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(125649344)))]; - tensor module_layers_9_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_9_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127746560)))]; - tensor linear_90_cast_fp16 = linear(bias = module_layers_9_feed_forward2_linear2_bias_to_fp16, weight = module_layers_9_feed_forward2_linear2_weight_to_fp16, x = input_531_cast_fp16)[name = tensor("linear_90_cast_fp16")]; - tensor var_2061_to_fp16 = const()[name = tensor("op_2061_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2062_cast_fp16 = mul(x = linear_90_cast_fp16, y = var_2061_to_fp16)[name = tensor("op_2062_cast_fp16")]; - tensor input_537_cast_fp16 = add(x = input_525_cast_fp16, y = var_2062_cast_fp16)[name = tensor("input_537_cast_fp16")]; - tensor input_539_axes_0 = const()[name = tensor("input_539_axes_0"), val = tensor([-1])]; - tensor module_layers_9_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_9_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127747648)))]; - tensor module_layers_9_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_9_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127748736)))]; - tensor input_539_cast_fp16 = layer_norm(axes = input_539_axes_0, beta = module_layers_9_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_9_norm_out_weight_to_fp16, x = input_537_cast_fp16)[name = tensor("input_539_cast_fp16")]; - tensor input_541_axes_0 = const()[name = tensor("input_541_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127749824)))]; - tensor module_layers_10_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127750912)))]; - tensor input_541_cast_fp16 = layer_norm(axes = input_541_axes_0, beta = module_layers_10_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward1_weight_to_fp16, x = input_539_cast_fp16)[name = tensor("input_541_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(127752000)))]; - tensor module_layers_10_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129849216)))]; - tensor linear_91_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear1_bias_to_fp16, weight = module_layers_10_feed_forward1_linear1_weight_to_fp16, x = input_541_cast_fp16)[name = tensor("linear_91_cast_fp16")]; - tensor input_545_cast_fp16 = silu(x = linear_91_cast_fp16)[name = tensor("input_545_cast_fp16")]; - tensor module_layers_10_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(129853376)))]; - tensor module_layers_10_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131950592)))]; - tensor linear_92_cast_fp16 = linear(bias = module_layers_10_feed_forward1_linear2_bias_to_fp16, weight = module_layers_10_feed_forward1_linear2_weight_to_fp16, x = input_545_cast_fp16)[name = tensor("linear_92_cast_fp16")]; - tensor var_2092_to_fp16 = const()[name = tensor("op_2092_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2093_cast_fp16 = mul(x = linear_92_cast_fp16, y = var_2092_to_fp16)[name = tensor("op_2093_cast_fp16")]; - tensor input_551_cast_fp16 = add(x = input_539_cast_fp16, y = var_2093_cast_fp16)[name = tensor("input_551_cast_fp16")]; - tensor query_21_axes_0 = const()[name = tensor("query_21_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131951680)))]; - tensor module_layers_10_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131952768)))]; - tensor query_21_cast_fp16 = layer_norm(axes = query_21_axes_0, beta = module_layers_10_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_self_att_weight_to_fp16, x = input_551_cast_fp16)[name = tensor("query_21_cast_fp16")]; - tensor module_layers_10_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(131953856)))]; - tensor module_layers_10_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132478208)))]; - tensor linear_93_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_q_bias_to_fp16, weight = module_layers_10_self_attn_linear_q_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_93_cast_fp16")]; - tensor var_2110 = const()[name = tensor("op_2110"), val = tensor([1, -1, 8, 64])]; - tensor q_61_cast_fp16 = reshape(shape = var_2110, x = linear_93_cast_fp16)[name = tensor("q_61_cast_fp16")]; - tensor module_layers_10_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(132479296)))]; - tensor module_layers_10_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133003648)))]; - tensor linear_94_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_k_bias_to_fp16, weight = module_layers_10_self_attn_linear_k_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_94_cast_fp16")]; - tensor var_2115 = const()[name = tensor("op_2115"), val = tensor([1, -1, 8, 64])]; - tensor k_41_cast_fp16 = reshape(shape = var_2115, x = linear_94_cast_fp16)[name = tensor("k_41_cast_fp16")]; - tensor module_layers_10_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133004736)))]; - tensor module_layers_10_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133529088)))]; - tensor linear_95_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_v_bias_to_fp16, weight = module_layers_10_self_attn_linear_v_weight_to_fp16, x = query_21_cast_fp16)[name = tensor("linear_95_cast_fp16")]; - tensor var_2120 = const()[name = tensor("op_2120"), val = tensor([1, -1, 8, 64])]; - tensor v_21_cast_fp16 = reshape(shape = var_2120, x = linear_95_cast_fp16)[name = tensor("v_21_cast_fp16")]; - tensor value_23_perm_0 = const()[name = tensor("value_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_10_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133530176)))]; - tensor var_2132_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2132_cast_fp16")]; - tensor module_layers_10_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_10_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133531264)))]; - tensor var_2134_cast_fp16 = add(x = q_61_cast_fp16, y = module_layers_10_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2134_cast_fp16")]; - tensor q_with_bias_v_21_perm_0 = const()[name = tensor("q_with_bias_v_21_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_227_transpose_x_0 = const()[name = tensor("x_227_transpose_x_0"), val = tensor(false)]; - tensor x_227_transpose_y_0 = const()[name = tensor("x_227_transpose_y_0"), val = tensor(false)]; - tensor var_2136_to_fp16 = const()[name = tensor("op_2136_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133532352)))]; - tensor q_with_bias_v_21_cast_fp16 = transpose(perm = q_with_bias_v_21_perm_0, x = var_2134_cast_fp16)[name = tensor("transpose_133")]; - tensor x_227_cast_fp16 = matmul(transpose_x = x_227_transpose_x_0, transpose_y = x_227_transpose_y_0, x = q_with_bias_v_21_cast_fp16, y = var_2136_to_fp16)[name = tensor("x_227_cast_fp16")]; - tensor x_229_pad_0 = const()[name = tensor("x_229_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_229_mode_0 = const()[name = tensor("x_229_mode_0"), val = tensor("constant")]; - tensor const_170_to_fp16 = const()[name = tensor("const_170_to_fp16"), val = tensor(0x0p+0)]; - tensor x_229_cast_fp16 = pad(constant_val = const_170_to_fp16, mode = x_229_mode_0, pad = x_229_pad_0, x = x_227_cast_fp16)[name = tensor("x_229_cast_fp16")]; - tensor var_2144 = const()[name = tensor("op_2144"), val = tensor([1, 8, -1, 188])]; - tensor x_231_cast_fp16 = reshape(shape = var_2144, x = x_229_cast_fp16)[name = tensor("x_231_cast_fp16")]; - tensor var_2148_begin_0 = const()[name = tensor("op_2148_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2148_end_0 = const()[name = tensor("op_2148_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2148_end_mask_0 = const()[name = tensor("op_2148_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2148_cast_fp16 = slice_by_index(begin = var_2148_begin_0, end = var_2148_end_0, end_mask = var_2148_end_mask_0, x = x_231_cast_fp16)[name = tensor("op_2148_cast_fp16")]; - tensor var_2149 = const()[name = tensor("op_2149"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_41_cast_fp16 = reshape(shape = var_2149, x = var_2148_cast_fp16)[name = tensor("matrix_bd_41_cast_fp16")]; - tensor matrix_ac_21_transpose_x_0 = const()[name = tensor("matrix_ac_21_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_21_transpose_y_0 = const()[name = tensor("matrix_ac_21_transpose_y_0"), val = tensor(false)]; - tensor transpose_71_perm_0 = const()[name = tensor("transpose_71_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_72_perm_0 = const()[name = tensor("transpose_72_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_72 = transpose(perm = transpose_72_perm_0, x = k_41_cast_fp16)[name = tensor("transpose_131")]; - tensor transpose_71 = transpose(perm = transpose_71_perm_0, x = var_2132_cast_fp16)[name = tensor("transpose_132")]; - tensor matrix_ac_21_cast_fp16 = matmul(transpose_x = matrix_ac_21_transpose_x_0, transpose_y = matrix_ac_21_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor("matrix_ac_21_cast_fp16")]; - tensor matrix_bd_43_begin_0 = const()[name = tensor("matrix_bd_43_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_43_end_0 = const()[name = tensor("matrix_bd_43_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_43_end_mask_0 = const()[name = tensor("matrix_bd_43_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_43_cast_fp16 = slice_by_index(begin = matrix_bd_43_begin_0, end = matrix_bd_43_end_0, end_mask = matrix_bd_43_end_mask_0, x = matrix_bd_41_cast_fp16)[name = tensor("matrix_bd_43_cast_fp16")]; - tensor var_2158_cast_fp16 = add(x = matrix_ac_21_cast_fp16, y = matrix_bd_43_cast_fp16)[name = tensor("op_2158_cast_fp16")]; - tensor _inversed_scores_41_y_0_to_fp16 = const()[name = tensor("_inversed_scores_41_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_41_cast_fp16 = mul(x = var_2158_cast_fp16, y = _inversed_scores_41_y_0_to_fp16)[name = tensor("_inversed_scores_41_cast_fp16")]; - tensor scores_43_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_41_cast_fp16, cond = mask_11)[name = tensor("scores_43_cast_fp16")]; - tensor var_2164_cast_fp16 = softmax(axis = var_23, x = scores_43_cast_fp16)[name = tensor("op_2164_cast_fp16")]; - tensor input_553_cast_fp16 = select(a = var_11_to_fp16, b = var_2164_cast_fp16, cond = mask_11)[name = tensor("input_553_cast_fp16")]; - tensor x_233_transpose_x_0 = const()[name = tensor("x_233_transpose_x_0"), val = tensor(false)]; - tensor x_233_transpose_y_0 = const()[name = tensor("x_233_transpose_y_0"), val = tensor(false)]; - tensor value_23_cast_fp16 = transpose(perm = value_23_perm_0, x = v_21_cast_fp16)[name = tensor("transpose_134")]; - tensor x_233_cast_fp16 = matmul(transpose_x = x_233_transpose_x_0, transpose_y = x_233_transpose_y_0, x = input_553_cast_fp16, y = value_23_cast_fp16)[name = tensor("x_233_cast_fp16")]; - tensor var_2168_perm_0 = const()[name = tensor("op_2168_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2169 = const()[name = tensor("op_2169"), val = tensor([1, -1, 512])]; - tensor var_2168_cast_fp16 = transpose(perm = var_2168_perm_0, x = x_233_cast_fp16)[name = tensor("transpose_130")]; - tensor input_555_cast_fp16 = reshape(shape = var_2169, x = var_2168_cast_fp16)[name = tensor("input_555_cast_fp16")]; - tensor module_layers_10_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(133916416)))]; - tensor module_layers_10_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_10_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134440768)))]; - tensor linear_97_cast_fp16 = linear(bias = module_layers_10_self_attn_linear_out_bias_to_fp16, weight = module_layers_10_self_attn_linear_out_weight_to_fp16, x = input_555_cast_fp16)[name = tensor("linear_97_cast_fp16")]; - tensor input_559_cast_fp16 = add(x = input_551_cast_fp16, y = linear_97_cast_fp16)[name = tensor("input_559_cast_fp16")]; - tensor x_237_axes_0 = const()[name = tensor("x_237_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134441856)))]; - tensor module_layers_10_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134442944)))]; - tensor x_237_cast_fp16 = layer_norm(axes = x_237_axes_0, beta = module_layers_10_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_conv_weight_to_fp16, x = input_559_cast_fp16)[name = tensor("x_237_cast_fp16")]; - tensor input_561_perm_0 = const()[name = tensor("input_561_perm_0"), val = tensor([0, 2, 1])]; - tensor input_563_pad_type_0 = const()[name = tensor("input_563_pad_type_0"), val = tensor("valid")]; - tensor input_563_strides_0 = const()[name = tensor("input_563_strides_0"), val = tensor([1])]; - tensor input_563_pad_0 = const()[name = tensor("input_563_pad_0"), val = tensor([0, 0])]; - tensor input_563_dilations_0 = const()[name = tensor("input_563_dilations_0"), val = tensor([1])]; - tensor input_563_groups_0 = const()[name = tensor("input_563_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(134444032)))]; - tensor module_layers_10_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135492672)))]; - tensor input_561_cast_fp16 = transpose(perm = input_561_perm_0, x = x_237_cast_fp16)[name = tensor("transpose_129")]; - tensor input_563_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv1_bias_to_fp16, dilations = input_563_dilations_0, groups = input_563_groups_0, pad = input_563_pad_0, pad_type = input_563_pad_type_0, strides = input_563_strides_0, weight = module_layers_10_conv_pointwise_conv1_weight_to_fp16, x = input_561_cast_fp16)[name = tensor("input_563_cast_fp16")]; - tensor x_239_split_num_splits_0 = const()[name = tensor("x_239_split_num_splits_0"), val = tensor(2)]; - tensor x_239_split_axis_0 = const()[name = tensor("x_239_split_axis_0"), val = tensor(1)]; - tensor x_239_split_cast_fp16_0, tensor x_239_split_cast_fp16_1 = split(axis = x_239_split_axis_0, num_splits = x_239_split_num_splits_0, x = input_563_cast_fp16)[name = tensor("x_239_split_cast_fp16")]; - tensor x_239_split_1_sigmoid_cast_fp16 = sigmoid(x = x_239_split_cast_fp16_1)[name = tensor("x_239_split_1_sigmoid_cast_fp16")]; - tensor x_239_cast_fp16 = mul(x = x_239_split_cast_fp16_0, y = x_239_split_1_sigmoid_cast_fp16)[name = tensor("x_239_cast_fp16")]; - tensor input_565_cast_fp16 = select(a = var_11_to_fp16, b = x_239_cast_fp16, cond = var_453)[name = tensor("input_565_cast_fp16")]; - tensor input_567_pad_0 = const()[name = tensor("input_567_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_567_mode_0 = const()[name = tensor("input_567_mode_0"), val = tensor("constant")]; - tensor const_173_to_fp16 = const()[name = tensor("const_173_to_fp16"), val = tensor(0x0p+0)]; - tensor input_567_cast_fp16 = pad(constant_val = const_173_to_fp16, mode = input_567_mode_0, pad = input_567_pad_0, x = input_565_cast_fp16)[name = tensor("input_567_cast_fp16")]; - tensor input_569_pad_type_0 = const()[name = tensor("input_569_pad_type_0"), val = tensor("valid")]; - tensor input_569_groups_0 = const()[name = tensor("input_569_groups_0"), val = tensor(512)]; - tensor input_569_strides_0 = const()[name = tensor("input_569_strides_0"), val = tensor([1])]; - tensor input_569_pad_0 = const()[name = tensor("input_569_pad_0"), val = tensor([0, 0])]; - tensor input_569_dilations_0 = const()[name = tensor("input_569_dilations_0"), val = tensor([1])]; - tensor const_254_to_fp16 = const()[name = tensor("const_254_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135494784)))]; - tensor const_255_to_fp16 = const()[name = tensor("const_255_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135504064)))]; - tensor input_571_cast_fp16 = conv(bias = const_255_to_fp16, dilations = input_569_dilations_0, groups = input_569_groups_0, pad = input_569_pad_0, pad_type = input_569_pad_type_0, strides = input_569_strides_0, weight = const_254_to_fp16, x = input_567_cast_fp16)[name = tensor("input_571_cast_fp16")]; - tensor input_573_cast_fp16 = silu(x = input_571_cast_fp16)[name = tensor("input_573_cast_fp16")]; - tensor x_241_pad_type_0 = const()[name = tensor("x_241_pad_type_0"), val = tensor("valid")]; - tensor x_241_strides_0 = const()[name = tensor("x_241_strides_0"), val = tensor([1])]; - tensor x_241_pad_0 = const()[name = tensor("x_241_pad_0"), val = tensor([0, 0])]; - tensor x_241_dilations_0 = const()[name = tensor("x_241_dilations_0"), val = tensor([1])]; - tensor x_241_groups_0 = const()[name = tensor("x_241_groups_0"), val = tensor(1)]; - tensor module_layers_10_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(135505152)))]; - tensor module_layers_10_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_10_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136029504)))]; - tensor x_241_cast_fp16 = conv(bias = module_layers_10_conv_pointwise_conv2_bias_to_fp16, dilations = x_241_dilations_0, groups = x_241_groups_0, pad = x_241_pad_0, pad_type = x_241_pad_type_0, strides = x_241_strides_0, weight = module_layers_10_conv_pointwise_conv2_weight_to_fp16, x = input_573_cast_fp16)[name = tensor("x_241_cast_fp16")]; - tensor input_575_perm_0 = const()[name = tensor("input_575_perm_0"), val = tensor([0, 2, 1])]; - tensor input_575_cast_fp16 = transpose(perm = input_575_perm_0, x = x_241_cast_fp16)[name = tensor("transpose_128")]; - tensor input_577_cast_fp16 = add(x = input_559_cast_fp16, y = input_575_cast_fp16)[name = tensor("input_577_cast_fp16")]; - tensor input_579_axes_0 = const()[name = tensor("input_579_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136030592)))]; - tensor module_layers_10_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136031680)))]; - tensor input_579_cast_fp16 = layer_norm(axes = input_579_axes_0, beta = module_layers_10_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_feed_forward2_weight_to_fp16, x = input_577_cast_fp16)[name = tensor("input_579_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(136032768)))]; - tensor module_layers_10_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138129984)))]; - tensor linear_98_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear1_bias_to_fp16, weight = module_layers_10_feed_forward2_linear1_weight_to_fp16, x = input_579_cast_fp16)[name = tensor("linear_98_cast_fp16")]; - tensor input_583_cast_fp16 = silu(x = linear_98_cast_fp16)[name = tensor("input_583_cast_fp16")]; - tensor module_layers_10_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(138134144)))]; - tensor module_layers_10_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_10_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140231360)))]; - tensor linear_99_cast_fp16 = linear(bias = module_layers_10_feed_forward2_linear2_bias_to_fp16, weight = module_layers_10_feed_forward2_linear2_weight_to_fp16, x = input_583_cast_fp16)[name = tensor("linear_99_cast_fp16")]; - tensor var_2235_to_fp16 = const()[name = tensor("op_2235_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2236_cast_fp16 = mul(x = linear_99_cast_fp16, y = var_2235_to_fp16)[name = tensor("op_2236_cast_fp16")]; - tensor input_589_cast_fp16 = add(x = input_577_cast_fp16, y = var_2236_cast_fp16)[name = tensor("input_589_cast_fp16")]; - tensor input_591_axes_0 = const()[name = tensor("input_591_axes_0"), val = tensor([-1])]; - tensor module_layers_10_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_10_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140232448)))]; - tensor module_layers_10_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_10_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140233536)))]; - tensor input_591_cast_fp16 = layer_norm(axes = input_591_axes_0, beta = module_layers_10_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_10_norm_out_weight_to_fp16, x = input_589_cast_fp16)[name = tensor("input_591_cast_fp16")]; - tensor input_593_axes_0 = const()[name = tensor("input_593_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140234624)))]; - tensor module_layers_11_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140235712)))]; - tensor input_593_cast_fp16 = layer_norm(axes = input_593_axes_0, beta = module_layers_11_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward1_weight_to_fp16, x = input_591_cast_fp16)[name = tensor("input_593_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(140236800)))]; - tensor module_layers_11_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142334016)))]; - tensor linear_100_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear1_bias_to_fp16, weight = module_layers_11_feed_forward1_linear1_weight_to_fp16, x = input_593_cast_fp16)[name = tensor("linear_100_cast_fp16")]; - tensor input_597_cast_fp16 = silu(x = linear_100_cast_fp16)[name = tensor("input_597_cast_fp16")]; - tensor module_layers_11_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(142338176)))]; - tensor module_layers_11_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144435392)))]; - tensor linear_101_cast_fp16 = linear(bias = module_layers_11_feed_forward1_linear2_bias_to_fp16, weight = module_layers_11_feed_forward1_linear2_weight_to_fp16, x = input_597_cast_fp16)[name = tensor("linear_101_cast_fp16")]; - tensor var_2266_to_fp16 = const()[name = tensor("op_2266_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2267_cast_fp16 = mul(x = linear_101_cast_fp16, y = var_2266_to_fp16)[name = tensor("op_2267_cast_fp16")]; - tensor input_603_cast_fp16 = add(x = input_591_cast_fp16, y = var_2267_cast_fp16)[name = tensor("input_603_cast_fp16")]; - tensor query_23_axes_0 = const()[name = tensor("query_23_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144436480)))]; - tensor module_layers_11_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144437568)))]; - tensor query_23_cast_fp16 = layer_norm(axes = query_23_axes_0, beta = module_layers_11_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_self_att_weight_to_fp16, x = input_603_cast_fp16)[name = tensor("query_23_cast_fp16")]; - tensor module_layers_11_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144438656)))]; - tensor module_layers_11_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144963008)))]; - tensor linear_102_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_q_bias_to_fp16, weight = module_layers_11_self_attn_linear_q_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_102_cast_fp16")]; - tensor var_2284 = const()[name = tensor("op_2284"), val = tensor([1, -1, 8, 64])]; - tensor q_67_cast_fp16 = reshape(shape = var_2284, x = linear_102_cast_fp16)[name = tensor("q_67_cast_fp16")]; - tensor module_layers_11_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(144964096)))]; - tensor module_layers_11_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145488448)))]; - tensor linear_103_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_k_bias_to_fp16, weight = module_layers_11_self_attn_linear_k_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_103_cast_fp16")]; - tensor var_2289 = const()[name = tensor("op_2289"), val = tensor([1, -1, 8, 64])]; - tensor k_45_cast_fp16 = reshape(shape = var_2289, x = linear_103_cast_fp16)[name = tensor("k_45_cast_fp16")]; - tensor module_layers_11_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(145489536)))]; - tensor module_layers_11_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146013888)))]; - tensor linear_104_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_v_bias_to_fp16, weight = module_layers_11_self_attn_linear_v_weight_to_fp16, x = query_23_cast_fp16)[name = tensor("linear_104_cast_fp16")]; - tensor var_2294 = const()[name = tensor("op_2294"), val = tensor([1, -1, 8, 64])]; - tensor v_23_cast_fp16 = reshape(shape = var_2294, x = linear_104_cast_fp16)[name = tensor("v_23_cast_fp16")]; - tensor value_25_perm_0 = const()[name = tensor("value_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_11_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146014976)))]; - tensor var_2306_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2306_cast_fp16")]; - tensor module_layers_11_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_11_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146016064)))]; - tensor var_2308_cast_fp16 = add(x = q_67_cast_fp16, y = module_layers_11_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2308_cast_fp16")]; - tensor q_with_bias_v_23_perm_0 = const()[name = tensor("q_with_bias_v_23_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_249_transpose_x_0 = const()[name = tensor("x_249_transpose_x_0"), val = tensor(false)]; - tensor x_249_transpose_y_0 = const()[name = tensor("x_249_transpose_y_0"), val = tensor(false)]; - tensor var_2310_to_fp16 = const()[name = tensor("op_2310_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146017152)))]; - tensor q_with_bias_v_23_cast_fp16 = transpose(perm = q_with_bias_v_23_perm_0, x = var_2308_cast_fp16)[name = tensor("transpose_126")]; - tensor x_249_cast_fp16 = matmul(transpose_x = x_249_transpose_x_0, transpose_y = x_249_transpose_y_0, x = q_with_bias_v_23_cast_fp16, y = var_2310_to_fp16)[name = tensor("x_249_cast_fp16")]; - tensor x_251_pad_0 = const()[name = tensor("x_251_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_251_mode_0 = const()[name = tensor("x_251_mode_0"), val = tensor("constant")]; - tensor const_180_to_fp16 = const()[name = tensor("const_180_to_fp16"), val = tensor(0x0p+0)]; - tensor x_251_cast_fp16 = pad(constant_val = const_180_to_fp16, mode = x_251_mode_0, pad = x_251_pad_0, x = x_249_cast_fp16)[name = tensor("x_251_cast_fp16")]; - tensor var_2318 = const()[name = tensor("op_2318"), val = tensor([1, 8, -1, 188])]; - tensor x_253_cast_fp16 = reshape(shape = var_2318, x = x_251_cast_fp16)[name = tensor("x_253_cast_fp16")]; - tensor var_2322_begin_0 = const()[name = tensor("op_2322_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2322_end_0 = const()[name = tensor("op_2322_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2322_end_mask_0 = const()[name = tensor("op_2322_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2322_cast_fp16 = slice_by_index(begin = var_2322_begin_0, end = var_2322_end_0, end_mask = var_2322_end_mask_0, x = x_253_cast_fp16)[name = tensor("op_2322_cast_fp16")]; - tensor var_2323 = const()[name = tensor("op_2323"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_45_cast_fp16 = reshape(shape = var_2323, x = var_2322_cast_fp16)[name = tensor("matrix_bd_45_cast_fp16")]; - tensor matrix_ac_23_transpose_x_0 = const()[name = tensor("matrix_ac_23_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_23_transpose_y_0 = const()[name = tensor("matrix_ac_23_transpose_y_0"), val = tensor(false)]; - tensor transpose_73_perm_0 = const()[name = tensor("transpose_73_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_74_perm_0 = const()[name = tensor("transpose_74_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_74 = transpose(perm = transpose_74_perm_0, x = k_45_cast_fp16)[name = tensor("transpose_124")]; - tensor transpose_73 = transpose(perm = transpose_73_perm_0, x = var_2306_cast_fp16)[name = tensor("transpose_125")]; - tensor matrix_ac_23_cast_fp16 = matmul(transpose_x = matrix_ac_23_transpose_x_0, transpose_y = matrix_ac_23_transpose_y_0, x = transpose_73, y = transpose_74)[name = tensor("matrix_ac_23_cast_fp16")]; - tensor matrix_bd_47_begin_0 = const()[name = tensor("matrix_bd_47_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_47_end_0 = const()[name = tensor("matrix_bd_47_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_47_end_mask_0 = const()[name = tensor("matrix_bd_47_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_47_cast_fp16 = slice_by_index(begin = matrix_bd_47_begin_0, end = matrix_bd_47_end_0, end_mask = matrix_bd_47_end_mask_0, x = matrix_bd_45_cast_fp16)[name = tensor("matrix_bd_47_cast_fp16")]; - tensor var_2332_cast_fp16 = add(x = matrix_ac_23_cast_fp16, y = matrix_bd_47_cast_fp16)[name = tensor("op_2332_cast_fp16")]; - tensor _inversed_scores_45_y_0_to_fp16 = const()[name = tensor("_inversed_scores_45_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_45_cast_fp16 = mul(x = var_2332_cast_fp16, y = _inversed_scores_45_y_0_to_fp16)[name = tensor("_inversed_scores_45_cast_fp16")]; - tensor scores_47_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_45_cast_fp16, cond = mask_11)[name = tensor("scores_47_cast_fp16")]; - tensor var_2338_cast_fp16 = softmax(axis = var_23, x = scores_47_cast_fp16)[name = tensor("op_2338_cast_fp16")]; - tensor input_605_cast_fp16 = select(a = var_11_to_fp16, b = var_2338_cast_fp16, cond = mask_11)[name = tensor("input_605_cast_fp16")]; - tensor x_255_transpose_x_0 = const()[name = tensor("x_255_transpose_x_0"), val = tensor(false)]; - tensor x_255_transpose_y_0 = const()[name = tensor("x_255_transpose_y_0"), val = tensor(false)]; - tensor value_25_cast_fp16 = transpose(perm = value_25_perm_0, x = v_23_cast_fp16)[name = tensor("transpose_127")]; - tensor x_255_cast_fp16 = matmul(transpose_x = x_255_transpose_x_0, transpose_y = x_255_transpose_y_0, x = input_605_cast_fp16, y = value_25_cast_fp16)[name = tensor("x_255_cast_fp16")]; - tensor var_2342_perm_0 = const()[name = tensor("op_2342_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2343 = const()[name = tensor("op_2343"), val = tensor([1, -1, 512])]; - tensor var_2342_cast_fp16 = transpose(perm = var_2342_perm_0, x = x_255_cast_fp16)[name = tensor("transpose_123")]; - tensor input_607_cast_fp16 = reshape(shape = var_2343, x = var_2342_cast_fp16)[name = tensor("input_607_cast_fp16")]; - tensor module_layers_11_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146401216)))]; - tensor module_layers_11_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_11_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146925568)))]; - tensor linear_106_cast_fp16 = linear(bias = module_layers_11_self_attn_linear_out_bias_to_fp16, weight = module_layers_11_self_attn_linear_out_weight_to_fp16, x = input_607_cast_fp16)[name = tensor("linear_106_cast_fp16")]; - tensor input_611_cast_fp16 = add(x = input_603_cast_fp16, y = linear_106_cast_fp16)[name = tensor("input_611_cast_fp16")]; - tensor x_259_axes_0 = const()[name = tensor("x_259_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146926656)))]; - tensor module_layers_11_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146927744)))]; - tensor x_259_cast_fp16 = layer_norm(axes = x_259_axes_0, beta = module_layers_11_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_conv_weight_to_fp16, x = input_611_cast_fp16)[name = tensor("x_259_cast_fp16")]; - tensor input_613_perm_0 = const()[name = tensor("input_613_perm_0"), val = tensor([0, 2, 1])]; - tensor input_615_pad_type_0 = const()[name = tensor("input_615_pad_type_0"), val = tensor("valid")]; - tensor input_615_strides_0 = const()[name = tensor("input_615_strides_0"), val = tensor([1])]; - tensor input_615_pad_0 = const()[name = tensor("input_615_pad_0"), val = tensor([0, 0])]; - tensor input_615_dilations_0 = const()[name = tensor("input_615_dilations_0"), val = tensor([1])]; - tensor input_615_groups_0 = const()[name = tensor("input_615_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(146928832)))]; - tensor module_layers_11_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147977472)))]; - tensor input_613_cast_fp16 = transpose(perm = input_613_perm_0, x = x_259_cast_fp16)[name = tensor("transpose_122")]; - tensor input_615_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv1_bias_to_fp16, dilations = input_615_dilations_0, groups = input_615_groups_0, pad = input_615_pad_0, pad_type = input_615_pad_type_0, strides = input_615_strides_0, weight = module_layers_11_conv_pointwise_conv1_weight_to_fp16, x = input_613_cast_fp16)[name = tensor("input_615_cast_fp16")]; - tensor x_261_split_num_splits_0 = const()[name = tensor("x_261_split_num_splits_0"), val = tensor(2)]; - tensor x_261_split_axis_0 = const()[name = tensor("x_261_split_axis_0"), val = tensor(1)]; - tensor x_261_split_cast_fp16_0, tensor x_261_split_cast_fp16_1 = split(axis = x_261_split_axis_0, num_splits = x_261_split_num_splits_0, x = input_615_cast_fp16)[name = tensor("x_261_split_cast_fp16")]; - tensor x_261_split_1_sigmoid_cast_fp16 = sigmoid(x = x_261_split_cast_fp16_1)[name = tensor("x_261_split_1_sigmoid_cast_fp16")]; - tensor x_261_cast_fp16 = mul(x = x_261_split_cast_fp16_0, y = x_261_split_1_sigmoid_cast_fp16)[name = tensor("x_261_cast_fp16")]; - tensor input_617_cast_fp16 = select(a = var_11_to_fp16, b = x_261_cast_fp16, cond = var_453)[name = tensor("input_617_cast_fp16")]; - tensor input_619_pad_0 = const()[name = tensor("input_619_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_619_mode_0 = const()[name = tensor("input_619_mode_0"), val = tensor("constant")]; - tensor const_183_to_fp16 = const()[name = tensor("const_183_to_fp16"), val = tensor(0x0p+0)]; - tensor input_619_cast_fp16 = pad(constant_val = const_183_to_fp16, mode = input_619_mode_0, pad = input_619_pad_0, x = input_617_cast_fp16)[name = tensor("input_619_cast_fp16")]; - tensor input_621_pad_type_0 = const()[name = tensor("input_621_pad_type_0"), val = tensor("valid")]; - tensor input_621_groups_0 = const()[name = tensor("input_621_groups_0"), val = tensor(512)]; - tensor input_621_strides_0 = const()[name = tensor("input_621_strides_0"), val = tensor([1])]; - tensor input_621_pad_0 = const()[name = tensor("input_621_pad_0"), val = tensor([0, 0])]; - tensor input_621_dilations_0 = const()[name = tensor("input_621_dilations_0"), val = tensor([1])]; - tensor const_256_to_fp16 = const()[name = tensor("const_256_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147979584)))]; - tensor const_257_to_fp16 = const()[name = tensor("const_257_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147988864)))]; - tensor input_623_cast_fp16 = conv(bias = const_257_to_fp16, dilations = input_621_dilations_0, groups = input_621_groups_0, pad = input_621_pad_0, pad_type = input_621_pad_type_0, strides = input_621_strides_0, weight = const_256_to_fp16, x = input_619_cast_fp16)[name = tensor("input_623_cast_fp16")]; - tensor input_625_cast_fp16 = silu(x = input_623_cast_fp16)[name = tensor("input_625_cast_fp16")]; - tensor x_263_pad_type_0 = const()[name = tensor("x_263_pad_type_0"), val = tensor("valid")]; - tensor x_263_strides_0 = const()[name = tensor("x_263_strides_0"), val = tensor([1])]; - tensor x_263_pad_0 = const()[name = tensor("x_263_pad_0"), val = tensor([0, 0])]; - tensor x_263_dilations_0 = const()[name = tensor("x_263_dilations_0"), val = tensor([1])]; - tensor x_263_groups_0 = const()[name = tensor("x_263_groups_0"), val = tensor(1)]; - tensor module_layers_11_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(147989952)))]; - tensor module_layers_11_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_11_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148514304)))]; - tensor x_263_cast_fp16 = conv(bias = module_layers_11_conv_pointwise_conv2_bias_to_fp16, dilations = x_263_dilations_0, groups = x_263_groups_0, pad = x_263_pad_0, pad_type = x_263_pad_type_0, strides = x_263_strides_0, weight = module_layers_11_conv_pointwise_conv2_weight_to_fp16, x = input_625_cast_fp16)[name = tensor("x_263_cast_fp16")]; - tensor input_627_perm_0 = const()[name = tensor("input_627_perm_0"), val = tensor([0, 2, 1])]; - tensor input_627_cast_fp16 = transpose(perm = input_627_perm_0, x = x_263_cast_fp16)[name = tensor("transpose_121")]; - tensor input_629_cast_fp16 = add(x = input_611_cast_fp16, y = input_627_cast_fp16)[name = tensor("input_629_cast_fp16")]; - tensor input_631_axes_0 = const()[name = tensor("input_631_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148515392)))]; - tensor module_layers_11_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148516480)))]; - tensor input_631_cast_fp16 = layer_norm(axes = input_631_axes_0, beta = module_layers_11_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_feed_forward2_weight_to_fp16, x = input_629_cast_fp16)[name = tensor("input_631_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(148517568)))]; - tensor module_layers_11_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150614784)))]; - tensor linear_107_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear1_bias_to_fp16, weight = module_layers_11_feed_forward2_linear1_weight_to_fp16, x = input_631_cast_fp16)[name = tensor("linear_107_cast_fp16")]; - tensor input_635_cast_fp16 = silu(x = linear_107_cast_fp16)[name = tensor("input_635_cast_fp16")]; - tensor module_layers_11_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(150618944)))]; - tensor module_layers_11_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_11_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152716160)))]; - tensor linear_108_cast_fp16 = linear(bias = module_layers_11_feed_forward2_linear2_bias_to_fp16, weight = module_layers_11_feed_forward2_linear2_weight_to_fp16, x = input_635_cast_fp16)[name = tensor("linear_108_cast_fp16")]; - tensor var_2409_to_fp16 = const()[name = tensor("op_2409_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2410_cast_fp16 = mul(x = linear_108_cast_fp16, y = var_2409_to_fp16)[name = tensor("op_2410_cast_fp16")]; - tensor input_641_cast_fp16 = add(x = input_629_cast_fp16, y = var_2410_cast_fp16)[name = tensor("input_641_cast_fp16")]; - tensor input_643_axes_0 = const()[name = tensor("input_643_axes_0"), val = tensor([-1])]; - tensor module_layers_11_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_11_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152717248)))]; - tensor module_layers_11_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_11_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152718336)))]; - tensor input_643_cast_fp16 = layer_norm(axes = input_643_axes_0, beta = module_layers_11_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_11_norm_out_weight_to_fp16, x = input_641_cast_fp16)[name = tensor("input_643_cast_fp16")]; - tensor input_645_axes_0 = const()[name = tensor("input_645_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152719424)))]; - tensor module_layers_12_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152720512)))]; - tensor input_645_cast_fp16 = layer_norm(axes = input_645_axes_0, beta = module_layers_12_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward1_weight_to_fp16, x = input_643_cast_fp16)[name = tensor("input_645_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(152721600)))]; - tensor module_layers_12_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154818816)))]; - tensor linear_109_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear1_bias_to_fp16, weight = module_layers_12_feed_forward1_linear1_weight_to_fp16, x = input_645_cast_fp16)[name = tensor("linear_109_cast_fp16")]; - tensor input_649_cast_fp16 = silu(x = linear_109_cast_fp16)[name = tensor("input_649_cast_fp16")]; - tensor module_layers_12_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(154822976)))]; - tensor module_layers_12_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156920192)))]; - tensor linear_110_cast_fp16 = linear(bias = module_layers_12_feed_forward1_linear2_bias_to_fp16, weight = module_layers_12_feed_forward1_linear2_weight_to_fp16, x = input_649_cast_fp16)[name = tensor("linear_110_cast_fp16")]; - tensor var_2440_to_fp16 = const()[name = tensor("op_2440_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2441_cast_fp16 = mul(x = linear_110_cast_fp16, y = var_2440_to_fp16)[name = tensor("op_2441_cast_fp16")]; - tensor input_655_cast_fp16 = add(x = input_643_cast_fp16, y = var_2441_cast_fp16)[name = tensor("input_655_cast_fp16")]; - tensor query_25_axes_0 = const()[name = tensor("query_25_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156921280)))]; - tensor module_layers_12_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156922368)))]; - tensor query_25_cast_fp16 = layer_norm(axes = query_25_axes_0, beta = module_layers_12_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_self_att_weight_to_fp16, x = input_655_cast_fp16)[name = tensor("query_25_cast_fp16")]; - tensor module_layers_12_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(156923456)))]; - tensor module_layers_12_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157447808)))]; - tensor linear_111_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_q_bias_to_fp16, weight = module_layers_12_self_attn_linear_q_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_111_cast_fp16")]; - tensor var_2458 = const()[name = tensor("op_2458"), val = tensor([1, -1, 8, 64])]; - tensor q_73_cast_fp16 = reshape(shape = var_2458, x = linear_111_cast_fp16)[name = tensor("q_73_cast_fp16")]; - tensor module_layers_12_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157448896)))]; - tensor module_layers_12_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157973248)))]; - tensor linear_112_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_k_bias_to_fp16, weight = module_layers_12_self_attn_linear_k_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_112_cast_fp16")]; - tensor var_2463 = const()[name = tensor("op_2463"), val = tensor([1, -1, 8, 64])]; - tensor k_49_cast_fp16 = reshape(shape = var_2463, x = linear_112_cast_fp16)[name = tensor("k_49_cast_fp16")]; - tensor module_layers_12_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(157974336)))]; - tensor module_layers_12_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158498688)))]; - tensor linear_113_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_v_bias_to_fp16, weight = module_layers_12_self_attn_linear_v_weight_to_fp16, x = query_25_cast_fp16)[name = tensor("linear_113_cast_fp16")]; - tensor var_2468 = const()[name = tensor("op_2468"), val = tensor([1, -1, 8, 64])]; - tensor v_25_cast_fp16 = reshape(shape = var_2468, x = linear_113_cast_fp16)[name = tensor("v_25_cast_fp16")]; - tensor value_27_perm_0 = const()[name = tensor("value_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_12_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158499776)))]; - tensor var_2480_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2480_cast_fp16")]; - tensor module_layers_12_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_12_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158500864)))]; - tensor var_2482_cast_fp16 = add(x = q_73_cast_fp16, y = module_layers_12_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2482_cast_fp16")]; - tensor q_with_bias_v_25_perm_0 = const()[name = tensor("q_with_bias_v_25_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_271_transpose_x_0 = const()[name = tensor("x_271_transpose_x_0"), val = tensor(false)]; - tensor x_271_transpose_y_0 = const()[name = tensor("x_271_transpose_y_0"), val = tensor(false)]; - tensor var_2484_to_fp16 = const()[name = tensor("op_2484_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158501952)))]; - tensor q_with_bias_v_25_cast_fp16 = transpose(perm = q_with_bias_v_25_perm_0, x = var_2482_cast_fp16)[name = tensor("transpose_119")]; - tensor x_271_cast_fp16 = matmul(transpose_x = x_271_transpose_x_0, transpose_y = x_271_transpose_y_0, x = q_with_bias_v_25_cast_fp16, y = var_2484_to_fp16)[name = tensor("x_271_cast_fp16")]; - tensor x_273_pad_0 = const()[name = tensor("x_273_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_273_mode_0 = const()[name = tensor("x_273_mode_0"), val = tensor("constant")]; - tensor const_190_to_fp16 = const()[name = tensor("const_190_to_fp16"), val = tensor(0x0p+0)]; - tensor x_273_cast_fp16 = pad(constant_val = const_190_to_fp16, mode = x_273_mode_0, pad = x_273_pad_0, x = x_271_cast_fp16)[name = tensor("x_273_cast_fp16")]; - tensor var_2492 = const()[name = tensor("op_2492"), val = tensor([1, 8, -1, 188])]; - tensor x_275_cast_fp16 = reshape(shape = var_2492, x = x_273_cast_fp16)[name = tensor("x_275_cast_fp16")]; - tensor var_2496_begin_0 = const()[name = tensor("op_2496_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2496_end_0 = const()[name = tensor("op_2496_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2496_end_mask_0 = const()[name = tensor("op_2496_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2496_cast_fp16 = slice_by_index(begin = var_2496_begin_0, end = var_2496_end_0, end_mask = var_2496_end_mask_0, x = x_275_cast_fp16)[name = tensor("op_2496_cast_fp16")]; - tensor var_2497 = const()[name = tensor("op_2497"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_49_cast_fp16 = reshape(shape = var_2497, x = var_2496_cast_fp16)[name = tensor("matrix_bd_49_cast_fp16")]; - tensor matrix_ac_25_transpose_x_0 = const()[name = tensor("matrix_ac_25_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_25_transpose_y_0 = const()[name = tensor("matrix_ac_25_transpose_y_0"), val = tensor(false)]; - tensor transpose_75_perm_0 = const()[name = tensor("transpose_75_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_76_perm_0 = const()[name = tensor("transpose_76_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_76 = transpose(perm = transpose_76_perm_0, x = k_49_cast_fp16)[name = tensor("transpose_117")]; - tensor transpose_75 = transpose(perm = transpose_75_perm_0, x = var_2480_cast_fp16)[name = tensor("transpose_118")]; - tensor matrix_ac_25_cast_fp16 = matmul(transpose_x = matrix_ac_25_transpose_x_0, transpose_y = matrix_ac_25_transpose_y_0, x = transpose_75, y = transpose_76)[name = tensor("matrix_ac_25_cast_fp16")]; - tensor matrix_bd_51_begin_0 = const()[name = tensor("matrix_bd_51_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_51_end_0 = const()[name = tensor("matrix_bd_51_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_51_end_mask_0 = const()[name = tensor("matrix_bd_51_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_51_cast_fp16 = slice_by_index(begin = matrix_bd_51_begin_0, end = matrix_bd_51_end_0, end_mask = matrix_bd_51_end_mask_0, x = matrix_bd_49_cast_fp16)[name = tensor("matrix_bd_51_cast_fp16")]; - tensor var_2506_cast_fp16 = add(x = matrix_ac_25_cast_fp16, y = matrix_bd_51_cast_fp16)[name = tensor("op_2506_cast_fp16")]; - tensor _inversed_scores_49_y_0_to_fp16 = const()[name = tensor("_inversed_scores_49_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_49_cast_fp16 = mul(x = var_2506_cast_fp16, y = _inversed_scores_49_y_0_to_fp16)[name = tensor("_inversed_scores_49_cast_fp16")]; - tensor scores_51_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_49_cast_fp16, cond = mask_11)[name = tensor("scores_51_cast_fp16")]; - tensor var_2512_cast_fp16 = softmax(axis = var_23, x = scores_51_cast_fp16)[name = tensor("op_2512_cast_fp16")]; - tensor input_657_cast_fp16 = select(a = var_11_to_fp16, b = var_2512_cast_fp16, cond = mask_11)[name = tensor("input_657_cast_fp16")]; - tensor x_277_transpose_x_0 = const()[name = tensor("x_277_transpose_x_0"), val = tensor(false)]; - tensor x_277_transpose_y_0 = const()[name = tensor("x_277_transpose_y_0"), val = tensor(false)]; - tensor value_27_cast_fp16 = transpose(perm = value_27_perm_0, x = v_25_cast_fp16)[name = tensor("transpose_120")]; - tensor x_277_cast_fp16 = matmul(transpose_x = x_277_transpose_x_0, transpose_y = x_277_transpose_y_0, x = input_657_cast_fp16, y = value_27_cast_fp16)[name = tensor("x_277_cast_fp16")]; - tensor var_2516_perm_0 = const()[name = tensor("op_2516_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2517 = const()[name = tensor("op_2517"), val = tensor([1, -1, 512])]; - tensor var_2516_cast_fp16 = transpose(perm = var_2516_perm_0, x = x_277_cast_fp16)[name = tensor("transpose_116")]; - tensor input_659_cast_fp16 = reshape(shape = var_2517, x = var_2516_cast_fp16)[name = tensor("input_659_cast_fp16")]; - tensor module_layers_12_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(158886016)))]; - tensor module_layers_12_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_12_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159410368)))]; - tensor linear_115_cast_fp16 = linear(bias = module_layers_12_self_attn_linear_out_bias_to_fp16, weight = module_layers_12_self_attn_linear_out_weight_to_fp16, x = input_659_cast_fp16)[name = tensor("linear_115_cast_fp16")]; - tensor input_663_cast_fp16 = add(x = input_655_cast_fp16, y = linear_115_cast_fp16)[name = tensor("input_663_cast_fp16")]; - tensor x_281_axes_0 = const()[name = tensor("x_281_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159411456)))]; - tensor module_layers_12_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159412544)))]; - tensor x_281_cast_fp16 = layer_norm(axes = x_281_axes_0, beta = module_layers_12_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_conv_weight_to_fp16, x = input_663_cast_fp16)[name = tensor("x_281_cast_fp16")]; - tensor input_665_perm_0 = const()[name = tensor("input_665_perm_0"), val = tensor([0, 2, 1])]; - tensor input_667_pad_type_0 = const()[name = tensor("input_667_pad_type_0"), val = tensor("valid")]; - tensor input_667_strides_0 = const()[name = tensor("input_667_strides_0"), val = tensor([1])]; - tensor input_667_pad_0 = const()[name = tensor("input_667_pad_0"), val = tensor([0, 0])]; - tensor input_667_dilations_0 = const()[name = tensor("input_667_dilations_0"), val = tensor([1])]; - tensor input_667_groups_0 = const()[name = tensor("input_667_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(159413632)))]; - tensor module_layers_12_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160462272)))]; - tensor input_665_cast_fp16 = transpose(perm = input_665_perm_0, x = x_281_cast_fp16)[name = tensor("transpose_115")]; - tensor input_667_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv1_bias_to_fp16, dilations = input_667_dilations_0, groups = input_667_groups_0, pad = input_667_pad_0, pad_type = input_667_pad_type_0, strides = input_667_strides_0, weight = module_layers_12_conv_pointwise_conv1_weight_to_fp16, x = input_665_cast_fp16)[name = tensor("input_667_cast_fp16")]; - tensor x_283_split_num_splits_0 = const()[name = tensor("x_283_split_num_splits_0"), val = tensor(2)]; - tensor x_283_split_axis_0 = const()[name = tensor("x_283_split_axis_0"), val = tensor(1)]; - tensor x_283_split_cast_fp16_0, tensor x_283_split_cast_fp16_1 = split(axis = x_283_split_axis_0, num_splits = x_283_split_num_splits_0, x = input_667_cast_fp16)[name = tensor("x_283_split_cast_fp16")]; - tensor x_283_split_1_sigmoid_cast_fp16 = sigmoid(x = x_283_split_cast_fp16_1)[name = tensor("x_283_split_1_sigmoid_cast_fp16")]; - tensor x_283_cast_fp16 = mul(x = x_283_split_cast_fp16_0, y = x_283_split_1_sigmoid_cast_fp16)[name = tensor("x_283_cast_fp16")]; - tensor input_669_cast_fp16 = select(a = var_11_to_fp16, b = x_283_cast_fp16, cond = var_453)[name = tensor("input_669_cast_fp16")]; - tensor input_671_pad_0 = const()[name = tensor("input_671_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_671_mode_0 = const()[name = tensor("input_671_mode_0"), val = tensor("constant")]; - tensor const_193_to_fp16 = const()[name = tensor("const_193_to_fp16"), val = tensor(0x0p+0)]; - tensor input_671_cast_fp16 = pad(constant_val = const_193_to_fp16, mode = input_671_mode_0, pad = input_671_pad_0, x = input_669_cast_fp16)[name = tensor("input_671_cast_fp16")]; - tensor input_673_pad_type_0 = const()[name = tensor("input_673_pad_type_0"), val = tensor("valid")]; - tensor input_673_groups_0 = const()[name = tensor("input_673_groups_0"), val = tensor(512)]; - tensor input_673_strides_0 = const()[name = tensor("input_673_strides_0"), val = tensor([1])]; - tensor input_673_pad_0 = const()[name = tensor("input_673_pad_0"), val = tensor([0, 0])]; - tensor input_673_dilations_0 = const()[name = tensor("input_673_dilations_0"), val = tensor([1])]; - tensor const_258_to_fp16 = const()[name = tensor("const_258_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160464384)))]; - tensor const_259_to_fp16 = const()[name = tensor("const_259_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160473664)))]; - tensor input_675_cast_fp16 = conv(bias = const_259_to_fp16, dilations = input_673_dilations_0, groups = input_673_groups_0, pad = input_673_pad_0, pad_type = input_673_pad_type_0, strides = input_673_strides_0, weight = const_258_to_fp16, x = input_671_cast_fp16)[name = tensor("input_675_cast_fp16")]; - tensor input_677_cast_fp16 = silu(x = input_675_cast_fp16)[name = tensor("input_677_cast_fp16")]; - tensor x_285_pad_type_0 = const()[name = tensor("x_285_pad_type_0"), val = tensor("valid")]; - tensor x_285_strides_0 = const()[name = tensor("x_285_strides_0"), val = tensor([1])]; - tensor x_285_pad_0 = const()[name = tensor("x_285_pad_0"), val = tensor([0, 0])]; - tensor x_285_dilations_0 = const()[name = tensor("x_285_dilations_0"), val = tensor([1])]; - tensor x_285_groups_0 = const()[name = tensor("x_285_groups_0"), val = tensor(1)]; - tensor module_layers_12_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160474752)))]; - tensor module_layers_12_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_12_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(160999104)))]; - tensor x_285_cast_fp16 = conv(bias = module_layers_12_conv_pointwise_conv2_bias_to_fp16, dilations = x_285_dilations_0, groups = x_285_groups_0, pad = x_285_pad_0, pad_type = x_285_pad_type_0, strides = x_285_strides_0, weight = module_layers_12_conv_pointwise_conv2_weight_to_fp16, x = input_677_cast_fp16)[name = tensor("x_285_cast_fp16")]; - tensor input_679_perm_0 = const()[name = tensor("input_679_perm_0"), val = tensor([0, 2, 1])]; - tensor input_679_cast_fp16 = transpose(perm = input_679_perm_0, x = x_285_cast_fp16)[name = tensor("transpose_114")]; - tensor input_681_cast_fp16 = add(x = input_663_cast_fp16, y = input_679_cast_fp16)[name = tensor("input_681_cast_fp16")]; - tensor input_683_axes_0 = const()[name = tensor("input_683_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161000192)))]; - tensor module_layers_12_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161001280)))]; - tensor input_683_cast_fp16 = layer_norm(axes = input_683_axes_0, beta = module_layers_12_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_feed_forward2_weight_to_fp16, x = input_681_cast_fp16)[name = tensor("input_683_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(161002368)))]; - tensor module_layers_12_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163099584)))]; - tensor linear_116_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear1_bias_to_fp16, weight = module_layers_12_feed_forward2_linear1_weight_to_fp16, x = input_683_cast_fp16)[name = tensor("linear_116_cast_fp16")]; - tensor input_687_cast_fp16 = silu(x = linear_116_cast_fp16)[name = tensor("input_687_cast_fp16")]; - tensor module_layers_12_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(163103744)))]; - tensor module_layers_12_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_12_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165200960)))]; - tensor linear_117_cast_fp16 = linear(bias = module_layers_12_feed_forward2_linear2_bias_to_fp16, weight = module_layers_12_feed_forward2_linear2_weight_to_fp16, x = input_687_cast_fp16)[name = tensor("linear_117_cast_fp16")]; - tensor var_2583_to_fp16 = const()[name = tensor("op_2583_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2584_cast_fp16 = mul(x = linear_117_cast_fp16, y = var_2583_to_fp16)[name = tensor("op_2584_cast_fp16")]; - tensor input_693_cast_fp16 = add(x = input_681_cast_fp16, y = var_2584_cast_fp16)[name = tensor("input_693_cast_fp16")]; - tensor input_695_axes_0 = const()[name = tensor("input_695_axes_0"), val = tensor([-1])]; - tensor module_layers_12_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_12_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165202048)))]; - tensor module_layers_12_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_12_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165203136)))]; - tensor input_695_cast_fp16 = layer_norm(axes = input_695_axes_0, beta = module_layers_12_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_12_norm_out_weight_to_fp16, x = input_693_cast_fp16)[name = tensor("input_695_cast_fp16")]; - tensor input_697_axes_0 = const()[name = tensor("input_697_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165204224)))]; - tensor module_layers_13_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165205312)))]; - tensor input_697_cast_fp16 = layer_norm(axes = input_697_axes_0, beta = module_layers_13_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward1_weight_to_fp16, x = input_695_cast_fp16)[name = tensor("input_697_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(165206400)))]; - tensor module_layers_13_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167303616)))]; - tensor linear_118_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear1_bias_to_fp16, weight = module_layers_13_feed_forward1_linear1_weight_to_fp16, x = input_697_cast_fp16)[name = tensor("linear_118_cast_fp16")]; - tensor input_701_cast_fp16 = silu(x = linear_118_cast_fp16)[name = tensor("input_701_cast_fp16")]; - tensor module_layers_13_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(167307776)))]; - tensor module_layers_13_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169404992)))]; - tensor linear_119_cast_fp16 = linear(bias = module_layers_13_feed_forward1_linear2_bias_to_fp16, weight = module_layers_13_feed_forward1_linear2_weight_to_fp16, x = input_701_cast_fp16)[name = tensor("linear_119_cast_fp16")]; - tensor var_2614_to_fp16 = const()[name = tensor("op_2614_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2615_cast_fp16 = mul(x = linear_119_cast_fp16, y = var_2614_to_fp16)[name = tensor("op_2615_cast_fp16")]; - tensor input_707_cast_fp16 = add(x = input_695_cast_fp16, y = var_2615_cast_fp16)[name = tensor("input_707_cast_fp16")]; - tensor query_27_axes_0 = const()[name = tensor("query_27_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169406080)))]; - tensor module_layers_13_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169407168)))]; - tensor query_27_cast_fp16 = layer_norm(axes = query_27_axes_0, beta = module_layers_13_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_self_att_weight_to_fp16, x = input_707_cast_fp16)[name = tensor("query_27_cast_fp16")]; - tensor module_layers_13_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169408256)))]; - tensor module_layers_13_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169932608)))]; - tensor linear_120_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_q_bias_to_fp16, weight = module_layers_13_self_attn_linear_q_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_120_cast_fp16")]; - tensor var_2632 = const()[name = tensor("op_2632"), val = tensor([1, -1, 8, 64])]; - tensor q_79_cast_fp16 = reshape(shape = var_2632, x = linear_120_cast_fp16)[name = tensor("q_79_cast_fp16")]; - tensor module_layers_13_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(169933696)))]; - tensor module_layers_13_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170458048)))]; - tensor linear_121_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_k_bias_to_fp16, weight = module_layers_13_self_attn_linear_k_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_121_cast_fp16")]; - tensor var_2637 = const()[name = tensor("op_2637"), val = tensor([1, -1, 8, 64])]; - tensor k_53_cast_fp16 = reshape(shape = var_2637, x = linear_121_cast_fp16)[name = tensor("k_53_cast_fp16")]; - tensor module_layers_13_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170459136)))]; - tensor module_layers_13_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170983488)))]; - tensor linear_122_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_v_bias_to_fp16, weight = module_layers_13_self_attn_linear_v_weight_to_fp16, x = query_27_cast_fp16)[name = tensor("linear_122_cast_fp16")]; - tensor var_2642 = const()[name = tensor("op_2642"), val = tensor([1, -1, 8, 64])]; - tensor v_27_cast_fp16 = reshape(shape = var_2642, x = linear_122_cast_fp16)[name = tensor("v_27_cast_fp16")]; - tensor value_29_perm_0 = const()[name = tensor("value_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_13_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170984576)))]; - tensor var_2654_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2654_cast_fp16")]; - tensor module_layers_13_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_13_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170985664)))]; - tensor var_2656_cast_fp16 = add(x = q_79_cast_fp16, y = module_layers_13_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2656_cast_fp16")]; - tensor q_with_bias_v_27_perm_0 = const()[name = tensor("q_with_bias_v_27_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_293_transpose_x_0 = const()[name = tensor("x_293_transpose_x_0"), val = tensor(false)]; - tensor x_293_transpose_y_0 = const()[name = tensor("x_293_transpose_y_0"), val = tensor(false)]; - tensor var_2658_to_fp16 = const()[name = tensor("op_2658_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(170986752)))]; - tensor q_with_bias_v_27_cast_fp16 = transpose(perm = q_with_bias_v_27_perm_0, x = var_2656_cast_fp16)[name = tensor("transpose_112")]; - tensor x_293_cast_fp16 = matmul(transpose_x = x_293_transpose_x_0, transpose_y = x_293_transpose_y_0, x = q_with_bias_v_27_cast_fp16, y = var_2658_to_fp16)[name = tensor("x_293_cast_fp16")]; - tensor x_295_pad_0 = const()[name = tensor("x_295_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_295_mode_0 = const()[name = tensor("x_295_mode_0"), val = tensor("constant")]; - tensor const_200_to_fp16 = const()[name = tensor("const_200_to_fp16"), val = tensor(0x0p+0)]; - tensor x_295_cast_fp16 = pad(constant_val = const_200_to_fp16, mode = x_295_mode_0, pad = x_295_pad_0, x = x_293_cast_fp16)[name = tensor("x_295_cast_fp16")]; - tensor var_2666 = const()[name = tensor("op_2666"), val = tensor([1, 8, -1, 188])]; - tensor x_297_cast_fp16 = reshape(shape = var_2666, x = x_295_cast_fp16)[name = tensor("x_297_cast_fp16")]; - tensor var_2670_begin_0 = const()[name = tensor("op_2670_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2670_end_0 = const()[name = tensor("op_2670_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2670_end_mask_0 = const()[name = tensor("op_2670_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2670_cast_fp16 = slice_by_index(begin = var_2670_begin_0, end = var_2670_end_0, end_mask = var_2670_end_mask_0, x = x_297_cast_fp16)[name = tensor("op_2670_cast_fp16")]; - tensor var_2671 = const()[name = tensor("op_2671"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_53_cast_fp16 = reshape(shape = var_2671, x = var_2670_cast_fp16)[name = tensor("matrix_bd_53_cast_fp16")]; - tensor matrix_ac_27_transpose_x_0 = const()[name = tensor("matrix_ac_27_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_27_transpose_y_0 = const()[name = tensor("matrix_ac_27_transpose_y_0"), val = tensor(false)]; - tensor transpose_77_perm_0 = const()[name = tensor("transpose_77_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_78_perm_0 = const()[name = tensor("transpose_78_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_78 = transpose(perm = transpose_78_perm_0, x = k_53_cast_fp16)[name = tensor("transpose_110")]; - tensor transpose_77 = transpose(perm = transpose_77_perm_0, x = var_2654_cast_fp16)[name = tensor("transpose_111")]; - tensor matrix_ac_27_cast_fp16 = matmul(transpose_x = matrix_ac_27_transpose_x_0, transpose_y = matrix_ac_27_transpose_y_0, x = transpose_77, y = transpose_78)[name = tensor("matrix_ac_27_cast_fp16")]; - tensor matrix_bd_55_begin_0 = const()[name = tensor("matrix_bd_55_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_55_end_0 = const()[name = tensor("matrix_bd_55_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_55_end_mask_0 = const()[name = tensor("matrix_bd_55_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_55_cast_fp16 = slice_by_index(begin = matrix_bd_55_begin_0, end = matrix_bd_55_end_0, end_mask = matrix_bd_55_end_mask_0, x = matrix_bd_53_cast_fp16)[name = tensor("matrix_bd_55_cast_fp16")]; - tensor var_2680_cast_fp16 = add(x = matrix_ac_27_cast_fp16, y = matrix_bd_55_cast_fp16)[name = tensor("op_2680_cast_fp16")]; - tensor _inversed_scores_53_y_0_to_fp16 = const()[name = tensor("_inversed_scores_53_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_53_cast_fp16 = mul(x = var_2680_cast_fp16, y = _inversed_scores_53_y_0_to_fp16)[name = tensor("_inversed_scores_53_cast_fp16")]; - tensor scores_55_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_53_cast_fp16, cond = mask_11)[name = tensor("scores_55_cast_fp16")]; - tensor var_2686_cast_fp16 = softmax(axis = var_23, x = scores_55_cast_fp16)[name = tensor("op_2686_cast_fp16")]; - tensor input_709_cast_fp16 = select(a = var_11_to_fp16, b = var_2686_cast_fp16, cond = mask_11)[name = tensor("input_709_cast_fp16")]; - tensor x_299_transpose_x_0 = const()[name = tensor("x_299_transpose_x_0"), val = tensor(false)]; - tensor x_299_transpose_y_0 = const()[name = tensor("x_299_transpose_y_0"), val = tensor(false)]; - tensor value_29_cast_fp16 = transpose(perm = value_29_perm_0, x = v_27_cast_fp16)[name = tensor("transpose_113")]; - tensor x_299_cast_fp16 = matmul(transpose_x = x_299_transpose_x_0, transpose_y = x_299_transpose_y_0, x = input_709_cast_fp16, y = value_29_cast_fp16)[name = tensor("x_299_cast_fp16")]; - tensor var_2690_perm_0 = const()[name = tensor("op_2690_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2691 = const()[name = tensor("op_2691"), val = tensor([1, -1, 512])]; - tensor var_2690_cast_fp16 = transpose(perm = var_2690_perm_0, x = x_299_cast_fp16)[name = tensor("transpose_109")]; - tensor input_711_cast_fp16 = reshape(shape = var_2691, x = var_2690_cast_fp16)[name = tensor("input_711_cast_fp16")]; - tensor module_layers_13_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171370816)))]; - tensor module_layers_13_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_13_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171895168)))]; - tensor linear_124_cast_fp16 = linear(bias = module_layers_13_self_attn_linear_out_bias_to_fp16, weight = module_layers_13_self_attn_linear_out_weight_to_fp16, x = input_711_cast_fp16)[name = tensor("linear_124_cast_fp16")]; - tensor input_715_cast_fp16 = add(x = input_707_cast_fp16, y = linear_124_cast_fp16)[name = tensor("input_715_cast_fp16")]; - tensor x_303_axes_0 = const()[name = tensor("x_303_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171896256)))]; - tensor module_layers_13_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171897344)))]; - tensor x_303_cast_fp16 = layer_norm(axes = x_303_axes_0, beta = module_layers_13_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_conv_weight_to_fp16, x = input_715_cast_fp16)[name = tensor("x_303_cast_fp16")]; - tensor input_717_perm_0 = const()[name = tensor("input_717_perm_0"), val = tensor([0, 2, 1])]; - tensor input_719_pad_type_0 = const()[name = tensor("input_719_pad_type_0"), val = tensor("valid")]; - tensor input_719_strides_0 = const()[name = tensor("input_719_strides_0"), val = tensor([1])]; - tensor input_719_pad_0 = const()[name = tensor("input_719_pad_0"), val = tensor([0, 0])]; - tensor input_719_dilations_0 = const()[name = tensor("input_719_dilations_0"), val = tensor([1])]; - tensor input_719_groups_0 = const()[name = tensor("input_719_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(171898432)))]; - tensor module_layers_13_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172947072)))]; - tensor input_717_cast_fp16 = transpose(perm = input_717_perm_0, x = x_303_cast_fp16)[name = tensor("transpose_108")]; - tensor input_719_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv1_bias_to_fp16, dilations = input_719_dilations_0, groups = input_719_groups_0, pad = input_719_pad_0, pad_type = input_719_pad_type_0, strides = input_719_strides_0, weight = module_layers_13_conv_pointwise_conv1_weight_to_fp16, x = input_717_cast_fp16)[name = tensor("input_719_cast_fp16")]; - tensor x_305_split_num_splits_0 = const()[name = tensor("x_305_split_num_splits_0"), val = tensor(2)]; - tensor x_305_split_axis_0 = const()[name = tensor("x_305_split_axis_0"), val = tensor(1)]; - tensor x_305_split_cast_fp16_0, tensor x_305_split_cast_fp16_1 = split(axis = x_305_split_axis_0, num_splits = x_305_split_num_splits_0, x = input_719_cast_fp16)[name = tensor("x_305_split_cast_fp16")]; - tensor x_305_split_1_sigmoid_cast_fp16 = sigmoid(x = x_305_split_cast_fp16_1)[name = tensor("x_305_split_1_sigmoid_cast_fp16")]; - tensor x_305_cast_fp16 = mul(x = x_305_split_cast_fp16_0, y = x_305_split_1_sigmoid_cast_fp16)[name = tensor("x_305_cast_fp16")]; - tensor input_721_cast_fp16 = select(a = var_11_to_fp16, b = x_305_cast_fp16, cond = var_453)[name = tensor("input_721_cast_fp16")]; - tensor input_723_pad_0 = const()[name = tensor("input_723_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_723_mode_0 = const()[name = tensor("input_723_mode_0"), val = tensor("constant")]; - tensor const_203_to_fp16 = const()[name = tensor("const_203_to_fp16"), val = tensor(0x0p+0)]; - tensor input_723_cast_fp16 = pad(constant_val = const_203_to_fp16, mode = input_723_mode_0, pad = input_723_pad_0, x = input_721_cast_fp16)[name = tensor("input_723_cast_fp16")]; - tensor input_725_pad_type_0 = const()[name = tensor("input_725_pad_type_0"), val = tensor("valid")]; - tensor input_725_groups_0 = const()[name = tensor("input_725_groups_0"), val = tensor(512)]; - tensor input_725_strides_0 = const()[name = tensor("input_725_strides_0"), val = tensor([1])]; - tensor input_725_pad_0 = const()[name = tensor("input_725_pad_0"), val = tensor([0, 0])]; - tensor input_725_dilations_0 = const()[name = tensor("input_725_dilations_0"), val = tensor([1])]; - tensor const_260_to_fp16 = const()[name = tensor("const_260_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172949184)))]; - tensor const_261_to_fp16 = const()[name = tensor("const_261_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172958464)))]; - tensor input_727_cast_fp16 = conv(bias = const_261_to_fp16, dilations = input_725_dilations_0, groups = input_725_groups_0, pad = input_725_pad_0, pad_type = input_725_pad_type_0, strides = input_725_strides_0, weight = const_260_to_fp16, x = input_723_cast_fp16)[name = tensor("input_727_cast_fp16")]; - tensor input_729_cast_fp16 = silu(x = input_727_cast_fp16)[name = tensor("input_729_cast_fp16")]; - tensor x_307_pad_type_0 = const()[name = tensor("x_307_pad_type_0"), val = tensor("valid")]; - tensor x_307_strides_0 = const()[name = tensor("x_307_strides_0"), val = tensor([1])]; - tensor x_307_pad_0 = const()[name = tensor("x_307_pad_0"), val = tensor([0, 0])]; - tensor x_307_dilations_0 = const()[name = tensor("x_307_dilations_0"), val = tensor([1])]; - tensor x_307_groups_0 = const()[name = tensor("x_307_groups_0"), val = tensor(1)]; - tensor module_layers_13_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(172959552)))]; - tensor module_layers_13_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_13_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173483904)))]; - tensor x_307_cast_fp16 = conv(bias = module_layers_13_conv_pointwise_conv2_bias_to_fp16, dilations = x_307_dilations_0, groups = x_307_groups_0, pad = x_307_pad_0, pad_type = x_307_pad_type_0, strides = x_307_strides_0, weight = module_layers_13_conv_pointwise_conv2_weight_to_fp16, x = input_729_cast_fp16)[name = tensor("x_307_cast_fp16")]; - tensor input_731_perm_0 = const()[name = tensor("input_731_perm_0"), val = tensor([0, 2, 1])]; - tensor input_731_cast_fp16 = transpose(perm = input_731_perm_0, x = x_307_cast_fp16)[name = tensor("transpose_107")]; - tensor input_733_cast_fp16 = add(x = input_715_cast_fp16, y = input_731_cast_fp16)[name = tensor("input_733_cast_fp16")]; - tensor input_735_axes_0 = const()[name = tensor("input_735_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173484992)))]; - tensor module_layers_13_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173486080)))]; - tensor input_735_cast_fp16 = layer_norm(axes = input_735_axes_0, beta = module_layers_13_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_feed_forward2_weight_to_fp16, x = input_733_cast_fp16)[name = tensor("input_735_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(173487168)))]; - tensor module_layers_13_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175584384)))]; - tensor linear_125_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear1_bias_to_fp16, weight = module_layers_13_feed_forward2_linear1_weight_to_fp16, x = input_735_cast_fp16)[name = tensor("linear_125_cast_fp16")]; - tensor input_739_cast_fp16 = silu(x = linear_125_cast_fp16)[name = tensor("input_739_cast_fp16")]; - tensor module_layers_13_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(175588544)))]; - tensor module_layers_13_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_13_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177685760)))]; - tensor linear_126_cast_fp16 = linear(bias = module_layers_13_feed_forward2_linear2_bias_to_fp16, weight = module_layers_13_feed_forward2_linear2_weight_to_fp16, x = input_739_cast_fp16)[name = tensor("linear_126_cast_fp16")]; - tensor var_2757_to_fp16 = const()[name = tensor("op_2757_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2758_cast_fp16 = mul(x = linear_126_cast_fp16, y = var_2757_to_fp16)[name = tensor("op_2758_cast_fp16")]; - tensor input_745_cast_fp16 = add(x = input_733_cast_fp16, y = var_2758_cast_fp16)[name = tensor("input_745_cast_fp16")]; - tensor input_747_axes_0 = const()[name = tensor("input_747_axes_0"), val = tensor([-1])]; - tensor module_layers_13_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_13_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177686848)))]; - tensor module_layers_13_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_13_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177687936)))]; - tensor input_747_cast_fp16 = layer_norm(axes = input_747_axes_0, beta = module_layers_13_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_13_norm_out_weight_to_fp16, x = input_745_cast_fp16)[name = tensor("input_747_cast_fp16")]; - tensor input_749_axes_0 = const()[name = tensor("input_749_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177689024)))]; - tensor module_layers_14_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177690112)))]; - tensor input_749_cast_fp16 = layer_norm(axes = input_749_axes_0, beta = module_layers_14_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward1_weight_to_fp16, x = input_747_cast_fp16)[name = tensor("input_749_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(177691200)))]; - tensor module_layers_14_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179788416)))]; - tensor linear_127_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear1_bias_to_fp16, weight = module_layers_14_feed_forward1_linear1_weight_to_fp16, x = input_749_cast_fp16)[name = tensor("linear_127_cast_fp16")]; - tensor input_753_cast_fp16 = silu(x = linear_127_cast_fp16)[name = tensor("input_753_cast_fp16")]; - tensor module_layers_14_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(179792576)))]; - tensor module_layers_14_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181889792)))]; - tensor linear_128_cast_fp16 = linear(bias = module_layers_14_feed_forward1_linear2_bias_to_fp16, weight = module_layers_14_feed_forward1_linear2_weight_to_fp16, x = input_753_cast_fp16)[name = tensor("linear_128_cast_fp16")]; - tensor var_2788_to_fp16 = const()[name = tensor("op_2788_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2789_cast_fp16 = mul(x = linear_128_cast_fp16, y = var_2788_to_fp16)[name = tensor("op_2789_cast_fp16")]; - tensor input_759_cast_fp16 = add(x = input_747_cast_fp16, y = var_2789_cast_fp16)[name = tensor("input_759_cast_fp16")]; - tensor query_29_axes_0 = const()[name = tensor("query_29_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181890880)))]; - tensor module_layers_14_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181891968)))]; - tensor query_29_cast_fp16 = layer_norm(axes = query_29_axes_0, beta = module_layers_14_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_self_att_weight_to_fp16, x = input_759_cast_fp16)[name = tensor("query_29_cast_fp16")]; - tensor module_layers_14_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(181893056)))]; - tensor module_layers_14_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182417408)))]; - tensor linear_129_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_q_bias_to_fp16, weight = module_layers_14_self_attn_linear_q_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_129_cast_fp16")]; - tensor var_2806 = const()[name = tensor("op_2806"), val = tensor([1, -1, 8, 64])]; - tensor q_85_cast_fp16 = reshape(shape = var_2806, x = linear_129_cast_fp16)[name = tensor("q_85_cast_fp16")]; - tensor module_layers_14_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182418496)))]; - tensor module_layers_14_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182942848)))]; - tensor linear_130_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_k_bias_to_fp16, weight = module_layers_14_self_attn_linear_k_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_130_cast_fp16")]; - tensor var_2811 = const()[name = tensor("op_2811"), val = tensor([1, -1, 8, 64])]; - tensor k_57_cast_fp16 = reshape(shape = var_2811, x = linear_130_cast_fp16)[name = tensor("k_57_cast_fp16")]; - tensor module_layers_14_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(182943936)))]; - tensor module_layers_14_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183468288)))]; - tensor linear_131_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_v_bias_to_fp16, weight = module_layers_14_self_attn_linear_v_weight_to_fp16, x = query_29_cast_fp16)[name = tensor("linear_131_cast_fp16")]; - tensor var_2816 = const()[name = tensor("op_2816"), val = tensor([1, -1, 8, 64])]; - tensor v_29_cast_fp16 = reshape(shape = var_2816, x = linear_131_cast_fp16)[name = tensor("v_29_cast_fp16")]; - tensor value_31_perm_0 = const()[name = tensor("value_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_14_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183469376)))]; - tensor var_2828_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_u_to_fp16)[name = tensor("op_2828_cast_fp16")]; - tensor module_layers_14_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_14_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183470464)))]; - tensor var_2830_cast_fp16 = add(x = q_85_cast_fp16, y = module_layers_14_self_attn_pos_bias_v_to_fp16)[name = tensor("op_2830_cast_fp16")]; - tensor q_with_bias_v_29_perm_0 = const()[name = tensor("q_with_bias_v_29_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_315_transpose_x_0 = const()[name = tensor("x_315_transpose_x_0"), val = tensor(false)]; - tensor x_315_transpose_y_0 = const()[name = tensor("x_315_transpose_y_0"), val = tensor(false)]; - tensor var_2832_to_fp16 = const()[name = tensor("op_2832_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183471552)))]; - tensor q_with_bias_v_29_cast_fp16 = transpose(perm = q_with_bias_v_29_perm_0, x = var_2830_cast_fp16)[name = tensor("transpose_105")]; - tensor x_315_cast_fp16 = matmul(transpose_x = x_315_transpose_x_0, transpose_y = x_315_transpose_y_0, x = q_with_bias_v_29_cast_fp16, y = var_2832_to_fp16)[name = tensor("x_315_cast_fp16")]; - tensor x_317_pad_0 = const()[name = tensor("x_317_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_317_mode_0 = const()[name = tensor("x_317_mode_0"), val = tensor("constant")]; - tensor const_210_to_fp16 = const()[name = tensor("const_210_to_fp16"), val = tensor(0x0p+0)]; - tensor x_317_cast_fp16 = pad(constant_val = const_210_to_fp16, mode = x_317_mode_0, pad = x_317_pad_0, x = x_315_cast_fp16)[name = tensor("x_317_cast_fp16")]; - tensor var_2840 = const()[name = tensor("op_2840"), val = tensor([1, 8, -1, 188])]; - tensor x_319_cast_fp16 = reshape(shape = var_2840, x = x_317_cast_fp16)[name = tensor("x_319_cast_fp16")]; - tensor var_2844_begin_0 = const()[name = tensor("op_2844_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_2844_end_0 = const()[name = tensor("op_2844_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_2844_end_mask_0 = const()[name = tensor("op_2844_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_2844_cast_fp16 = slice_by_index(begin = var_2844_begin_0, end = var_2844_end_0, end_mask = var_2844_end_mask_0, x = x_319_cast_fp16)[name = tensor("op_2844_cast_fp16")]; - tensor var_2845 = const()[name = tensor("op_2845"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_57_cast_fp16 = reshape(shape = var_2845, x = var_2844_cast_fp16)[name = tensor("matrix_bd_57_cast_fp16")]; - tensor matrix_ac_29_transpose_x_0 = const()[name = tensor("matrix_ac_29_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_29_transpose_y_0 = const()[name = tensor("matrix_ac_29_transpose_y_0"), val = tensor(false)]; - tensor transpose_79_perm_0 = const()[name = tensor("transpose_79_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_80_perm_0 = const()[name = tensor("transpose_80_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_80 = transpose(perm = transpose_80_perm_0, x = k_57_cast_fp16)[name = tensor("transpose_103")]; - tensor transpose_79 = transpose(perm = transpose_79_perm_0, x = var_2828_cast_fp16)[name = tensor("transpose_104")]; - tensor matrix_ac_29_cast_fp16 = matmul(transpose_x = matrix_ac_29_transpose_x_0, transpose_y = matrix_ac_29_transpose_y_0, x = transpose_79, y = transpose_80)[name = tensor("matrix_ac_29_cast_fp16")]; - tensor matrix_bd_59_begin_0 = const()[name = tensor("matrix_bd_59_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_59_end_0 = const()[name = tensor("matrix_bd_59_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_59_end_mask_0 = const()[name = tensor("matrix_bd_59_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_59_cast_fp16 = slice_by_index(begin = matrix_bd_59_begin_0, end = matrix_bd_59_end_0, end_mask = matrix_bd_59_end_mask_0, x = matrix_bd_57_cast_fp16)[name = tensor("matrix_bd_59_cast_fp16")]; - tensor var_2854_cast_fp16 = add(x = matrix_ac_29_cast_fp16, y = matrix_bd_59_cast_fp16)[name = tensor("op_2854_cast_fp16")]; - tensor _inversed_scores_57_y_0_to_fp16 = const()[name = tensor("_inversed_scores_57_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_57_cast_fp16 = mul(x = var_2854_cast_fp16, y = _inversed_scores_57_y_0_to_fp16)[name = tensor("_inversed_scores_57_cast_fp16")]; - tensor scores_59_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_57_cast_fp16, cond = mask_11)[name = tensor("scores_59_cast_fp16")]; - tensor var_2860_cast_fp16 = softmax(axis = var_23, x = scores_59_cast_fp16)[name = tensor("op_2860_cast_fp16")]; - tensor input_761_cast_fp16 = select(a = var_11_to_fp16, b = var_2860_cast_fp16, cond = mask_11)[name = tensor("input_761_cast_fp16")]; - tensor x_321_transpose_x_0 = const()[name = tensor("x_321_transpose_x_0"), val = tensor(false)]; - tensor x_321_transpose_y_0 = const()[name = tensor("x_321_transpose_y_0"), val = tensor(false)]; - tensor value_31_cast_fp16 = transpose(perm = value_31_perm_0, x = v_29_cast_fp16)[name = tensor("transpose_106")]; - tensor x_321_cast_fp16 = matmul(transpose_x = x_321_transpose_x_0, transpose_y = x_321_transpose_y_0, x = input_761_cast_fp16, y = value_31_cast_fp16)[name = tensor("x_321_cast_fp16")]; - tensor var_2864_perm_0 = const()[name = tensor("op_2864_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_2865 = const()[name = tensor("op_2865"), val = tensor([1, -1, 512])]; - tensor var_2864_cast_fp16 = transpose(perm = var_2864_perm_0, x = x_321_cast_fp16)[name = tensor("transpose_102")]; - tensor input_763_cast_fp16 = reshape(shape = var_2865, x = var_2864_cast_fp16)[name = tensor("input_763_cast_fp16")]; - tensor module_layers_14_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(183855616)))]; - tensor module_layers_14_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_14_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184379968)))]; - tensor linear_133_cast_fp16 = linear(bias = module_layers_14_self_attn_linear_out_bias_to_fp16, weight = module_layers_14_self_attn_linear_out_weight_to_fp16, x = input_763_cast_fp16)[name = tensor("linear_133_cast_fp16")]; - tensor input_767_cast_fp16 = add(x = input_759_cast_fp16, y = linear_133_cast_fp16)[name = tensor("input_767_cast_fp16")]; - tensor x_325_axes_0 = const()[name = tensor("x_325_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184381056)))]; - tensor module_layers_14_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184382144)))]; - tensor x_325_cast_fp16 = layer_norm(axes = x_325_axes_0, beta = module_layers_14_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_conv_weight_to_fp16, x = input_767_cast_fp16)[name = tensor("x_325_cast_fp16")]; - tensor input_769_perm_0 = const()[name = tensor("input_769_perm_0"), val = tensor([0, 2, 1])]; - tensor input_771_pad_type_0 = const()[name = tensor("input_771_pad_type_0"), val = tensor("valid")]; - tensor input_771_strides_0 = const()[name = tensor("input_771_strides_0"), val = tensor([1])]; - tensor input_771_pad_0 = const()[name = tensor("input_771_pad_0"), val = tensor([0, 0])]; - tensor input_771_dilations_0 = const()[name = tensor("input_771_dilations_0"), val = tensor([1])]; - tensor input_771_groups_0 = const()[name = tensor("input_771_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(184383232)))]; - tensor module_layers_14_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185431872)))]; - tensor input_769_cast_fp16 = transpose(perm = input_769_perm_0, x = x_325_cast_fp16)[name = tensor("transpose_101")]; - tensor input_771_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv1_bias_to_fp16, dilations = input_771_dilations_0, groups = input_771_groups_0, pad = input_771_pad_0, pad_type = input_771_pad_type_0, strides = input_771_strides_0, weight = module_layers_14_conv_pointwise_conv1_weight_to_fp16, x = input_769_cast_fp16)[name = tensor("input_771_cast_fp16")]; - tensor x_327_split_num_splits_0 = const()[name = tensor("x_327_split_num_splits_0"), val = tensor(2)]; - tensor x_327_split_axis_0 = const()[name = tensor("x_327_split_axis_0"), val = tensor(1)]; - tensor x_327_split_cast_fp16_0, tensor x_327_split_cast_fp16_1 = split(axis = x_327_split_axis_0, num_splits = x_327_split_num_splits_0, x = input_771_cast_fp16)[name = tensor("x_327_split_cast_fp16")]; - tensor x_327_split_1_sigmoid_cast_fp16 = sigmoid(x = x_327_split_cast_fp16_1)[name = tensor("x_327_split_1_sigmoid_cast_fp16")]; - tensor x_327_cast_fp16 = mul(x = x_327_split_cast_fp16_0, y = x_327_split_1_sigmoid_cast_fp16)[name = tensor("x_327_cast_fp16")]; - tensor input_773_cast_fp16 = select(a = var_11_to_fp16, b = x_327_cast_fp16, cond = var_453)[name = tensor("input_773_cast_fp16")]; - tensor input_775_pad_0 = const()[name = tensor("input_775_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_775_mode_0 = const()[name = tensor("input_775_mode_0"), val = tensor("constant")]; - tensor const_213_to_fp16 = const()[name = tensor("const_213_to_fp16"), val = tensor(0x0p+0)]; - tensor input_775_cast_fp16 = pad(constant_val = const_213_to_fp16, mode = input_775_mode_0, pad = input_775_pad_0, x = input_773_cast_fp16)[name = tensor("input_775_cast_fp16")]; - tensor input_777_pad_type_0 = const()[name = tensor("input_777_pad_type_0"), val = tensor("valid")]; - tensor input_777_groups_0 = const()[name = tensor("input_777_groups_0"), val = tensor(512)]; - tensor input_777_strides_0 = const()[name = tensor("input_777_strides_0"), val = tensor([1])]; - tensor input_777_pad_0 = const()[name = tensor("input_777_pad_0"), val = tensor([0, 0])]; - tensor input_777_dilations_0 = const()[name = tensor("input_777_dilations_0"), val = tensor([1])]; - tensor const_262_to_fp16 = const()[name = tensor("const_262_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185433984)))]; - tensor const_263_to_fp16 = const()[name = tensor("const_263_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185443264)))]; - tensor input_779_cast_fp16 = conv(bias = const_263_to_fp16, dilations = input_777_dilations_0, groups = input_777_groups_0, pad = input_777_pad_0, pad_type = input_777_pad_type_0, strides = input_777_strides_0, weight = const_262_to_fp16, x = input_775_cast_fp16)[name = tensor("input_779_cast_fp16")]; - tensor input_781_cast_fp16 = silu(x = input_779_cast_fp16)[name = tensor("input_781_cast_fp16")]; - tensor x_329_pad_type_0 = const()[name = tensor("x_329_pad_type_0"), val = tensor("valid")]; - tensor x_329_strides_0 = const()[name = tensor("x_329_strides_0"), val = tensor([1])]; - tensor x_329_pad_0 = const()[name = tensor("x_329_pad_0"), val = tensor([0, 0])]; - tensor x_329_dilations_0 = const()[name = tensor("x_329_dilations_0"), val = tensor([1])]; - tensor x_329_groups_0 = const()[name = tensor("x_329_groups_0"), val = tensor(1)]; - tensor module_layers_14_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185444352)))]; - tensor module_layers_14_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_14_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185968704)))]; - tensor x_329_cast_fp16 = conv(bias = module_layers_14_conv_pointwise_conv2_bias_to_fp16, dilations = x_329_dilations_0, groups = x_329_groups_0, pad = x_329_pad_0, pad_type = x_329_pad_type_0, strides = x_329_strides_0, weight = module_layers_14_conv_pointwise_conv2_weight_to_fp16, x = input_781_cast_fp16)[name = tensor("x_329_cast_fp16")]; - tensor input_783_perm_0 = const()[name = tensor("input_783_perm_0"), val = tensor([0, 2, 1])]; - tensor input_783_cast_fp16 = transpose(perm = input_783_perm_0, x = x_329_cast_fp16)[name = tensor("transpose_100")]; - tensor input_785_cast_fp16 = add(x = input_767_cast_fp16, y = input_783_cast_fp16)[name = tensor("input_785_cast_fp16")]; - tensor input_787_axes_0 = const()[name = tensor("input_787_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185969792)))]; - tensor module_layers_14_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185970880)))]; - tensor input_787_cast_fp16 = layer_norm(axes = input_787_axes_0, beta = module_layers_14_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_feed_forward2_weight_to_fp16, x = input_785_cast_fp16)[name = tensor("input_787_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(185971968)))]; - tensor module_layers_14_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188069184)))]; - tensor linear_134_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear1_bias_to_fp16, weight = module_layers_14_feed_forward2_linear1_weight_to_fp16, x = input_787_cast_fp16)[name = tensor("linear_134_cast_fp16")]; - tensor input_791_cast_fp16 = silu(x = linear_134_cast_fp16)[name = tensor("input_791_cast_fp16")]; - tensor module_layers_14_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(188073344)))]; - tensor module_layers_14_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_14_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190170560)))]; - tensor linear_135_cast_fp16 = linear(bias = module_layers_14_feed_forward2_linear2_bias_to_fp16, weight = module_layers_14_feed_forward2_linear2_weight_to_fp16, x = input_791_cast_fp16)[name = tensor("linear_135_cast_fp16")]; - tensor var_2931_to_fp16 = const()[name = tensor("op_2931_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2932_cast_fp16 = mul(x = linear_135_cast_fp16, y = var_2931_to_fp16)[name = tensor("op_2932_cast_fp16")]; - tensor input_797_cast_fp16 = add(x = input_785_cast_fp16, y = var_2932_cast_fp16)[name = tensor("input_797_cast_fp16")]; - tensor input_799_axes_0 = const()[name = tensor("input_799_axes_0"), val = tensor([-1])]; - tensor module_layers_14_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_14_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190171648)))]; - tensor module_layers_14_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_14_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190172736)))]; - tensor input_799_cast_fp16 = layer_norm(axes = input_799_axes_0, beta = module_layers_14_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_14_norm_out_weight_to_fp16, x = input_797_cast_fp16)[name = tensor("input_799_cast_fp16")]; - tensor input_801_axes_0 = const()[name = tensor("input_801_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190173824)))]; - tensor module_layers_15_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190174912)))]; - tensor input_801_cast_fp16 = layer_norm(axes = input_801_axes_0, beta = module_layers_15_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward1_weight_to_fp16, x = input_799_cast_fp16)[name = tensor("input_801_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(190176000)))]; - tensor module_layers_15_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192273216)))]; - tensor linear_136_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear1_bias_to_fp16, weight = module_layers_15_feed_forward1_linear1_weight_to_fp16, x = input_801_cast_fp16)[name = tensor("linear_136_cast_fp16")]; - tensor input_805_cast_fp16 = silu(x = linear_136_cast_fp16)[name = tensor("input_805_cast_fp16")]; - tensor module_layers_15_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(192277376)))]; - tensor module_layers_15_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194374592)))]; - tensor linear_137_cast_fp16 = linear(bias = module_layers_15_feed_forward1_linear2_bias_to_fp16, weight = module_layers_15_feed_forward1_linear2_weight_to_fp16, x = input_805_cast_fp16)[name = tensor("linear_137_cast_fp16")]; - tensor var_2962_to_fp16 = const()[name = tensor("op_2962_to_fp16"), val = tensor(0x1p-1)]; - tensor var_2963_cast_fp16 = mul(x = linear_137_cast_fp16, y = var_2962_to_fp16)[name = tensor("op_2963_cast_fp16")]; - tensor input_811_cast_fp16 = add(x = input_799_cast_fp16, y = var_2963_cast_fp16)[name = tensor("input_811_cast_fp16")]; - tensor query_31_axes_0 = const()[name = tensor("query_31_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194375680)))]; - tensor module_layers_15_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194376768)))]; - tensor query_31_cast_fp16 = layer_norm(axes = query_31_axes_0, beta = module_layers_15_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_self_att_weight_to_fp16, x = input_811_cast_fp16)[name = tensor("query_31_cast_fp16")]; - tensor module_layers_15_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194377856)))]; - tensor module_layers_15_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194902208)))]; - tensor linear_138_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_q_bias_to_fp16, weight = module_layers_15_self_attn_linear_q_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_138_cast_fp16")]; - tensor var_2980 = const()[name = tensor("op_2980"), val = tensor([1, -1, 8, 64])]; - tensor q_91_cast_fp16 = reshape(shape = var_2980, x = linear_138_cast_fp16)[name = tensor("q_91_cast_fp16")]; - tensor module_layers_15_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(194903296)))]; - tensor module_layers_15_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195427648)))]; - tensor linear_139_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_k_bias_to_fp16, weight = module_layers_15_self_attn_linear_k_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_139_cast_fp16")]; - tensor var_2985 = const()[name = tensor("op_2985"), val = tensor([1, -1, 8, 64])]; - tensor k_61_cast_fp16 = reshape(shape = var_2985, x = linear_139_cast_fp16)[name = tensor("k_61_cast_fp16")]; - tensor module_layers_15_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195428736)))]; - tensor module_layers_15_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195953088)))]; - tensor linear_140_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_v_bias_to_fp16, weight = module_layers_15_self_attn_linear_v_weight_to_fp16, x = query_31_cast_fp16)[name = tensor("linear_140_cast_fp16")]; - tensor var_2990 = const()[name = tensor("op_2990"), val = tensor([1, -1, 8, 64])]; - tensor v_31_cast_fp16 = reshape(shape = var_2990, x = linear_140_cast_fp16)[name = tensor("v_31_cast_fp16")]; - tensor value_33_perm_0 = const()[name = tensor("value_33_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_15_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195954176)))]; - tensor var_3002_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3002_cast_fp16")]; - tensor module_layers_15_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_15_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195955264)))]; - tensor var_3004_cast_fp16 = add(x = q_91_cast_fp16, y = module_layers_15_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3004_cast_fp16")]; - tensor q_with_bias_v_31_perm_0 = const()[name = tensor("q_with_bias_v_31_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_337_transpose_x_0 = const()[name = tensor("x_337_transpose_x_0"), val = tensor(false)]; - tensor x_337_transpose_y_0 = const()[name = tensor("x_337_transpose_y_0"), val = tensor(false)]; - tensor var_3006_to_fp16 = const()[name = tensor("op_3006_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(195956352)))]; - tensor q_with_bias_v_31_cast_fp16 = transpose(perm = q_with_bias_v_31_perm_0, x = var_3004_cast_fp16)[name = tensor("transpose_98")]; - tensor x_337_cast_fp16 = matmul(transpose_x = x_337_transpose_x_0, transpose_y = x_337_transpose_y_0, x = q_with_bias_v_31_cast_fp16, y = var_3006_to_fp16)[name = tensor("x_337_cast_fp16")]; - tensor x_339_pad_0 = const()[name = tensor("x_339_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_339_mode_0 = const()[name = tensor("x_339_mode_0"), val = tensor("constant")]; - tensor const_220_to_fp16 = const()[name = tensor("const_220_to_fp16"), val = tensor(0x0p+0)]; - tensor x_339_cast_fp16 = pad(constant_val = const_220_to_fp16, mode = x_339_mode_0, pad = x_339_pad_0, x = x_337_cast_fp16)[name = tensor("x_339_cast_fp16")]; - tensor var_3014 = const()[name = tensor("op_3014"), val = tensor([1, 8, -1, 188])]; - tensor x_341_cast_fp16 = reshape(shape = var_3014, x = x_339_cast_fp16)[name = tensor("x_341_cast_fp16")]; - tensor var_3018_begin_0 = const()[name = tensor("op_3018_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3018_end_0 = const()[name = tensor("op_3018_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3018_end_mask_0 = const()[name = tensor("op_3018_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3018_cast_fp16 = slice_by_index(begin = var_3018_begin_0, end = var_3018_end_0, end_mask = var_3018_end_mask_0, x = x_341_cast_fp16)[name = tensor("op_3018_cast_fp16")]; - tensor var_3019 = const()[name = tensor("op_3019"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_61_cast_fp16 = reshape(shape = var_3019, x = var_3018_cast_fp16)[name = tensor("matrix_bd_61_cast_fp16")]; - tensor matrix_ac_31_transpose_x_0 = const()[name = tensor("matrix_ac_31_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_31_transpose_y_0 = const()[name = tensor("matrix_ac_31_transpose_y_0"), val = tensor(false)]; - tensor transpose_81_perm_0 = const()[name = tensor("transpose_81_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_82_perm_0 = const()[name = tensor("transpose_82_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_82 = transpose(perm = transpose_82_perm_0, x = k_61_cast_fp16)[name = tensor("transpose_96")]; - tensor transpose_81 = transpose(perm = transpose_81_perm_0, x = var_3002_cast_fp16)[name = tensor("transpose_97")]; - tensor matrix_ac_31_cast_fp16 = matmul(transpose_x = matrix_ac_31_transpose_x_0, transpose_y = matrix_ac_31_transpose_y_0, x = transpose_81, y = transpose_82)[name = tensor("matrix_ac_31_cast_fp16")]; - tensor matrix_bd_63_begin_0 = const()[name = tensor("matrix_bd_63_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_63_end_0 = const()[name = tensor("matrix_bd_63_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_63_end_mask_0 = const()[name = tensor("matrix_bd_63_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_63_cast_fp16 = slice_by_index(begin = matrix_bd_63_begin_0, end = matrix_bd_63_end_0, end_mask = matrix_bd_63_end_mask_0, x = matrix_bd_61_cast_fp16)[name = tensor("matrix_bd_63_cast_fp16")]; - tensor var_3028_cast_fp16 = add(x = matrix_ac_31_cast_fp16, y = matrix_bd_63_cast_fp16)[name = tensor("op_3028_cast_fp16")]; - tensor _inversed_scores_61_y_0_to_fp16 = const()[name = tensor("_inversed_scores_61_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_61_cast_fp16 = mul(x = var_3028_cast_fp16, y = _inversed_scores_61_y_0_to_fp16)[name = tensor("_inversed_scores_61_cast_fp16")]; - tensor scores_63_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_61_cast_fp16, cond = mask_11)[name = tensor("scores_63_cast_fp16")]; - tensor var_3034_cast_fp16 = softmax(axis = var_23, x = scores_63_cast_fp16)[name = tensor("op_3034_cast_fp16")]; - tensor input_813_cast_fp16 = select(a = var_11_to_fp16, b = var_3034_cast_fp16, cond = mask_11)[name = tensor("input_813_cast_fp16")]; - tensor x_343_transpose_x_0 = const()[name = tensor("x_343_transpose_x_0"), val = tensor(false)]; - tensor x_343_transpose_y_0 = const()[name = tensor("x_343_transpose_y_0"), val = tensor(false)]; - tensor value_33_cast_fp16 = transpose(perm = value_33_perm_0, x = v_31_cast_fp16)[name = tensor("transpose_99")]; - tensor x_343_cast_fp16 = matmul(transpose_x = x_343_transpose_x_0, transpose_y = x_343_transpose_y_0, x = input_813_cast_fp16, y = value_33_cast_fp16)[name = tensor("x_343_cast_fp16")]; - tensor var_3038_perm_0 = const()[name = tensor("op_3038_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3039 = const()[name = tensor("op_3039"), val = tensor([1, -1, 512])]; - tensor var_3038_cast_fp16 = transpose(perm = var_3038_perm_0, x = x_343_cast_fp16)[name = tensor("transpose_95")]; - tensor input_815_cast_fp16 = reshape(shape = var_3039, x = var_3038_cast_fp16)[name = tensor("input_815_cast_fp16")]; - tensor module_layers_15_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196340416)))]; - tensor module_layers_15_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_15_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196864768)))]; - tensor linear_142_cast_fp16 = linear(bias = module_layers_15_self_attn_linear_out_bias_to_fp16, weight = module_layers_15_self_attn_linear_out_weight_to_fp16, x = input_815_cast_fp16)[name = tensor("linear_142_cast_fp16")]; - tensor input_819_cast_fp16 = add(x = input_811_cast_fp16, y = linear_142_cast_fp16)[name = tensor("input_819_cast_fp16")]; - tensor x_347_axes_0 = const()[name = tensor("x_347_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196865856)))]; - tensor module_layers_15_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196866944)))]; - tensor x_347_cast_fp16 = layer_norm(axes = x_347_axes_0, beta = module_layers_15_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_conv_weight_to_fp16, x = input_819_cast_fp16)[name = tensor("x_347_cast_fp16")]; - tensor input_821_perm_0 = const()[name = tensor("input_821_perm_0"), val = tensor([0, 2, 1])]; - tensor input_823_pad_type_0 = const()[name = tensor("input_823_pad_type_0"), val = tensor("valid")]; - tensor input_823_strides_0 = const()[name = tensor("input_823_strides_0"), val = tensor([1])]; - tensor input_823_pad_0 = const()[name = tensor("input_823_pad_0"), val = tensor([0, 0])]; - tensor input_823_dilations_0 = const()[name = tensor("input_823_dilations_0"), val = tensor([1])]; - tensor input_823_groups_0 = const()[name = tensor("input_823_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(196868032)))]; - tensor module_layers_15_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197916672)))]; - tensor input_821_cast_fp16 = transpose(perm = input_821_perm_0, x = x_347_cast_fp16)[name = tensor("transpose_94")]; - tensor input_823_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv1_bias_to_fp16, dilations = input_823_dilations_0, groups = input_823_groups_0, pad = input_823_pad_0, pad_type = input_823_pad_type_0, strides = input_823_strides_0, weight = module_layers_15_conv_pointwise_conv1_weight_to_fp16, x = input_821_cast_fp16)[name = tensor("input_823_cast_fp16")]; - tensor x_349_split_num_splits_0 = const()[name = tensor("x_349_split_num_splits_0"), val = tensor(2)]; - tensor x_349_split_axis_0 = const()[name = tensor("x_349_split_axis_0"), val = tensor(1)]; - tensor x_349_split_cast_fp16_0, tensor x_349_split_cast_fp16_1 = split(axis = x_349_split_axis_0, num_splits = x_349_split_num_splits_0, x = input_823_cast_fp16)[name = tensor("x_349_split_cast_fp16")]; - tensor x_349_split_1_sigmoid_cast_fp16 = sigmoid(x = x_349_split_cast_fp16_1)[name = tensor("x_349_split_1_sigmoid_cast_fp16")]; - tensor x_349_cast_fp16 = mul(x = x_349_split_cast_fp16_0, y = x_349_split_1_sigmoid_cast_fp16)[name = tensor("x_349_cast_fp16")]; - tensor input_825_cast_fp16 = select(a = var_11_to_fp16, b = x_349_cast_fp16, cond = var_453)[name = tensor("input_825_cast_fp16")]; - tensor input_827_pad_0 = const()[name = tensor("input_827_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_827_mode_0 = const()[name = tensor("input_827_mode_0"), val = tensor("constant")]; - tensor const_223_to_fp16 = const()[name = tensor("const_223_to_fp16"), val = tensor(0x0p+0)]; - tensor input_827_cast_fp16 = pad(constant_val = const_223_to_fp16, mode = input_827_mode_0, pad = input_827_pad_0, x = input_825_cast_fp16)[name = tensor("input_827_cast_fp16")]; - tensor input_829_pad_type_0 = const()[name = tensor("input_829_pad_type_0"), val = tensor("valid")]; - tensor input_829_groups_0 = const()[name = tensor("input_829_groups_0"), val = tensor(512)]; - tensor input_829_strides_0 = const()[name = tensor("input_829_strides_0"), val = tensor([1])]; - tensor input_829_pad_0 = const()[name = tensor("input_829_pad_0"), val = tensor([0, 0])]; - tensor input_829_dilations_0 = const()[name = tensor("input_829_dilations_0"), val = tensor([1])]; - tensor const_264_to_fp16 = const()[name = tensor("const_264_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197918784)))]; - tensor const_265_to_fp16 = const()[name = tensor("const_265_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197928064)))]; - tensor input_831_cast_fp16 = conv(bias = const_265_to_fp16, dilations = input_829_dilations_0, groups = input_829_groups_0, pad = input_829_pad_0, pad_type = input_829_pad_type_0, strides = input_829_strides_0, weight = const_264_to_fp16, x = input_827_cast_fp16)[name = tensor("input_831_cast_fp16")]; - tensor input_833_cast_fp16 = silu(x = input_831_cast_fp16)[name = tensor("input_833_cast_fp16")]; - tensor x_351_pad_type_0 = const()[name = tensor("x_351_pad_type_0"), val = tensor("valid")]; - tensor x_351_strides_0 = const()[name = tensor("x_351_strides_0"), val = tensor([1])]; - tensor x_351_pad_0 = const()[name = tensor("x_351_pad_0"), val = tensor([0, 0])]; - tensor x_351_dilations_0 = const()[name = tensor("x_351_dilations_0"), val = tensor([1])]; - tensor x_351_groups_0 = const()[name = tensor("x_351_groups_0"), val = tensor(1)]; - tensor module_layers_15_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(197929152)))]; - tensor module_layers_15_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_15_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198453504)))]; - tensor x_351_cast_fp16 = conv(bias = module_layers_15_conv_pointwise_conv2_bias_to_fp16, dilations = x_351_dilations_0, groups = x_351_groups_0, pad = x_351_pad_0, pad_type = x_351_pad_type_0, strides = x_351_strides_0, weight = module_layers_15_conv_pointwise_conv2_weight_to_fp16, x = input_833_cast_fp16)[name = tensor("x_351_cast_fp16")]; - tensor input_835_perm_0 = const()[name = tensor("input_835_perm_0"), val = tensor([0, 2, 1])]; - tensor input_835_cast_fp16 = transpose(perm = input_835_perm_0, x = x_351_cast_fp16)[name = tensor("transpose_93")]; - tensor input_837_cast_fp16 = add(x = input_819_cast_fp16, y = input_835_cast_fp16)[name = tensor("input_837_cast_fp16")]; - tensor input_839_axes_0 = const()[name = tensor("input_839_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198454592)))]; - tensor module_layers_15_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198455680)))]; - tensor input_839_cast_fp16 = layer_norm(axes = input_839_axes_0, beta = module_layers_15_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_feed_forward2_weight_to_fp16, x = input_837_cast_fp16)[name = tensor("input_839_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(198456768)))]; - tensor module_layers_15_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200553984)))]; - tensor linear_143_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear1_bias_to_fp16, weight = module_layers_15_feed_forward2_linear1_weight_to_fp16, x = input_839_cast_fp16)[name = tensor("linear_143_cast_fp16")]; - tensor input_843_cast_fp16 = silu(x = linear_143_cast_fp16)[name = tensor("input_843_cast_fp16")]; - tensor module_layers_15_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(200558144)))]; - tensor module_layers_15_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_15_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202655360)))]; - tensor linear_144_cast_fp16 = linear(bias = module_layers_15_feed_forward2_linear2_bias_to_fp16, weight = module_layers_15_feed_forward2_linear2_weight_to_fp16, x = input_843_cast_fp16)[name = tensor("linear_144_cast_fp16")]; - tensor var_3105_to_fp16 = const()[name = tensor("op_3105_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3106_cast_fp16 = mul(x = linear_144_cast_fp16, y = var_3105_to_fp16)[name = tensor("op_3106_cast_fp16")]; - tensor input_849_cast_fp16 = add(x = input_837_cast_fp16, y = var_3106_cast_fp16)[name = tensor("input_849_cast_fp16")]; - tensor input_851_axes_0 = const()[name = tensor("input_851_axes_0"), val = tensor([-1])]; - tensor module_layers_15_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_15_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202656448)))]; - tensor module_layers_15_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_15_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202657536)))]; - tensor input_851_cast_fp16 = layer_norm(axes = input_851_axes_0, beta = module_layers_15_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_15_norm_out_weight_to_fp16, x = input_849_cast_fp16)[name = tensor("input_851_cast_fp16")]; - tensor input_853_axes_0 = const()[name = tensor("input_853_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward1_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202658624)))]; - tensor module_layers_16_norm_feed_forward1_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202659712)))]; - tensor input_853_cast_fp16 = layer_norm(axes = input_853_axes_0, beta = module_layers_16_norm_feed_forward1_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward1_weight_to_fp16, x = input_851_cast_fp16)[name = tensor("input_853_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(202660800)))]; - tensor module_layers_16_feed_forward1_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204758016)))]; - tensor linear_145_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear1_bias_to_fp16, weight = module_layers_16_feed_forward1_linear1_weight_to_fp16, x = input_853_cast_fp16)[name = tensor("linear_145_cast_fp16")]; - tensor input_857_cast_fp16 = silu(x = linear_145_cast_fp16)[name = tensor("input_857_cast_fp16")]; - tensor module_layers_16_feed_forward1_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(204762176)))]; - tensor module_layers_16_feed_forward1_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward1_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206859392)))]; - tensor linear_146_cast_fp16 = linear(bias = module_layers_16_feed_forward1_linear2_bias_to_fp16, weight = module_layers_16_feed_forward1_linear2_weight_to_fp16, x = input_857_cast_fp16)[name = tensor("linear_146_cast_fp16")]; - tensor var_3136_to_fp16 = const()[name = tensor("op_3136_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3137_cast_fp16 = mul(x = linear_146_cast_fp16, y = var_3136_to_fp16)[name = tensor("op_3137_cast_fp16")]; - tensor input_863_cast_fp16 = add(x = input_851_cast_fp16, y = var_3137_cast_fp16)[name = tensor("input_863_cast_fp16")]; - tensor query_axes_0 = const()[name = tensor("query_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_self_att_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206860480)))]; - tensor module_layers_16_norm_self_att_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_self_att_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206861568)))]; - tensor query_cast_fp16 = layer_norm(axes = query_axes_0, beta = module_layers_16_norm_self_att_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_self_att_weight_to_fp16, x = input_863_cast_fp16)[name = tensor("query_cast_fp16")]; - tensor module_layers_16_self_attn_linear_q_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(206862656)))]; - tensor module_layers_16_self_attn_linear_q_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_q_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207387008)))]; - tensor linear_147_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_q_bias_to_fp16, weight = module_layers_16_self_attn_linear_q_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_147_cast_fp16")]; - tensor var_3154 = const()[name = tensor("op_3154"), val = tensor([1, -1, 8, 64])]; - tensor q_97_cast_fp16 = reshape(shape = var_3154, x = linear_147_cast_fp16)[name = tensor("q_97_cast_fp16")]; - tensor module_layers_16_self_attn_linear_k_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207388096)))]; - tensor module_layers_16_self_attn_linear_k_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_k_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207912448)))]; - tensor linear_148_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_k_bias_to_fp16, weight = module_layers_16_self_attn_linear_k_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_148_cast_fp16")]; - tensor var_3159 = const()[name = tensor("op_3159"), val = tensor([1, -1, 8, 64])]; - tensor k_65_cast_fp16 = reshape(shape = var_3159, x = linear_148_cast_fp16)[name = tensor("k_65_cast_fp16")]; - tensor module_layers_16_self_attn_linear_v_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(207913536)))]; - tensor module_layers_16_self_attn_linear_v_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_v_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208437888)))]; - tensor linear_149_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_v_bias_to_fp16, weight = module_layers_16_self_attn_linear_v_weight_to_fp16, x = query_cast_fp16)[name = tensor("linear_149_cast_fp16")]; - tensor var_3164 = const()[name = tensor("op_3164"), val = tensor([1, -1, 8, 64])]; - tensor v_cast_fp16 = reshape(shape = var_3164, x = linear_149_cast_fp16)[name = tensor("v_cast_fp16")]; - tensor value_perm_0 = const()[name = tensor("value_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor module_layers_16_self_attn_pos_bias_u_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_u_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208438976)))]; - tensor var_3176_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_u_to_fp16)[name = tensor("op_3176_cast_fp16")]; - tensor module_layers_16_self_attn_pos_bias_v_to_fp16 = const()[name = tensor("module_layers_16_self_attn_pos_bias_v_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208440064)))]; - tensor var_3178_cast_fp16 = add(x = q_97_cast_fp16, y = module_layers_16_self_attn_pos_bias_v_to_fp16)[name = tensor("op_3178_cast_fp16")]; - tensor q_with_bias_v_perm_0 = const()[name = tensor("q_with_bias_v_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor x_359_transpose_x_0 = const()[name = tensor("x_359_transpose_x_0"), val = tensor(false)]; - tensor x_359_transpose_y_0 = const()[name = tensor("x_359_transpose_y_0"), val = tensor(false)]; - tensor var_3180_to_fp16 = const()[name = tensor("op_3180_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208441152)))]; - tensor q_with_bias_v_cast_fp16 = transpose(perm = q_with_bias_v_perm_0, x = var_3178_cast_fp16)[name = tensor("transpose_91")]; - tensor x_359_cast_fp16 = matmul(transpose_x = x_359_transpose_x_0, transpose_y = x_359_transpose_y_0, x = q_with_bias_v_cast_fp16, y = var_3180_to_fp16)[name = tensor("x_359_cast_fp16")]; - tensor x_361_pad_0 = const()[name = tensor("x_361_pad_0"), val = tensor([0, 0, 0, 0, 0, 0, 1, 0])]; - tensor x_361_mode_0 = const()[name = tensor("x_361_mode_0"), val = tensor("constant")]; - tensor const_230_to_fp16 = const()[name = tensor("const_230_to_fp16"), val = tensor(0x0p+0)]; - tensor x_361_cast_fp16 = pad(constant_val = const_230_to_fp16, mode = x_361_mode_0, pad = x_361_pad_0, x = x_359_cast_fp16)[name = tensor("x_361_cast_fp16")]; - tensor var_3188 = const()[name = tensor("op_3188"), val = tensor([1, 8, -1, 188])]; - tensor x_363_cast_fp16 = reshape(shape = var_3188, x = x_361_cast_fp16)[name = tensor("x_363_cast_fp16")]; - tensor var_3192_begin_0 = const()[name = tensor("op_3192_begin_0"), val = tensor([0, 0, 1, 0])]; - tensor var_3192_end_0 = const()[name = tensor("op_3192_end_0"), val = tensor([1, 8, 376, 188])]; - tensor var_3192_end_mask_0 = const()[name = tensor("op_3192_end_mask_0"), val = tensor([true, true, true, true])]; - tensor var_3192_cast_fp16 = slice_by_index(begin = var_3192_begin_0, end = var_3192_end_0, end_mask = var_3192_end_mask_0, x = x_363_cast_fp16)[name = tensor("op_3192_cast_fp16")]; - tensor var_3193 = const()[name = tensor("op_3193"), val = tensor([1, 8, 188, 375])]; - tensor matrix_bd_65_cast_fp16 = reshape(shape = var_3193, x = var_3192_cast_fp16)[name = tensor("matrix_bd_65_cast_fp16")]; - tensor matrix_ac_transpose_x_0 = const()[name = tensor("matrix_ac_transpose_x_0"), val = tensor(false)]; - tensor matrix_ac_transpose_y_0 = const()[name = tensor("matrix_ac_transpose_y_0"), val = tensor(false)]; - tensor transpose_83_perm_0 = const()[name = tensor("transpose_83_perm_0"), val = tensor([0, 2, -3, -1])]; - tensor transpose_84_perm_0 = const()[name = tensor("transpose_84_perm_0"), val = tensor([0, 2, -1, -3])]; - tensor transpose_84 = transpose(perm = transpose_84_perm_0, x = k_65_cast_fp16)[name = tensor("transpose_89")]; - tensor transpose_83 = transpose(perm = transpose_83_perm_0, x = var_3176_cast_fp16)[name = tensor("transpose_90")]; - tensor matrix_ac_cast_fp16 = matmul(transpose_x = matrix_ac_transpose_x_0, transpose_y = matrix_ac_transpose_y_0, x = transpose_83, y = transpose_84)[name = tensor("matrix_ac_cast_fp16")]; - tensor matrix_bd_begin_0 = const()[name = tensor("matrix_bd_begin_0"), val = tensor([0, 0, 0, 0])]; - tensor matrix_bd_end_0 = const()[name = tensor("matrix_bd_end_0"), val = tensor([1, 8, 188, 188])]; - tensor matrix_bd_end_mask_0 = const()[name = tensor("matrix_bd_end_mask_0"), val = tensor([true, true, true, false])]; - tensor matrix_bd_cast_fp16 = slice_by_index(begin = matrix_bd_begin_0, end = matrix_bd_end_0, end_mask = matrix_bd_end_mask_0, x = matrix_bd_65_cast_fp16)[name = tensor("matrix_bd_cast_fp16")]; - tensor var_3202_cast_fp16 = add(x = matrix_ac_cast_fp16, y = matrix_bd_cast_fp16)[name = tensor("op_3202_cast_fp16")]; - tensor _inversed_scores_65_y_0_to_fp16 = const()[name = tensor("_inversed_scores_65_y_0_to_fp16"), val = tensor(0x1p-3)]; - tensor _inversed_scores_65_cast_fp16 = mul(x = var_3202_cast_fp16, y = _inversed_scores_65_y_0_to_fp16)[name = tensor("_inversed_scores_65_cast_fp16")]; - tensor scores_cast_fp16 = select(a = var_12_to_fp16, b = _inversed_scores_65_cast_fp16, cond = mask_11)[name = tensor("scores_cast_fp16")]; - tensor var_3208_cast_fp16 = softmax(axis = var_23, x = scores_cast_fp16)[name = tensor("op_3208_cast_fp16")]; - tensor input_865_cast_fp16 = select(a = var_11_to_fp16, b = var_3208_cast_fp16, cond = mask_11)[name = tensor("input_865_cast_fp16")]; - tensor x_365_transpose_x_0 = const()[name = tensor("x_365_transpose_x_0"), val = tensor(false)]; - tensor x_365_transpose_y_0 = const()[name = tensor("x_365_transpose_y_0"), val = tensor(false)]; - tensor value_cast_fp16 = transpose(perm = value_perm_0, x = v_cast_fp16)[name = tensor("transpose_92")]; - tensor x_365_cast_fp16 = matmul(transpose_x = x_365_transpose_x_0, transpose_y = x_365_transpose_y_0, x = input_865_cast_fp16, y = value_cast_fp16)[name = tensor("x_365_cast_fp16")]; - tensor var_3212_perm_0 = const()[name = tensor("op_3212_perm_0"), val = tensor([0, 2, 1, 3])]; - tensor var_3213 = const()[name = tensor("op_3213"), val = tensor([1, -1, 512])]; - tensor var_3212_cast_fp16 = transpose(perm = var_3212_perm_0, x = x_365_cast_fp16)[name = tensor("transpose_88")]; - tensor input_867_cast_fp16 = reshape(shape = var_3213, x = var_3212_cast_fp16)[name = tensor("input_867_cast_fp16")]; - tensor module_layers_16_self_attn_linear_out_weight_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(208825216)))]; - tensor module_layers_16_self_attn_linear_out_bias_to_fp16 = const()[name = tensor("module_layers_16_self_attn_linear_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209349568)))]; - tensor linear_151_cast_fp16 = linear(bias = module_layers_16_self_attn_linear_out_bias_to_fp16, weight = module_layers_16_self_attn_linear_out_weight_to_fp16, x = input_867_cast_fp16)[name = tensor("linear_151_cast_fp16")]; - tensor input_871_cast_fp16 = add(x = input_863_cast_fp16, y = linear_151_cast_fp16)[name = tensor("input_871_cast_fp16")]; - tensor x_369_axes_0 = const()[name = tensor("x_369_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_conv_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209350656)))]; - tensor module_layers_16_norm_conv_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_conv_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209351744)))]; - tensor x_369_cast_fp16 = layer_norm(axes = x_369_axes_0, beta = module_layers_16_norm_conv_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_conv_weight_to_fp16, x = input_871_cast_fp16)[name = tensor("x_369_cast_fp16")]; - tensor input_873_perm_0 = const()[name = tensor("input_873_perm_0"), val = tensor([0, 2, 1])]; - tensor input_875_pad_type_0 = const()[name = tensor("input_875_pad_type_0"), val = tensor("valid")]; - tensor input_875_strides_0 = const()[name = tensor("input_875_strides_0"), val = tensor([1])]; - tensor input_875_pad_0 = const()[name = tensor("input_875_pad_0"), val = tensor([0, 0])]; - tensor input_875_dilations_0 = const()[name = tensor("input_875_dilations_0"), val = tensor([1])]; - tensor input_875_groups_0 = const()[name = tensor("input_875_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv1_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(209352832)))]; - tensor module_layers_16_conv_pointwise_conv1_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210401472)))]; - tensor input_873_cast_fp16 = transpose(perm = input_873_perm_0, x = x_369_cast_fp16)[name = tensor("transpose_87")]; - tensor input_875_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv1_bias_to_fp16, dilations = input_875_dilations_0, groups = input_875_groups_0, pad = input_875_pad_0, pad_type = input_875_pad_type_0, strides = input_875_strides_0, weight = module_layers_16_conv_pointwise_conv1_weight_to_fp16, x = input_873_cast_fp16)[name = tensor("input_875_cast_fp16")]; - tensor x_371_split_num_splits_0 = const()[name = tensor("x_371_split_num_splits_0"), val = tensor(2)]; - tensor x_371_split_axis_0 = const()[name = tensor("x_371_split_axis_0"), val = tensor(1)]; - tensor x_371_split_cast_fp16_0, tensor x_371_split_cast_fp16_1 = split(axis = x_371_split_axis_0, num_splits = x_371_split_num_splits_0, x = input_875_cast_fp16)[name = tensor("x_371_split_cast_fp16")]; - tensor x_371_split_1_sigmoid_cast_fp16 = sigmoid(x = x_371_split_cast_fp16_1)[name = tensor("x_371_split_1_sigmoid_cast_fp16")]; - tensor x_371_cast_fp16 = mul(x = x_371_split_cast_fp16_0, y = x_371_split_1_sigmoid_cast_fp16)[name = tensor("x_371_cast_fp16")]; - tensor input_877_cast_fp16 = select(a = var_11_to_fp16, b = x_371_cast_fp16, cond = var_453)[name = tensor("input_877_cast_fp16")]; - tensor input_879_pad_0 = const()[name = tensor("input_879_pad_0"), val = tensor([0, 0, 0, 0, 4, 4])]; - tensor input_879_mode_0 = const()[name = tensor("input_879_mode_0"), val = tensor("constant")]; - tensor const_233_to_fp16 = const()[name = tensor("const_233_to_fp16"), val = tensor(0x0p+0)]; - tensor input_879_cast_fp16 = pad(constant_val = const_233_to_fp16, mode = input_879_mode_0, pad = input_879_pad_0, x = input_877_cast_fp16)[name = tensor("input_879_cast_fp16")]; - tensor input_881_pad_type_0 = const()[name = tensor("input_881_pad_type_0"), val = tensor("valid")]; - tensor input_881_groups_0 = const()[name = tensor("input_881_groups_0"), val = tensor(512)]; - tensor input_881_strides_0 = const()[name = tensor("input_881_strides_0"), val = tensor([1])]; - tensor input_881_pad_0 = const()[name = tensor("input_881_pad_0"), val = tensor([0, 0])]; - tensor input_881_dilations_0 = const()[name = tensor("input_881_dilations_0"), val = tensor([1])]; - tensor const_266_to_fp16 = const()[name = tensor("const_266_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210403584)))]; - tensor const_267_to_fp16 = const()[name = tensor("const_267_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210412864)))]; - tensor input_883_cast_fp16 = conv(bias = const_267_to_fp16, dilations = input_881_dilations_0, groups = input_881_groups_0, pad = input_881_pad_0, pad_type = input_881_pad_type_0, strides = input_881_strides_0, weight = const_266_to_fp16, x = input_879_cast_fp16)[name = tensor("input_883_cast_fp16")]; - tensor input_885_cast_fp16 = silu(x = input_883_cast_fp16)[name = tensor("input_885_cast_fp16")]; - tensor x_373_pad_type_0 = const()[name = tensor("x_373_pad_type_0"), val = tensor("valid")]; - tensor x_373_strides_0 = const()[name = tensor("x_373_strides_0"), val = tensor([1])]; - tensor x_373_pad_0 = const()[name = tensor("x_373_pad_0"), val = tensor([0, 0])]; - tensor x_373_dilations_0 = const()[name = tensor("x_373_dilations_0"), val = tensor([1])]; - tensor x_373_groups_0 = const()[name = tensor("x_373_groups_0"), val = tensor(1)]; - tensor module_layers_16_conv_pointwise_conv2_weight_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210413952)))]; - tensor module_layers_16_conv_pointwise_conv2_bias_to_fp16 = const()[name = tensor("module_layers_16_conv_pointwise_conv2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210938304)))]; - tensor x_373_cast_fp16 = conv(bias = module_layers_16_conv_pointwise_conv2_bias_to_fp16, dilations = x_373_dilations_0, groups = x_373_groups_0, pad = x_373_pad_0, pad_type = x_373_pad_type_0, strides = x_373_strides_0, weight = module_layers_16_conv_pointwise_conv2_weight_to_fp16, x = input_885_cast_fp16)[name = tensor("x_373_cast_fp16")]; - tensor input_887_perm_0 = const()[name = tensor("input_887_perm_0"), val = tensor([0, 2, 1])]; - tensor input_887_cast_fp16 = transpose(perm = input_887_perm_0, x = x_373_cast_fp16)[name = tensor("transpose_86")]; - tensor input_889_cast_fp16 = add(x = input_871_cast_fp16, y = input_887_cast_fp16)[name = tensor("input_889_cast_fp16")]; - tensor input_891_axes_0 = const()[name = tensor("input_891_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_feed_forward2_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210939392)))]; - tensor module_layers_16_norm_feed_forward2_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_feed_forward2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210940480)))]; - tensor input_891_cast_fp16 = layer_norm(axes = input_891_axes_0, beta = module_layers_16_norm_feed_forward2_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_feed_forward2_weight_to_fp16, x = input_889_cast_fp16)[name = tensor("input_891_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear1_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(210941568)))]; - tensor module_layers_16_feed_forward2_linear1_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear1_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213038784)))]; - tensor linear_152_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear1_bias_to_fp16, weight = module_layers_16_feed_forward2_linear1_weight_to_fp16, x = input_891_cast_fp16)[name = tensor("linear_152_cast_fp16")]; - tensor input_895_cast_fp16 = silu(x = linear_152_cast_fp16)[name = tensor("input_895_cast_fp16")]; - tensor module_layers_16_feed_forward2_linear2_weight_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(213042944)))]; - tensor module_layers_16_feed_forward2_linear2_bias_to_fp16 = const()[name = tensor("module_layers_16_feed_forward2_linear2_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215140160)))]; - tensor linear_153_cast_fp16 = linear(bias = module_layers_16_feed_forward2_linear2_bias_to_fp16, weight = module_layers_16_feed_forward2_linear2_weight_to_fp16, x = input_895_cast_fp16)[name = tensor("linear_153_cast_fp16")]; - tensor var_3279_to_fp16 = const()[name = tensor("op_3279_to_fp16"), val = tensor(0x1p-1)]; - tensor var_3280_cast_fp16 = mul(x = linear_153_cast_fp16, y = var_3279_to_fp16)[name = tensor("op_3280_cast_fp16")]; - tensor input_cast_fp16 = add(x = input_889_cast_fp16, y = var_3280_cast_fp16)[name = tensor("input_cast_fp16")]; - tensor audio_signal_axes_0 = const()[name = tensor("audio_signal_axes_0"), val = tensor([-1])]; - tensor module_layers_16_norm_out_weight_to_fp16 = const()[name = tensor("module_layers_16_norm_out_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215141248)))]; - tensor module_layers_16_norm_out_bias_to_fp16 = const()[name = tensor("module_layers_16_norm_out_bias_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(215142336)))]; - tensor audio_signal_cast_fp16 = layer_norm(axes = audio_signal_axes_0, beta = module_layers_16_norm_out_bias_to_fp16, epsilon = var_9_to_fp16, gamma = module_layers_16_norm_out_weight_to_fp16, x = input_cast_fp16)[name = tensor("audio_signal_cast_fp16")]; - tensor obj_1_perm_0 = const()[name = tensor("obj_1_perm_0"), val = tensor([0, 2, 1])]; - tensor obj_1_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_1_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; - tensor obj_1_cast_fp16 = transpose(perm = obj_1_perm_0, x = audio_signal_cast_fp16)[name = tensor("transpose_85")]; - tensor encoder_output = cast(dtype = obj_1_cast_fp16_to_fp32_dtype_0, x = obj_1_cast_fp16)[name = tensor("cast_172")]; - } -> (encoder_output, encoder_length); -} \ No newline at end of file