johnkillington's picture
Automated CoreML conversion
91d62eb verified
program(1.0)
[buildInfo = dict<tensor<string, []>, tensor<string, []>>({{"coremlc-component-MIL", "3510.2.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.8.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})]
{
func main<ios16>(tensor<fp32, [1, 77]> input_ids) {
tensor<string, []> cast_1_dtype_0 = const()[name = tensor<string, []>("cast_1_dtype_0"), val = tensor<string, []>("int32")];
tensor<int32, []> inputs_embeds_axis_0 = const()[name = tensor<string, []>("inputs_embeds_axis_0"), val = tensor<int32, []>(0)];
tensor<int32, []> inputs_embeds_batch_dims_0 = const()[name = tensor<string, []>("inputs_embeds_batch_dims_0"), val = tensor<int32, []>(0)];
tensor<fp16, [49408, 768]> text_encoder_text_model_embeddings_token_embedding_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_embeddings_token_embedding_weight_to_fp16"), val = tensor<fp16, [49408, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(64)))];
tensor<int32, [1, 77]> cast_1 = cast(dtype = cast_1_dtype_0, x = input_ids)[name = tensor<string, []>("cast_56")];
tensor<fp16, [1, 77, 768]> inputs_embeds_cast_fp16 = gather(axis = inputs_embeds_axis_0, batch_dims = inputs_embeds_batch_dims_0, indices = cast_1, x = text_encoder_text_model_embeddings_token_embedding_weight_to_fp16)[name = tensor<string, []>("inputs_embeds_cast_fp16")];
tensor<fp16, [1, 77, 768]> position_embeddings_to_fp16 = const()[name = tensor<string, []>("position_embeddings_to_fp16"), val = tensor<fp16, [1, 77, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(75890816)))];
tensor<fp16, [1, 77, 768]> input_3_cast_fp16 = add(x = inputs_embeds_cast_fp16, y = position_embeddings_to_fp16)[name = tensor<string, []>("input_3_cast_fp16")];
tensor<int32, [1]> hidden_states_1_axes_0 = const()[name = tensor<string, []>("hidden_states_1_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(76009152)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(76010752)))];
tensor<fp16, []> var_15_to_fp16 = const()[name = tensor<string, []>("op_15_to_fp16"), val = tensor<fp16, []>(0x1.5p-17)];
tensor<fp16, [1, 77, 768]> hidden_states_1_cast_fp16 = layer_norm(axes = hidden_states_1_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm1_weight_to_fp16, x = input_3_cast_fp16)[name = tensor<string, []>("hidden_states_1_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(76012352)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(77192064)))];
tensor<fp16, [1, 77, 768]> linear_0_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_q_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor<string, []>("linear_0_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(77193664)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(78373376)))];
tensor<fp16, [1, 77, 768]> linear_1_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_k_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor<string, []>("linear_1_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(78374976)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(79554688)))];
tensor<fp16, [1, 77, 768]> linear_2_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_v_proj_weight_to_fp16, x = hidden_states_1_cast_fp16)[name = tensor<string, []>("linear_2_cast_fp16")];
tensor<int32, [4]> var_113 = const()[name = tensor<string, []>("op_113"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_114_cast_fp16 = reshape(shape = var_113, x = linear_0_cast_fp16)[name = tensor<string, []>("op_114_cast_fp16")];
tensor<int32, [4]> var_116 = const()[name = tensor<string, []>("op_116"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_117_cast_fp16 = reshape(shape = var_116, x = linear_1_cast_fp16)[name = tensor<string, []>("op_117_cast_fp16")];
tensor<int32, [4]> var_119 = const()[name = tensor<string, []>("op_119"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_120_cast_fp16 = reshape(shape = var_119, x = linear_2_cast_fp16)[name = tensor<string, []>("op_120_cast_fp16")];
tensor<int32, [4]> value_states_3_perm_0 = const()[name = tensor<string, []>("value_states_3_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, []> var_17_to_fp16 = const()[name = tensor<string, []>("op_17_to_fp16"), val = tensor<fp16, []>(0x1p-3)];
tensor<fp16, [1, 77, 12, 64]> mul_0_cast_fp16 = mul(x = var_114_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_0_cast_fp16")];
tensor<bool, []> matmul_0_transpose_y_0 = const()[name = tensor<string, []>("matmul_0_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_0_transpose_x_0 = const()[name = tensor<string, []>("matmul_0_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_49_perm_0 = const()[name = tensor<string, []>("transpose_49_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_50_perm_0 = const()[name = tensor<string, []>("transpose_50_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_50 = transpose(perm = transpose_50_perm_0, x = var_117_cast_fp16)[name = tensor<string, []>("transpose_118")];
tensor<fp16, [1, 12, 77, 64]> transpose_49 = transpose(perm = transpose_49_perm_0, x = mul_0_cast_fp16)[name = tensor<string, []>("transpose_119")];
tensor<fp16, [1, 12, 77, 77]> matmul_0_cast_fp16 = matmul(transpose_x = matmul_0_transpose_x_0, transpose_y = matmul_0_transpose_y_0, x = transpose_49, y = transpose_50)[name = tensor<string, []>("matmul_0_cast_fp16")];
tensor<fp16, [1, 1, 77, 77]> var_57_to_fp16 = const()[name = tensor<string, []>("op_57_to_fp16"), val = tensor<fp16, [1, 1, 77, 77]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(79556288)))];
tensor<fp16, [1, 12, 77, 77]> add_0_cast_fp16 = add(x = matmul_0_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_0_cast_fp16")];
tensor<int32, []> softmax_0_axis_0 = const()[name = tensor<string, []>("softmax_0_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_0_cast_fp16 = softmax(axis = softmax_0_axis_0, x = add_0_cast_fp16)[name = tensor<string, []>("softmax_0_cast_fp16")];
tensor<bool, []> attn_output_1_transpose_x_0 = const()[name = tensor<string, []>("attn_output_1_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_1_transpose_y_0 = const()[name = tensor<string, []>("attn_output_1_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_3_cast_fp16 = transpose(perm = value_states_3_perm_0, x = var_120_cast_fp16)[name = tensor<string, []>("transpose_120")];
tensor<fp16, [1, 12, 77, 64]> attn_output_1_cast_fp16 = matmul(transpose_x = attn_output_1_transpose_x_0, transpose_y = attn_output_1_transpose_y_0, x = softmax_0_cast_fp16, y = value_states_3_cast_fp16)[name = tensor<string, []>("attn_output_1_cast_fp16")];
tensor<int32, [4]> attn_output_3_perm_0 = const()[name = tensor<string, []>("attn_output_3_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_124 = const()[name = tensor<string, []>("op_124"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_3_cast_fp16 = transpose(perm = attn_output_3_perm_0, x = attn_output_1_cast_fp16)[name = tensor<string, []>("transpose_117")];
tensor<fp16, [1, 77, 768]> input_5_cast_fp16 = reshape(shape = var_124, x = attn_output_3_cast_fp16)[name = tensor<string, []>("input_5_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(79568256)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(80747968)))];
tensor<fp16, [1, 77, 768]> linear_3_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_self_attn_out_proj_weight_to_fp16, x = input_5_cast_fp16)[name = tensor<string, []>("linear_3_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_7_cast_fp16 = add(x = input_3_cast_fp16, y = linear_3_cast_fp16)[name = tensor<string, []>("input_7_cast_fp16")];
tensor<int32, [1]> input_9_axes_0 = const()[name = tensor<string, []>("input_9_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(80749568)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(80751168)))];
tensor<fp16, [1, 77, 768]> input_9_cast_fp16 = layer_norm(axes = input_9_axes_0, beta = text_encoder_text_model_encoder_layers_0_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_0_layer_norm2_weight_to_fp16, x = input_7_cast_fp16)[name = tensor<string, []>("input_9_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(80752768)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(85471424)))];
tensor<fp16, [1, 77, 3072]> linear_4_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc1_weight_to_fp16, x = input_9_cast_fp16)[name = tensor<string, []>("linear_4_cast_fp16")];
tensor<fp16, []> var_139_to_fp16 = const()[name = tensor<string, []>("op_139_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_140_cast_fp16 = mul(x = linear_4_cast_fp16, y = var_139_to_fp16)[name = tensor<string, []>("op_140_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_141_cast_fp16 = sigmoid(x = var_140_cast_fp16)[name = tensor<string, []>("op_141_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_13_cast_fp16 = mul(x = linear_4_cast_fp16, y = var_141_cast_fp16)[name = tensor<string, []>("input_13_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(85477632)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(90196288)))];
tensor<fp16, [1, 77, 768]> linear_5_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_0_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_0_mlp_fc2_weight_to_fp16, x = input_13_cast_fp16)[name = tensor<string, []>("linear_5_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_15_cast_fp16 = add(x = input_7_cast_fp16, y = linear_5_cast_fp16)[name = tensor<string, []>("input_15_cast_fp16")];
tensor<int32, [1]> hidden_states_7_axes_0 = const()[name = tensor<string, []>("hidden_states_7_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(90197888)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(90199488)))];
tensor<fp16, [1, 77, 768]> hidden_states_7_cast_fp16 = layer_norm(axes = hidden_states_7_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm1_weight_to_fp16, x = input_15_cast_fp16)[name = tensor<string, []>("hidden_states_7_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(90201088)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(91380800)))];
tensor<fp16, [1, 77, 768]> linear_6_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_q_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor<string, []>("linear_6_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(91382400)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(92562112)))];
tensor<fp16, [1, 77, 768]> linear_7_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_k_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor<string, []>("linear_7_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(92563712)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(93743424)))];
tensor<fp16, [1, 77, 768]> linear_8_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_v_proj_weight_to_fp16, x = hidden_states_7_cast_fp16)[name = tensor<string, []>("linear_8_cast_fp16")];
tensor<int32, [4]> var_171 = const()[name = tensor<string, []>("op_171"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_172_cast_fp16 = reshape(shape = var_171, x = linear_6_cast_fp16)[name = tensor<string, []>("op_172_cast_fp16")];
tensor<int32, [4]> var_174 = const()[name = tensor<string, []>("op_174"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_175_cast_fp16 = reshape(shape = var_174, x = linear_7_cast_fp16)[name = tensor<string, []>("op_175_cast_fp16")];
tensor<int32, [4]> var_177 = const()[name = tensor<string, []>("op_177"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_178_cast_fp16 = reshape(shape = var_177, x = linear_8_cast_fp16)[name = tensor<string, []>("op_178_cast_fp16")];
tensor<int32, [4]> value_states_7_perm_0 = const()[name = tensor<string, []>("value_states_7_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_1_cast_fp16 = mul(x = var_172_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_1_cast_fp16")];
tensor<bool, []> matmul_1_transpose_y_0 = const()[name = tensor<string, []>("matmul_1_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_1_transpose_x_0 = const()[name = tensor<string, []>("matmul_1_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_51_perm_0 = const()[name = tensor<string, []>("transpose_51_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_52_perm_0 = const()[name = tensor<string, []>("transpose_52_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_52 = transpose(perm = transpose_52_perm_0, x = var_175_cast_fp16)[name = tensor<string, []>("transpose_114")];
tensor<fp16, [1, 12, 77, 64]> transpose_51 = transpose(perm = transpose_51_perm_0, x = mul_1_cast_fp16)[name = tensor<string, []>("transpose_115")];
tensor<fp16, [1, 12, 77, 77]> matmul_1_cast_fp16 = matmul(transpose_x = matmul_1_transpose_x_0, transpose_y = matmul_1_transpose_y_0, x = transpose_51, y = transpose_52)[name = tensor<string, []>("matmul_1_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_1_cast_fp16 = add(x = matmul_1_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_1_cast_fp16")];
tensor<int32, []> softmax_1_axis_0 = const()[name = tensor<string, []>("softmax_1_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_1_cast_fp16 = softmax(axis = softmax_1_axis_0, x = add_1_cast_fp16)[name = tensor<string, []>("softmax_1_cast_fp16")];
tensor<bool, []> attn_output_5_transpose_x_0 = const()[name = tensor<string, []>("attn_output_5_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_5_transpose_y_0 = const()[name = tensor<string, []>("attn_output_5_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_7_cast_fp16 = transpose(perm = value_states_7_perm_0, x = var_178_cast_fp16)[name = tensor<string, []>("transpose_116")];
tensor<fp16, [1, 12, 77, 64]> attn_output_5_cast_fp16 = matmul(transpose_x = attn_output_5_transpose_x_0, transpose_y = attn_output_5_transpose_y_0, x = softmax_1_cast_fp16, y = value_states_7_cast_fp16)[name = tensor<string, []>("attn_output_5_cast_fp16")];
tensor<int32, [4]> attn_output_7_perm_0 = const()[name = tensor<string, []>("attn_output_7_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_182 = const()[name = tensor<string, []>("op_182"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_7_cast_fp16 = transpose(perm = attn_output_7_perm_0, x = attn_output_5_cast_fp16)[name = tensor<string, []>("transpose_113")];
tensor<fp16, [1, 77, 768]> input_17_cast_fp16 = reshape(shape = var_182, x = attn_output_7_cast_fp16)[name = tensor<string, []>("input_17_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(93745024)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(94924736)))];
tensor<fp16, [1, 77, 768]> linear_9_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_self_attn_out_proj_weight_to_fp16, x = input_17_cast_fp16)[name = tensor<string, []>("linear_9_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_19_cast_fp16 = add(x = input_15_cast_fp16, y = linear_9_cast_fp16)[name = tensor<string, []>("input_19_cast_fp16")];
tensor<int32, [1]> input_21_axes_0 = const()[name = tensor<string, []>("input_21_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(94926336)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(94927936)))];
tensor<fp16, [1, 77, 768]> input_21_cast_fp16 = layer_norm(axes = input_21_axes_0, beta = text_encoder_text_model_encoder_layers_1_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_1_layer_norm2_weight_to_fp16, x = input_19_cast_fp16)[name = tensor<string, []>("input_21_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(94929536)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(99648192)))];
tensor<fp16, [1, 77, 3072]> linear_10_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc1_weight_to_fp16, x = input_21_cast_fp16)[name = tensor<string, []>("linear_10_cast_fp16")];
tensor<fp16, []> var_197_to_fp16 = const()[name = tensor<string, []>("op_197_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_198_cast_fp16 = mul(x = linear_10_cast_fp16, y = var_197_to_fp16)[name = tensor<string, []>("op_198_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_199_cast_fp16 = sigmoid(x = var_198_cast_fp16)[name = tensor<string, []>("op_199_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_25_cast_fp16 = mul(x = linear_10_cast_fp16, y = var_199_cast_fp16)[name = tensor<string, []>("input_25_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(99654400)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(104373056)))];
tensor<fp16, [1, 77, 768]> linear_11_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_1_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_1_mlp_fc2_weight_to_fp16, x = input_25_cast_fp16)[name = tensor<string, []>("linear_11_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_27_cast_fp16 = add(x = input_19_cast_fp16, y = linear_11_cast_fp16)[name = tensor<string, []>("input_27_cast_fp16")];
tensor<int32, [1]> hidden_states_13_axes_0 = const()[name = tensor<string, []>("hidden_states_13_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(104374656)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(104376256)))];
tensor<fp16, [1, 77, 768]> hidden_states_13_cast_fp16 = layer_norm(axes = hidden_states_13_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm1_weight_to_fp16, x = input_27_cast_fp16)[name = tensor<string, []>("hidden_states_13_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(104377856)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(105557568)))];
tensor<fp16, [1, 77, 768]> linear_12_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_q_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor<string, []>("linear_12_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(105559168)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(106738880)))];
tensor<fp16, [1, 77, 768]> linear_13_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_k_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor<string, []>("linear_13_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(106740480)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(107920192)))];
tensor<fp16, [1, 77, 768]> linear_14_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_v_proj_weight_to_fp16, x = hidden_states_13_cast_fp16)[name = tensor<string, []>("linear_14_cast_fp16")];
tensor<int32, [4]> var_229 = const()[name = tensor<string, []>("op_229"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_230_cast_fp16 = reshape(shape = var_229, x = linear_12_cast_fp16)[name = tensor<string, []>("op_230_cast_fp16")];
tensor<int32, [4]> var_232 = const()[name = tensor<string, []>("op_232"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_233_cast_fp16 = reshape(shape = var_232, x = linear_13_cast_fp16)[name = tensor<string, []>("op_233_cast_fp16")];
tensor<int32, [4]> var_235 = const()[name = tensor<string, []>("op_235"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_236_cast_fp16 = reshape(shape = var_235, x = linear_14_cast_fp16)[name = tensor<string, []>("op_236_cast_fp16")];
tensor<int32, [4]> value_states_11_perm_0 = const()[name = tensor<string, []>("value_states_11_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_2_cast_fp16 = mul(x = var_230_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_2_cast_fp16")];
tensor<bool, []> matmul_2_transpose_y_0 = const()[name = tensor<string, []>("matmul_2_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_2_transpose_x_0 = const()[name = tensor<string, []>("matmul_2_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_53_perm_0 = const()[name = tensor<string, []>("transpose_53_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_54_perm_0 = const()[name = tensor<string, []>("transpose_54_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_54 = transpose(perm = transpose_54_perm_0, x = var_233_cast_fp16)[name = tensor<string, []>("transpose_110")];
tensor<fp16, [1, 12, 77, 64]> transpose_53 = transpose(perm = transpose_53_perm_0, x = mul_2_cast_fp16)[name = tensor<string, []>("transpose_111")];
tensor<fp16, [1, 12, 77, 77]> matmul_2_cast_fp16 = matmul(transpose_x = matmul_2_transpose_x_0, transpose_y = matmul_2_transpose_y_0, x = transpose_53, y = transpose_54)[name = tensor<string, []>("matmul_2_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_2_cast_fp16 = add(x = matmul_2_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_2_cast_fp16")];
tensor<int32, []> softmax_2_axis_0 = const()[name = tensor<string, []>("softmax_2_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_2_cast_fp16 = softmax(axis = softmax_2_axis_0, x = add_2_cast_fp16)[name = tensor<string, []>("softmax_2_cast_fp16")];
tensor<bool, []> attn_output_9_transpose_x_0 = const()[name = tensor<string, []>("attn_output_9_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_9_transpose_y_0 = const()[name = tensor<string, []>("attn_output_9_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_11_cast_fp16 = transpose(perm = value_states_11_perm_0, x = var_236_cast_fp16)[name = tensor<string, []>("transpose_112")];
tensor<fp16, [1, 12, 77, 64]> attn_output_9_cast_fp16 = matmul(transpose_x = attn_output_9_transpose_x_0, transpose_y = attn_output_9_transpose_y_0, x = softmax_2_cast_fp16, y = value_states_11_cast_fp16)[name = tensor<string, []>("attn_output_9_cast_fp16")];
tensor<int32, [4]> attn_output_11_perm_0 = const()[name = tensor<string, []>("attn_output_11_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_240 = const()[name = tensor<string, []>("op_240"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_11_cast_fp16 = transpose(perm = attn_output_11_perm_0, x = attn_output_9_cast_fp16)[name = tensor<string, []>("transpose_109")];
tensor<fp16, [1, 77, 768]> input_29_cast_fp16 = reshape(shape = var_240, x = attn_output_11_cast_fp16)[name = tensor<string, []>("input_29_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(107921792)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(109101504)))];
tensor<fp16, [1, 77, 768]> linear_15_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_self_attn_out_proj_weight_to_fp16, x = input_29_cast_fp16)[name = tensor<string, []>("linear_15_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_31_cast_fp16 = add(x = input_27_cast_fp16, y = linear_15_cast_fp16)[name = tensor<string, []>("input_31_cast_fp16")];
tensor<int32, [1]> input_33_axes_0 = const()[name = tensor<string, []>("input_33_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(109103104)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(109104704)))];
tensor<fp16, [1, 77, 768]> input_33_cast_fp16 = layer_norm(axes = input_33_axes_0, beta = text_encoder_text_model_encoder_layers_2_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_2_layer_norm2_weight_to_fp16, x = input_31_cast_fp16)[name = tensor<string, []>("input_33_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(109106304)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(113824960)))];
tensor<fp16, [1, 77, 3072]> linear_16_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc1_weight_to_fp16, x = input_33_cast_fp16)[name = tensor<string, []>("linear_16_cast_fp16")];
tensor<fp16, []> var_255_to_fp16 = const()[name = tensor<string, []>("op_255_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_256_cast_fp16 = mul(x = linear_16_cast_fp16, y = var_255_to_fp16)[name = tensor<string, []>("op_256_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_257_cast_fp16 = sigmoid(x = var_256_cast_fp16)[name = tensor<string, []>("op_257_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_37_cast_fp16 = mul(x = linear_16_cast_fp16, y = var_257_cast_fp16)[name = tensor<string, []>("input_37_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(113831168)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(118549824)))];
tensor<fp16, [1, 77, 768]> linear_17_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_2_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_2_mlp_fc2_weight_to_fp16, x = input_37_cast_fp16)[name = tensor<string, []>("linear_17_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_39_cast_fp16 = add(x = input_31_cast_fp16, y = linear_17_cast_fp16)[name = tensor<string, []>("input_39_cast_fp16")];
tensor<int32, [1]> hidden_states_19_axes_0 = const()[name = tensor<string, []>("hidden_states_19_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(118551424)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(118553024)))];
tensor<fp16, [1, 77, 768]> hidden_states_19_cast_fp16 = layer_norm(axes = hidden_states_19_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm1_weight_to_fp16, x = input_39_cast_fp16)[name = tensor<string, []>("hidden_states_19_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(118554624)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(119734336)))];
tensor<fp16, [1, 77, 768]> linear_18_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_q_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor<string, []>("linear_18_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(119735936)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(120915648)))];
tensor<fp16, [1, 77, 768]> linear_19_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_k_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor<string, []>("linear_19_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(120917248)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(122096960)))];
tensor<fp16, [1, 77, 768]> linear_20_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_v_proj_weight_to_fp16, x = hidden_states_19_cast_fp16)[name = tensor<string, []>("linear_20_cast_fp16")];
tensor<int32, [4]> var_287 = const()[name = tensor<string, []>("op_287"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_288_cast_fp16 = reshape(shape = var_287, x = linear_18_cast_fp16)[name = tensor<string, []>("op_288_cast_fp16")];
tensor<int32, [4]> var_290 = const()[name = tensor<string, []>("op_290"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_291_cast_fp16 = reshape(shape = var_290, x = linear_19_cast_fp16)[name = tensor<string, []>("op_291_cast_fp16")];
tensor<int32, [4]> var_293 = const()[name = tensor<string, []>("op_293"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_294_cast_fp16 = reshape(shape = var_293, x = linear_20_cast_fp16)[name = tensor<string, []>("op_294_cast_fp16")];
tensor<int32, [4]> value_states_15_perm_0 = const()[name = tensor<string, []>("value_states_15_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_3_cast_fp16 = mul(x = var_288_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_3_cast_fp16")];
tensor<bool, []> matmul_3_transpose_y_0 = const()[name = tensor<string, []>("matmul_3_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_3_transpose_x_0 = const()[name = tensor<string, []>("matmul_3_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_55_perm_0 = const()[name = tensor<string, []>("transpose_55_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_56_perm_0 = const()[name = tensor<string, []>("transpose_56_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_56 = transpose(perm = transpose_56_perm_0, x = var_291_cast_fp16)[name = tensor<string, []>("transpose_106")];
tensor<fp16, [1, 12, 77, 64]> transpose_55 = transpose(perm = transpose_55_perm_0, x = mul_3_cast_fp16)[name = tensor<string, []>("transpose_107")];
tensor<fp16, [1, 12, 77, 77]> matmul_3_cast_fp16 = matmul(transpose_x = matmul_3_transpose_x_0, transpose_y = matmul_3_transpose_y_0, x = transpose_55, y = transpose_56)[name = tensor<string, []>("matmul_3_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_3_cast_fp16 = add(x = matmul_3_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_3_cast_fp16")];
tensor<int32, []> softmax_3_axis_0 = const()[name = tensor<string, []>("softmax_3_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_3_cast_fp16 = softmax(axis = softmax_3_axis_0, x = add_3_cast_fp16)[name = tensor<string, []>("softmax_3_cast_fp16")];
tensor<bool, []> attn_output_13_transpose_x_0 = const()[name = tensor<string, []>("attn_output_13_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_13_transpose_y_0 = const()[name = tensor<string, []>("attn_output_13_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_15_cast_fp16 = transpose(perm = value_states_15_perm_0, x = var_294_cast_fp16)[name = tensor<string, []>("transpose_108")];
tensor<fp16, [1, 12, 77, 64]> attn_output_13_cast_fp16 = matmul(transpose_x = attn_output_13_transpose_x_0, transpose_y = attn_output_13_transpose_y_0, x = softmax_3_cast_fp16, y = value_states_15_cast_fp16)[name = tensor<string, []>("attn_output_13_cast_fp16")];
tensor<int32, [4]> attn_output_15_perm_0 = const()[name = tensor<string, []>("attn_output_15_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_298 = const()[name = tensor<string, []>("op_298"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_15_cast_fp16 = transpose(perm = attn_output_15_perm_0, x = attn_output_13_cast_fp16)[name = tensor<string, []>("transpose_105")];
tensor<fp16, [1, 77, 768]> input_41_cast_fp16 = reshape(shape = var_298, x = attn_output_15_cast_fp16)[name = tensor<string, []>("input_41_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(122098560)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(123278272)))];
tensor<fp16, [1, 77, 768]> linear_21_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_self_attn_out_proj_weight_to_fp16, x = input_41_cast_fp16)[name = tensor<string, []>("linear_21_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_43_cast_fp16 = add(x = input_39_cast_fp16, y = linear_21_cast_fp16)[name = tensor<string, []>("input_43_cast_fp16")];
tensor<int32, [1]> input_45_axes_0 = const()[name = tensor<string, []>("input_45_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(123279872)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(123281472)))];
tensor<fp16, [1, 77, 768]> input_45_cast_fp16 = layer_norm(axes = input_45_axes_0, beta = text_encoder_text_model_encoder_layers_3_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_3_layer_norm2_weight_to_fp16, x = input_43_cast_fp16)[name = tensor<string, []>("input_45_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(123283072)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(128001728)))];
tensor<fp16, [1, 77, 3072]> linear_22_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc1_weight_to_fp16, x = input_45_cast_fp16)[name = tensor<string, []>("linear_22_cast_fp16")];
tensor<fp16, []> var_313_to_fp16 = const()[name = tensor<string, []>("op_313_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_314_cast_fp16 = mul(x = linear_22_cast_fp16, y = var_313_to_fp16)[name = tensor<string, []>("op_314_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_315_cast_fp16 = sigmoid(x = var_314_cast_fp16)[name = tensor<string, []>("op_315_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_49_cast_fp16 = mul(x = linear_22_cast_fp16, y = var_315_cast_fp16)[name = tensor<string, []>("input_49_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(128007936)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(132726592)))];
tensor<fp16, [1, 77, 768]> linear_23_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_3_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_3_mlp_fc2_weight_to_fp16, x = input_49_cast_fp16)[name = tensor<string, []>("linear_23_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_51_cast_fp16 = add(x = input_43_cast_fp16, y = linear_23_cast_fp16)[name = tensor<string, []>("input_51_cast_fp16")];
tensor<int32, [1]> hidden_states_25_axes_0 = const()[name = tensor<string, []>("hidden_states_25_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(132728192)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(132729792)))];
tensor<fp16, [1, 77, 768]> hidden_states_25_cast_fp16 = layer_norm(axes = hidden_states_25_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm1_weight_to_fp16, x = input_51_cast_fp16)[name = tensor<string, []>("hidden_states_25_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(132731392)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(133911104)))];
tensor<fp16, [1, 77, 768]> linear_24_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_q_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor<string, []>("linear_24_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(133912704)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(135092416)))];
tensor<fp16, [1, 77, 768]> linear_25_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_k_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor<string, []>("linear_25_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(135094016)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(136273728)))];
tensor<fp16, [1, 77, 768]> linear_26_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_v_proj_weight_to_fp16, x = hidden_states_25_cast_fp16)[name = tensor<string, []>("linear_26_cast_fp16")];
tensor<int32, [4]> var_345 = const()[name = tensor<string, []>("op_345"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_346_cast_fp16 = reshape(shape = var_345, x = linear_24_cast_fp16)[name = tensor<string, []>("op_346_cast_fp16")];
tensor<int32, [4]> var_348 = const()[name = tensor<string, []>("op_348"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_349_cast_fp16 = reshape(shape = var_348, x = linear_25_cast_fp16)[name = tensor<string, []>("op_349_cast_fp16")];
tensor<int32, [4]> var_351 = const()[name = tensor<string, []>("op_351"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_352_cast_fp16 = reshape(shape = var_351, x = linear_26_cast_fp16)[name = tensor<string, []>("op_352_cast_fp16")];
tensor<int32, [4]> value_states_19_perm_0 = const()[name = tensor<string, []>("value_states_19_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_4_cast_fp16 = mul(x = var_346_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_4_cast_fp16")];
tensor<bool, []> matmul_4_transpose_y_0 = const()[name = tensor<string, []>("matmul_4_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_4_transpose_x_0 = const()[name = tensor<string, []>("matmul_4_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_57_perm_0 = const()[name = tensor<string, []>("transpose_57_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_58_perm_0 = const()[name = tensor<string, []>("transpose_58_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_58 = transpose(perm = transpose_58_perm_0, x = var_349_cast_fp16)[name = tensor<string, []>("transpose_102")];
tensor<fp16, [1, 12, 77, 64]> transpose_57 = transpose(perm = transpose_57_perm_0, x = mul_4_cast_fp16)[name = tensor<string, []>("transpose_103")];
tensor<fp16, [1, 12, 77, 77]> matmul_4_cast_fp16 = matmul(transpose_x = matmul_4_transpose_x_0, transpose_y = matmul_4_transpose_y_0, x = transpose_57, y = transpose_58)[name = tensor<string, []>("matmul_4_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_4_cast_fp16 = add(x = matmul_4_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_4_cast_fp16")];
tensor<int32, []> softmax_4_axis_0 = const()[name = tensor<string, []>("softmax_4_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_4_cast_fp16 = softmax(axis = softmax_4_axis_0, x = add_4_cast_fp16)[name = tensor<string, []>("softmax_4_cast_fp16")];
tensor<bool, []> attn_output_17_transpose_x_0 = const()[name = tensor<string, []>("attn_output_17_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_17_transpose_y_0 = const()[name = tensor<string, []>("attn_output_17_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_19_cast_fp16 = transpose(perm = value_states_19_perm_0, x = var_352_cast_fp16)[name = tensor<string, []>("transpose_104")];
tensor<fp16, [1, 12, 77, 64]> attn_output_17_cast_fp16 = matmul(transpose_x = attn_output_17_transpose_x_0, transpose_y = attn_output_17_transpose_y_0, x = softmax_4_cast_fp16, y = value_states_19_cast_fp16)[name = tensor<string, []>("attn_output_17_cast_fp16")];
tensor<int32, [4]> attn_output_19_perm_0 = const()[name = tensor<string, []>("attn_output_19_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_356 = const()[name = tensor<string, []>("op_356"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_19_cast_fp16 = transpose(perm = attn_output_19_perm_0, x = attn_output_17_cast_fp16)[name = tensor<string, []>("transpose_101")];
tensor<fp16, [1, 77, 768]> input_53_cast_fp16 = reshape(shape = var_356, x = attn_output_19_cast_fp16)[name = tensor<string, []>("input_53_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(136275328)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(137455040)))];
tensor<fp16, [1, 77, 768]> linear_27_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_self_attn_out_proj_weight_to_fp16, x = input_53_cast_fp16)[name = tensor<string, []>("linear_27_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_55_cast_fp16 = add(x = input_51_cast_fp16, y = linear_27_cast_fp16)[name = tensor<string, []>("input_55_cast_fp16")];
tensor<int32, [1]> input_57_axes_0 = const()[name = tensor<string, []>("input_57_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(137456640)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(137458240)))];
tensor<fp16, [1, 77, 768]> input_57_cast_fp16 = layer_norm(axes = input_57_axes_0, beta = text_encoder_text_model_encoder_layers_4_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_4_layer_norm2_weight_to_fp16, x = input_55_cast_fp16)[name = tensor<string, []>("input_57_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(137459840)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(142178496)))];
tensor<fp16, [1, 77, 3072]> linear_28_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc1_weight_to_fp16, x = input_57_cast_fp16)[name = tensor<string, []>("linear_28_cast_fp16")];
tensor<fp16, []> var_371_to_fp16 = const()[name = tensor<string, []>("op_371_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_372_cast_fp16 = mul(x = linear_28_cast_fp16, y = var_371_to_fp16)[name = tensor<string, []>("op_372_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_373_cast_fp16 = sigmoid(x = var_372_cast_fp16)[name = tensor<string, []>("op_373_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_61_cast_fp16 = mul(x = linear_28_cast_fp16, y = var_373_cast_fp16)[name = tensor<string, []>("input_61_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(142184704)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(146903360)))];
tensor<fp16, [1, 77, 768]> linear_29_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_4_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_4_mlp_fc2_weight_to_fp16, x = input_61_cast_fp16)[name = tensor<string, []>("linear_29_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_63_cast_fp16 = add(x = input_55_cast_fp16, y = linear_29_cast_fp16)[name = tensor<string, []>("input_63_cast_fp16")];
tensor<int32, [1]> hidden_states_31_axes_0 = const()[name = tensor<string, []>("hidden_states_31_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(146904960)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(146906560)))];
tensor<fp16, [1, 77, 768]> hidden_states_31_cast_fp16 = layer_norm(axes = hidden_states_31_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm1_weight_to_fp16, x = input_63_cast_fp16)[name = tensor<string, []>("hidden_states_31_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(146908160)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(148087872)))];
tensor<fp16, [1, 77, 768]> linear_30_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_q_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor<string, []>("linear_30_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(148089472)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(149269184)))];
tensor<fp16, [1, 77, 768]> linear_31_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_k_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor<string, []>("linear_31_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(149270784)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(150450496)))];
tensor<fp16, [1, 77, 768]> linear_32_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_v_proj_weight_to_fp16, x = hidden_states_31_cast_fp16)[name = tensor<string, []>("linear_32_cast_fp16")];
tensor<int32, [4]> var_403 = const()[name = tensor<string, []>("op_403"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_404_cast_fp16 = reshape(shape = var_403, x = linear_30_cast_fp16)[name = tensor<string, []>("op_404_cast_fp16")];
tensor<int32, [4]> var_406 = const()[name = tensor<string, []>("op_406"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_407_cast_fp16 = reshape(shape = var_406, x = linear_31_cast_fp16)[name = tensor<string, []>("op_407_cast_fp16")];
tensor<int32, [4]> var_409 = const()[name = tensor<string, []>("op_409"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_410_cast_fp16 = reshape(shape = var_409, x = linear_32_cast_fp16)[name = tensor<string, []>("op_410_cast_fp16")];
tensor<int32, [4]> value_states_23_perm_0 = const()[name = tensor<string, []>("value_states_23_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_5_cast_fp16 = mul(x = var_404_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_5_cast_fp16")];
tensor<bool, []> matmul_5_transpose_y_0 = const()[name = tensor<string, []>("matmul_5_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_5_transpose_x_0 = const()[name = tensor<string, []>("matmul_5_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_59_perm_0 = const()[name = tensor<string, []>("transpose_59_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_60_perm_0 = const()[name = tensor<string, []>("transpose_60_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_60 = transpose(perm = transpose_60_perm_0, x = var_407_cast_fp16)[name = tensor<string, []>("transpose_98")];
tensor<fp16, [1, 12, 77, 64]> transpose_59 = transpose(perm = transpose_59_perm_0, x = mul_5_cast_fp16)[name = tensor<string, []>("transpose_99")];
tensor<fp16, [1, 12, 77, 77]> matmul_5_cast_fp16 = matmul(transpose_x = matmul_5_transpose_x_0, transpose_y = matmul_5_transpose_y_0, x = transpose_59, y = transpose_60)[name = tensor<string, []>("matmul_5_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_5_cast_fp16 = add(x = matmul_5_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_5_cast_fp16")];
tensor<int32, []> softmax_5_axis_0 = const()[name = tensor<string, []>("softmax_5_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_5_cast_fp16 = softmax(axis = softmax_5_axis_0, x = add_5_cast_fp16)[name = tensor<string, []>("softmax_5_cast_fp16")];
tensor<bool, []> attn_output_21_transpose_x_0 = const()[name = tensor<string, []>("attn_output_21_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_21_transpose_y_0 = const()[name = tensor<string, []>("attn_output_21_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_23_cast_fp16 = transpose(perm = value_states_23_perm_0, x = var_410_cast_fp16)[name = tensor<string, []>("transpose_100")];
tensor<fp16, [1, 12, 77, 64]> attn_output_21_cast_fp16 = matmul(transpose_x = attn_output_21_transpose_x_0, transpose_y = attn_output_21_transpose_y_0, x = softmax_5_cast_fp16, y = value_states_23_cast_fp16)[name = tensor<string, []>("attn_output_21_cast_fp16")];
tensor<int32, [4]> attn_output_23_perm_0 = const()[name = tensor<string, []>("attn_output_23_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_414 = const()[name = tensor<string, []>("op_414"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_23_cast_fp16 = transpose(perm = attn_output_23_perm_0, x = attn_output_21_cast_fp16)[name = tensor<string, []>("transpose_97")];
tensor<fp16, [1, 77, 768]> input_65_cast_fp16 = reshape(shape = var_414, x = attn_output_23_cast_fp16)[name = tensor<string, []>("input_65_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(150452096)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(151631808)))];
tensor<fp16, [1, 77, 768]> linear_33_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_self_attn_out_proj_weight_to_fp16, x = input_65_cast_fp16)[name = tensor<string, []>("linear_33_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_67_cast_fp16 = add(x = input_63_cast_fp16, y = linear_33_cast_fp16)[name = tensor<string, []>("input_67_cast_fp16")];
tensor<int32, [1]> input_69_axes_0 = const()[name = tensor<string, []>("input_69_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(151633408)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(151635008)))];
tensor<fp16, [1, 77, 768]> input_69_cast_fp16 = layer_norm(axes = input_69_axes_0, beta = text_encoder_text_model_encoder_layers_5_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_5_layer_norm2_weight_to_fp16, x = input_67_cast_fp16)[name = tensor<string, []>("input_69_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(151636608)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(156355264)))];
tensor<fp16, [1, 77, 3072]> linear_34_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc1_weight_to_fp16, x = input_69_cast_fp16)[name = tensor<string, []>("linear_34_cast_fp16")];
tensor<fp16, []> var_429_to_fp16 = const()[name = tensor<string, []>("op_429_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_430_cast_fp16 = mul(x = linear_34_cast_fp16, y = var_429_to_fp16)[name = tensor<string, []>("op_430_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_431_cast_fp16 = sigmoid(x = var_430_cast_fp16)[name = tensor<string, []>("op_431_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_73_cast_fp16 = mul(x = linear_34_cast_fp16, y = var_431_cast_fp16)[name = tensor<string, []>("input_73_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(156361472)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(161080128)))];
tensor<fp16, [1, 77, 768]> linear_35_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_5_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_5_mlp_fc2_weight_to_fp16, x = input_73_cast_fp16)[name = tensor<string, []>("linear_35_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_75_cast_fp16 = add(x = input_67_cast_fp16, y = linear_35_cast_fp16)[name = tensor<string, []>("input_75_cast_fp16")];
tensor<int32, [1]> hidden_states_37_axes_0 = const()[name = tensor<string, []>("hidden_states_37_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(161081728)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(161083328)))];
tensor<fp16, [1, 77, 768]> hidden_states_37_cast_fp16 = layer_norm(axes = hidden_states_37_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm1_weight_to_fp16, x = input_75_cast_fp16)[name = tensor<string, []>("hidden_states_37_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(161084928)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(162264640)))];
tensor<fp16, [1, 77, 768]> linear_36_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_q_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor<string, []>("linear_36_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(162266240)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(163445952)))];
tensor<fp16, [1, 77, 768]> linear_37_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_k_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor<string, []>("linear_37_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(163447552)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(164627264)))];
tensor<fp16, [1, 77, 768]> linear_38_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_v_proj_weight_to_fp16, x = hidden_states_37_cast_fp16)[name = tensor<string, []>("linear_38_cast_fp16")];
tensor<int32, [4]> var_461 = const()[name = tensor<string, []>("op_461"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_462_cast_fp16 = reshape(shape = var_461, x = linear_36_cast_fp16)[name = tensor<string, []>("op_462_cast_fp16")];
tensor<int32, [4]> var_464 = const()[name = tensor<string, []>("op_464"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_465_cast_fp16 = reshape(shape = var_464, x = linear_37_cast_fp16)[name = tensor<string, []>("op_465_cast_fp16")];
tensor<int32, [4]> var_467 = const()[name = tensor<string, []>("op_467"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_468_cast_fp16 = reshape(shape = var_467, x = linear_38_cast_fp16)[name = tensor<string, []>("op_468_cast_fp16")];
tensor<int32, [4]> value_states_27_perm_0 = const()[name = tensor<string, []>("value_states_27_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_6_cast_fp16 = mul(x = var_462_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_6_cast_fp16")];
tensor<bool, []> matmul_6_transpose_y_0 = const()[name = tensor<string, []>("matmul_6_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_6_transpose_x_0 = const()[name = tensor<string, []>("matmul_6_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_61_perm_0 = const()[name = tensor<string, []>("transpose_61_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_62_perm_0 = const()[name = tensor<string, []>("transpose_62_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_62 = transpose(perm = transpose_62_perm_0, x = var_465_cast_fp16)[name = tensor<string, []>("transpose_94")];
tensor<fp16, [1, 12, 77, 64]> transpose_61 = transpose(perm = transpose_61_perm_0, x = mul_6_cast_fp16)[name = tensor<string, []>("transpose_95")];
tensor<fp16, [1, 12, 77, 77]> matmul_6_cast_fp16 = matmul(transpose_x = matmul_6_transpose_x_0, transpose_y = matmul_6_transpose_y_0, x = transpose_61, y = transpose_62)[name = tensor<string, []>("matmul_6_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_6_cast_fp16 = add(x = matmul_6_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_6_cast_fp16")];
tensor<int32, []> softmax_6_axis_0 = const()[name = tensor<string, []>("softmax_6_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_6_cast_fp16 = softmax(axis = softmax_6_axis_0, x = add_6_cast_fp16)[name = tensor<string, []>("softmax_6_cast_fp16")];
tensor<bool, []> attn_output_25_transpose_x_0 = const()[name = tensor<string, []>("attn_output_25_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_25_transpose_y_0 = const()[name = tensor<string, []>("attn_output_25_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_27_cast_fp16 = transpose(perm = value_states_27_perm_0, x = var_468_cast_fp16)[name = tensor<string, []>("transpose_96")];
tensor<fp16, [1, 12, 77, 64]> attn_output_25_cast_fp16 = matmul(transpose_x = attn_output_25_transpose_x_0, transpose_y = attn_output_25_transpose_y_0, x = softmax_6_cast_fp16, y = value_states_27_cast_fp16)[name = tensor<string, []>("attn_output_25_cast_fp16")];
tensor<int32, [4]> attn_output_27_perm_0 = const()[name = tensor<string, []>("attn_output_27_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_472 = const()[name = tensor<string, []>("op_472"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_27_cast_fp16 = transpose(perm = attn_output_27_perm_0, x = attn_output_25_cast_fp16)[name = tensor<string, []>("transpose_93")];
tensor<fp16, [1, 77, 768]> input_77_cast_fp16 = reshape(shape = var_472, x = attn_output_27_cast_fp16)[name = tensor<string, []>("input_77_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(164628864)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(165808576)))];
tensor<fp16, [1, 77, 768]> linear_39_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_self_attn_out_proj_weight_to_fp16, x = input_77_cast_fp16)[name = tensor<string, []>("linear_39_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_79_cast_fp16 = add(x = input_75_cast_fp16, y = linear_39_cast_fp16)[name = tensor<string, []>("input_79_cast_fp16")];
tensor<int32, [1]> input_81_axes_0 = const()[name = tensor<string, []>("input_81_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(165810176)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(165811776)))];
tensor<fp16, [1, 77, 768]> input_81_cast_fp16 = layer_norm(axes = input_81_axes_0, beta = text_encoder_text_model_encoder_layers_6_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_6_layer_norm2_weight_to_fp16, x = input_79_cast_fp16)[name = tensor<string, []>("input_81_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(165813376)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(170532032)))];
tensor<fp16, [1, 77, 3072]> linear_40_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc1_weight_to_fp16, x = input_81_cast_fp16)[name = tensor<string, []>("linear_40_cast_fp16")];
tensor<fp16, []> var_487_to_fp16 = const()[name = tensor<string, []>("op_487_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_488_cast_fp16 = mul(x = linear_40_cast_fp16, y = var_487_to_fp16)[name = tensor<string, []>("op_488_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_489_cast_fp16 = sigmoid(x = var_488_cast_fp16)[name = tensor<string, []>("op_489_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_85_cast_fp16 = mul(x = linear_40_cast_fp16, y = var_489_cast_fp16)[name = tensor<string, []>("input_85_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(170538240)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(175256896)))];
tensor<fp16, [1, 77, 768]> linear_41_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_6_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_6_mlp_fc2_weight_to_fp16, x = input_85_cast_fp16)[name = tensor<string, []>("linear_41_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_87_cast_fp16 = add(x = input_79_cast_fp16, y = linear_41_cast_fp16)[name = tensor<string, []>("input_87_cast_fp16")];
tensor<int32, [1]> hidden_states_43_axes_0 = const()[name = tensor<string, []>("hidden_states_43_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(175258496)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(175260096)))];
tensor<fp16, [1, 77, 768]> hidden_states_43_cast_fp16 = layer_norm(axes = hidden_states_43_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm1_weight_to_fp16, x = input_87_cast_fp16)[name = tensor<string, []>("hidden_states_43_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(175261696)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(176441408)))];
tensor<fp16, [1, 77, 768]> linear_42_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_q_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor<string, []>("linear_42_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(176443008)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(177622720)))];
tensor<fp16, [1, 77, 768]> linear_43_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_k_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor<string, []>("linear_43_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(177624320)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(178804032)))];
tensor<fp16, [1, 77, 768]> linear_44_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_v_proj_weight_to_fp16, x = hidden_states_43_cast_fp16)[name = tensor<string, []>("linear_44_cast_fp16")];
tensor<int32, [4]> var_519 = const()[name = tensor<string, []>("op_519"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_520_cast_fp16 = reshape(shape = var_519, x = linear_42_cast_fp16)[name = tensor<string, []>("op_520_cast_fp16")];
tensor<int32, [4]> var_522 = const()[name = tensor<string, []>("op_522"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_523_cast_fp16 = reshape(shape = var_522, x = linear_43_cast_fp16)[name = tensor<string, []>("op_523_cast_fp16")];
tensor<int32, [4]> var_525 = const()[name = tensor<string, []>("op_525"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_526_cast_fp16 = reshape(shape = var_525, x = linear_44_cast_fp16)[name = tensor<string, []>("op_526_cast_fp16")];
tensor<int32, [4]> value_states_31_perm_0 = const()[name = tensor<string, []>("value_states_31_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_7_cast_fp16 = mul(x = var_520_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_7_cast_fp16")];
tensor<bool, []> matmul_7_transpose_y_0 = const()[name = tensor<string, []>("matmul_7_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_7_transpose_x_0 = const()[name = tensor<string, []>("matmul_7_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_63_perm_0 = const()[name = tensor<string, []>("transpose_63_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_64_perm_0 = const()[name = tensor<string, []>("transpose_64_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_64 = transpose(perm = transpose_64_perm_0, x = var_523_cast_fp16)[name = tensor<string, []>("transpose_90")];
tensor<fp16, [1, 12, 77, 64]> transpose_63 = transpose(perm = transpose_63_perm_0, x = mul_7_cast_fp16)[name = tensor<string, []>("transpose_91")];
tensor<fp16, [1, 12, 77, 77]> matmul_7_cast_fp16 = matmul(transpose_x = matmul_7_transpose_x_0, transpose_y = matmul_7_transpose_y_0, x = transpose_63, y = transpose_64)[name = tensor<string, []>("matmul_7_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_7_cast_fp16 = add(x = matmul_7_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_7_cast_fp16")];
tensor<int32, []> softmax_7_axis_0 = const()[name = tensor<string, []>("softmax_7_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_7_cast_fp16 = softmax(axis = softmax_7_axis_0, x = add_7_cast_fp16)[name = tensor<string, []>("softmax_7_cast_fp16")];
tensor<bool, []> attn_output_29_transpose_x_0 = const()[name = tensor<string, []>("attn_output_29_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_29_transpose_y_0 = const()[name = tensor<string, []>("attn_output_29_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_31_cast_fp16 = transpose(perm = value_states_31_perm_0, x = var_526_cast_fp16)[name = tensor<string, []>("transpose_92")];
tensor<fp16, [1, 12, 77, 64]> attn_output_29_cast_fp16 = matmul(transpose_x = attn_output_29_transpose_x_0, transpose_y = attn_output_29_transpose_y_0, x = softmax_7_cast_fp16, y = value_states_31_cast_fp16)[name = tensor<string, []>("attn_output_29_cast_fp16")];
tensor<int32, [4]> attn_output_31_perm_0 = const()[name = tensor<string, []>("attn_output_31_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_530 = const()[name = tensor<string, []>("op_530"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_31_cast_fp16 = transpose(perm = attn_output_31_perm_0, x = attn_output_29_cast_fp16)[name = tensor<string, []>("transpose_89")];
tensor<fp16, [1, 77, 768]> input_89_cast_fp16 = reshape(shape = var_530, x = attn_output_31_cast_fp16)[name = tensor<string, []>("input_89_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(178805632)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(179985344)))];
tensor<fp16, [1, 77, 768]> linear_45_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_self_attn_out_proj_weight_to_fp16, x = input_89_cast_fp16)[name = tensor<string, []>("linear_45_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_91_cast_fp16 = add(x = input_87_cast_fp16, y = linear_45_cast_fp16)[name = tensor<string, []>("input_91_cast_fp16")];
tensor<int32, [1]> input_93_axes_0 = const()[name = tensor<string, []>("input_93_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(179986944)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(179988544)))];
tensor<fp16, [1, 77, 768]> input_93_cast_fp16 = layer_norm(axes = input_93_axes_0, beta = text_encoder_text_model_encoder_layers_7_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_7_layer_norm2_weight_to_fp16, x = input_91_cast_fp16)[name = tensor<string, []>("input_93_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(179990144)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(184708800)))];
tensor<fp16, [1, 77, 3072]> linear_46_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc1_weight_to_fp16, x = input_93_cast_fp16)[name = tensor<string, []>("linear_46_cast_fp16")];
tensor<fp16, []> var_545_to_fp16 = const()[name = tensor<string, []>("op_545_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_546_cast_fp16 = mul(x = linear_46_cast_fp16, y = var_545_to_fp16)[name = tensor<string, []>("op_546_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_547_cast_fp16 = sigmoid(x = var_546_cast_fp16)[name = tensor<string, []>("op_547_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_97_cast_fp16 = mul(x = linear_46_cast_fp16, y = var_547_cast_fp16)[name = tensor<string, []>("input_97_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(184715008)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(189433664)))];
tensor<fp16, [1, 77, 768]> linear_47_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_7_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_7_mlp_fc2_weight_to_fp16, x = input_97_cast_fp16)[name = tensor<string, []>("linear_47_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_99_cast_fp16 = add(x = input_91_cast_fp16, y = linear_47_cast_fp16)[name = tensor<string, []>("input_99_cast_fp16")];
tensor<int32, [1]> hidden_states_49_axes_0 = const()[name = tensor<string, []>("hidden_states_49_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(189435264)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(189436864)))];
tensor<fp16, [1, 77, 768]> hidden_states_49_cast_fp16 = layer_norm(axes = hidden_states_49_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm1_weight_to_fp16, x = input_99_cast_fp16)[name = tensor<string, []>("hidden_states_49_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(189438464)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(190618176)))];
tensor<fp16, [1, 77, 768]> linear_48_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_q_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor<string, []>("linear_48_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(190619776)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(191799488)))];
tensor<fp16, [1, 77, 768]> linear_49_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_k_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor<string, []>("linear_49_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(191801088)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(192980800)))];
tensor<fp16, [1, 77, 768]> linear_50_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_v_proj_weight_to_fp16, x = hidden_states_49_cast_fp16)[name = tensor<string, []>("linear_50_cast_fp16")];
tensor<int32, [4]> var_577 = const()[name = tensor<string, []>("op_577"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_578_cast_fp16 = reshape(shape = var_577, x = linear_48_cast_fp16)[name = tensor<string, []>("op_578_cast_fp16")];
tensor<int32, [4]> var_580 = const()[name = tensor<string, []>("op_580"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_581_cast_fp16 = reshape(shape = var_580, x = linear_49_cast_fp16)[name = tensor<string, []>("op_581_cast_fp16")];
tensor<int32, [4]> var_583 = const()[name = tensor<string, []>("op_583"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_584_cast_fp16 = reshape(shape = var_583, x = linear_50_cast_fp16)[name = tensor<string, []>("op_584_cast_fp16")];
tensor<int32, [4]> value_states_35_perm_0 = const()[name = tensor<string, []>("value_states_35_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_8_cast_fp16 = mul(x = var_578_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_8_cast_fp16")];
tensor<bool, []> matmul_8_transpose_y_0 = const()[name = tensor<string, []>("matmul_8_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_8_transpose_x_0 = const()[name = tensor<string, []>("matmul_8_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_65_perm_0 = const()[name = tensor<string, []>("transpose_65_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_66_perm_0 = const()[name = tensor<string, []>("transpose_66_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_66 = transpose(perm = transpose_66_perm_0, x = var_581_cast_fp16)[name = tensor<string, []>("transpose_86")];
tensor<fp16, [1, 12, 77, 64]> transpose_65 = transpose(perm = transpose_65_perm_0, x = mul_8_cast_fp16)[name = tensor<string, []>("transpose_87")];
tensor<fp16, [1, 12, 77, 77]> matmul_8_cast_fp16 = matmul(transpose_x = matmul_8_transpose_x_0, transpose_y = matmul_8_transpose_y_0, x = transpose_65, y = transpose_66)[name = tensor<string, []>("matmul_8_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_8_cast_fp16 = add(x = matmul_8_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_8_cast_fp16")];
tensor<int32, []> softmax_8_axis_0 = const()[name = tensor<string, []>("softmax_8_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_8_cast_fp16 = softmax(axis = softmax_8_axis_0, x = add_8_cast_fp16)[name = tensor<string, []>("softmax_8_cast_fp16")];
tensor<bool, []> attn_output_33_transpose_x_0 = const()[name = tensor<string, []>("attn_output_33_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_33_transpose_y_0 = const()[name = tensor<string, []>("attn_output_33_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_35_cast_fp16 = transpose(perm = value_states_35_perm_0, x = var_584_cast_fp16)[name = tensor<string, []>("transpose_88")];
tensor<fp16, [1, 12, 77, 64]> attn_output_33_cast_fp16 = matmul(transpose_x = attn_output_33_transpose_x_0, transpose_y = attn_output_33_transpose_y_0, x = softmax_8_cast_fp16, y = value_states_35_cast_fp16)[name = tensor<string, []>("attn_output_33_cast_fp16")];
tensor<int32, [4]> attn_output_35_perm_0 = const()[name = tensor<string, []>("attn_output_35_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_588 = const()[name = tensor<string, []>("op_588"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_35_cast_fp16 = transpose(perm = attn_output_35_perm_0, x = attn_output_33_cast_fp16)[name = tensor<string, []>("transpose_85")];
tensor<fp16, [1, 77, 768]> input_101_cast_fp16 = reshape(shape = var_588, x = attn_output_35_cast_fp16)[name = tensor<string, []>("input_101_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(192982400)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(194162112)))];
tensor<fp16, [1, 77, 768]> linear_51_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_self_attn_out_proj_weight_to_fp16, x = input_101_cast_fp16)[name = tensor<string, []>("linear_51_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_103_cast_fp16 = add(x = input_99_cast_fp16, y = linear_51_cast_fp16)[name = tensor<string, []>("input_103_cast_fp16")];
tensor<int32, [1]> input_105_axes_0 = const()[name = tensor<string, []>("input_105_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(194163712)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(194165312)))];
tensor<fp16, [1, 77, 768]> input_105_cast_fp16 = layer_norm(axes = input_105_axes_0, beta = text_encoder_text_model_encoder_layers_8_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_8_layer_norm2_weight_to_fp16, x = input_103_cast_fp16)[name = tensor<string, []>("input_105_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(194166912)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(198885568)))];
tensor<fp16, [1, 77, 3072]> linear_52_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc1_weight_to_fp16, x = input_105_cast_fp16)[name = tensor<string, []>("linear_52_cast_fp16")];
tensor<fp16, []> var_603_to_fp16 = const()[name = tensor<string, []>("op_603_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_604_cast_fp16 = mul(x = linear_52_cast_fp16, y = var_603_to_fp16)[name = tensor<string, []>("op_604_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_605_cast_fp16 = sigmoid(x = var_604_cast_fp16)[name = tensor<string, []>("op_605_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_109_cast_fp16 = mul(x = linear_52_cast_fp16, y = var_605_cast_fp16)[name = tensor<string, []>("input_109_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(198891776)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(203610432)))];
tensor<fp16, [1, 77, 768]> linear_53_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_8_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_8_mlp_fc2_weight_to_fp16, x = input_109_cast_fp16)[name = tensor<string, []>("linear_53_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_111_cast_fp16 = add(x = input_103_cast_fp16, y = linear_53_cast_fp16)[name = tensor<string, []>("input_111_cast_fp16")];
tensor<int32, [1]> hidden_states_55_axes_0 = const()[name = tensor<string, []>("hidden_states_55_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(203612032)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(203613632)))];
tensor<fp16, [1, 77, 768]> hidden_states_55_cast_fp16 = layer_norm(axes = hidden_states_55_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm1_weight_to_fp16, x = input_111_cast_fp16)[name = tensor<string, []>("hidden_states_55_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(203615232)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(204794944)))];
tensor<fp16, [1, 77, 768]> linear_54_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_q_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor<string, []>("linear_54_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(204796544)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(205976256)))];
tensor<fp16, [1, 77, 768]> linear_55_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_k_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor<string, []>("linear_55_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(205977856)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(207157568)))];
tensor<fp16, [1, 77, 768]> linear_56_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_v_proj_weight_to_fp16, x = hidden_states_55_cast_fp16)[name = tensor<string, []>("linear_56_cast_fp16")];
tensor<int32, [4]> var_635 = const()[name = tensor<string, []>("op_635"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_636_cast_fp16 = reshape(shape = var_635, x = linear_54_cast_fp16)[name = tensor<string, []>("op_636_cast_fp16")];
tensor<int32, [4]> var_638 = const()[name = tensor<string, []>("op_638"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_639_cast_fp16 = reshape(shape = var_638, x = linear_55_cast_fp16)[name = tensor<string, []>("op_639_cast_fp16")];
tensor<int32, [4]> var_641 = const()[name = tensor<string, []>("op_641"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_642_cast_fp16 = reshape(shape = var_641, x = linear_56_cast_fp16)[name = tensor<string, []>("op_642_cast_fp16")];
tensor<int32, [4]> value_states_39_perm_0 = const()[name = tensor<string, []>("value_states_39_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_9_cast_fp16 = mul(x = var_636_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_9_cast_fp16")];
tensor<bool, []> matmul_9_transpose_y_0 = const()[name = tensor<string, []>("matmul_9_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_9_transpose_x_0 = const()[name = tensor<string, []>("matmul_9_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_67_perm_0 = const()[name = tensor<string, []>("transpose_67_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_68_perm_0 = const()[name = tensor<string, []>("transpose_68_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_68 = transpose(perm = transpose_68_perm_0, x = var_639_cast_fp16)[name = tensor<string, []>("transpose_82")];
tensor<fp16, [1, 12, 77, 64]> transpose_67 = transpose(perm = transpose_67_perm_0, x = mul_9_cast_fp16)[name = tensor<string, []>("transpose_83")];
tensor<fp16, [1, 12, 77, 77]> matmul_9_cast_fp16 = matmul(transpose_x = matmul_9_transpose_x_0, transpose_y = matmul_9_transpose_y_0, x = transpose_67, y = transpose_68)[name = tensor<string, []>("matmul_9_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_9_cast_fp16 = add(x = matmul_9_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_9_cast_fp16")];
tensor<int32, []> softmax_9_axis_0 = const()[name = tensor<string, []>("softmax_9_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_9_cast_fp16 = softmax(axis = softmax_9_axis_0, x = add_9_cast_fp16)[name = tensor<string, []>("softmax_9_cast_fp16")];
tensor<bool, []> attn_output_37_transpose_x_0 = const()[name = tensor<string, []>("attn_output_37_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_37_transpose_y_0 = const()[name = tensor<string, []>("attn_output_37_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_39_cast_fp16 = transpose(perm = value_states_39_perm_0, x = var_642_cast_fp16)[name = tensor<string, []>("transpose_84")];
tensor<fp16, [1, 12, 77, 64]> attn_output_37_cast_fp16 = matmul(transpose_x = attn_output_37_transpose_x_0, transpose_y = attn_output_37_transpose_y_0, x = softmax_9_cast_fp16, y = value_states_39_cast_fp16)[name = tensor<string, []>("attn_output_37_cast_fp16")];
tensor<int32, [4]> attn_output_39_perm_0 = const()[name = tensor<string, []>("attn_output_39_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_646 = const()[name = tensor<string, []>("op_646"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_39_cast_fp16 = transpose(perm = attn_output_39_perm_0, x = attn_output_37_cast_fp16)[name = tensor<string, []>("transpose_81")];
tensor<fp16, [1, 77, 768]> input_113_cast_fp16 = reshape(shape = var_646, x = attn_output_39_cast_fp16)[name = tensor<string, []>("input_113_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(207159168)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(208338880)))];
tensor<fp16, [1, 77, 768]> linear_57_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_self_attn_out_proj_weight_to_fp16, x = input_113_cast_fp16)[name = tensor<string, []>("linear_57_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_115_cast_fp16 = add(x = input_111_cast_fp16, y = linear_57_cast_fp16)[name = tensor<string, []>("input_115_cast_fp16")];
tensor<int32, [1]> input_117_axes_0 = const()[name = tensor<string, []>("input_117_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(208340480)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(208342080)))];
tensor<fp16, [1, 77, 768]> input_117_cast_fp16 = layer_norm(axes = input_117_axes_0, beta = text_encoder_text_model_encoder_layers_9_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_9_layer_norm2_weight_to_fp16, x = input_115_cast_fp16)[name = tensor<string, []>("input_117_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(208343680)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(213062336)))];
tensor<fp16, [1, 77, 3072]> linear_58_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc1_weight_to_fp16, x = input_117_cast_fp16)[name = tensor<string, []>("linear_58_cast_fp16")];
tensor<fp16, []> var_661_to_fp16 = const()[name = tensor<string, []>("op_661_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_662_cast_fp16 = mul(x = linear_58_cast_fp16, y = var_661_to_fp16)[name = tensor<string, []>("op_662_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_663_cast_fp16 = sigmoid(x = var_662_cast_fp16)[name = tensor<string, []>("op_663_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_121_cast_fp16 = mul(x = linear_58_cast_fp16, y = var_663_cast_fp16)[name = tensor<string, []>("input_121_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(213068544)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(217787200)))];
tensor<fp16, [1, 77, 768]> linear_59_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_9_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_9_mlp_fc2_weight_to_fp16, x = input_121_cast_fp16)[name = tensor<string, []>("linear_59_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_123_cast_fp16 = add(x = input_115_cast_fp16, y = linear_59_cast_fp16)[name = tensor<string, []>("input_123_cast_fp16")];
tensor<int32, [1]> hidden_states_61_axes_0 = const()[name = tensor<string, []>("hidden_states_61_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(217788800)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(217790400)))];
tensor<fp16, [1, 77, 768]> hidden_states_61_cast_fp16 = layer_norm(axes = hidden_states_61_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm1_weight_to_fp16, x = input_123_cast_fp16)[name = tensor<string, []>("hidden_states_61_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(217792000)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(218971712)))];
tensor<fp16, [1, 77, 768]> linear_60_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_q_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor<string, []>("linear_60_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(218973312)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(220153024)))];
tensor<fp16, [1, 77, 768]> linear_61_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_k_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor<string, []>("linear_61_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(220154624)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(221334336)))];
tensor<fp16, [1, 77, 768]> linear_62_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_v_proj_weight_to_fp16, x = hidden_states_61_cast_fp16)[name = tensor<string, []>("linear_62_cast_fp16")];
tensor<int32, [4]> var_693 = const()[name = tensor<string, []>("op_693"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_694_cast_fp16 = reshape(shape = var_693, x = linear_60_cast_fp16)[name = tensor<string, []>("op_694_cast_fp16")];
tensor<int32, [4]> var_696 = const()[name = tensor<string, []>("op_696"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_697_cast_fp16 = reshape(shape = var_696, x = linear_61_cast_fp16)[name = tensor<string, []>("op_697_cast_fp16")];
tensor<int32, [4]> var_699 = const()[name = tensor<string, []>("op_699"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_700_cast_fp16 = reshape(shape = var_699, x = linear_62_cast_fp16)[name = tensor<string, []>("op_700_cast_fp16")];
tensor<int32, [4]> value_states_43_perm_0 = const()[name = tensor<string, []>("value_states_43_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_10_cast_fp16 = mul(x = var_694_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_10_cast_fp16")];
tensor<bool, []> matmul_10_transpose_y_0 = const()[name = tensor<string, []>("matmul_10_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_10_transpose_x_0 = const()[name = tensor<string, []>("matmul_10_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_69_perm_0 = const()[name = tensor<string, []>("transpose_69_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_70_perm_0 = const()[name = tensor<string, []>("transpose_70_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_70 = transpose(perm = transpose_70_perm_0, x = var_697_cast_fp16)[name = tensor<string, []>("transpose_78")];
tensor<fp16, [1, 12, 77, 64]> transpose_69 = transpose(perm = transpose_69_perm_0, x = mul_10_cast_fp16)[name = tensor<string, []>("transpose_79")];
tensor<fp16, [1, 12, 77, 77]> matmul_10_cast_fp16 = matmul(transpose_x = matmul_10_transpose_x_0, transpose_y = matmul_10_transpose_y_0, x = transpose_69, y = transpose_70)[name = tensor<string, []>("matmul_10_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_10_cast_fp16 = add(x = matmul_10_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_10_cast_fp16")];
tensor<int32, []> softmax_10_axis_0 = const()[name = tensor<string, []>("softmax_10_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_10_cast_fp16 = softmax(axis = softmax_10_axis_0, x = add_10_cast_fp16)[name = tensor<string, []>("softmax_10_cast_fp16")];
tensor<bool, []> attn_output_41_transpose_x_0 = const()[name = tensor<string, []>("attn_output_41_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_41_transpose_y_0 = const()[name = tensor<string, []>("attn_output_41_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_43_cast_fp16 = transpose(perm = value_states_43_perm_0, x = var_700_cast_fp16)[name = tensor<string, []>("transpose_80")];
tensor<fp16, [1, 12, 77, 64]> attn_output_41_cast_fp16 = matmul(transpose_x = attn_output_41_transpose_x_0, transpose_y = attn_output_41_transpose_y_0, x = softmax_10_cast_fp16, y = value_states_43_cast_fp16)[name = tensor<string, []>("attn_output_41_cast_fp16")];
tensor<int32, [4]> attn_output_43_perm_0 = const()[name = tensor<string, []>("attn_output_43_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_704 = const()[name = tensor<string, []>("op_704"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_43_cast_fp16 = transpose(perm = attn_output_43_perm_0, x = attn_output_41_cast_fp16)[name = tensor<string, []>("transpose_77")];
tensor<fp16, [1, 77, 768]> input_125_cast_fp16 = reshape(shape = var_704, x = attn_output_43_cast_fp16)[name = tensor<string, []>("input_125_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(221335936)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(222515648)))];
tensor<fp16, [1, 77, 768]> linear_63_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_self_attn_out_proj_weight_to_fp16, x = input_125_cast_fp16)[name = tensor<string, []>("linear_63_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_127_cast_fp16 = add(x = input_123_cast_fp16, y = linear_63_cast_fp16)[name = tensor<string, []>("input_127_cast_fp16")];
tensor<int32, [1]> input_129_axes_0 = const()[name = tensor<string, []>("input_129_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(222517248)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(222518848)))];
tensor<fp16, [1, 77, 768]> input_129_cast_fp16 = layer_norm(axes = input_129_axes_0, beta = text_encoder_text_model_encoder_layers_10_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_10_layer_norm2_weight_to_fp16, x = input_127_cast_fp16)[name = tensor<string, []>("input_129_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(222520448)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(227239104)))];
tensor<fp16, [1, 77, 3072]> linear_64_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc1_weight_to_fp16, x = input_129_cast_fp16)[name = tensor<string, []>("linear_64_cast_fp16")];
tensor<fp16, []> var_719_to_fp16 = const()[name = tensor<string, []>("op_719_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_720_cast_fp16 = mul(x = linear_64_cast_fp16, y = var_719_to_fp16)[name = tensor<string, []>("op_720_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_721_cast_fp16 = sigmoid(x = var_720_cast_fp16)[name = tensor<string, []>("op_721_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_133_cast_fp16 = mul(x = linear_64_cast_fp16, y = var_721_cast_fp16)[name = tensor<string, []>("input_133_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(227245312)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(231963968)))];
tensor<fp16, [1, 77, 768]> linear_65_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_10_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_10_mlp_fc2_weight_to_fp16, x = input_133_cast_fp16)[name = tensor<string, []>("linear_65_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_135_cast_fp16 = add(x = input_127_cast_fp16, y = linear_65_cast_fp16)[name = tensor<string, []>("input_135_cast_fp16")];
tensor<int32, [1]> hidden_states_67_axes_0 = const()[name = tensor<string, []>("hidden_states_67_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(231965568)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(231967168)))];
tensor<fp16, [1, 77, 768]> hidden_states_67_cast_fp16 = layer_norm(axes = hidden_states_67_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm1_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm1_weight_to_fp16, x = input_135_cast_fp16)[name = tensor<string, []>("hidden_states_67_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(231968768)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(233148480)))];
tensor<fp16, [1, 77, 768]> linear_66_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_q_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor<string, []>("linear_66_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(233150080)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(234329792)))];
tensor<fp16, [1, 77, 768]> linear_67_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_k_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor<string, []>("linear_67_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(234331392)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(235511104)))];
tensor<fp16, [1, 77, 768]> linear_68_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_v_proj_weight_to_fp16, x = hidden_states_67_cast_fp16)[name = tensor<string, []>("linear_68_cast_fp16")];
tensor<int32, [4]> var_751 = const()[name = tensor<string, []>("op_751"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_752_cast_fp16 = reshape(shape = var_751, x = linear_66_cast_fp16)[name = tensor<string, []>("op_752_cast_fp16")];
tensor<int32, [4]> var_754 = const()[name = tensor<string, []>("op_754"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_755_cast_fp16 = reshape(shape = var_754, x = linear_67_cast_fp16)[name = tensor<string, []>("op_755_cast_fp16")];
tensor<int32, [4]> var_757 = const()[name = tensor<string, []>("op_757"), val = tensor<int32, [4]>([1, -1, 12, 64])];
tensor<fp16, [1, 77, 12, 64]> var_758_cast_fp16 = reshape(shape = var_757, x = linear_68_cast_fp16)[name = tensor<string, []>("op_758_cast_fp16")];
tensor<int32, [4]> value_states_perm_0 = const()[name = tensor<string, []>("value_states_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<fp16, [1, 77, 12, 64]> mul_11_cast_fp16 = mul(x = var_752_cast_fp16, y = var_17_to_fp16)[name = tensor<string, []>("mul_11_cast_fp16")];
tensor<bool, []> matmul_11_transpose_y_0 = const()[name = tensor<string, []>("matmul_11_transpose_y_0"), val = tensor<bool, []>(true)];
tensor<bool, []> matmul_11_transpose_x_0 = const()[name = tensor<string, []>("matmul_11_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<int32, [4]> transpose_71_perm_0 = const()[name = tensor<string, []>("transpose_71_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<int32, [4]> transpose_72_perm_0 = const()[name = tensor<string, []>("transpose_72_perm_0"), val = tensor<int32, [4]>([0, 2, -3, -1])];
tensor<fp16, [1, 12, 77, 64]> transpose_72 = transpose(perm = transpose_72_perm_0, x = var_755_cast_fp16)[name = tensor<string, []>("transpose_74")];
tensor<fp16, [1, 12, 77, 64]> transpose_71 = transpose(perm = transpose_71_perm_0, x = mul_11_cast_fp16)[name = tensor<string, []>("transpose_75")];
tensor<fp16, [1, 12, 77, 77]> matmul_11_cast_fp16 = matmul(transpose_x = matmul_11_transpose_x_0, transpose_y = matmul_11_transpose_y_0, x = transpose_71, y = transpose_72)[name = tensor<string, []>("matmul_11_cast_fp16")];
tensor<fp16, [1, 12, 77, 77]> add_11_cast_fp16 = add(x = matmul_11_cast_fp16, y = var_57_to_fp16)[name = tensor<string, []>("add_11_cast_fp16")];
tensor<int32, []> softmax_11_axis_0 = const()[name = tensor<string, []>("softmax_11_axis_0"), val = tensor<int32, []>(-1)];
tensor<fp16, [1, 12, 77, 77]> softmax_11_cast_fp16 = softmax(axis = softmax_11_axis_0, x = add_11_cast_fp16)[name = tensor<string, []>("softmax_11_cast_fp16")];
tensor<bool, []> attn_output_45_transpose_x_0 = const()[name = tensor<string, []>("attn_output_45_transpose_x_0"), val = tensor<bool, []>(false)];
tensor<bool, []> attn_output_45_transpose_y_0 = const()[name = tensor<string, []>("attn_output_45_transpose_y_0"), val = tensor<bool, []>(false)];
tensor<fp16, [1, 12, 77, 64]> value_states_cast_fp16 = transpose(perm = value_states_perm_0, x = var_758_cast_fp16)[name = tensor<string, []>("transpose_76")];
tensor<fp16, [1, 12, 77, 64]> attn_output_45_cast_fp16 = matmul(transpose_x = attn_output_45_transpose_x_0, transpose_y = attn_output_45_transpose_y_0, x = softmax_11_cast_fp16, y = value_states_cast_fp16)[name = tensor<string, []>("attn_output_45_cast_fp16")];
tensor<int32, [4]> attn_output_perm_0 = const()[name = tensor<string, []>("attn_output_perm_0"), val = tensor<int32, [4]>([0, 2, 1, 3])];
tensor<int32, [3]> var_762 = const()[name = tensor<string, []>("op_762"), val = tensor<int32, [3]>([1, 77, 768])];
tensor<fp16, [1, 77, 12, 64]> attn_output_cast_fp16 = transpose(perm = attn_output_perm_0, x = attn_output_45_cast_fp16)[name = tensor<string, []>("transpose_73")];
tensor<fp16, [1, 77, 768]> input_137_cast_fp16 = reshape(shape = var_762, x = attn_output_cast_fp16)[name = tensor<string, []>("input_137_cast_fp16")];
tensor<fp16, [768, 768]> text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16"), val = tensor<fp16, [768, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(235512704)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(236692416)))];
tensor<fp16, [1, 77, 768]> linear_69_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_self_attn_out_proj_weight_to_fp16, x = input_137_cast_fp16)[name = tensor<string, []>("linear_69_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_139_cast_fp16 = add(x = input_135_cast_fp16, y = linear_69_cast_fp16)[name = tensor<string, []>("input_139_cast_fp16")];
tensor<int32, [1]> input_141_axes_0 = const()[name = tensor<string, []>("input_141_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(236694016)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(236695616)))];
tensor<fp16, [1, 77, 768]> input_141_cast_fp16 = layer_norm(axes = input_141_axes_0, beta = text_encoder_text_model_encoder_layers_11_layer_norm2_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_encoder_layers_11_layer_norm2_weight_to_fp16, x = input_139_cast_fp16)[name = tensor<string, []>("input_141_cast_fp16")];
tensor<fp16, [3072, 768]> text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16"), val = tensor<fp16, [3072, 768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(236697216)))];
tensor<fp16, [3072]> text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16"), val = tensor<fp16, [3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(241415872)))];
tensor<fp16, [1, 77, 3072]> linear_70_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc1_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc1_weight_to_fp16, x = input_141_cast_fp16)[name = tensor<string, []>("linear_70_cast_fp16")];
tensor<fp16, []> var_777_to_fp16 = const()[name = tensor<string, []>("op_777_to_fp16"), val = tensor<fp16, []>(0x1.b3cp+0)];
tensor<fp16, [1, 77, 3072]> var_778_cast_fp16 = mul(x = linear_70_cast_fp16, y = var_777_to_fp16)[name = tensor<string, []>("op_778_cast_fp16")];
tensor<fp16, [1, 77, 3072]> var_779_cast_fp16 = sigmoid(x = var_778_cast_fp16)[name = tensor<string, []>("op_779_cast_fp16")];
tensor<fp16, [1, 77, 3072]> input_145_cast_fp16 = mul(x = linear_70_cast_fp16, y = var_779_cast_fp16)[name = tensor<string, []>("input_145_cast_fp16")];
tensor<fp16, [768, 3072]> text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16"), val = tensor<fp16, [768, 3072]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(241422080)))];
tensor<fp16, [768]> text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(246140736)))];
tensor<fp16, [1, 77, 768]> linear_71_cast_fp16 = linear(bias = text_encoder_text_model_encoder_layers_11_mlp_fc2_bias_to_fp16, weight = text_encoder_text_model_encoder_layers_11_mlp_fc2_weight_to_fp16, x = input_145_cast_fp16)[name = tensor<string, []>("linear_71_cast_fp16")];
tensor<fp16, [1, 77, 768]> input_cast_fp16 = add(x = input_139_cast_fp16, y = linear_71_cast_fp16)[name = tensor<string, []>("input_cast_fp16")];
tensor<int32, [1]> last_hidden_state_axes_0 = const()[name = tensor<string, []>("last_hidden_state_axes_0"), val = tensor<int32, [1]>([-1])];
tensor<fp16, [768]> text_encoder_text_model_final_layer_norm_weight_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_final_layer_norm_weight_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(246142336)))];
tensor<fp16, [768]> text_encoder_text_model_final_layer_norm_bias_to_fp16 = const()[name = tensor<string, []>("text_encoder_text_model_final_layer_norm_bias_to_fp16"), val = tensor<fp16, [768]>(BLOBFILE(path = tensor<string, []>("@model_path/weights/weight.bin"), offset = tensor<uint64, []>(246143936)))];
tensor<fp16, [1, 77, 768]> last_hidden_state_cast_fp16 = layer_norm(axes = last_hidden_state_axes_0, beta = text_encoder_text_model_final_layer_norm_bias_to_fp16, epsilon = var_15_to_fp16, gamma = text_encoder_text_model_final_layer_norm_weight_to_fp16, x = input_cast_fp16)[name = tensor<string, []>("last_hidden_state_cast_fp16")];
tensor<string, []> last_hidden_state_cast_fp16_to_fp32_dtype_0 = const()[name = tensor<string, []>("last_hidden_state_cast_fp16_to_fp32_dtype_0"), val = tensor<string, []>("fp32")];
tensor<int32, [1]> var_790 = const()[name = tensor<string, []>("op_790"), val = tensor<int32, [1]>([0])];
tensor<int32, []> var_792_axis_0 = const()[name = tensor<string, []>("op_792_axis_0"), val = tensor<int32, []>(-1)];
tensor<bool, []> var_792_keep_dims_0 = const()[name = tensor<string, []>("op_792_keep_dims_0"), val = tensor<bool, []>(false)];
tensor<int32, [1]> var_792 = reduce_argmax(axis = var_792_axis_0, keep_dims = var_792_keep_dims_0, x = cast_1)[name = tensor<string, []>("op_792")];
tensor<int32, []> stack_0_axis_0 = const()[name = tensor<string, []>("stack_0_axis_0"), val = tensor<int32, []>(1)];
tensor<int32, [1, 2]> stack_0 = stack(axis = stack_0_axis_0, values = (var_790, var_792))[name = tensor<string, []>("stack_0")];
tensor<int32, []> var_794_transpose_batch_dims_0 = const()[name = tensor<string, []>("op_794_transpose_batch_dims_0"), val = tensor<int32, []>(0)];
tensor<fp16, [1, 768]> var_794_transpose_cast_fp16 = gather_nd(batch_dims = var_794_transpose_batch_dims_0, indices = stack_0, x = last_hidden_state_cast_fp16)[name = tensor<string, []>("op_794_transpose_cast_fp16")];
tensor<string, []> var_794_cast_fp16_to_fp32_dtype_0 = const()[name = tensor<string, []>("op_794_cast_fp16_to_fp32_dtype_0"), val = tensor<string, []>("fp32")];
tensor<fp32, [1, 768]> pooled_outputs = cast(dtype = var_794_cast_fp16_to_fp32_dtype_0, x = var_794_transpose_cast_fp16)[name = tensor<string, []>("cast_54")];
tensor<fp32, [1, 77, 768]> last_hidden_state = cast(dtype = last_hidden_state_cast_fp16_to_fp32_dtype_0, x = last_hidden_state_cast_fp16)[name = tensor<string, []>("cast_55")];
} -> (last_hidden_state, pooled_outputs);
}