program(1.0) [buildInfo = dict, tensor>({{"coremlc-component-MIL", "3500.14.1"}, {"coremlc-version", "3500.32.1"}, {"coremltools-component-torch", "2.4.0"}, {"coremltools-source-dialect", "TorchScript"}, {"coremltools-version", "8.3.0"}})] { func main(tensor c_in, tensor h_in, tensor target_length, tensor targets) { tensor y_axis_0 = const()[name = tensor("y_axis_0"), val = tensor(0)]; tensor y_batch_dims_0 = const()[name = tensor("y_batch_dims_0"), val = tensor(0)]; tensor y_validate_indices_0 = const()[name = tensor("y_validate_indices_0"), val = tensor(false)]; tensor module_prediction_embed_weight_to_fp16 = const()[name = tensor("module_prediction_embed_weight_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(64)))]; tensor targets_to_int16_dtype_0 = const()[name = tensor("targets_to_int16_dtype_0"), val = tensor("int16")]; tensor targets_to_int16 = cast(dtype = targets_to_int16_dtype_0, x = targets)[name = tensor("cast_8")]; tensor y_cast_fp16_cast_uint16 = gather(axis = y_axis_0, batch_dims = y_batch_dims_0, indices = targets_to_int16, validate_indices = y_validate_indices_0, x = module_prediction_embed_weight_to_fp16)[name = tensor("y_cast_fp16_cast_uint16")]; tensor input_3_perm_0 = const()[name = tensor("input_3_perm_0"), val = tensor([1, 0, 2])]; tensor input_lstm_h0_squeeze_axes_0 = const()[name = tensor("input_lstm_h0_squeeze_axes_0"), val = tensor([0])]; tensor h_in_to_fp16_dtype_0 = const()[name = tensor("h_in_to_fp16_dtype_0"), val = tensor("fp16")]; tensor h_in_to_fp16 = cast(dtype = h_in_to_fp16_dtype_0, x = h_in)[name = tensor("cast_7")]; tensor input_lstm_h0_squeeze_cast_fp16 = squeeze(axes = input_lstm_h0_squeeze_axes_0, x = h_in_to_fp16)[name = tensor("input_lstm_h0_squeeze_cast_fp16")]; tensor input_lstm_c0_squeeze_axes_0 = const()[name = tensor("input_lstm_c0_squeeze_axes_0"), val = tensor([0])]; tensor c_in_to_fp16_dtype_0 = const()[name = tensor("c_in_to_fp16_dtype_0"), val = tensor("fp16")]; tensor c_in_to_fp16 = cast(dtype = c_in_to_fp16_dtype_0, x = c_in)[name = tensor("cast_6")]; tensor input_lstm_c0_squeeze_cast_fp16 = squeeze(axes = input_lstm_c0_squeeze_axes_0, x = c_in_to_fp16)[name = tensor("input_lstm_c0_squeeze_cast_fp16")]; tensor input_direction_0 = const()[name = tensor("input_direction_0"), val = tensor("forward")]; tensor input_output_sequence_0 = const()[name = tensor("input_output_sequence_0"), val = tensor(true)]; tensor input_recurrent_activation_0 = const()[name = tensor("input_recurrent_activation_0"), val = tensor("sigmoid")]; tensor input_cell_activation_0 = const()[name = tensor("input_cell_activation_0"), val = tensor("tanh")]; tensor input_activation_0 = const()[name = tensor("input_activation_0"), val = tensor("tanh")]; tensor concat_1_to_fp16 = const()[name = tensor("concat_1_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(1314688)))]; tensor concat_2_to_fp16 = const()[name = tensor("concat_2_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(4591552)))]; tensor concat_0_to_fp16 = const()[name = tensor("concat_0_to_fp16"), val = tensor(BLOBFILE(path = tensor("@model_path/weights/weight.bin"), offset = tensor(7868416)))]; tensor input_3_cast_fp16 = transpose(perm = input_3_perm_0, x = y_cast_fp16_cast_uint16)[name = tensor("transpose_2")]; tensor input_cast_fp16_0, tensor input_cast_fp16_1, tensor input_cast_fp16_2 = lstm(activation = input_activation_0, bias = concat_0_to_fp16, cell_activation = input_cell_activation_0, direction = input_direction_0, initial_c = input_lstm_c0_squeeze_cast_fp16, initial_h = input_lstm_h0_squeeze_cast_fp16, output_sequence = input_output_sequence_0, recurrent_activation = input_recurrent_activation_0, weight_hh = concat_2_to_fp16, weight_ih = concat_1_to_fp16, x = input_3_cast_fp16)[name = tensor("input_cast_fp16")]; tensor obj_3_axes_0 = const()[name = tensor("obj_3_axes_0"), val = tensor([0])]; tensor obj_3_cast_fp16 = expand_dims(axes = obj_3_axes_0, x = input_cast_fp16_1)[name = tensor("obj_3_cast_fp16")]; tensor obj_3_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_3_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor obj_axes_0 = const()[name = tensor("obj_axes_0"), val = tensor([0])]; tensor obj_cast_fp16 = expand_dims(axes = obj_axes_0, x = input_cast_fp16_2)[name = tensor("obj_cast_fp16")]; tensor obj_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("obj_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor transpose_0_perm_0 = const()[name = tensor("transpose_0_perm_0"), val = tensor([1, 2, 0])]; tensor transpose_0_cast_fp16_to_fp32_dtype_0 = const()[name = tensor("transpose_0_cast_fp16_to_fp32_dtype_0"), val = tensor("fp32")]; tensor transpose_0_cast_fp16 = transpose(perm = transpose_0_perm_0, x = input_cast_fp16_0)[name = tensor("transpose_1")]; tensor decoder = cast(dtype = transpose_0_cast_fp16_to_fp32_dtype_0, x = transpose_0_cast_fp16)[name = tensor("cast_3")]; tensor c_out = cast(dtype = obj_cast_fp16_to_fp32_dtype_0, x = obj_cast_fp16)[name = tensor("cast_4")]; tensor h_out = cast(dtype = obj_3_cast_fp16_to_fp32_dtype_0, x = obj_3_cast_fp16)[name = tensor("cast_5")]; tensor target_length_tmp = identity(x = target_length)[name = tensor("target_length_tmp")]; } -> (decoder, h_out, c_out); }