diff --git "a/MobileViT_DeepLabV3.mlpackage/Data/com.apple.CoreML/model.mlmodel" "b/MobileViT_DeepLabV3.mlpackage/Data/com.apple.CoreML/model.mlmodel" --- "a/MobileViT_DeepLabV3.mlpackage/Data/com.apple.CoreML/model.mlmodel" +++ "b/MobileViT_DeepLabV3.mlpackage/Data/com.apple.CoreML/model.mlmodel" @@ -3,20 +3,20 @@ image Image input "��R. classLabelsSegmentation map * -�������# -transformers_version 4.20.0.dev0�� -classes�background,aeroplane,bicycle,bird,boat,bottle,bus,car,cat,chair,cow,diningtable,dog,horse,motorbike,person,pottedplant,sheep,sofa,train,tvmonitor�4 +�������� +classes�background,aeroplane,bicycle,bird,boat,bottle,bus,car,cat,chair,cow,diningtable,dog,horse,motorbike,person,pottedplant,sheep,sofa,train,tvmonitor�# +transformers_version 4.20.0.dev0�4 #com.github.apple.coremltools.source torch==1.11.0�- -$com.github.apple.coremltools.version5.2.0����� -main�� +$com.github.apple.coremltools.version5.2.0����� +main�� ) image     � -�CoreML5�� -CoreML5�� classLabelsj +�CoreML5�� +CoreML5�� classLabelsj const image__scaled___y_0  * @@ -42,1969 +42,2145 @@ $com.github.apple.coremltools.version5.2.0   � -�� -constY -6model_mobilevit_encoder_layer_2_1_conv_1x1_conv_weight -  -� -` +�� +constX +6model_mobilevit_encoder_layer_2_1_conv_1x1_conv_weight +  +@ +0  -*L +*K +valD +  +@ +0 + +*" +@model_path/weights/weight.bin@*L nameD < :"8 -6model_mobilevit_encoder_layer_2_1_conv_1x1_conv_weight*L -valE -  -� -` - -*" -@model_path/weights/weight.bin@� -constV -Emodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*[ +6model_mobilevit_encoder_layer_2_1_conv_1x1_conv_weight� +constU +Emodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_bias + +  +@*: +val3 + +  +@*# +@model_path/weights/weight.bin�a*[ nameS K I"G -Emodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_bias� -constX -Gmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_weight -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*] +Emodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_bias� +constW +Gmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_weight + +  +@*: +val3 + +  +@*# +@model_path/weights/weight.bin�c*] nameU M K"I -Gmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_weight� -const_ -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*d +Gmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_bias + +  +@*d name\ T R"P -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_bias*: +val3 + +  +@*# +@model_path/weights/weight.bin�f� +constf +Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_weight +  +@ +@*@ +val9 +  +@ +@*# +@model_path/weights/weight.bin�h*f name^ V T"R -Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��� -const] -Lmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*b +Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_weight� +const\ +Lmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin��*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_bias� -constf -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_bias� +constd +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight +  +@ +@*d name\ T R"P -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��� -const_ -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin�� *d +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight*A +val: +  +@ +@*$ +@model_path/weights/weight.bin��� +const^ +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias + +  +@*d name\ T R"P -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin�� *f +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias*; +val4 + +  +@*$ +@model_path/weights/weight.bin��� +constf +Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_weight +  +@ +@*A +val: +  +@ +@*$ +@model_path/weights/weight.bin��*f name^ V T"R -Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_weight� -const\ -Kmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_weight� +const[ +Kmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin��*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��� -conste -Mmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_weight -  -� -�*c +Kmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_bias� +constc +Mmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_weight +  +@ +@*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��� -constU -Dmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*Z +Mmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_weight*A +val: +  +@ +@*$ +@model_path/weights/weight.bin��� +constT +Dmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin��*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_bias� -constW -Fmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_weight -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*\ +Dmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_bias� +constV +Fmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_weight + +  +@*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_weight� +Fmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_weight*; +val4 + +  +@*$ +@model_path/weights/weight.bin��� constX Gmodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_bias   -�*] +�*< +val5 +   +�*$ +@model_path/weights/weight.bin��*] nameU M K"I -Gmodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��� -consta -Imodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��*_ +Gmodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_bias� +const` +Imodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_weight +  +� +@*B +val; +  +� +@*$ +@model_path/weights/weight.bin��*_ nameW O M"K -Imodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_weight� -constR -Amodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_bias -   -�*W +Imodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_weight� +constQ +Amodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_bias + +  +@*W nameO G E"C -Amodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��"� -const[ -Cmodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��"*Y +Amodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_bias*; +val4 + +  +@*$ +@model_path/weights/weight.bin��� +constZ +Cmodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_weight +  +@ +�*B +val; +  +@ +�*$ +@model_path/weights/weight.bin��*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_weight� -constV -Emodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_bias -   -�*[ +Cmodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_weight� +constU +Emodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin��*[ nameS K I"G -Emodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��,� -constX -Gmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_weight -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��,*] +Emodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_bias� +constW +Gmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_weight + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� *] nameU M K"I -Gmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_weight� -const_ -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��,*d +Gmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� *d name\ T R"P -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_bias� +constf +Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_weight +  +@ +@*A +val: +  +@ +@*$ +@model_path/weights/weight.bin�� *f name^ V T"R -Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��,� -const] -Lmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��1*b +Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_weight� +const\ +Lmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_bias + +  +@*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_bias� -constf -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_bias*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� +� +constd +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_weight +  +@ +@*d name\ T R"P -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��1� -const_ -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��6*d +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_weight*A +val: +  +@ +@*$ +@model_path/weights/weight.bin�� +� +const^ +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� *d name\ T R"P -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��6*f +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_bias� +constf +Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weight +  +@ +@*A +val: +  +@ +@*$ +@model_path/weights/weight.bin�� *f name^ V T"R -Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weight� -const\ -Kmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weight� +const[ +Kmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_bias + +  +@*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��;� -conste -Mmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��;*c +Kmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_bias*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� � +constc +Mmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_weight +  +@ +@*A +val: +  +@ +@*$ +@model_path/weights/weight.bin�� *c name[ S Q"O -Mmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_weight� -constU -Dmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��@*Z +Mmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_weight� +constT +Dmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_bias + +  +@*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� *Z nameR J H"F -Dmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_bias� -constW -Fmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_weight -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��@*\ +Dmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_bias� +constV +Fmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_weight + +  +@*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_weight� +Fmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_weight*; +val4 + +  +@*$ +@model_path/weights/weight.bin�� � constX Gmodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_bias   -�*] +�*] nameU M K"I Gmodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_bias*< val5   -�*$ -@model_path/weights/weight.bin��@� -consta -Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight -  -� -�*_ +�*$ +@model_path/weights/weight.bin�� � +const` +Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight +  +� +@*B +val; +  +� +@*$ +@model_path/weights/weight.bin�� *_ nameW O M"K -Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��@� -constR -Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��K*W +Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight� +constQ +Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias + +  +@*W nameO G E"C -Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias� -const[ -Cmodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��K*Y +Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias*; +val4 + +  +@*$ +@model_path/weights/weight.bin��� +constZ +Cmodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_weight +  +@ +�*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_weight� -constA -0model_mobilevit_encoder_layer_2_1_layernorm_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��U*F +Cmodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_weight*B +val; +  +@ +�*$ +@model_path/weights/weight.bin��� +const@ +0model_mobilevit_encoder_layer_2_1_layernorm_bias + +  +@*F name> 6 4"2 -0model_mobilevit_encoder_layer_2_1_layernorm_bias� -constC -2model_mobilevit_encoder_layer_2_1_layernorm_weight -   -�*H +0model_mobilevit_encoder_layer_2_1_layernorm_bias*; +val4 + +  +@*$ +@model_path/weights/weight.bin��� +constB +2model_mobilevit_encoder_layer_2_1_layernorm_weight + +  +@*H name@ 8 6"4 -2model_mobilevit_encoder_layer_2_1_layernorm_weight*< -val5 -   -�*$ -@model_path/weights/weight.bin��U� -constZ -6model_mobilevit_encoder_layer_3_1_conv_1x1_conv_weight -  -� -� +2model_mobilevit_encoder_layer_2_1_layernorm_weight*; +val4 + +  +@*$ +@model_path/weights/weight.bin��� +constX +6model_mobilevit_encoder_layer_3_1_conv_1x1_conv_weight +  +P +@  -*L +*M +valF +  +P +@ + +*$ +@model_path/weights/weight.bin��*L nameD < :"8 -6model_mobilevit_encoder_layer_3_1_conv_1x1_conv_weight*O -valH -  -� -� - -*$ -@model_path/weights/weight.bin��U� -constV -Emodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_bias -   -�*[ +6model_mobilevit_encoder_layer_3_1_conv_1x1_conv_weight� +constU +Emodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��*[ nameS K I"G -Emodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��[� -constX -Gmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_weight -   -�*] +Emodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_bias� +constW +Gmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_weight + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_weight*< -val5 -   -�*$ -@model_path/weights/weight.bin��[� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��[*d +Gmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_bias + +  +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_weight +  +P +P*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_weight*C -val< -  -� -�*$ -@model_path/weights/weight.bin��[� -const] -Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��� +const\ +Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��d� -constf -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��d*d +Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias� +constd +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_weight +  +P +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_weight� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��m*d +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��m*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_weight +  +P +P*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_weight� -const\ -Kmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��� +const[ +Kmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_bias + +  +P*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��v� -conste -Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��v*c +Kmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��� +constc +Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight +  +P +P*A +val: +  +P +P*$ +@model_path/weights/weight.bin��*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight� -constU -Dmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_bias -   -�*Z +Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight� +constT +Dmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_bias*< -val5 -   -�*$ -@model_path/weights/weight.bin��� -constW -Fmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_weight -   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*\ +Dmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_bias� +constV +Fmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_weight + +  +P*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_weight� +Fmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_weight*; +val4 + +  +P*$ +@model_path/weights/weight.bin��� constX Gmodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_bias   -�*< -val5 -   -�*$ -@model_path/weights/weight.bin��*] +�*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_bias� -consta -Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weight -  -� -�*C -val< -  -� -�*$ -@model_path/weights/weight.bin��*_ +Gmodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_bias*< +val5 +   +�*$ +@model_path/weights/weight.bin��� +const` +Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weight +  +� +P*B +val; +  +� +P*$ +@model_path/weights/weight.bin��*_ nameW O M"K -Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weight� -constR -Amodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_bias -   -�*W +Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weight� +constQ +Amodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��*W nameO G E"C -Amodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin��� -const[ -Cmodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*Y +Amodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_bias� +constZ +Cmodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_weight +  +P +�*B +val; +  +P +�*$ +@model_path/weights/weight.bin��*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_weight� -constV -Emodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_bias -   -�*[ +Cmodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_weight� +constU +Emodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_bias + +  +P*[ nameS K I"G -Emodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constX -Gmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_weight -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*] +Emodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��� +constW +Gmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_weight + +  +P*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_weight� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_bias -   -�*d +Gmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_weight*; +val4 + +  +P*$ +@model_path/weights/weight.bin��� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_bias + +  +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_weight +  +P +P*A +val: +  +P +P*$ +@model_path/weights/weight.bin��*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const] -Lmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_weight� +const\ +Lmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_bias + +  +P*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constf -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin�� � +constd +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_weight +  +P +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*d +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin�� � +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_bias + +  +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��"� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weight +  +P +P*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weight� -const\ -Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��"� +const[ +Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��$*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -conste -Mmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*c +Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias� +constc +Mmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_weight +  +P +P*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_weight� -constU -Dmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*Z +Mmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��$� +constT +Dmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_bias + +  +P*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_bias� -constW -Fmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_weight -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*\ +Dmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��%� +constV +Fmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_weight + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��%*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_weight� +Fmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_weight� constX Gmodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_bias   -�*] +�*< +val5 +   +�*$ +@model_path/weights/weight.bin��%*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -consta -Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*_ +Gmodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_bias� +const` +Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weight +  +� +P*B +val; +  +� +P*$ +@model_path/weights/weight.bin��%*_ nameW O M"K -Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weight� -constR -Amodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*W +Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weight� +constQ +Amodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_bias + +  +P*W nameO G E"C -Amodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_bias� -const[ -Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight -  -� -�*Y +Amodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��(� +constZ +Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight +  +P +�*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constV -Emodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_bias -   -�*[ +Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight*B +val; +  +P +�*$ +@model_path/weights/weight.bin��(� +constU +Emodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��,*[ nameS K I"G -Emodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constX -Gmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_weight -   -�*] +Emodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_bias� +constW +Gmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_weight + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��,*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_bias -   -�*d +Gmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��,*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_bias� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_weight +  +P +P*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const] -Lmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��,� +const\ +Lmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_bias + +  +P*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constf -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*d +Lmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��-� +constd +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_weight +  +P +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_weight� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias -   -�*d +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��-� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias + +  +P*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��/� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_weight +  +P +P*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_weight� -const\ -Kmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��/� +const[ +Kmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��0*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin��� -conste -Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight -  -� -�*c +Kmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_bias� +constc +Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight +  +P +P*A +val: +  +P +P*$ +@model_path/weights/weight.bin��0*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin��� -constU -Dmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin��*Z +Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight� +constT +Dmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��2*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_bias� -constW -Fmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_weight -   -�*\ +Dmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_bias� +constV +Fmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_weight + +  +P*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_weight*= -val6 -   -�*% -@model_path/weights/weight.bin��� +Fmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_weight*; +val4 + +  +P*$ +@model_path/weights/weight.bin��2� constX Gmodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_bias   -�*] +�*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_bias*= -val6 +Gmodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_bias*< +val5   -�*% -@model_path/weights/weight.bin���� -consta -Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight -  -� -�*_ +�*$ +@model_path/weights/weight.bin��2� +const` +Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight +  +� +P*_ nameW O M"K -Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constR -Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias -   -�*W +Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight*B +val; +  +� +P*$ +@model_path/weights/weight.bin��2� +constQ +Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��5*W nameO G E"C -Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -const[ -Cmodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*Y +Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias� +constZ +Cmodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_weight +  +P +�*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_weight� -constV -Emodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*[ +Cmodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_weight*B +val; +  +P +�*$ +@model_path/weights/weight.bin��5� +constU +Emodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��8*[ nameS K I"G -Emodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_bias� -constX -Gmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_weight -   -�*] +Emodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_bias� +constW +Gmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_weight + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��8*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*d +Gmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��8*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_bias� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_weight +  +P +P*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const] -Lmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_weight*A +val: +  +P +P*$ +@model_path/weights/weight.bin��8� +const\ +Lmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_bias + +  +P*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constf -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��:� +constd +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight +  +P +P*A +val: +  +P +P*$ +@model_path/weights/weight.bin��:*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_bias -   -�*d +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight� +const^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��<*d name\ T R"P -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -consth -Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_bias� +constf +Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weight +  +P +P*A +val: +  +P +P*$ +@model_path/weights/weight.bin��<*f name^ V T"R -Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const\ -Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weight� +const[ +Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias + +  +P*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -conste -Mmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*c +Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��=� +constc +Mmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_weight +  +P +P*A +val: +  +P +P*$ +@model_path/weights/weight.bin��=*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_weight� -constU -Dmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*Z +Mmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_weight� +constT +Dmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_bias + +  +P*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_bias� -constW -Fmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_weight -   -�*\ +Dmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_bias*; +val4 + +  +P*$ +@model_path/weights/weight.bin��?� +constV +Fmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_weight + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��?*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� +Fmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_weight� constX Gmodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_bias   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*] +�*] nameU M K"I -Gmodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_bias� -consta -Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight -  -� -�*_ +Gmodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_bias*< +val5 +   +�*$ +@model_path/weights/weight.bin��?� +const` +Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight +  +� +P*_ nameW O M"K -Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constR -Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*W +Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight*B +val; +  +� +P*$ +@model_path/weights/weight.bin��?� +constQ +Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��B*W nameO G E"C -Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_bias� -const[ -Cmodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_weight -  -� -�*Y +Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_bias� +constZ +Cmodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_weight +  +P +�*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constA -0model_mobilevit_encoder_layer_3_1_layernorm_bias -   -�*F +Cmodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_weight*B +val; +  +P +�*$ +@model_path/weights/weight.bin��B� +const@ +0model_mobilevit_encoder_layer_3_1_layernorm_bias + +  +P*; +val4 + +  +P*$ +@model_path/weights/weight.bin��E*F name> 6 4"2 -0model_mobilevit_encoder_layer_3_1_layernorm_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constC -2model_mobilevit_encoder_layer_3_1_layernorm_weight -   -�*H +0model_mobilevit_encoder_layer_3_1_layernorm_bias� +constB +2model_mobilevit_encoder_layer_3_1_layernorm_weight + +  +P*H name@ 8 6"4 -2model_mobilevit_encoder_layer_3_1_layernorm_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constZ -6model_mobilevit_encoder_layer_4_1_conv_1x1_conv_weight -  -� -� +2model_mobilevit_encoder_layer_3_1_layernorm_weight*; +val4 + +  +P*$ +@model_path/weights/weight.bin��E� +constX +6model_mobilevit_encoder_layer_4_1_conv_1x1_conv_weight +  +` +P  *L nameD < :"8 -6model_mobilevit_encoder_layer_4_1_conv_1x1_conv_weight*P -valI -  -� -� +6model_mobilevit_encoder_layer_4_1_conv_1x1_conv_weight*M +valF +  +` +P  -*% -@model_path/weights/weight.bin���� -constV -Emodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*[ +*$ +@model_path/weights/weight.bin��E� +constU +Emodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_bias + +  +`*[ nameS K I"G -Emodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_bias� -constX -Gmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_weight -   -�*] +Emodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��G� +constW +Gmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_weight + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��G*] nameU -M -K"I -Gmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*d +M +K"I +Gmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_bias + +  +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*f +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��G� +constf +Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_weight +  +` +`*A +val: +  +` +`*$ +@model_path/weights/weight.bin��G*f name^ V T"R -Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_weight� -const] -Lmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_weight� +const\ +Lmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��I*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constf -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_bias� +constd +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_weight +  +` +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin�ţ*d +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��I� +const^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias + +  +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin�ͣ*f +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��L� +constf +Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_weight +  +` +`*f name^ V T"R -Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_weight� -const\ -Kmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��L� +const[ +Kmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��N*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin�ձ� -conste -Mmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin�ݱ*c +Kmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_bias� +constc +Mmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_weight +  +` +`*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_weight� -constU -Dmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_bias -   -�*Z +Mmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��N� +constT +Dmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��P*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_bias*= -val6 -   -�*% -@model_path/weights/weight.bin��� -constW -Fmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_weight -   -�*\ +Dmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_bias� +constV +Fmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_weight + +  +`*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_weight*= -val6 -   -�*% -@model_path/weights/weight.bin��� +Fmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_weight*; +val4 + +  +`*$ +@model_path/weights/weight.bin��P� constX Gmodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_bias   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*] +�*] nameU M K"I -Gmodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_bias� -consta -Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight -  -� -�*_ +Gmodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_bias*< +val5 +   +�*$ +@model_path/weights/weight.bin��P� +const` +Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight +  +� +`*B +val; +  +� +`*$ +@model_path/weights/weight.bin��P*_ nameW O M"K -Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constR -Amodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*W +Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight� +constQ +Amodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��U*W nameO G E"C -Amodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_bias� -const[ -Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weight -  -� -�*Y +Amodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_bias� +constZ +Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weight +  +` +�*B +val; +  +` +�*$ +@model_path/weights/weight.bin��U*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constV -Emodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*[ +Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weight� +constU +Emodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��Y*[ nameS K I"G -Emodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_bias� -constX -Gmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_weight -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*] +Emodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_bias� +constW +Gmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_weight + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��Y*] nameU M K"I -Gmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_weight� -const_ -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*d +Gmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_bias + +  +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��Y� +constf +Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_weight +  +` +`*A +val: +  +` +`*$ +@model_path/weights/weight.bin��Y*f name^ V T"R -Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const] -Lmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_weight� +const\ +Lmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��\*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_bias*= -val6 -   -�*% -@model_path/weights/weight.bin�φ� -constf -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_bias� +constd +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight +  +` +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin�׆� -const_ -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin�ߔ*d +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��\� +const^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_bias + +  +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��^� +constf +Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weight +  +` +`*f name^ V T"R -Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin��� -const\ -Kmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_bias -   -�*a +Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��^� +const[ +Kmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��`*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin��� -conste -Mmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_weight -  -� -�*c +Kmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_bias� +constc +Mmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_weight +  +` +`*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constU -Dmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*Z +Mmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��`� +constT +Dmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_bias + +  +`*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_bias� -constW -Fmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_weight -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*\ +Dmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��c� +constV +Fmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_weight + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��c*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_weight� +Fmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_weight� constX Gmodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_bias   -�*] +�*< +val5 +   +�*$ +@model_path/weights/weight.bin��c*] nameU M K"I -Gmodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -consta -Imodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_weight -  -� -�*_ +Gmodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_bias� +const` +Imodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_weight +  +� +`*B +val; +  +� +`*$ +@model_path/weights/weight.bin��c*_ nameW O M"K -Imodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constR -Amodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*W +Imodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_weight� +constQ +Amodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_bias + +  +`*W nameO G E"C -Amodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_bias� -const[ -Cmodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_weight -  -� -�*Y +Amodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��g� +constZ +Cmodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_weight +  +` +�*B +val; +  +` +�*$ +@model_path/weights/weight.bin��g*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constV -Emodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_bias -   -�*[ +Cmodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_weight� +constU +Emodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_bias + +  +`*[ nameS K I"G -Emodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constX -Gmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_weight -   -�*] +Emodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��l� +constW +Gmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_weight + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��l*] nameU M K"I -Gmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*d +Gmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_weight� +const^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��l*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_bias� -consth -Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*f +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_bias� +constf +Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_weight +  +` +`*f name^ V T"R -Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_weight� -const] -Lmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_bias -   -�*b +Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��l� +const\ +Lmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_bias + +  +`*b nameZ R P"N -Lmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constf -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_weight -  -� -�*d +Lmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��n� +constd +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_weight +  +` +`*A +val: +  +` +`*$ +@model_path/weights/weight.bin��n*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const_ -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*d +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_weight� +const^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_bias + +  +`*d name\ T R"P -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_bias� -consth -Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weight -  -� -�*f +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��p� +constf +Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weight +  +` +`*f name^ V T"R -Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -const\ -Kmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*a +Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��p� +const[ +Kmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��s*a nameY Q O"M -Kmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_bias� -conste -Mmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_weight -  -� -�*c +Kmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_bias� +constc +Mmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_weight +  +` +`*c name[ S Q"O -Mmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_weight*D -val= -  -� -�*% -@model_path/weights/weight.bin���� -constU -Dmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_bias -   -�*Z +Mmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_weight*A +val: +  +` +`*$ +@model_path/weights/weight.bin��s� +constT +Dmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��u*Z nameR J H"F -Dmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_bias*= -val6 -   -�*% -@model_path/weights/weight.bin���� -constW -Fmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_weight -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*\ +Dmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_bias� +constV +Fmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_weight + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��u*\ nameT L J"H -Fmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_weight� +Fmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_weight� constX Gmodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_bias   -�*= -val6 +�*< +val5   -�*% -@model_path/weights/weight.bin���*] +�*$ +@model_path/weights/weight.bin��u*] nameU M K"I -Gmodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_bias� -consta -Imodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin���*_ +Gmodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_bias� +const` +Imodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_weight +  +� +`*_ nameW O M"K -Imodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_weight� -constR -Amodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin�˾*W +Imodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_weight*B +val; +  +� +`*$ +@model_path/weights/weight.bin��u� +constQ +Amodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_bias + +  +`*W nameO G E"C -Amodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_bias� -const[ -Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight -  -� -�*D -val= -  -� -�*% -@model_path/weights/weight.bin�Ӿ*Y +Amodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_bias*; +val4 + +  +`*$ +@model_path/weights/weight.bin��y� +constZ +Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight +  +` +�*Y nameQ I G"E -Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight� -constA -0model_mobilevit_encoder_layer_4_1_layernorm_bias -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*F +Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight*B +val; +  +` +�*$ +@model_path/weights/weight.bin��z� +const@ +0model_mobilevit_encoder_layer_4_1_layernorm_bias + +  +`*; +val4 + +  +`*$ +@model_path/weights/weight.bin��~*F name> 6 4"2 -0model_mobilevit_encoder_layer_4_1_layernorm_bias� -constC -2model_mobilevit_encoder_layer_4_1_layernorm_weight -   -�*H +0model_mobilevit_encoder_layer_4_1_layernorm_bias� +constB +2model_mobilevit_encoder_layer_4_1_layernorm_weight + +  +`*H name@ 8 6"4 -2model_mobilevit_encoder_layer_4_1_layernorm_weight*= -val6 -   -�*% -@model_path/weights/weight.bin���� +2model_mobilevit_encoder_layer_4_1_layernorm_weight*; +val4 + +  +`*$ +@model_path/weights/weight.bin��~� const> .model_seg_head_classifier_classifier_conv_bias @@ -2013,30 +2189,30 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight name< 4 2"0 -.model_seg_head_classifier_classifier_conv_bias*< -val5 +.model_seg_head_classifier_classifier_conv_bias*; +val4   -*% -@model_path/weights/weight.bin���� +*$ +@model_path/weights/weight.bin��~� constS 0model_seg_head_classifier_classifier_conv_weight    �  -*O -valH +*F +name> +6 +4"2 +0model_seg_head_classifier_classifier_conv_weight*N +valG    �  -*% -@model_path/weights/weight.bin���*F -name> -6 -4"2 -0model_seg_head_classifier_classifier_conv_weightM +*$ +@model_path/weights/weight.bin��~M const var_7  * @@ -2068,22 +2244,22 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight val   -�* +�* name  " -op_14M +op_14L const -var_17 -* -val - - -�* +var_16 +* name  " -op_17U +op_16* +val + + +`U const var_25 * @@ -2095,42 +2271,31 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight   -���������M -const -var_30 -* -name - - " -op_30* -val - - -�M +���������L const -var_31 -* +var_29 +* +val + + +0* name  " -op_31* -val - - -�L +op_29L const -var_32 +var_30 * val   -@* + * name  " -op_32] +op_30] const -var_40 +var_38  * @@ -2143,22 +2308,22 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight name  " -op_40] +op_38] const -var_42 +var_40  -* +* +name + + " +op_40* val    -* -name - - " -op_42j +j const input_1_pad_type_0 *( @@ -2186,25 +2351,25 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight   -� +� const2 input_3_weight_0      -*N -valG +*& +name + +" +input_3_weight_0*M +valF      -*% -@model_path/weights/weight.bin���*& -name - -" -input_3_weight_0� +*$ +@model_path/weights/weight.bin��� const input_3_bias_0 @@ -2218,35 +2383,35 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight   *% -@model_path/weights/weight.bin���� -conv" -pad_type - -input_1_pad_type_0 -groups - - -var_10 +@model_path/weights/weight.bin���� +conv +bias + +input_3_bias_0 strides  -var_40 -bias - -input_3_bias_0 +var_38 pad  - input_1_pad_0 + input_1_pad_0 +x + +image__scaled__ +weight + +input_3_weight_0 +groups + + +var_10" +pad_type + +input_1_pad_type_0 dilations  -var_42 -weight - -input_3_weight_0 -x - -image__scaled__+ +var_40+ input_3    @@ -2264,67 +2429,67 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight � �] const -var_68 +var_66  -* +* +name + + " +op_66* val    -* -name - - " -op_68] +] const -var_70 +var_68  -* -name - - " -op_70* +* val    -j +* +name + + " +op_68j const input_7_pad_type_0 -*( -name  - -" -input_7_pad_type_0* +* val  " -customn +custom*( +name  + +" +input_7_pad_type_0n const input_7_pad_0  -*! +*# +name + +" + input_7_pad_0*! val    -*# -name - -" - input_7_pad_0� +� const2 input_9_weight_0   -@ +    *& @@ -2334,57 +2499,57 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_9_weight_0*N valG   -@ +    *% -@model_path/weights/weight.bin���� +@model_path/weights/weight.bin���� const input_9_bias_0   -@*< -val5 - -  -@*% -@model_path/weights/weight.bin���*$ + *$ name  " -input_9_bias_0� -conv" -pad_type - -input_7_pad_type_0 -bias - -input_9_bias_0 +input_9_bias_0*< +val5 + +  + *% +@model_path/weights/weight.bin���� +conv weight  -input_9_weight_0 - dilations +input_9_weight_0 +x + +input_5 +bias + +input_9_bias_0" +pad_type + +input_7_pad_type_0 +groups  -var_70 +var_10 strides + +var_66 + dilations +  var_68 pad  - input_7_pad_0 -x - -input_5 -groups - - -var_10+ + input_7_pad_0+ input_9    -@ +  � �F silu @@ -2394,33 +2559,33 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_11    -@ +  � �] const -var_82 +var_80  -* +* +name + + " +op_80* val    -* -name - - " -op_82] +] const -var_84 +var_82  * name  " -op_84* +op_82* val  @@ -2429,36 +2594,36 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight l const input_13_pad_type_0 -* +*) +name! + +" +input_13_pad_type_0* val  " -custom*) -name! - -" -input_13_pad_type_0p +customp const input_13_pad_0  -*$ -name - -" -input_13_pad_0*! +*! val    -� +*$ +name + +" +input_13_pad_0� const3 input_15_weight_0   -@ +    *' @@ -2468,58 +2633,58 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_15_weight_0*N valG   -@ +    *% -@model_path/weights/weight.bin���� +@model_path/weights/weight.bin���� const input_15_bias_0   -@*< + *< val5   -@*% -@model_path/weights/weight.bin���*% + *% +@model_path/weights/weight.bin���*% name  " input_15_bias_0� -conv -pad - -input_13_pad_0 -weight - -input_15_weight_0 - dilations - - -var_84 -strides - - -var_82 +conv bias  input_15_bias_0# pad_type  -input_13_pad_type_0 -groups +input_13_pad_type_0 +x + + +input_11 +strides  -var_32 -x +var_80 + dilations + +var_82 +groups -input_11, + +var_30 +pad + +input_13_pad_0 +weight + +input_15_weight_0, input_15    -@ +  � �G silu @@ -2530,11 +2695,11 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_17    -@ +  � �] const -var_96 +var_94  * @@ -2547,16 +2712,16 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight name  " -op_96] +op_94] const -var_98 +var_96  * name  " -op_98* +op_96* val  @@ -2590,92 +2755,107 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight   -� -const3 -input_21_weight_0 +� +const5 +features_1_weight_0   +   -@  -*' -name - -" -input_21_weight_0*N +*N valG   +   -@  *% -@model_path/weights/weight.bin���� -const -input_21_bias_0 +@model_path/weights/weight.bin���*) +name! + +" +features_1_weight_0� +const! +features_1_bias_0   - *< +*' +name + +" +features_1_bias_0*< val5   - *% -@model_path/weights/weight.bin���*% -name - -" -input_21_bias_0� -conv - dilations +*% +@model_path/weights/weight.bin���� +conv +x - -var_98 -strides - -var_96 +input_17 pad  -input_19_pad_0 -weight - -input_21_weight_0 -groups +input_19_pad_0 +strides  -var_10 -x +var_94# +pad_type + +input_19_pad_type_0 +groups + +var_10 +bias + +features_1_bias_0 + dilations -input_17 -bias - -input_21_bias_0# -pad_type + +var_96! +weight  -input_19_pad_type_0, +features_1_weight_0. + +features_1 +  + + +� +�Z +add +x + +input_5 +y + + +features_1, input_21    -  + � �_ const -var_115 +var_114  -* +* +name + + +" +op_114* val    -* -name - - -" -op_115_ +_ const -var_117 +var_116  * @@ -2683,7 +2863,7 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_117* +op_116* val  @@ -2706,58 +2886,55 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_23_pad_0  -*! +*$ +name + +" +input_23_pad_0*! val    -*$ -name - -" -input_23_pad_0� -const4 -input_25_weight_0 -  -� -  - -*O -valH -  -� +� +const3 +input_25_weight_0 +    +  -*% -@model_path/weights/weight.bin���*' +*' name  " -input_25_weight_0� -const -input_25_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*% +input_25_weight_0*N +valG +  +  + + +*% +@model_path/weights/weight.bin���� +const +input_25_bias_0 + +  + *< +val5 + +  + *% +@model_path/weights/weight.bin���*% name  " -input_25_bias_0� -conv - dilations - -var_117 -weight - -input_25_weight_0 -bias - -input_25_bias_0# +input_25_bias_0� +conv +x + + +input_21# pad_type  input_23_pad_type_0 @@ -2767,49 +2944,54 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight groups  -var_10 -x - - -input_21 +var_10 +weight + +input_25_weight_0 +bias + +input_25_bias_0 + dilations + +var_116 strides -var_115- -input_25! -  - -� +var_114, +input_25 +  + +  � -�H +�G silu x -input_25- -input_27! -  - -� +input_25, +input_27 +  + +  � �_ const -var_129 +var_128  -* +* +name + + +" +op_128* val    -* -name - - -" -op_129_ +_ const -var_131 +var_130  * @@ -2823,7 +3005,7 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_131l +op_130l const input_29_pad_type_0 * @@ -2840,94 +3022,96 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_29_pad_0  -*$ -name - -" -input_29_pad_0*! +*! val    -� -const4 -input_31_weight_0 -  -� +*$ +name + +" +input_29_pad_0� +const3 +input_31_weight_0 +  +    -*' -name - -" -input_31_weight_0*O -valH -  -� +*N +valG +  +    *% -@model_path/weights/weight.bin���� -const -input_31_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*% +@model_path/weights/weight.bin�€*' +name + +" +input_31_weight_0� +const +input_31_bias_0 + +  + *< +val5 + +  + *% +@model_path/weights/weight.bin�̀*% name  " -input_31_bias_0� -conv - dilations +input_31_bias_0� +conv +weight + +input_31_weight_0 +strides -var_131 -pad - -input_29_pad_0 +var_128 bias  -input_31_bias_0# +input_31_bias_0 +groups + + +var_30# pad_type  input_29_pad_type_0 x -input_27 -weight - -input_31_weight_0 -strides +input_27 + dilations -var_129 -groups - - -var_30- -input_31! -  - -� +var_130 +pad + +input_29_pad_0, +input_31 +  + +  � -�H +�G silu x -input_31- -input_33! -  - -� +input_31, +input_33 +  + +  � �_ const -var_143 +var_142  * @@ -2935,7 +3119,7 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_143* +op_142* val  @@ -2943,7 +3127,7 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  _ const -var_145 +var_144  * @@ -2957,58 +3141,58 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_145l -const -input_35_pad_type_0 -* -val - - -" -custom*) +op_144l +const +input_35_pad_type_0 +*) name!  " -input_35_pad_type_0p +input_35_pad_type_0* +val + + +" +customp const input_35_pad_0  -*! +*$ +name + +" +input_35_pad_0*! val    -*$ -name - -" -input_35_pad_0� -const4 -input_37_weight_0 -  -@ -� +� +const3 +input_37_weight_0 +  + +   *' name  " -input_37_weight_0*O -valH -  -@ -� +input_37_weight_0*N +valG +  + +   *% -@model_path/weights/weight.bin���� +@model_path/weights/weight.bin�̀� const input_37_bias_0   -@*% +*% name  " @@ -3016,59 +3200,59 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight val5   -@*% -@model_path/weights/weight.bin���� -conv -weight - -input_37_weight_0 -groups - - -var_10 - dilations - -var_145 +*% +@model_path/weights/weight.bin��� +conv +pad + +input_35_pad_0 strides -var_143# -pad_type - -input_35_pad_type_0 +var_142 +weight + +input_37_weight_0 x input_33 bias  -input_37_bias_0 -pad - -input_35_pad_0, +input_37_bias_0# +pad_type + +input_35_pad_type_0 +groups + + +var_10 + dilations + +var_144, input_37    -@ + � �_ const -var_159 +var_158  -* -name - - -" -op_159* +* val    -_ +* +name + + +" +op_158_ const -var_161 +var_160  * @@ -3082,19 +3266,19 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_161l +op_160l const input_39_pad_type_0 -* +*) +name! + +" +input_39_pad_type_0* val  " -custom*) -name! - -" -input_39_pad_type_0p +customp const input_39_pad_0 @@ -3110,113 +3294,115 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight   -� -const4 -input_41_weight_0 -  -� -@ +� +const3 +input_41_weight_0 +  +0 +  *' name  " -input_41_weight_0*O -valH -  -� -@ +input_41_weight_0*N +valG +  +0 +  *% -@model_path/weights/weight.bin���� -const -input_41_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*% +@model_path/weights/weight.bin��� +const +input_41_bias_0 + +  +0*< +val5 + +  +0*% +@model_path/weights/weight.bin���*% name  " -input_41_bias_0� -conv -bias - -input_41_bias_0 +input_41_bias_0� +conv pad  -input_39_pad_0 +input_39_pad_0 +x + + +input_37 +weight + +input_41_weight_0 + dilations + +var_160 groups  var_10 strides -var_159 - dilations - -var_161# +var_158 +bias + +input_41_bias_0# pad_type  -input_39_pad_type_0 -weight - -input_41_weight_0 -x - - -input_37- -input_41! -  - -� +input_39_pad_type_0, +input_41 +  + +0 � -�H +�G silu x -input_41- -input_43! -  - -� +input_41, +input_43 +  + +0 � �_ const -var_173 +var_172  -* +* +name + + +" +op_172* val    -* +_ +const +var_174 + + +* name  " -op_173_ -const -var_175 - - -* +op_174* val    -* -name - - -" -op_175l +l const input_45_pad_type_0 *) @@ -3233,94 +3419,96 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_45_pad_0  -*$ -name - -" -input_45_pad_0*! +*! val    -� -const4 -input_47_weight_0 -  -� +*$ +name + +" +input_45_pad_0� +const3 +input_47_weight_0 +  +0   -*' -name - -" -input_47_weight_0*O -valH -  -� +*N +valG +  +0   *% -@model_path/weights/weight.bin���� -const -input_47_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*% +@model_path/weights/weight.bin���*' +name + +" +input_47_weight_0� +const +input_47_bias_0 + +  +0*< +val5 + +  +0*% +@model_path/weights/weight.bin���*% name  " -input_47_bias_0� -conv -pad - -input_45_pad_0 -weight - -input_47_weight_0# -pad_type - -input_45_pad_type_0 - dilations - -var_175 -strides - -var_173 +input_47_bias_0� +conv x -input_43 +input_43 +bias + +input_47_bias_0# +pad_type + +input_45_pad_type_0 groups  -var_31 -bias - -input_47_bias_0- -input_47! -  - -� +var_29 +strides + +var_172 +pad + +input_45_pad_0 + dilations + +var_174 +weight + +input_47_weight_0, +input_47 +  + +0 � -�H +�G silu x -input_47- -input_49! -  - -� +input_47, +input_49 +  + +0 � �_ const -var_187 +var_186  * @@ -3334,9 +3522,9 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_187_ +op_186_ const -var_189 +var_188  * @@ -3350,7 +3538,7 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight  " -op_189l +op_188l const input_51_pad_type_0 * @@ -3367,101 +3555,101 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight input_51_pad_0  -*! +*$ +name + +" +input_51_pad_0*! val    -*$ -name - -" -input_51_pad_0� -const6 -features_1_weight_0 -  -@ -� +� +const5 +features_3_weight_0 +  + +0 + +*N +valG +  + +0  -*) +*% +@model_path/weights/weight.bin���*) name!  " -features_1_weight_0*O -valH -  -@ -� - -*% -@model_path/weights/weight.bin���� +features_3_weight_0� const! -features_1_bias_0 +features_3_bias_0   -@*' +*' name  " -features_1_bias_0*< +features_3_bias_0*< val5   -@*% -@model_path/weights/weight.bin���� -conv -x - - -input_49 +*% +@model_path/weights/weight.bin�� +conv + dilations + +var_188 strides -var_187 -groups - - -var_10# +var_186# pad_type  -input_51_pad_type_0 - dilations - -var_189 -pad - -input_51_pad_0! +input_51_pad_type_0! weight  -features_1_weight_0 +features_3_weight_0 +x + + +input_49 +pad + +input_51_pad_0 +groups + + +var_10 bias  -features_1_bias_0. +features_3_bias_0. -features_1 +features_3    -@ + � �[ -add -x - - -input_37 +add y -features_1, +features_3 +x + + +input_37, input_53    -@ + � �_ const -var_204 +var_203  * @@ -3469,7 +3657,7 @@ features_1,  " -op_204* +op_203* val  @@ -3477,7 +3665,7 @@ features_1,  _ const -var_206 +var_205  * @@ -3491,111 +3679,113 @@ features_1,  " -op_206l +op_205l const input_55_pad_type_0 -*) -name! - -" -input_55_pad_type_0* +* val  " -customp +custom*) +name! + +" +input_55_pad_type_0p const input_55_pad_0  -*$ -name - -" -input_55_pad_0*! +*! val    -� -const4 -input_57_weight_0 -  -� -@ - -*O -valH -  -� -@ +*$ +name + +" +input_55_pad_0� +const3 +input_57_weight_0 +  +0 +  -*% -@model_path/weights/weight.bin���*' +*' name  " -input_57_weight_0� -const -input_57_bias_0 -   -�*% +input_57_weight_0*N +valG +  +0 + + +*% +@model_path/weights/weight.bin�ā� +const +input_57_bias_0 + +  +0*< +val5 + +  +0*% +@model_path/weights/weight.bin��*% name  " -input_57_bias_0*= -val6 -   -�*% -@model_path/weights/weight.bin���� -conv - dilations - -var_206 +input_57_bias_0� +conv x -input_53 +input_53# +pad_type + +input_55_pad_type_0 weight  -input_57_weight_0 -groups - - -var_10 +input_57_weight_0 pad  -input_55_pad_0 +input_55_pad_0 + dilations + +var_205 bias  -input_57_bias_0# -pad_type - -input_55_pad_type_0 +input_57_bias_0 strides -var_204- -input_57! -  - -� +var_203 +groups + + +var_10, +input_57 +  + +0 � -�H +�G silu x -input_57- -input_59! -  - -� +input_57, +input_59 +  + +0 � �_ const -var_218 +var_217  * @@ -3603,7 +3793,7 @@ features_1,  " -op_218* +op_217* val  @@ -3611,7 +3801,7 @@ features_1,  _ const -var_220 +var_219  * @@ -3625,19 +3815,19 @@ features_1,  " -op_220l +op_219l const input_61_pad_type_0 -* +*) +name! + +" +input_61_pad_type_0* val  " -custom*) -name! - -" -input_61_pad_type_0p +customp const input_61_pad_0 @@ -3653,83 +3843,85 @@ features_1,   -� -const4 -input_63_weight_0 -  -� +� +const3 +input_63_weight_0 +  +0   -*O -valH -  -� +*N +valG +  +0   *% -@model_path/weights/weight.bin���*' +@model_path/weights/weight.bin��*' name  " -input_63_weight_0� -const -input_63_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*% +input_63_weight_0� +const +input_63_bias_0 + +  +0*< +val5 + +  +0*% +@model_path/weights/weight.bin���*% name  " -input_63_bias_0� -conv# +input_63_bias_0� +conv +pad + +input_61_pad_0 +strides + +var_217 +weight + +input_63_weight_0 +groups + + +var_29# pad_type  input_61_pad_type_0 dilations -var_220 -weight - -input_63_weight_0 +var_219 x input_59 bias  -input_63_bias_0 -strides - -var_218 -pad - -input_61_pad_0 -groups - - -var_31- -input_63! -  - -� +input_63_bias_0, +input_63 +  + +0 � -�H +�G silu x -input_63- -input_65! -  - -� +input_63, +input_65 +  + +0 � �_ const -var_232 +var_231  * @@ -3737,7 +3929,7 @@ features_1,  " -op_232* +op_231* val  @@ -3745,7 +3937,7 @@ features_1,  _ const -var_234 +var_233  * @@ -3759,7 +3951,7 @@ features_1,  " -op_234l +op_233l const input_67_pad_type_0 * @@ -3787,90 +3979,90 @@ features_1,   -� -const6 -features_3_weight_0 -  -@ -� +� +const5 +features_5_weight_0 +  + +0  -*O -valH -  -@ -� +*N +valG +  + +0  *% -@model_path/weights/weight.bin���*) +@model_path/weights/weight.bin���*) name!  " -features_3_weight_0� +features_5_weight_0� const! -features_3_bias_0 - -  -@*< -val5 +features_5_bias_0   -@*% -@model_path/weights/weight.bin���*' +*' name  " -features_3_bias_0� -conv -groups +features_5_bias_0*< +val5 - -var_10 +  +*% +@model_path/weights/weight.bin���� +conv strides -var_232 +var_231 dilations -var_234 +var_233 bias  -features_3_bias_0! +features_5_bias_0! weight  -features_3_weight_0 -x - - -input_65 +features_5_weight_0 pad  -input_67_pad_0# +input_67_pad_0 +groups + + +var_10# pad_type  -input_67_pad_type_0. +input_67_pad_type_0 +x -features_3 + +input_65. + +features_5    -@ + � �[ -add -x - - -input_53 +add y -features_3, +features_5 +x + + +input_53, input_69    -@ + � �_ const -var_251 +var_250  * @@ -3878,7 +4070,7 @@ features_3,  " -op_251* +op_250* val  @@ -3886,167 +4078,169 @@ features_3,  _ const -var_253 +var_252  -* -name - - -" -op_253* +* val    -l +* +name + + +" +op_252l const -input_71_pad_type_0 -*) -name! - -" -input_71_pad_type_0* +input_71_pad_type_0 +* val  " -customp +custom*) +name! + +" +input_71_pad_type_0p const input_71_pad_0  -*$ -name - -" -input_71_pad_0*! +*! val    -� -const4 -input_73_weight_0 -  -� -@ +*$ +name + +" +input_71_pad_0� +const3 +input_73_weight_0 +  +0 +  *' name  " -input_73_weight_0*O -valH -  -� -@ +input_73_weight_0*N +valG +  +0 +  *% -@model_path/weights/weight.bin���� -const -input_73_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*% +@model_path/weights/weight.bin���� +const +input_73_bias_0 + +  +0*% name  " -input_73_bias_0� -conv - dilations +input_73_bias_0*< +val5 + +  +0*% +@model_path/weights/weight.bin�ł� +conv +strides -var_253 +var_250# +pad_type + +input_71_pad_type_0 +pad + +input_71_pad_0 +x + + +input_69 groups  var_10 weight  -input_73_weight_0 -pad - -input_71_pad_0 +input_73_weight_0 bias  -input_73_bias_0 -strides +input_73_bias_0 + dilations -var_251# -pad_type - -input_71_pad_type_0 -x - - -input_69- -input_73! -  - -� +var_252, +input_73 +  + +0 � -�H +�G silu x -input_73- -input_75! -  - -� +input_73, +input_75 +  + +0 � �_ const -var_265 +var_264  -* -name - - -" -op_265* +* val    -_ +* +name + + +" +op_264_ const -var_267 +var_266  -* +* +name + + +" +op_266* val    -* -name - - -" -op_267l +l const input_77_pad_type_0 -* +*) +name! + +" +input_77_pad_type_0* val  " -custom*) -name! - -" -input_77_pad_type_0p +customp const input_77_pad_0 @@ -4062,37 +4256,39 @@ features_3,   -� -const4 -input_79_weight_0 -  -� +� +const3 +input_79_weight_0 +  +0   -*O -valH -  -� +*N +valG +  +0   *% -@model_path/weights/weight.bin���*' +@model_path/weights/weight.bin�ǂ*' name  " -input_79_weight_0� -const -input_79_bias_0 -   -�*% +input_79_weight_0� +const +input_79_bias_0 + +  +0*< +val5 + +  +0*% +@model_path/weights/weight.bin�Ղ*% name  " -input_79_bias_0*= -val6 -   -�*% -@model_path/weights/weight.bin���� +input_79_bias_0� conv bias  @@ -4102,43 +4298,43 @@ features_3, input_79_weight_0 pad  -input_77_pad_0 -x - - -input_75 - dilations - -var_267 -strides - -var_265# +input_77_pad_0# pad_type  input_77_pad_type_0 groups  -var_31+ -input_79 -  - -� +var_29 +strides + +var_264 + dilations + +var_266 +x + + +input_75* +input_79 +  + +0 @ -@F +@E silu x -input_79+ -input_81 -  - -� +input_79* +input_81 +  + +0 @ @_ const -var_279 +var_278  * @@ -4146,7 +4342,7 @@ features_3,  " -op_279* +op_278* val  @@ -4154,7 +4350,7 @@ features_3,  _ const -var_281 +var_280  * @@ -4162,7 +4358,7 @@ features_3,  " -op_281* +op_280* val  @@ -4196,90 +4392,90 @@ features_3, name  " -input_83_pad_0� -const4 -input_85_weight_0 -  -` -� +input_83_pad_0� +const3 +input_85_weight_0 +  +0 +0  -*' +*N +valG +  +0 +0 + +*% +@model_path/weights/weight.bin�ׂ*' name  " -input_85_weight_0*O -valH -  -` -� - -*% -@model_path/weights/weight.bin���� +input_85_weight_0� const input_85_bias_0   -`*< -val5 - -  -`*% -@model_path/weights/weight.bin���*% +0*% name  " -input_85_bias_0� -conv +input_85_bias_0*< +val5 + +  +0*% +@model_path/weights/weight.bin���� +conv +bias + +input_85_bias_0 weight  -input_85_weight_0 -strides - -var_279 -pad - -input_83_pad_0 +input_85_weight_0# +pad_type + +input_83_pad_type_0 groups  -var_10 -bias - -input_85_bias_0 - dilations +var_10 +strides -var_281 +var_278 x -input_81# -pad_type - -input_83_pad_type_0* +input_81 + dilations + +var_280 +pad + +input_83_pad_0* input_85    -` +0 @ @_ const -var_298 +var_297  -* -name - - -" -op_298* +* val    -_ +* +name + + +" +op_297_ const -var_300 +var_299  * @@ -4287,7 +4483,7 @@ features_3,  " -op_300* +op_299* val  @@ -4296,95 +4492,95 @@ features_3, l const input_87_pad_type_0 -*) -name! - -" -input_87_pad_type_0* +* val  " -customp +custom*) +name! + +" +input_87_pad_type_0p const input_87_pad_0  -*$ -name - -" -input_87_pad_0*! +*! val    -� +*$ +name + +" +input_87_pad_0� const3 input_89_weight_0   -` -` +0 +0  -*' -name - -" -input_89_weight_0*N +*N valG   -` -` +0 +0  *% -@model_path/weights/weight.bin���� +@model_path/weights/weight.bin���*' +name + +" +input_89_weight_0� const input_89_bias_0   -`*< -val5 - -  -`*% -@model_path/weights/weight.bin���*% +0*% name  " -input_89_bias_0� -conv -weight - -input_89_weight_0 -x - +input_89_bias_0*< +val5 -input_85# -pad_type - -input_87_pad_type_0 +  +0*% +@model_path/weights/weight.bin���� +conv groups  -var_10 -strides - -var_298 +var_10 bias  -input_89_bias_0 - dilations - -var_300 +input_89_bias_0 pad  -input_87_pad_0* +input_87_pad_0 +x + + +input_85 + dilations + +var_299# +pad_type + +input_87_pad_type_0 +weight + +input_89_weight_0 +strides + +var_297* input_89    -` +0 @ @E silu @@ -4395,102 +4591,102 @@ features_3, input_91    -` +0 @ @_ const -var_311 +var_310  -* +* +name + + +" +op_310* val    -* +_ +const +var_312 + + +* name  " -op_311_ -const -var_313 - - -* +op_312* val    -* -name +p +const +features_7_pad_type_0 +* +val  " -op_313p -const -features_5_pad_type_0 -*+ +custom*+ name#  " -features_5_pad_type_0* -val - - -" -customt +features_7_pad_type_0t const -features_5_pad_0 +features_7_pad_0  *& name  " -features_5_pad_0*! +features_7_pad_0*! val    -� -convD +� +conv +strides + +var_310 +x + + +input_91D weight: 8 -6model_mobilevit_encoder_layer_2_1_conv_1x1_conv_weight -pad - -features_5_pad_0 +6model_mobilevit_encoder_layer_2_1_conv_1x1_conv_weight groups  -var_10 -x - - -input_91 -strides +var_10 + dilations -var_311% +var_312 +pad + +features_7_pad_0% pad_type  -features_5_pad_type_0 - dilations - -var_313- +features_7_pad_type_0, -features_5 -  - -� +features_7 +  + +@ @ @b const -var_324 +var_323  * @@ -4498,24 +4694,24 @@ features_5  " -op_324*" +op_323*" val    -�$ b +� b reshape shape -var_324 +var_323 x -features_5, +features_7, patches_1   -�$ +�    t @@ -4534,9 +4730,9 @@ features_5, name  " -patches_3_perm_0c +patches_3_perm_0b const -var_327 +var_326  * @@ -4544,14 +4740,13 @@ features_5,  " -op_327*# -val +op_326*" +val  - - - -��n + +  +@�n transpose perm  @@ -4561,171 +4756,171 @@ features_5, patches_1/ transpose_56   -�$ +�    -e -reshape -x - - transpose_56 +d +reshape shape -var_327- - patches_5 -  - -� +var_326 +x + + transpose_56, + patches_5 +  + +@ � t const patches_7_perm_0  -*! +*& +name + +" +patches_7_perm_0*! val    -*& -name - -" -patches_7_perm_0j +j const -var_333 +var_332  -** +* +name + + +" +op_332** val#    - ����������* -name - - -" -op_333o - transpose -x - - patches_5 + ����������n + transpose perm  -patches_7_perm_00 - transpose_55 -  +patches_7_perm_0 +x + + patches_5/ + transpose_55 +    -� -�^ -reshape -x - - transpose_55 +� +@] +reshape shape -var_333& -input_93 -  +var_332 +x + + transpose_55% +input_93 +   -� -�x +� +@x const input_95_axes_0  -*' +*% +name + +" +input_95_axes_0*' val     -���������*% -name - -" -input_95_axes_0� +���������� -layer_norm +layer_norm +epsilon + +var_7T +gammaK +I +Gmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_weight axes  input_95_axes_0 x -input_93 -epsilon - -var_7T -gammaK -I -Gmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_weightQ +input_93Q betaI G -Emodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_bias& -input_95 -  +Emodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_before_bias% +input_95 +   -� -�� +� +@� linear^ weightT R -Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_weightZ -biasR -P -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_bias +Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_weight x -input_95! -x_9 -  +input_95Z +biasR +P +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_query_bias +x_9 +   -� -�� +� +@� linear x -input_95X -biasP -N -Lmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_bias\ +input_95\ weightR P -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight! -x_1 -  +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weightX +biasP +N +Lmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_bias +x_1 +   -� -�b +� +@b const -var_359 +var_358  -*" +* +name + + +" +op_358*" val    -�$* -name - - -" -op_359U +�U reshape shape -var_359 +var_358 x  x_1& @@ -4734,9 +4929,9 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight  �  -$a +a const -var_361 +var_360  *! @@ -4751,43 +4946,43 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_key_weight  " -op_361� -linear -x - - -input_95^ +op_360� +linear^ weightT R Pmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_weightZ biasR P -Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias! -x_5 -  +Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias +x + + +input_95 +x_5 +   -� -�b +� +@b const -var_368 +var_367  -* -name - - -" -op_368*" +*" val    -�$U +�* +name + + +" +op_367U reshape shape -var_368 +var_367 x  x_5& @@ -4796,26 +4991,26 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias  �  -$a +a const -var_370 +var_369  -* -name - - -" -op_370*! +*! val    -b +* +name + + +" +op_369b const -var_374 +var_373  * @@ -4823,17 +5018,17 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias  " -op_374*" +op_373*" val    -�$V +�V reshape shape -var_374 +var_373 x  x_9' @@ -4842,9 +5037,9 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias  �  -$a +a const -var_376 +var_375  * @@ -4852,7 +5047,7 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias  " -op_376*! +op_375*! val  @@ -4861,14 +5056,14 @@ Nmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_attention_value_bias  � const -var_378_perm_0 +var_377_perm_0  *# name  " - op_378_perm_0*3 + op_377_perm_0*3 val,  @@ -4888,19 +5083,19 @@ $"" attention_scores_1_transpose_x_0� const( attention_scores_1_transpose_y_0 -* -val - - -*6 +*6 name. & $"" - attention_scores_1_transpose_y_0_ + attention_scores_1_transpose_y_0* +val + + +_ transpose perm -var_361 +var_360 x  x_3/ @@ -4909,19 +5104,19 @@ $""   � -$o - transpose -perm - -var_378_perm_0 +o + transpose x  - transpose_54/ + transpose_54 +perm + +var_377_perm_0/ transpose_51     -$ + �` transpose x @@ -4929,26 +5124,26 @@ $"" x_11 perm -var_376/ +var_375/ transpose_52     � -$� +� matmul +x + + transpose_52 y  transpose_513 - transpose_x$ -" - attention_scores_1_transpose_x_0 -x - - transpose_523 transpose_y$ " - attention_scores_1_transpose_y_06 + attention_scores_1_transpose_y_03 + transpose_x$ +" + attention_scores_1_transpose_x_06 attention_scores_1    @@ -4957,17 +5152,17 @@ $"" �p const _inversed_input_97_y_0 - * + *, +name$ + +" +_inversed_input_97_y_0* val     -��*>*, -name$ - -" -_inversed_input_97_y_0{ +�>{ mul y  @@ -5008,39 +5203,39 @@ $"" context_layer_1_transpose_x_0{ const% context_layer_1_transpose_y_0 -* -val - - -*3 +*3 name+ # !" -context_layer_1_transpose_y_0_ - transpose -x - -x_7 +context_layer_1_transpose_y_0* +val + + +_ + transpose perm -var_370/ +var_369 +x + +x_7/ transpose_53     -� -$� -matmul -y - - transpose_530 +� +� +matmul0 transpose_y!  context_layer_1_transpose_y_0 x -input_990 +input_99 +y + + transpose_530 transpose_x!  context_layer_1_transpose_x_02 @@ -5049,65 +5244,66 @@ $""   � -$a +a const -var_385 +var_384  -* -name - - -" -op_385*! +*! val    -b +* +name + + +" +op_384a const -var_390 +var_389  -*" -val +*! +val  - -  -��* + + + +�@* name  " -op_390k - transpose -perm - -var_385 +op_389k + transpose x  -context_layer_1/ +context_layer_1 +perm + +var_384/ transpose_50    �  -$_ -reshape -shape - -var_390 +^ +reshape x  - transpose_50' - input_101 -  + transpose_50 +shape + +var_389& + input_101 +   -� -�� +� +@� linearW biasO M @@ -5117,78 +5313,78 @@ O Mmodel_mobilevit_encoder_layer_2_1_transformer_0_attention_output_dense_weight x - input_101' - input_103 -  + input_101& + input_103 +   -� -�U -add -x - - input_103 +� +@T +add y -input_93' - input_105 -  +input_93 +x + + input_103& + input_105 +   -� -�z +� +@z const input_107_axes_0  -*& -name - -" -input_107_axes_0*' +*' val     -���������� +���������*& +name + +" +input_107_axes_0� -layer_norm -axes - -input_107_axes_0S +layer_norm +x + + input_105P +betaH +F +Dmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_biasS gammaJ H Fmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_weight epsilon  -var_7P -betaH -F -Dmodel_mobilevit_encoder_layer_2_1_transformer_0_layernorm_after_bias -x - - input_105' - input_107 -  +var_7 +axes + +input_107_axes_0& + input_107 +   -� -�� -linear -x - - input_107W -weightM -K -Imodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_weightS +� +@� +linearS biasK I -Gmodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_bias' +Gmodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_biasW +weightM +K +Imodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_weight +x + + input_107' input_109    � -�C +�C silu x @@ -5197,87 +5393,87 @@ Gmodel_mobilevit_encoder_layer_2_1_transformer_0_intermediate_dense_bias'    � -�� -linear -x - - input_111M +�� +linearM biasE C Amodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_biasQ weightG E -Cmodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_weight' - input_113 -  +Cmodel_mobilevit_encoder_layer_2_1_transformer_0_output_dense_weight +x + + input_111& + input_113 +   -� -�V +� +@U add y input_105 x - input_113' - input_115 -  + input_113& + input_115 +   -� -�z +� +@z const input_117_axes_0  -*& -name - -" -input_117_axes_0*' +*' val     -���������� +���������*& +name + +" +input_117_axes_0� -layer_normT -gammaK -I -Gmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_weight -x - - input_115 +layer_norm axes  -input_117_axes_0Q +input_117_axes_0 +x + + input_115Q betaI G Emodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_bias epsilon  -var_7' - input_117 -  +var_7T +gammaK +I +Gmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_before_weight& + input_117 +   -� -�� -linear^ -weightT -R -Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_weightZ +� +@� +linearZ biasR P Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_bias x - input_117" -x_21 -  + input_117^ +weightT +R +Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_query_weight! +x_21 +   -� -�� +� +@� linear x @@ -5287,121 +5483,121 @@ P Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_weightX biasP N -Lmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_bias" -x_13 -  +Lmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_key_bias! +x_13 +   -� -�b +� +@b const -var_435 +var_434  -* -name - - -" -op_435*" +*" val    -�$W -reshape -x - -x_13 +�* +name + + +" +op_434W +reshape shape -var_435' +var_434 +x + +x_13' x_15    �  -$a +a const -var_437 +var_436  -*! +* +name + + +" +op_436*! val    -* -name - - -" -op_437� -linearZ -biasR -P -Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_bias +� +linear x - input_117^ + input_117Z +biasR +P +Nmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_bias^ weightT R -Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weight" -x_17 -  +Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weight! +x_17 +   -� -�b +� +@b const -var_444 +var_443  -*" +* +name + + +" +op_443*" val    -�$* -name - - -" -op_444W -reshape -x - -x_17 +�W +reshape shape -var_444' +var_443 +x + +x_17' x_19    �  -$a +a const -var_446 +var_445  -*! +* +name + + +" +op_445*! val    -* -name - - -" -op_446b +b const -var_450 +var_449  * @@ -5409,28 +5605,28 @@ Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weigh  " -op_450*" +op_449*" val    -�$W -reshape -shape - -var_450 +�W +reshape x  -x_21' +x_21 +shape + +var_449' x_23    �  -$a +a const -var_452 +var_451  * @@ -5438,7 +5634,7 @@ Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weigh  " -op_452*! +op_451*! val  @@ -5447,94 +5643,94 @@ Pmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_attention_value_weigh  � const -var_454_perm_0 +var_453_perm_0  -*3 +*# +name + +" + op_453_perm_0*3 val,    -������������������*# -name - -" - op_454_perm_0� +������������������� const( attention_scores_3_transpose_x_0 -*6 -name. -& -$"" - attention_scores_3_transpose_x_0* +* val   -� -const( - attention_scores_3_transpose_y_0 -*6 +*6 name. & $"" - attention_scores_3_transpose_y_0* + attention_scores_3_transpose_x_0� +const( + attention_scores_3_transpose_y_0 +* val   -` - transpose -perm - -var_437 +*6 +name. +& +$"" + attention_scores_3_transpose_y_0` + transpose x  -x_15/ +x_15 +perm + +var_436/ transpose_49     � -$o +o transpose x  transpose_49 perm  -var_454_perm_0/ +var_453_perm_0/ transpose_46     -$ + �` - transpose -perm - -var_452 + transpose x  -x_23/ +x_23 +perm + +var_451/ transpose_47    - -� -$� -matmul3 - transpose_y$ -" - attention_scores_3_transpose_y_03 - transpose_x$ -" - attention_scores_3_transpose_x_0 + +� +� +matmul y  transpose_46 x  - transpose_476 + transpose_473 + transpose_x$ +" + attention_scores_3_transpose_x_03 + transpose_y$ +" + attention_scores_3_transpose_y_06 attention_scores_3    @@ -5543,38 +5739,38 @@ $"" �r const _inversed_input_119_y_0 - * + *- +name% + +" +_inversed_input_119_y_0* val     -��*>*- -name% - -" -_inversed_input_119_y_0} -mul -y - -_inversed_input_119_y_0 +�>} +mul x  -attention_scores_37 +attention_scores_3 +y + +_inversed_input_119_y_07 _inversed_input_119     � �j -softmax +softmax +x + +_inversed_input_119 axis  -var_25 -x - -_inversed_input_119- +var_25- input_121    @@ -5594,35 +5790,35 @@ $"" context_layer_5_transpose_x_0{ const% context_layer_5_transpose_y_0 -*3 -name+ -# -!" -context_layer_5_transpose_y_0* +* val   -` - transpose -perm - -var_446 +*3 +name+ +# +!" +context_layer_5_transpose_y_0` + transpose x  -x_19/ +x_19 +perm + +var_445/ transpose_48     � -$� -matmul0 - transpose_x! - -context_layer_5_transpose_x_0 +� +matmul x input_1210 + transpose_x! + +context_layer_5_transpose_x_00 transpose_y!  context_layer_5_transpose_y_0 @@ -5634,26 +5830,26 @@ $""   � -$a +a const -var_461 +var_460  -*! +* +name + + +" +op_460*! val    -* -name - - -" -op_461b +a const -var_466 +var_465  * @@ -5661,110 +5857,111 @@ $""  " -op_466*" -val +op_465*! +val  - -  -��k - transpose -perm - -var_461 + + + +�@k + transpose x  -context_layer_5/ +context_layer_5 +perm + +var_460/ transpose_45    �  -$_ +^ reshape x  transpose_45 shape -var_466' - input_123 -  +var_465& + input_123 +   -� -�� -linear -x - - input_123W +� +@� +linearW biasO M Kmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_bias[ weightQ O -Mmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_weight' - input_125 -  - -� -�V -add +Mmodel_mobilevit_encoder_layer_2_1_transformer_1_attention_output_dense_weight x - input_125 + input_123& + input_125 +  + +� +@U +add y - input_115' - input_127 -  + input_115 +x + + input_125& + input_127 +   -� -�z +� +@z const input_129_axes_0  -*' +*& +name + +" +input_129_axes_0*' val     -���������*& -name - -" -input_129_axes_0� +���������� -layer_norm +layer_normS +gammaJ +H +Fmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_weight x - input_127 -epsilon - -var_7P + input_127P betaH F -Dmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_biasS -gammaJ -H -Fmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_weight +Dmodel_mobilevit_encoder_layer_2_1_transformer_1_layernorm_after_bias axes  -input_129_axes_0' - input_129 -  +input_129_axes_0 +epsilon + +var_7& + input_129 +   -� -�� -linearS -biasK -I -Gmodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_biasW +� +@� +linearW weightM K -Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight +Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weightS +biasK +I +Gmodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_bias x input_129' @@ -5772,7 +5969,7 @@ Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight    � -�C +�C silu x @@ -5781,34 +5978,34 @@ Imodel_mobilevit_encoder_layer_2_1_transformer_1_intermediate_dense_weight    � -�� -linear +�� +linearM +biasE +C +Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias x input_133Q weightG E -Cmodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_weightM -biasE -C -Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias' - input_135 -  +Cmodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_weight& + input_135 +   -� -�V +� +@U add x input_135 y - input_127' - input_137 -  + input_127& + input_137 +   -� -�z +� +@z const patches_9_axes_0 @@ -5824,61 +6021,61 @@ Amodel_mobilevit_encoder_layer_2_1_transformer_1_output_dense_bias' name  " -patches_9_axes_0� +patches_9_axes_0� -layer_norm< -beta4 -2 -0model_mobilevit_encoder_layer_2_1_layernorm_bias? +layer_norm? gamma6 4 -2model_mobilevit_encoder_layer_2_1_layernorm_weight +2model_mobilevit_encoder_layer_2_1_layernorm_weight +x + + input_137< +beta4 +2 +0model_mobilevit_encoder_layer_2_1_layernorm_bias epsilon  var_7 axes  -patches_9_axes_0 -x - - input_137' - patches_9 -  +patches_9_axes_0& + patches_9 +   -� -�k +� +@k const -var_494 +var_493  -* -name - - -" -op_494*+ +*+ val$    -����������c -reshape -shape - -var_494 +����������* +name + + +" +op_493b +reshape x - patches_9. + patches_9 +shape + +var_493- -patches_13 -  +patches_13 +    -� -�v +� +@v const! -features_7_perm_0 +features_9_perm_0  *! @@ -5892,184 +6089,183 @@ patches_13 name  " -features_7_perm_0b +features_9_perm_0b const -var_505 +var_504  -*" +* +name + + +" +op_504*" val    -�$ * -name - - -" -op_505q - transpose +� p + transpose +perm + +features_9_perm_0 x -patches_13 -perm - -features_7_perm_00 - transpose_44 -  - -� +patches_13/ + transpose_44 +  + +@ � -e -reshape -x - - transpose_44 +f +reshape shape -var_505- - -features_9 +var_504 +x + + transpose_44. + features_11   -�$ +�    x const" -features_11_perm_0 +features_13_perm_0  *( name   " -features_11_perm_0*! +features_13_perm_0*! val    -b +a const -var_508 +var_507  -* +*! +val + + + + + +@@@* name  " -op_508*" -val - - - -  -�@@q - transpose +op_507r + transpose +x + + features_11 perm  -features_11_perm_0 -x - - -features_9/ +features_13_perm_0/ transpose_43   -�$ +�    -d -reshape -x - - transpose_43 +c +reshape shape -var_508, - input_139 -  - -� +var_507 +x + + transpose_43+ + input_139 +  + +@ @ @_ const -var_513 +var_512  -* -name - - -" -op_513* +* val    -_ -const -var_515 - - -* +* name  " -op_515* +op_512_ +const +var_514 + + +* val    -n -const -input_141_pad_type_0 -* -val +* +name  " -custom** +op_514n +const +input_141_pad_type_0 +** name"  " -input_141_pad_type_0r +input_141_pad_type_0* +val + + +" +customr const input_141_pad_0  -*% -name - -" -input_141_pad_0*! +*! val    -� -const5 -input_143_weight_0 -  -` -� +*% +name + +" +input_141_pad_0� +const4 +input_143_weight_0 +  +0 +@  -*O -valH -  -` -� +*N +valG +  +0 +@  *% -@model_path/weights/weight.bin���*( +@model_path/weights/weight.bin���*( name   " @@ -6078,164 +6274,164 @@ features_9/ input_143_bias_0   -`*& -name - -" -input_143_bias_0*< +0*< val5   -`*% -@model_path/weights/weight.bin�Ĕ� -conv$ -pad_type - -input_141_pad_type_0 -weight - -input_143_weight_0 +0*% +@model_path/weights/weight.bin���*& +name + +" +input_143_bias_0� +conv +x + + input_139 +bias + +input_143_bias_0 pad  -input_141_pad_0 +input_141_pad_0 +strides + +var_512 groups  var_10 dilations -var_515 -x - - input_139 -strides - -var_513 -bias - -input_143_bias_0+ +var_514 +weight + +input_143_weight_0$ +pad_type + +input_141_pad_type_0+ input_143    -` +0 @ @I silu x input_143- - features_13 + features_15    -` +0 @ @m const input_145_interleave_0 -*, -name$ - -" -input_145_interleave_0* +* val   -� -concat -axis - - -var_10( +*, +name$ + +" +input_145_interleave_0� +concat( interleave  -input_145_interleave_0% +input_145_interleave_0 +axis + + +var_10% values input_85 - features_13, - input_145 -  - -� + features_15+ + input_145 +  + +` @ @_ const -var_529 +var_528  -* +* +name + + +" +op_528* val    -* +_ +const +var_530 + + +* name  " -op_529_ -const -var_531 - - -* +op_530* val    -* -name - - -" -op_531n +n const input_147_pad_type_0 -* +** +name" + +" +input_147_pad_type_0* val  " -custom** -name" - -" -input_147_pad_type_0r +customr const input_147_pad_0  -*% -name - -" -input_147_pad_0*! +*! val    -� -const5 -input_149_weight_0 -  -` -� +*% +name + +" +input_147_pad_0� +const4 +input_149_weight_0 +  +0 +`  -*O -valH -  -` -� +*N +valG +  +0 +`  *% -@model_path/weights/weight.bin�ǔ*( +@model_path/weights/weight.bin���*( name   " @@ -6244,46 +6440,46 @@ interleave input_149_bias_0   -`*& -name - -" -input_149_bias_0*< +0*< val5   -`*% -@model_path/weights/weight.bin���� -conv - dilations - -var_531$ -pad_type - -input_147_pad_type_0 +0*% +@model_path/weights/weight.bin���*& +name + +" +input_149_bias_0� +conv +weight + +input_149_weight_0 pad  -input_147_pad_0 +input_147_pad_0 +strides + +var_528 bias  -input_149_bias_0 -weight - -input_149_weight_0 -strides +input_149_bias_0 + dilations -var_529 -x - - input_145 +var_530$ +pad_type + +input_147_pad_type_0 groups  -var_10+ +var_10 +x + + input_145+ input_149    -` +0 @ @G silu @@ -6293,305 +6489,309 @@ interleave input_151    -` +0 @ @_ const -var_548 +var_547  -* -name - - -" -op_548* +* val    -_ -const -var_550 - - -* +* name  " -op_550* +op_547_ +const +var_549 + + +* val    -n +* +name + + +" +op_549n const input_153_pad_type_0 -** -name" - -" -input_153_pad_type_0* +* val  " -customr +custom** +name" + +" +input_153_pad_type_0r const input_153_pad_0  -*! +*% +name + +" +input_153_pad_0*! val    -*% -name - -" -input_153_pad_0� -const5 -input_155_weight_0 -  -� +� +const4 +input_155_weight_0 +  ` +0  -*( -name  - -" -input_155_weight_0*O -valH -  -� +*N +valG +  ` +0  *% -@model_path/weights/weight.bin���� -const! -input_155_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*& +@model_path/weights/weight.bin���*( +name  + +" +input_155_weight_0� +const +input_155_bias_0 + +  +`*< +val5 + +  +`*% +@model_path/weights/weight.bin���*& name  " -input_155_bias_0� -conv -pad - -input_153_pad_0$ -pad_type - -input_153_pad_type_0 -x - - input_151 +input_155_bias_0� +conv groups  -var_10 - dilations - -var_550 -bias - -input_155_bias_0 +var_10 +x + + input_151$ +pad_type + +input_153_pad_type_0 weight  -input_155_weight_0 +input_155_weight_0 +bias + +input_155_bias_0 strides -var_548, - input_155 -  - -� +var_547 + dilations + +var_549 +pad + +input_153_pad_0+ + input_155 +  + +` @ -@H +@G silu x - input_155, - input_157 -  - -� + input_155+ + input_157 +  + +` @ @_ const -var_562 +var_561  -* -name - - -" -op_562* +* val    -_ +* +name + + +" +op_561_ const -var_564 +var_563  -* +* +name + + +" +op_563* val    -* -name - - -" -op_564n +n const input_159_pad_type_0 -** -name" - -" -input_159_pad_type_0* +* val  " -customr +custom** +name" + +" +input_159_pad_type_0r const input_159_pad_0  -*% -name - -" -input_159_pad_0*! +*! val    -� -const5 -input_161_weight_0 -  -� - - -*O -valH -  -� +*% +name + +" +input_159_pad_0� +const4 +input_161_weight_0 +  +`   -*% -@model_path/weights/weight.bin���*( +*( name   " -input_161_weight_0� -const! -input_161_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*& +input_161_weight_0*N +valG +  +` + + +*% +@model_path/weights/weight.bin���� +const +input_161_bias_0 + +  +`*< +val5 + +  +`*% +@model_path/weights/weight.bin�Д*& name  " -input_161_bias_0� -conv$ -pad_type - -input_159_pad_type_0 - dilations - -var_564 -strides - -var_562 +input_161_bias_0� +conv bias  -input_161_bias_0 -weight - -input_161_weight_0 -x - - input_157 +input_161_bias_0 pad  input_159_pad_0 groups  -var_17, - input_161 -  - -� +var_16 +strides + +var_561 + dilations + +var_563 +weight + +input_161_weight_0$ +pad_type + +input_159_pad_type_0 +x + + input_157+ + input_161 +  + +`   - H + G silu x - input_161, - input_163 -  - -� + input_161+ + input_163 +  + +`    _ const -var_576 +var_575  -* +* +name + + +" +op_575* val    -* -name - - -" -op_576_ +_ const -var_578 +var_577  -* -name - - -" -op_578* +* val    -n +* +name + + +" +op_577n const input_165_pad_type_0 ** @@ -6619,71 +6819,73 @@ interleave name  " -input_165_pad_0� -const6 -input_167_weight_0 -  -� -� +input_165_pad_0� +const4 +input_167_weight_0 +  +@ +`  -*P -valI -  -� -� +*N +valG +  +@ +`  *% -@model_path/weights/weight.bin���*( +@model_path/weights/weight.bin�Ԕ*( name   " -input_167_weight_0� -const! -input_167_bias_0 -   -�*& +input_167_weight_0� +const +input_167_bias_0 + +  +@*& name  " -input_167_bias_0*= -val6 -   -�*% -@model_path/weights/weight.bin���� -conv$ +input_167_bias_0*< +val5 + +  +@*% +@model_path/weights/weight.bin���� +conv +strides + +var_575$ pad_type  -input_165_pad_type_0 +input_165_pad_type_0 +x + + input_163 groups  -var_10 - dilations - -var_578 -bias - -input_167_bias_0 -x - - input_163 +var_10 weight  -input_167_weight_0 -strides +input_167_weight_0 +bias + +input_167_bias_0 + dilations -var_576 +var_577 pad  -input_165_pad_0, - input_167 -  - -� +input_165_pad_0+ + input_167 +  + +@    _ const -var_595 +var_594  * @@ -6697,141 +6899,143 @@ interleave  " -op_595_ +op_594_ const -var_597 +var_596  -* +* +name + + +" +op_596* val    -* -name - - -" -op_597n +n const input_169_pad_type_0 -** -name" - -" -input_169_pad_type_0* +* val  " -customr +custom** +name" + +" +input_169_pad_type_0r const input_169_pad_0  -*! +*% +name + +" +input_169_pad_0*! val    -*% -name - -" -input_169_pad_0� -const6 -input_171_weight_0 -  -� -� - -*( -name  - -" -input_171_weight_0*P -valI -  -� -� +� +const4 +input_171_weight_0 +  +@ +@  -*% -@model_path/weights/weight.bin���� -const! -input_171_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*& +*N +valG +  +@ +@ + +*% +@model_path/weights/weight.bin���*( +name  + +" +input_171_weight_0� +const +input_171_bias_0 + +  +@*& name  " -input_171_bias_0� -conv - dilations - -var_597 -strides - -var_595 -x - - input_167 +input_171_bias_0*< +val5 + +  +@*% +@model_path/weights/weight.bin���� +conv +weight + +input_171_weight_0 +bias + +input_171_bias_0 pad  -input_169_pad_0$ +input_169_pad_0 +strides + +var_594$ pad_type  -input_169_pad_type_0 -weight - -input_171_weight_0 +input_169_pad_type_0 +x + + input_167 groups  -var_10 -bias - -input_171_bias_0, - input_171 -  - -� +var_10 + dilations + +var_596+ + input_171 +  + +@   - H + G silu x - input_171, - input_173 -  - -� + input_171+ + input_173 +  + +@    _ const -var_608 +var_607  -* +* +name + + +" +op_607* val    -* -name - - -" -op_608_ +_ const -var_610 +var_609  * @@ -6839,7 +7043,7 @@ interleave  " -op_610* +op_609* val  @@ -6847,89 +7051,91 @@ interleave  r const -features_15_pad_type_0 -* +features_17_pad_type_0 +*, +name$ + +" +features_17_pad_type_0* val  " -custom*, -name$ - -" -features_15_pad_type_0v +customv const! -features_15_pad_0 +features_17_pad_0  -*! +*' +name + +" +features_17_pad_0*! val    -*' -name - -" -features_15_pad_0� -convD +� +conv +x + + input_173D weight: 8 6model_mobilevit_encoder_layer_3_1_conv_1x1_conv_weight strides -var_608& -pad_type - -features_15_pad_type_0 -x - - input_173 +var_607 groups  var_10 pad  -features_15_pad_0 +features_17_pad_0& +pad_type + +features_17_pad_type_0 dilations -var_610. - features_15 -  - -� +var_609- + features_17 +  + +P    b const -var_621 +var_620  -*" +* +name + + +" +op_620*" val    -�* -name - - -" -op_621d +� +d reshape -x - - features_15 shape -var_621- +var_620 +x + + features_17- patches_15   -� +� +   v @@ -6948,24 +7154,23 @@ patches_15 name  " -patches_17_perm_0c +patches_17_perm_0b const -var_624 +var_623  -*# -val +*" +val  - - - -��* + +  +P�* name  " -op_624p +op_623p transpose x @@ -6976,42 +7181,43 @@ patches_15 patches_17_perm_0/ transpose_42   -� +� +   -f -reshape -x - - transpose_42 +e +reshape shape -var_624. +var_623 +x + + transpose_42- -patches_19 -  - -� +patches_19 +  + +P � v const! patches_21_perm_0  -*! +*' +name + +" +patches_21_perm_0*! val    -*' -name - -" -patches_21_perm_0j +j const -var_630 +var_629  * @@ -7019,13 +7225,13 @@ patches_19  " -op_630** +op_629** val#    - ����������q + ����������p transpose x @@ -7033,25 +7239,25 @@ patches_19 patches_19 perm  -patches_21_perm_00 - transpose_41 -  +patches_21_perm_0/ + transpose_41 +    -� -�_ +� +P^ reshape shape -var_630 +var_629 x  - transpose_41' - input_175 -  + transpose_41& + input_175 +   -� -�z +� +Pz const input_177_axes_0 @@ -7067,61 +7273,61 @@ patches_19   -���������� +���������� layer_normT gammaK I -Gmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_weight -x - - input_175 +Gmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_weight epsilon  -var_7 +var_7 +x + + input_175 axes  input_177_axes_0Q betaI G -Emodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_bias' - input_177 -  +Emodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_before_bias& + input_177 +   -� -�� -linear^ -weightT -R -Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_weight +� +P� +linear x - input_177Z + input_177^ +weightT +R +Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_weightZ biasR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_bias" -x_33 -  +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_query_bias! +x_33 +   -� -�� -linear -x - - input_177\ +� +P� +linear\ weightR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_weightX +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_weight +x + + input_177X biasP N -Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias" -x_25 -  +Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias! +x_25 +   -� -�b +� +Pb const -var_658 +var_657  * @@ -7129,28 +7335,28 @@ Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias"  " -op_658*" +op_657*" val    -�0W -reshape -shape - -var_658 +�W +reshape x  -x_25' +x_25 +shape + +var_657' x_27    �  -0a +a const -var_660 +var_659  *! @@ -7165,116 +7371,116 @@ Lmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_key_bias"  " -op_660� -linear^ -weightT -R -Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_weight +op_659� +linear x input_177Z biasR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias" -x_29 -  +Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias^ +weightT +R +Pmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_weight! +x_29 +   -� -�b +� +Pb const -var_667 +var_666  -* -name - - -" -op_667*" +*" val    -�0W -reshape -shape - -var_667 +�* +name + + +" +op_666W +reshape x  -x_29' +x_29 +shape + +var_666' x_31    �  -0a +a const -var_669 +var_668  -*! +* +name + + +" +op_668*! val    -* +b +const +var_672 + + +* name  " -op_669b -const -var_673 - - -*" +op_672*" val    -�0* -name - - -" -op_673W -reshape -x - -x_33 +�W +reshape shape -var_673' +var_672 +x + +x_33' x_35    �  -0a +a const -var_675 +var_674  -* -name - - -" -op_675*! +*! val    -� +* +name + + +" +op_674� const -var_677_perm_0 +var_676_perm_0  *3 @@ -7287,18 +7493,18 @@ Nmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_attention_value_bias name  " - op_677_perm_0� + op_676_perm_0� const( attention_scores_5_transpose_x_0 -*6 -name. -& -$"" - attention_scores_5_transpose_x_0* +* val   -� +*6 +name. +& +$"" + attention_scores_5_transpose_x_0� const( attention_scores_5_transpose_y_0 *6 @@ -7313,7 +7519,7 @@ $"" transpose perm -var_660 +var_659 x  x_27/ @@ -7322,11 +7528,11 @@ $""   � -0o +o transpose perm  -var_677_perm_0 +var_676_perm_0 x  transpose_40/ @@ -7334,7 +7540,7 @@ $""     -0 + �` transpose x @@ -7342,20 +7548,20 @@ $"" x_35 perm -var_675/ +var_674/ transpose_38     � -0� -matmul -x - - transpose_383 +� +matmul3 transpose_x$ " attention_scores_5_transpose_x_0 +x + + transpose_38 y  transpose_373 @@ -7370,17 +7576,17 @@ $"" �r const _inversed_input_179_y_0 - * + *- +name% + +" +_inversed_input_179_y_0* val     -:�>*- -name% - -" -_inversed_input_179_y_0} +.�d>} mul y  @@ -7410,49 +7616,49 @@ $"" �{ const% context_layer_9_transpose_x_0 -* -val - - -*3 +*3 name+ # !" -context_layer_9_transpose_x_0{ -const% -context_layer_9_transpose_y_0 -* +context_layer_9_transpose_x_0* val   -*3 +{ +const% +context_layer_9_transpose_y_0 +*3 name+ # !" -context_layer_9_transpose_y_0` +context_layer_9_transpose_y_0* +val + + +` transpose x  x_31 perm -var_669/ +var_668/ transpose_39     � -0� +� matmul0 transpose_x!  -context_layer_9_transpose_x_00 - transpose_y! - -context_layer_9_transpose_y_0 +context_layer_9_transpose_x_0 x - input_181 + input_1810 + transpose_y! + +context_layer_9_transpose_y_0 y  transpose_392 @@ -7461,92 +7667,93 @@ $""   � -0a +a const -var_684 +var_683  -* -name - - -" -op_684*! +*! val    -b -const -var_689 +* +name + - -*" -val +" +op_683a +const +var_688  - -  -��* +* name  " -op_689k +op_688*! +val + + + + + +�Pk transpose x  context_layer_9 perm -var_684/ +var_683/ transpose_36    �  -0_ -reshape -shape - -var_689 +^ +reshape x  - transpose_36' - input_183 -  + transpose_36 +shape + +var_688& + input_183 +   -� -�� -linearW -biasO -M -Kmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_bias +� +P� +linear x - input_183[ + input_183W +biasO +M +Kmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_bias[ weightQ O -Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight' - input_185 -  +Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight& + input_185 +   -� -�V +� +PU add -y - - input_175 x - input_185' - input_187 -  + input_185 +y + + input_175& + input_187 +   -� -�z +� +Pz const input_189_axes_0 @@ -7562,44 +7769,44 @@ Mmodel_mobilevit_encoder_layer_3_1_transformer_0_attention_output_dense_weight'   -���������� +���������� -layer_norm -epsilon - -var_7S +layer_normP +betaH +F +Dmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_biasS gammaJ H -Fmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_weight -axes - -input_189_axes_0 +Fmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_weight x - input_187P -betaH -F -Dmodel_mobilevit_encoder_layer_3_1_transformer_0_layernorm_after_bias' - input_189 -  + input_187 +epsilon + +var_7 +axes + +input_189_axes_0& + input_189 +   -� -�� -linear -x - - input_189S -biasK -I -Gmodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_biasW +� +P� +linearW weightM K -Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weight' +Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weightS +biasK +I +Gmodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_bias +x + + input_189' input_191    � -�C +�C silu x @@ -7608,7 +7815,7 @@ Imodel_mobilevit_encoder_layer_3_1_transformer_0_intermediate_dense_weight'    � -�� +�� linearM biasE C @@ -7618,24 +7825,24 @@ E Cmodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_weight x - input_193' - input_195 -  + input_193& + input_195 +   -� -�V +� +PU add y input_187 x - input_195' - input_197 -  + input_195& + input_197 +   -� -�z +� +Pz const input_199_axes_0 @@ -7651,44 +7858,44 @@ Cmodel_mobilevit_encoder_layer_3_1_transformer_0_output_dense_weight   -���������� +���������� -layer_norm +layer_norm +axes + +input_199_axes_0 x input_197 epsilon  -var_7 -axes - -input_199_axes_0Q +var_7Q betaI G Emodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_biasT gammaK I -Gmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_weight' - input_199 -  +Gmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_before_weight& + input_199 +   -� -�� -linear^ -weightT -R -Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_weight -x - - input_199Z +� +P� +linearZ biasR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_bias" -x_45 -  +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_bias +x + + input_199^ +weightT +R +Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_query_weight! +x_45 +   -� -�� +� +P� linear x @@ -7698,121 +7905,121 @@ N Lmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_bias\ weightR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_weight" -x_37 -  +Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_key_weight! +x_37 +   -� -�b +� +Pb const -var_734 +var_733  -*" +* +name + + +" +op_733*" val    -�0* -name - - -" -op_734W +�W reshape x  x_37 shape -var_734' +var_733' x_39    �  -0a +a const -var_736 +var_735  -* -name - - -" -op_736*! +*! val    -� -linearZ +* +name + + +" +op_735� +linear +x + + input_199Z biasR P Nmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_bias^ weightT R -Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weight -x - - input_199" -x_41 -  +Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weight! +x_41 +   -� -�b +� +Pb const -var_743 +var_742  -*" +* +name + + +" +op_742*" val    -�0* -name - - -" -op_743W -reshape -x - -x_41 +�W +reshape shape -var_743' +var_742 +x + +x_41' x_43    �  -0a +a const -var_745 +var_744  -*! +* +name + + +" +op_744*! val    -* -name - - -" -op_745b +b const -var_749 +var_748  * @@ -7820,52 +8027,52 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_attention_value_weigh  " -op_749*" +op_748*" val    -�0W -reshape -shape - -var_749 +�W +reshape x  -x_45' +x_45 +shape + +var_748' x_47    �  -0a +a const -var_751 +var_750  -*! +* +name + + +" +op_750*! val    -* -name - - -" -op_751� +� const -var_753_perm_0 +var_752_perm_0  *# name  " - op_753_perm_0*3 + op_752_perm_0*3 val,  @@ -7885,40 +8092,40 @@ $"" attention_scores_7_transpose_x_0� const( attention_scores_7_transpose_y_0 -*6 -name. -& -$"" - attention_scores_7_transpose_y_0* +* val   -` - transpose -perm - -var_736 +*6 +name. +& +$"" + attention_scores_7_transpose_y_0` + transpose x  -x_39/ +x_39 +perm + +var_735/ transpose_35     � -0o +o transpose x  transpose_35 perm  -var_753_perm_0/ +var_752_perm_0/ transpose_32     -0 + �` transpose x @@ -7926,66 +8133,66 @@ $"" x_47 perm -var_751/ +var_750/ transpose_33     � -0� +� matmul3 - transpose_x$ -" - attention_scores_7_transpose_x_03 transpose_y$ " - attention_scores_7_transpose_y_0 -x - - transpose_33 + attention_scores_7_transpose_y_03 + transpose_x$ +" + attention_scores_7_transpose_x_0 y  - transpose_326 + transpose_32 +x + + transpose_336 attention_scores_7     � �r -const -_inversed_input_201_y_0 - *- -name% - -" -_inversed_input_201_y_0* +const +_inversed_input_201_y_0 + * val     -:�>} -mul -x - -attention_scores_7 +.�d>*- +name% + +" +_inversed_input_201_y_0} +mul y  -_inversed_input_201_y_07 +_inversed_input_201_y_0 +x + +attention_scores_77 _inversed_input_201     � �j -softmax +softmax +x + +_inversed_input_201 axis  -var_25 -x - -_inversed_input_201- +var_25- input_203    @@ -7994,15 +8201,15 @@ $"" �} const& context_layer_13_transpose_x_0 -*4 -name, -$ -"" -context_layer_13_transpose_x_0* +* val   -} +*4 +name, +$ +"" +context_layer_13_transpose_x_0} const& context_layer_13_transpose_y_0 *4 @@ -8014,123 +8221,124 @@ $""   ` - transpose -perm - -var_745 + transpose x  -x_43/ +x_43 +perm + +var_744/ transpose_34     � -0� +� matmul y  - transpose_34 -x - - input_2031 + transpose_341 transpose_y" context_layer_13_transpose_y_01 transpose_x" -context_layer_13_transpose_x_03 +context_layer_13_transpose_x_0 +x + + input_2033 context_layer_13     � -0a +a const -var_760 +var_759  -*! +* +name + + +" +op_759*! val    -* -name - - -" -op_760b +a const -var_765 +var_764  -*" -val - - - -  -��* +* name  " -op_765l - transpose -perm - -var_760 +op_764*! +val + + + + + +�Pl + transpose x  -context_layer_13/ +context_layer_13 +perm + +var_759/ transpose_31    �  -0_ +^ reshape x  transpose_31 shape -var_765' - input_205 -  +var_764& + input_205 +   -� -�� -linear[ -weightQ -O -Mmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_weightW +� +P� +linearW biasO M -Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias +Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias[ +weightQ +O +Mmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_weight x - input_205' - input_207 -  + input_205& + input_207 +   -� -�V +� +PU add -x - - input_207 y - input_197' - input_209 -  + input_197 +x + + input_207& + input_209 +   -� -�z +� +Pz const input_211_axes_0 @@ -8146,44 +8354,44 @@ Kmodel_mobilevit_encoder_layer_3_1_transformer_1_attention_output_dense_bias   -���������� +���������� -layer_norm -x - - input_209 +layer_norm epsilon  -var_7P +var_7S +gammaJ +H +Fmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_weight +x + + input_209P betaH F Dmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_bias axes  -input_211_axes_0S -gammaJ -H -Fmodel_mobilevit_encoder_layer_3_1_transformer_1_layernorm_after_weight' - input_211 -  +input_211_axes_0& + input_211 +   -� -�� -linearS -biasK -I -Gmodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_bias +� +P� +linear x input_211W weightM K -Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weight' +Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weightS +biasK +I +Gmodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_bias' input_213    � -�C +�C silu x @@ -8192,7 +8400,7 @@ Imodel_mobilevit_encoder_layer_3_1_transformer_1_intermediate_dense_weight'    � -�� +�� linear x @@ -8202,24 +8410,24 @@ C Amodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_biasQ weightG E -Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight' - input_217 -  +Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight& + input_217 +   -� -�V +� +PU add -y - - input_209 x - input_217' - input_219 -  + input_217 +y + + input_209& + input_219 +   -� -�z +� +Pz const input_221_axes_0 @@ -8235,122 +8443,122 @@ Cmodel_mobilevit_encoder_layer_3_1_transformer_1_output_dense_weight' name  " -input_221_axes_0� +input_221_axes_0� layer_normQ betaI G -Emodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_bias -epsilon - -var_7 -axes - -input_221_axes_0T +Emodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_biasT gammaK I -Gmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_weight +Gmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_before_weight +axes + +input_221_axes_0 x - input_219' - input_221 -  + input_219 +epsilon + +var_7& + input_221 +   -� -�� +� +P� linear^ weightT R -Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_weight -x - - input_221Z +Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_weightZ biasR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_bias" -x_57 -  +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_query_bias +x + + input_221! +x_57 +   -� -�� +� +P� linear\ weightR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_weightX -biasP -N -Lmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_bias +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_weight x - input_221" -x_49 -  + input_221X +biasP +N +Lmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_key_bias! +x_49 +   -� -�b +� +Pb const -var_810 +var_809  -* -name - - -" -op_810*" +*" val    -�0W -reshape -shape - -var_810 +�* +name + + +" +op_809W +reshape x  -x_49' +x_49 +shape + +var_809' x_51    �  -0a +a const -var_812 +var_811  -* -name - - -" -op_812*! +*! val    -� -linear +* +name + + +" +op_811� +linearZ +biasR +P +Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias x input_221^ weightT R -Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_weightZ -biasR -P -Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias" -x_53 -  +Pmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_weight! +x_53 +   -� -�b +� +Pb const -var_819 +var_818  * @@ -8358,45 +8566,45 @@ Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias  " -op_819*" +op_818*" val    -�0W -reshape -x - -x_53 +�W +reshape shape -var_819' +var_818 +x + +x_53' x_55    �  -0a +a const -var_821 +var_820  -*! +* +name + + +" +op_820*! val    -* -name - - -" -op_821b +b const -var_825 +var_824  *" @@ -8405,27 +8613,27 @@ Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias    -�0* +�* name  " -op_825W -reshape -x - -x_57 +op_824W +reshape shape -var_825' +var_824 +x + +x_57' x_59    �  -0a +a const -var_827 +var_826  * @@ -8433,7 +8641,7 @@ Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias  " -op_827*! +op_826*! val  @@ -8442,14 +8650,14 @@ Nmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_attention_value_bias  � const -var_829_perm_0 +var_828_perm_0  *# name  " - op_829_perm_0*3 + op_828_perm_0*3 val,  @@ -8484,17 +8692,17 @@ $"" x_51 perm -var_812/ +var_811/ transpose_30     � -0o +o transpose perm  -var_829_perm_0 +var_828_perm_0 x  transpose_30/ @@ -8502,28 +8710,28 @@ $""     -0 + �` - transpose -perm - -var_827 + transpose x  -x_59/ +x_59 +perm + +var_826/ transpose_28     � -0� -matmul3 +� +matmul +y + + transpose_273 transpose_x$ " attention_scores_9_transpose_x_0 -y - - transpose_27 x  transpose_283 @@ -8538,24 +8746,24 @@ $"" �r const _inversed_input_223_y_0 - *- -name% - -" -_inversed_input_223_y_0* + * val     -:�>} -mul -y - -_inversed_input_223_y_0 +.�d>*- +name% + +" +_inversed_input_223_y_0} +mul x  -attention_scores_97 +attention_scores_9 +y + +_inversed_input_223_y_07 _inversed_input_223    @@ -8578,30 +8786,30 @@ $"" �} const& context_layer_17_transpose_x_0 -* -val - - -*4 +*4 name, $ "" -context_layer_17_transpose_x_0} -const& -context_layer_17_transpose_y_0 -* +context_layer_17_transpose_x_0* val   -*4 +} +const& +context_layer_17_transpose_y_0 +*4 name, $ "" -context_layer_17_transpose_y_0` +context_layer_17_transpose_y_0* +val + + +` transpose perm -var_821 +var_820 x  x_55/ @@ -8610,28 +8818,28 @@ $""   � -0� -matmul -y - - transpose_29 +� +matmul1 + transpose_y" + +context_layer_17_transpose_y_0 x - input_2251 + input_225 +y + + transpose_291 transpose_x" -context_layer_17_transpose_x_01 - transpose_y" - -context_layer_17_transpose_y_03 +context_layer_17_transpose_x_03 context_layer_17     � -0a +a const -var_836 +var_835  * @@ -8639,16 +8847,16 @@ $""  " -op_836*! +op_835*! val    -b +a const -var_841 +var_840  * @@ -8656,17 +8864,18 @@ $""  " -op_841*" -val +op_840*! +val  - -  -��l + + + +�Pl transpose perm -var_836 +var_835 x  context_layer_17/ @@ -8675,19 +8884,19 @@ $""  �  -0_ -reshape -shape - -var_841 +^ +reshape x  - transpose_26' - input_227 -  + transpose_26 +shape + +var_840& + input_227 +   -� -�� +� +P� linearW biasO M @@ -8697,24 +8906,24 @@ Kmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_bias input_227[ weightQ O -Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight' - input_229 -  +Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight& + input_229 +   -� -�V +� +PU add x input_229 y - input_219' - input_231 -  + input_219& + input_231 +   -� -�z +� +Pz const input_233_axes_0 @@ -8730,36 +8939,36 @@ Mmodel_mobilevit_encoder_layer_3_1_transformer_2_attention_output_dense_weight' name  " -input_233_axes_0� +input_233_axes_0� -layer_normP -betaH -F -Dmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_bias +layer_norm x - input_231 -axes - -input_233_axes_0S + input_231S gammaJ H -Fmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_weight +Fmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_weight +axes + +input_233_axes_0 epsilon  -var_7' - input_233 -  +var_7P +betaH +F +Dmodel_mobilevit_encoder_layer_3_1_transformer_2_layernorm_after_bias& + input_233 +   -� -�� -linearS -biasK -I -Gmodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_biasW +� +P� +linearW weightM K -Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight +Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weightS +biasK +I +Gmodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_bias x input_233' @@ -8767,7 +8976,7 @@ Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight    � -�C +�C silu x @@ -8776,34 +8985,34 @@ Imodel_mobilevit_encoder_layer_3_1_transformer_2_intermediate_dense_weight    � -�� -linearQ -weightG -E -Cmodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_weight -x - - input_237M +�� +linearM biasE C -Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias' - input_239 -  - -� -�V -add +Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias x - input_239 + input_237Q +weightG +E +Cmodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_weight& + input_239 +  + +� +PU +add y - input_231' - input_241 -  + input_231 +x + + input_239& + input_241 +   -� -�z +� +Pz const input_243_axes_0 @@ -8819,44 +9028,44 @@ Amodel_mobilevit_encoder_layer_3_1_transformer_2_output_dense_bias'   -���������� +���������� -layer_normQ +layer_norm +axes + +input_243_axes_0 +x + + input_241Q betaI G Emodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_biasT gammaK I -Gmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_weight -x - - input_241 +Gmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_before_weight epsilon  -var_7 -axes - -input_243_axes_0' - input_243 -  +var_7& + input_243 +   -� -�� -linearZ -biasR -P -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_bias +� +P� +linear x - input_243^ + input_243Z +biasR +P +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_bias^ weightT R -Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_weight" -x_69 -  +Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_query_weight! +x_69 +   -� -�� +� +P� linear x @@ -8866,32 +9075,32 @@ N Lmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_bias\ weightR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight" -x_61 -  +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight! +x_61 +   -� -�b +� +Pb const -var_886 +var_885  -*" +* +name + + +" +op_885*" val    -�0* -name - - -" -op_886W +�W reshape shape -var_886 +var_885 x  x_61' @@ -8900,70 +9109,70 @@ Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_key_weight  �  -0a +a const -var_888 +var_887  -* -name - - -" -op_888*! +*! val    -� -linear -x - - input_243Z +* +name + + +" +op_887� +linearZ biasR P -Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_bias^ +Nmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_bias +x + + input_243^ weightT R -Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weight" -x_65 -  +Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weight! +x_65 +   -� -�b +� +Pb const -var_895 +var_894  -* -name - - -" -op_895*" +*" val    -�0W -reshape -x - -x_65 +�* +name + + +" +op_894W +reshape shape -var_895' +var_894 +x + +x_65' x_67    �  -0a +a const -var_897 +var_896  *! @@ -8978,9 +9187,9 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weigh  " -op_897b +op_896b const -var_901 +var_900  *" @@ -8989,16 +9198,16 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weigh    -�0* +�* name  " -op_901W +op_900W reshape shape -var_901 +var_900 x  x_69' @@ -9007,50 +9216,50 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weigh  �  -0a +a const -var_903 +var_902  -*! +* +name + + +" +op_902*! val    -* -name - - -" -op_903� +� const -var_905_perm_0 +var_904_perm_0  -*3 +*# +name + +" + op_904_perm_0*3 val,    -������������������*# -name - -" - op_905_perm_0� +������������������� const) !attention_scores_11_transpose_x_0 -* -val - - -*7 +*7 name/ ' %"# -!attention_scores_11_transpose_x_0� +!attention_scores_11_transpose_x_0* +val + + +� const) !attention_scores_11_transpose_y_0 *7 @@ -9062,55 +9271,55 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weigh   ` - transpose -x - -x_63 + transpose perm -var_888/ +var_887 +x + +x_63/ transpose_25     � -0o - transpose -x - - transpose_25 +o + transpose perm  -var_905_perm_0/ +var_904_perm_0 +x + + transpose_25/ transpose_22     -0 + �` - transpose -perm - -var_903 + transpose x  -x_71/ +x_71 +perm + +var_902/ transpose_23     � -0� -matmul4 - transpose_x% -# -!attention_scores_11_transpose_x_0 +� +matmul y  transpose_22 x  transpose_234 + transpose_x% +# +!attention_scores_11_transpose_x_04 transpose_y% # !attention_scores_11_transpose_y_07 @@ -9122,38 +9331,38 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weigh �r const _inversed_input_245_y_0 - *- -name% - -" -_inversed_input_245_y_0* + * val     -:�>~ -mul -x - -attention_scores_11 +.�d>*- +name% + +" +_inversed_input_245_y_0~ +mul y  -_inversed_input_245_y_07 +_inversed_input_245_y_0 +x + +attention_scores_117 _inversed_input_245     � �j -softmax -x - -_inversed_input_245 +softmax axis  -var_25- +var_25 +x + +_inversed_input_245- input_247    @@ -9182,123 +9391,124 @@ Pmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_attention_value_weigh   ` - transpose -perm - -var_897 + transpose x  -x_67/ +x_67 +perm + +var_896/ transpose_24     � -0� -matmul1 - transpose_x" - -context_layer_21_transpose_x_0 -x - - input_247 +� +matmul y  - transpose_241 + transpose_24 +x + + input_2471 transpose_y" -context_layer_21_transpose_y_03 +context_layer_21_transpose_y_01 + transpose_x" + +context_layer_21_transpose_x_03 context_layer_21     � -0a +a const -var_912 +var_911  -* -name - - -" -op_912*! +*! val    -b -const -var_917 +* +name + - -*" -val +" +op_911a +const +var_916  - -  -��* +* name  " -op_917l - transpose -x - -context_layer_21 +op_916*! +val + + + + + +�Pl + transpose perm -var_912/ +var_911 +x + +context_layer_21/ transpose_21    �  -0_ +^ reshape shape -var_917 +var_916 x  - transpose_21' - input_249 -  + transpose_21& + input_249 +   -� -�� -linear[ -weightQ -O -Mmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_weight -x - - input_249W +� +P� +linearW biasO M -Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias' - input_251 -  +Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias +x + + input_249[ +weightQ +O +Mmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_weight& + input_251 +   -� -�V +� +PU add y input_241 x - input_251' - input_253 -  + input_251& + input_253 +   -� -�z +� +Pz const input_255_axes_0 @@ -9314,36 +9524,36 @@ Kmodel_mobilevit_encoder_layer_3_1_transformer_3_attention_output_dense_bias' name  " -input_255_axes_0� +input_255_axes_0� layer_normS gammaJ H -Fmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_weightP +Fmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_weight +axes + +input_255_axes_0 +x + + input_253P betaH F Dmodel_mobilevit_encoder_layer_3_1_transformer_3_layernorm_after_bias epsilon  -var_7 -axes - -input_255_axes_0 +var_7& + input_255 +  + +� +P� +linear x - input_253' - input_255 -  - -� -�� -linearS + input_255S biasK I -Gmodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_bias -x - - input_255W +Gmodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_biasW weightM K Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight' @@ -9351,7 +9561,7 @@ Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight'    � -�C +�C silu x @@ -9360,34 +9570,34 @@ Imodel_mobilevit_encoder_layer_3_1_transformer_3_intermediate_dense_weight'    � -�� +�� linear x - input_259Q -weightG -E -Cmodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_weightM + input_259M biasE C -Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_bias' - input_261 -  +Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_biasQ +weightG +E +Cmodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_weight& + input_261 +   -� -�V +� +PU add -x - - input_261 y - input_253' - input_263 -  + input_253 +x + + input_261& + input_263 +   -� -�| +� +P| const! patches_23_axes_0 @@ -9403,46 +9613,46 @@ Amodel_mobilevit_encoder_layer_3_1_transformer_3_output_dense_bias'   -���������� +���������� -layer_norm? -gamma6 -4 -2model_mobilevit_encoder_layer_3_1_layernorm_weight -x - - input_263< +layer_norm +axes + +patches_23_axes_0< beta4 2 0model_mobilevit_encoder_layer_3_1_layernorm_bias epsilon  -var_7 -axes - -patches_23_axes_0( +var_7? +gamma6 +4 +2model_mobilevit_encoder_layer_3_1_layernorm_weight +x + + input_263' -patches_23 -  +patches_23 +   -� -�k +� +Pk const -var_945 +var_944  -*+ +* +name + + +" +op_944*+ val$    -����������* -name - - -" -op_945d +����������c reshape x @@ -9450,58 +9660,59 @@ patches_23 patches_23 shape -var_945. +var_944- -patches_27 -  +patches_27 +    -� -�x +� +Px const" -features_17_perm_0 +features_19_perm_0  -*! +*( +name  + +" +features_19_perm_0*! val    -*( -name  - -" -features_17_perm_0b +b const -var_956 +var_955  -*" +* +name + + +" +op_955*" val    -�* -name - - -" -op_956r +� +q transpose perm  -features_17_perm_0 +features_19_perm_0 x -patches_270 - transpose_20 -  - -� +patches_27/ + transpose_20 +  + +P � f reshape @@ -9510,31 +9721,32 @@ patches_270 transpose_20 shape -var_956. - features_19 +var_955. + features_21   -� +� +   x const" -features_21_perm_0 +features_23_perm_0  -*( -name  - -" -features_21_perm_0*! +*! val    -b +*( +name  + +" +features_23_perm_0a const -var_959 +var_958  * @@ -9542,57 +9754,59 @@ patches_270  " -op_959*" -val +op_958*! +val  - -  -� r - transpose -x - - features_19 + + + +P r + transpose perm  -features_21_perm_0/ +features_23_perm_0 +x + + features_21/ transpose_19   -� +� +   -d -reshape -x - - transpose_19 +c +reshape shape -var_959, - input_265 -  - -� +var_958 +x + + transpose_19+ + input_265 +  + +P    _ const -var_964 +var_963  -* +* +name + + +" +op_963* val    -* -name - - -" -op_964_ +_ const -var_966 +var_965  * @@ -9600,7 +9814,7 @@ patches_270  " -op_966* +op_965* val  @@ -9609,153 +9823,155 @@ patches_270 n const input_267_pad_type_0 -** -name" - -" -input_267_pad_type_0* +* val  " -customr +custom** +name" + +" +input_267_pad_type_0r const input_267_pad_0  -*! +*% +name + +" +input_267_pad_0*! val    -*% -name - -" -input_267_pad_0� -const6 -input_269_weight_0 -  -� -� +� +const4 +input_269_weight_0 +  +@ +P  *( name   " -input_269_weight_0*P -valI -  -� -� +input_269_weight_0*N +valG +  +@ +P  *% -@model_path/weights/weight.bin���� -const! -input_269_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*& +@model_path/weights/weight.bin���� +const +input_269_bias_0 + +  +@*& name  " -input_269_bias_0� -conv - dilations - -var_966$ +input_269_bias_0*< +val5 + +  +@*% +@model_path/weights/weight.bin���� +conv$ pad_type  input_267_pad_type_0 weight  -input_269_weight_0 +input_269_weight_0 + dilations + +var_965 +groups + + +var_10 pad  input_267_pad_0 strides -var_964 -bias - -input_269_bias_0 -groups - - -var_10 +var_963 x - input_265, - input_269 -  - -� + input_265 +bias + +input_269_bias_0+ + input_269 +  + +@   - J + I silu x - input_269. - features_23 -  - -� + input_269- + features_25 +  + +@    m const input_271_interleave_0 -* -val - - -*, +*, name$  -" -input_271_interleave_0� -concat -axis - - -var_10( +" +input_271_interleave_0* +val + + +� +concat( interleave  -input_271_interleave_0& +input_271_interleave_0 +axis + + +var_10& values input_167 - features_23, + features_25, input_271    -� +�    _ const -var_980 +var_979  -* -name - - -" -op_980* +* val    -_ +* +name + + +" +op_979_ const -var_982 +var_981  * @@ -9763,7 +9979,7 @@ interleave  " -op_982* +op_981* val  @@ -9772,16 +9988,16 @@ interleave n const input_273_pad_type_0 -** -name" - -" -input_273_pad_type_0* +* val  " -customr +custom** +name" + +" +input_273_pad_type_0r const input_273_pad_0 @@ -9797,81 +10013,83 @@ interleave name  " -input_273_pad_0� -const6 -input_275_weight_0 -  -� -� +input_273_pad_0� +const5 +input_275_weight_0 +  +@ +�  *( name   " -input_275_weight_0*P -valI -  -� -� +input_275_weight_0*O +valH +  +@ +�  *% -@model_path/weights/weight.bin���� -const! -input_275_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*& +@model_path/weights/weight.bin���� +const +input_275_bias_0 + +  +@*< +val5 + +  +@*% +@model_path/weights/weight.bin���*& name  " -input_275_bias_0� -conv -x - - input_271$ +input_275_bias_0� +conv +bias + +input_275_bias_0$ pad_type  input_273_pad_type_0 weight  -input_275_weight_0 +input_275_weight_0 +strides + +var_979 groups  -var_10 - dilations - -var_982 -bias - -input_275_bias_0 -strides - -var_980 +var_10 pad  -input_273_pad_0, - input_275 -  - -� +input_273_pad_0 + dilations + +var_981 +x + + input_271+ + input_275 +  + +@   - H + G silu x - input_275, - input_277 -  - -� + input_275+ + input_277 +  + +@    _ const -var_999 +var_998  * @@ -9885,22 +10103,22 @@ interleave  " -op_999a +op_998a const -var_1001 +var_1000  -* +* +name + + " +op_1000* val    -* -name - - " -op_1001n +n const input_279_pad_type_0 * @@ -9928,33 +10146,33 @@ interleave name  " -input_279_pad_0� -const6 -input_281_weight_0 -  -� -� - -*P -valI -  -� +input_279_pad_0� +const5 +input_281_weight_0 +  � +@  -*% -@model_path/weights/weight.bin���*( +*( name   " -input_281_weight_0� +input_281_weight_0*O +valH +  +� +@ + +*% +@model_path/weights/weight.bin���� const! input_281_bias_0   -�*= +�*= val6   -�*% -@model_path/weights/weight.bin���*& +�*% +@model_path/weights/weight.bin���*& name  " @@ -9968,28 +10186,28 @@ interleave input_281_bias_0 strides -var_999$ -pad_type - -input_279_pad_type_0 +var_998 +pad + +input_279_pad_0 dilations -var_1001 -pad - -input_279_pad_0 -weight - -input_281_weight_0 +var_1000 groups  -var_10, +var_10 +weight + +input_281_weight_0$ +pad_type + +input_279_pad_type_0, input_281    -� +�    H silu @@ -9999,51 +10217,51 @@ interleave input_283    -� +�    a const -var_1013 +var_1012  -* -name - - " -op_1013* +* val    -a +* +name + + " +op_1012a const -var_1015 +var_1014  -* +* +name + + " +op_1014* val    -* -name - - " -op_1015n +n const input_285_pad_type_0 -* +** +name" + +" +input_285_pad_type_0* val  " -custom** -name" - -" -input_285_pad_type_0r +customr const input_285_pad_0 @@ -10063,40 +10281,50 @@ interleave const5 input_287_weight_0   -� +�   -*O +*( +name  + +" +input_287_weight_0*O valH   -� +�   *% -@model_path/weights/weight.bin���*( -name  - -" -input_287_weight_0� +@model_path/weights/weight.bin�Ŵ� const! input_287_bias_0   -�*& +�*& name  " input_287_bias_0*= val6   -�*% -@model_path/weights/weight.bin���� -conv -weight - -input_287_weight_0 +�*% +@model_path/weights/weight.bin��� +conv +pad + +input_285_pad_0 bias  -input_287_bias_0 +input_287_bias_0 +x + + input_283 + dilations + + +var_1014$ +pad_type + +input_285_pad_type_0 groups  @@ -10104,24 +10332,14 @@ interleave strides -var_1013$ -pad_type - -input_285_pad_type_0 -x - - input_283 - dilations - - -var_1015 -pad - -input_285_pad_0, +var_1012 +weight + +input_287_weight_0, input_287    -� +�    H silu @@ -10131,18 +10349,18 @@ interleave input_289    -� +�    a const -var_1027 +var_1026  * name  " -op_1027* +op_1026* val  @@ -10150,32 +10368,32 @@ interleave  a const -var_1029 +var_1028  -* +* +name + + " +op_1028* val    -* -name - - " -op_1029n +n const input_291_pad_type_0 -** -name" - -" -input_291_pad_type_0* +* val  " -customr +custom** +name" + +" +input_291_pad_type_0r const input_291_pad_0 @@ -10191,80 +10409,82 @@ interleave name  " -input_291_pad_0� -const6 -input_293_weight_0 -  -� -� +input_291_pad_0� +const5 +input_293_weight_0 +  +P +�  *( name   " -input_293_weight_0*P -valI -  -� -� +input_293_weight_0*O +valH +  +P +�  *% -@model_path/weights/weight.bin���� -const! -input_293_bias_0 -   -�*& +@model_path/weights/weight.bin��� +const +input_293_bias_0 + +  +P*& name  " -input_293_bias_0*= -val6 -   -�*% -@model_path/weights/weight.bin���� +input_293_bias_0*< +val5 + +  +P*% +@model_path/weights/weight.bin���� conv bias  -input_293_bias_0 +input_293_bias_0$ +pad_type + +input_291_pad_type_0 x - input_289 -strides - + input_289 +groups -var_1027 - dilations + +var_10 +strides -var_1029 +var_1026 weight  -input_293_weight_0 -groups +input_293_weight_0 + dilations - -var_10 + +var_1028 pad  -input_291_pad_0$ -pad_type - -input_291_pad_type_0, - input_293 -  - -� +input_291_pad_0+ + input_293 +  + +P    a const -var_1046 +var_1045  * name  " -op_1046* +op_1045* val  @@ -10272,14 +10492,14 @@ interleave  a const -var_1048 +var_1047  * name  " -op_1048* +op_1047* val  @@ -10288,16 +10508,16 @@ interleave n const input_295_pad_type_0 -* +** +name" + +" +input_295_pad_type_0* val  " -custom** -name" - -" -input_295_pad_type_0r +customr const input_295_pad_0 @@ -10313,83 +10533,85 @@ interleave name  " -input_295_pad_0� -const6 -input_297_weight_0 -  -� -� +input_295_pad_0� +const4 +input_297_weight_0 +  +P +P  -*P -valI -  -� -� +*N +valG +  +P +P  *% -@model_path/weights/weight.bin���*( +@model_path/weights/weight.bin���*( name   " -input_297_weight_0� -const! -input_297_bias_0 -   -�*& +input_297_weight_0� +const +input_297_bias_0 + +  +P*< +val5 + +  +P*% +@model_path/weights/weight.bin���*& name  " -input_297_bias_0*= -val6 -   -�*% -@model_path/weights/weight.bin���� -conv +input_297_bias_0� +conv +pad + +input_295_pad_0 +groups + + +var_10$ +pad_type + +input_295_pad_type_0 bias  -input_297_bias_0 - dilations +input_297_bias_0 +strides -var_1048 +var_1045 weight  -input_297_weight_0 -strides +input_297_weight_0 + dilations -var_1046 +var_1047 x - input_293 -pad - -input_295_pad_0 -groups - - -var_10$ -pad_type - -input_295_pad_type_0, - input_297 -  - -� + input_293+ + input_297 +  + +P   - H + G silu x - input_297, - input_299 -  - -� + input_297+ + input_299 +  + +P    a const -var_1059 +var_1058  * @@ -10402,9 +10624,9 @@ interleave name  " -op_1059a +op_1058a const -var_1061 +var_1060  * @@ -10417,68 +10639,68 @@ interleave name  " -op_1061r +op_1060r const -features_25_pad_type_0 -*, -name$ - -" -features_25_pad_type_0* +features_27_pad_type_0 +* val  " -customv +custom*, +name$ + +" +features_27_pad_type_0v const! -features_25_pad_0 +features_27_pad_0  *' name  " -features_25_pad_0*! +features_27_pad_0*! val    -� -conv -x - - input_299& -pad_type - -features_25_pad_type_0 +� +conv strides -var_1059 -pad - -features_25_pad_0 - dilations - - -var_1061D +var_1058D weight: 8 6model_mobilevit_encoder_layer_4_1_conv_1x1_conv_weight groups  -var_10. - features_25 -  - -� +var_10 + dilations + + +var_1060 +pad + +features_27_pad_0 +x + + input_299& +pad_type + +features_27_pad_type_0- + features_27 +  + +`    d const -var_1072 +var_1071  *" @@ -10487,23 +10709,23 @@ interleave    -�* +� * name  " -op_1072e +op_1071e reshape shape -var_1072 +var_1071 x - features_25- + features_27- patches_29   -� +�    v @@ -10511,94 +10733,93 @@ patches_29 patches_31_perm_0  -*! +*' +name + +" +patches_31_perm_0*! val    -*' -name - -" -patches_31_perm_0e +d const -var_1075 +var_1074  -* +*" +val + + + +  +`�* name  " -op_1075*# -val - - - - - -��p - transpose -perm - -patches_31_perm_0 +op_1074p + transpose x -patches_29/ +patches_29 +perm + +patches_31_perm_0/ transpose_18   -� +�    -g +f reshape shape -var_1075 +var_1074 x  - transpose_18. + transpose_18- -patches_33 -  - -� +patches_33 +  + +` � v const! patches_35_perm_0  -*' -name - -" -patches_35_perm_0*! +*! val    -l +*' +name + +" +patches_35_perm_0l const -var_1081 +var_1080  -* -name - - " -op_1081** +** val#    - ����������q + ����������* +name + + " +op_1080p transpose perm  @@ -10606,26 +10827,26 @@ patches_33 x -patches_330 - transpose_17 -  +patches_33/ + transpose_17 +    -� -�` +� +`_ reshape -x - - transpose_17 shape -var_1081' - input_301 -  +var_1080 +x + + transpose_17& + input_301 +   -� -�z +� +`z const input_303_axes_0 @@ -10641,29 +10862,29 @@ patches_330 name  " -input_303_axes_0� +input_303_axes_0� -layer_normT -gammaK -I -Gmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_weightQ +layer_normQ betaI G -Emodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_bias +Emodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_biasT +gammaK +I +Gmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_before_weight +epsilon + +var_7 x input_301 axes  -input_303_axes_0 -epsilon - -var_7' - input_303 -  +input_303_axes_0& + input_303 +   -� -�� +� +`� linear x @@ -10673,89 +10894,89 @@ R Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_weightZ biasR P -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_bias" -x_81 -  +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_query_bias! +x_81 +   -� -�� -linearX +� +`� +linear +x + + input_303X biasP N Lmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_bias\ weightR P -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_weight -x - - input_303" -x_73 -  +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_key_weight! +x_73 +   -� -�d +� +`d const -var_1108 +var_1107  * name  " -op_1108*" +op_1107*" val    -�<X -reshape -x - -x_73 +�X +reshape shape -var_1108' +var_1107 +x + +x_73' x_75    �  -<c +c const -var_1110 +var_1109  -* -name - - " -op_1110*! +*! val    -� -linear^ -weightT -R -Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_weight +* +name + + " +op_1109� +linear x - input_303Z + input_303^ +weightT +R +Pmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_weightZ biasR P -Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias" -x_77 -  +Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias! +x_77 +   -� -�d +� +`d const -var_1117 +var_1116  *" @@ -10764,16 +10985,16 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias    -�<* +�* name  " -op_1117X +op_1116X reshape shape -var_1117 +var_1116 x  x_77' @@ -10782,16 +11003,16 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias  �  -<c +c const -var_1119 +var_1118  * name  " -op_1119*! +op_1118*! val  @@ -10800,7 +11021,7 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias  d const -var_1123 +var_1122  *" @@ -10809,11 +11030,11 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias    -�<* +�* name  " -op_1123X +op_1122X reshape x  @@ -10821,22 +11042,22 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias shape -var_1123' +var_1122' x_83    �  -<c +c const -var_1125 +var_1124  * name  " -op_1125*! +op_1124*! val  @@ -10845,31 +11066,31 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias  � const -var_1127_perm_0 +var_1126_perm_0  -*$ -name - -" -op_1127_perm_0*3 +*3 val,    -������������������� +������������������*$ +name + +" +op_1126_perm_0� const) !attention_scores_13_transpose_x_0 -*7 -name/ -' -%"# -!attention_scores_13_transpose_x_0* +* val   -� +*7 +name/ +' +%"# +!attention_scores_13_transpose_x_0� const) !attention_scores_13_transpose_y_0 *7 @@ -10881,24 +11102,24 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias   a - transpose + transpose +x + +x_75 perm -var_1110 -x - -x_75/ +var_1109/ transpose_16     � -<p +p transpose perm  -var_1127_perm_0 +var_1126_perm_0 x  transpose_16/ @@ -10906,7 +11127,7 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias     -< + �a transpose x @@ -10915,23 +11136,23 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias perm -var_1125/ +var_1124/ transpose_14     � -<� -matmul4 - transpose_x% -# -!attention_scores_13_transpose_x_04 - transpose_y% -# -!attention_scores_13_transpose_y_0 +� +matmul x  - transpose_14 + transpose_144 + transpose_y% +# +!attention_scores_13_transpose_y_04 + transpose_x% +# +!attention_scores_13_transpose_x_0 y  transpose_137 @@ -10943,38 +11164,38 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias �r const _inversed_input_305_y_0 - *- -name% - -" -_inversed_input_305_y_0* + * val     -�2>~ -mul -y - -_inversed_input_305_y_0 +�Q>*- +name% + +" +_inversed_input_305_y_0~ +mul x  -attention_scores_137 +attention_scores_13 +y + +_inversed_input_305_y_07 _inversed_input_305     � �j -softmax +softmax +x + +_inversed_input_305 axis  -var_25 -x - -_inversed_input_305- +var_25- input_307    @@ -10994,20 +11215,20 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias context_layer_25_transpose_x_0} const& context_layer_25_transpose_y_0 -* -val - - -*4 +*4 name, $ "" -context_layer_25_transpose_y_0a +context_layer_25_transpose_y_0* +val + + +a transpose perm -var_1119 +var_1118 x  x_79/ @@ -11016,17 +11237,17 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias   � -<� +� matmul x - input_3071 - transpose_x" - -context_layer_25_transpose_x_0 + input_307 y  transpose_151 + transpose_x" + +context_layer_25_transpose_x_01 transpose_y" context_layer_25_transpose_y_03 @@ -11035,38 +11256,39 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias   � -<c +c const -var_1134 +var_1133  * name  " -op_1134*! +op_1133*! val    -d +c const -var_1139 +var_1138  -*" -val +*! +val  - -  -��* + + + +�`* name  " -op_1139m +op_1138m transpose x  @@ -11074,69 +11296,69 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_attention_value_bias perm -var_1134/ +var_1133/ transpose_12    �  -<` +_ reshape shape -var_1139 +var_1138 x  - transpose_12' - input_309 -  + transpose_12& + input_309 +   -� -�� +� +`� linearW biasO M -Kmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_bias -x - - input_309[ +Kmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_bias[ weightQ O -Mmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_weight' - input_311 -  +Mmodel_mobilevit_encoder_layer_4_1_transformer_0_attention_output_dense_weight +x + + input_309& + input_311 +   -� -�V +� +`U add x input_311 y - input_301' - input_313 -  + input_301& + input_313 +   -� -�z +� +`z const input_315_axes_0  -*' +*& +name + +" +input_315_axes_0*' val     -���������*& -name - -" -input_315_axes_0� +���������� layer_norm axes @@ -11144,36 +11366,36 @@ layer_norm input_315_axes_0 x - input_313S -gammaJ -H -Fmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_weightP + input_313P betaH F -Dmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_bias +Dmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_biasS +gammaJ +H +Fmodel_mobilevit_encoder_layer_4_1_transformer_0_layernorm_after_weight epsilon  -var_7' - input_315 -  +var_7& + input_315 +   -� -�� +� +`� linearS biasK I -Gmodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_bias -x - - input_315W +Gmodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_biasW weightM K -Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight' +Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight +x + + input_315' input_317    � -�C +�C silu x @@ -11182,34 +11404,34 @@ Imodel_mobilevit_encoder_layer_4_1_transformer_0_intermediate_dense_weight'    � -�� -linearM -biasE -C -Amodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_bias +�� +linear x input_319Q weightG E -Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weight' - input_321 -  +Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weightM +biasE +C +Amodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_bias& + input_321 +   -� -�V +� +`U add -x - - input_321 y - input_313' - input_323 -  + input_313 +x + + input_321& + input_323 +   -� -�z +� +`z const input_325_axes_0 @@ -11225,74 +11447,74 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_0_output_dense_weight' name  " -input_325_axes_0� +input_325_axes_0� -layer_normT -gammaK -I -Gmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_weight -x - - input_323 -axes - -input_325_axes_0 +layer_norm epsilon  -var_7Q +var_7 +x + + input_323Q betaI G -Emodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_bias' - input_325 -  +Emodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_bias +axes + +input_325_axes_0T +gammaK +I +Gmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_before_weight& + input_325 +   -� -�� -linearZ -biasR -P -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_bias^ +� +`� +linear^ weightT R Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_weight x - input_325" -x_93 -  + input_325Z +biasR +P +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_query_bias! +x_93 +   -� -�� -linear -x - - input_325X +� +`� +linearX biasP N -Lmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_bias\ +Lmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_bias +x + + input_325\ weightR P -Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight" -x_85 -  +Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight! +x_85 +   -� -�d +� +`d const -var_1184 +var_1183  * name  " -op_1184*" +op_1183*" val    -�<X +�X reshape x  @@ -11300,46 +11522,46 @@ Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_key_weight shape -var_1184' +var_1183' x_87    �  -<c +c const -var_1186 +var_1185  -* -name - - " -op_1186*! +*! val    -� -linear -x - - input_325Z +* +name + + " +op_1185� +linearZ biasR P Nmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_bias^ weightT R -Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weight" -x_89 -  +Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weight +x + + input_325! +x_89 +   -� -�d +� +`d const -var_1193 +var_1192  *" @@ -11348,34 +11570,34 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh    -�<* +�* name  " -op_1193X -reshape +op_1192X +reshape +x + +x_89 shape -var_1193 -x - -x_89' +var_1192' x_91    �  -<c +c const -var_1195 +var_1194  * name  " -op_1195*! +op_1194*! val  @@ -11384,20 +11606,20 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh  d const -var_1199 +var_1198  -* -name - - " -op_1199*" +*" val    -�<X +�* +name + + " +op_1198X reshape x  @@ -11405,57 +11627,46 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh shape -var_1199' +var_1198' x_95    �  -<c +c const -var_1201 +var_1200  -* -name - - " -op_1201*! +*! val    -� +* +name + + " +op_1200� const -var_1203_perm_0 +var_1202_perm_0  -*3 +*$ +name + +" +op_1202_perm_0*3 val,    -������������������*$ -name - -" -op_1203_perm_0� +������������������� const) !attention_scores_15_transpose_x_0 -*7 -name/ -' -%"# -!attention_scores_15_transpose_x_0* -val - - -� -const) -!attention_scores_15_transpose_y_0 * val  @@ -11464,39 +11675,50 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh name/ ' %"# -!attention_scores_15_transpose_y_0a - transpose +!attention_scores_15_transpose_x_0� +const) +!attention_scores_15_transpose_y_0 +*7 +name/ +' +%"# +!attention_scores_15_transpose_y_0* +val + + +a + transpose +x + +x_87 perm -var_1186 -x - -x_87/ +var_1185/ transpose_11     � -<o - transpose -x - - transpose_11 +o + transpose perm  -var_1203_perm_0. +var_1202_perm_0 +x + + transpose_11. transpose_8     -< + �` transpose perm -var_1201 +var_1200 x  x_95. @@ -11505,20 +11727,20 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh   � -<� +� matmul x - transpose_9 -y - - transpose_84 + transpose_94 transpose_y% # !attention_scores_15_transpose_y_04 transpose_x% # -!attention_scores_15_transpose_x_07 +!attention_scores_15_transpose_x_0 +y + + transpose_87 attention_scores_15    @@ -11537,14 +11759,14 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh   -�2>~ -mul -x - -attention_scores_15 +�Q>~ +mul y  -_inversed_input_327_y_07 +_inversed_input_327_y_0 +x + +attention_scores_157 _inversed_input_327    @@ -11591,7 +11813,7 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh perm -var_1195 +var_1194 x  x_91/ @@ -11600,71 +11822,72 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh   � -<� -matmul1 - transpose_y" - -context_layer_29_transpose_y_01 - transpose_x" - -context_layer_29_transpose_x_0 +� +matmul x - input_329 + input_3291 + transpose_y" + +context_layer_29_transpose_y_0 y  - transpose_103 + transpose_101 + transpose_x" + +context_layer_29_transpose_x_03 context_layer_29     � -<c +c const -var_1210 +var_1209  -*! +* +name + + " +op_1209*! val    -* -name - - " -op_1210d +c const -var_1215 - - -*" -val +var_1214  - -  -��* +* name  " -op_1215l - transpose -x - -context_layer_29 +op_1214*! +val + + + + + +�`l + transpose perm -var_1210. +var_1209 +x + +context_layer_29. transpose_7    �  -<_ +^ reshape x @@ -11672,63 +11895,63 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_attention_value_weigh shape -var_1215' - input_331 -  +var_1214& + input_331 +   -� -�� -linear -x - - input_331[ +� +`� +linear[ weightQ O -Mmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_weightW +Mmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_weight +x + + input_331W biasO M -Kmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_bias' - input_333 -  +Kmodel_mobilevit_encoder_layer_4_1_transformer_1_attention_output_dense_bias& + input_333 +   -� -�V +� +`U add y input_323 x - input_333' - input_335 -  + input_333& + input_335 +   -� -�z +� +`z const input_337_axes_0  -*& -name - -" -input_337_axes_0*' +*' val     -���������� +���������*& +name + +" +input_337_axes_0� -layer_normP -betaH -F -Dmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_bias +layer_norm epsilon  -var_7S +var_7P +betaH +F +Dmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_biasS gammaJ H Fmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_weight @@ -11737,12 +11960,12 @@ Fmodel_mobilevit_encoder_layer_4_1_transformer_1_layernorm_after_weight input_337_axes_0 x - input_335' - input_337 -  + input_335& + input_337 +   -� -�� +� +`� linearS biasK I @@ -11757,7 +11980,7 @@ Imodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_weight    � -�C +�C silu x @@ -11766,34 +11989,34 @@ Imodel_mobilevit_encoder_layer_4_1_transformer_1_intermediate_dense_weight    � -�� -linearM -biasE -C -Amodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_biasQ +�� +linearQ weightG E Cmodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_weight x - input_341' - input_343 -  + input_341M +biasE +C +Amodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_bias& + input_343 +   -� -�V +� +`U add -y - - input_335 x - input_343' - input_345 -  + input_343 +y + + input_335& + input_345 +   -� -�z +� +`z const input_347_axes_0 @@ -11809,184 +12032,184 @@ Cmodel_mobilevit_encoder_layer_4_1_transformer_1_output_dense_weight name  " -input_347_axes_0� +input_347_axes_0� -layer_norm +layer_normQ +betaI +G +Emodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_bias +axes + +input_347_axes_0T +gammaK +I +Gmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_weight epsilon  var_7 x - input_345Q -betaI -G -Emodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_biasT -gammaK -I -Gmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_before_weight -axes - -input_347_axes_0' - input_347 -  + input_345& + input_347 +   -� -�� +� +`� linearZ biasR P -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_bias -x - - input_347^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_bias^ weightT R -Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_weight# -x_105 -  - -� -�� -linear +Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_query_weight x - input_347\ -weightR -P -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_weightX + input_347" +x_105 +  + +� +`� +linearX biasP N -Lmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_bias" -x_97 -  +Lmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_bias\ +weightR +P +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_key_weight +x + + input_347! +x_97 +   -� -�d +� +`d const -var_1260 +var_1259  -* -name - - " -op_1260*" +*" val    -�<X -reshape -x - -x_97 +�* +name + + " +op_1259X +reshape shape -var_1260' +var_1259 +x + +x_97' x_99    �  -<c +c const -var_1262 +var_1261  * name  " -op_1262*! +op_1261*! val    -� +� linearZ biasR P -Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_bias -x - - input_347^ +Nmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_bias^ weightT R -Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weight# -x_101 -  +Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weight +x + + input_347" +x_101 +   -� -�d +� +`d const -var_1269 +var_1268  * name  " -op_1269*" +op_1268*" val    -�<Z -reshape -x - -x_101 +�Z +reshape shape -var_1269( +var_1268 +x + +x_101( x_103    �  -<c +c const -var_1271 +var_1270  -* -name - - " -op_1271*! +*! val    -d +* +name + + " +op_1270d const -var_1275 +var_1274  * name  " -op_1275*" +op_1274*" val    -�<Z +�Z reshape shape -var_1275 +var_1274 x  x_105( @@ -11995,25 +12218,25 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh  �  -<c +c const -var_1277 +var_1276  -*! +* +name + + " +op_1276*! val    -* -name - - " -op_1277� +� const -var_1279_perm_0 +var_1278_perm_0  *3 @@ -12026,7 +12249,7 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh name  " -op_1279_perm_0} +op_1278_perm_0} const& attention_scores_transpose_x_0 * @@ -12040,41 +12263,41 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh attention_scores_transpose_x_0} const& attention_scores_transpose_y_0 -* -val - - -*4 +*4 name, $ "" -attention_scores_transpose_y_0` - transpose +attention_scores_transpose_y_0* +val + + +` + transpose +x + +x_99 perm -var_1262 -x - -x_99. +var_1261. transpose_6     � -<n +n transpose x transpose_6 perm  -var_1279_perm_0. +var_1278_perm_0. transpose_3     -< + �a transpose x @@ -12083,14 +12306,17 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh perm -var_1277. +var_1276. transpose_4     � -<� -matmul1 +� +matmul +y + + transpose_31 transpose_x" attention_scores_transpose_x_01 @@ -12099,10 +12325,7 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh attention_scores_transpose_y_0 x - transpose_4 -y - - transpose_34 + transpose_44 attention_scores    @@ -12111,38 +12334,38 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh �r const _inversed_input_349_y_0 - * + *- +name% + +" +_inversed_input_349_y_0* val     -�2>*- -name% - -" -_inversed_input_349_y_0{ -mul -y - -_inversed_input_349_y_0 +�Q>{ +mul x  -attention_scores7 +attention_scores +y + +_inversed_input_349_y_07 _inversed_input_349     � �j -softmax +softmax +x + +_inversed_input_349 axis  -var_25 -x - -_inversed_input_349- +var_25- input_351    @@ -12175,7 +12398,7 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh perm -var_1271 +var_1270 x  x_103. @@ -12184,17 +12407,17 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh   � -<� -matmul1 - transpose_y" - -context_layer_33_transpose_y_0 +� +matmul x input_3511 transpose_x" -context_layer_33_transpose_x_0 +context_layer_33_transpose_x_01 + transpose_y" + +context_layer_33_transpose_y_0 y transpose_53 @@ -12203,38 +12426,39 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh   � -<c +c const -var_1286 +var_1285  -* -name - - " -op_1286*! +*! val    -d +* +name + + " +op_1285c const -var_1291 +var_1290  -* +*! +val + + + + + +�`* name  " -op_1291*" -val - - - -  -��l +op_1290l transpose x  @@ -12242,13 +12466,13 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh perm -var_1286. +var_1285. transpose_2    �  -<_ +^ reshape x @@ -12256,12 +12480,12 @@ Pmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_attention_value_weigh shape -var_1291' - input_353 -  +var_1290& + input_353 +   -� -�� +� +`� linearW biasO M @@ -12271,62 +12495,62 @@ O Mmodel_mobilevit_encoder_layer_4_1_transformer_2_attention_output_dense_weight x - input_353' - input_355 -  + input_353& + input_355 +   -� -�V +� +`U add -x - - input_355 y - input_345' - input_357 -  + input_345 +x + + input_355& + input_357 +   -� -�z +� +`z const input_359_axes_0  -*& -name - -" -input_359_axes_0*' +*' val     -���������� - -layer_normP -betaH -F -Dmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_bias -epsilon - -var_7S -gammaJ -H -Fmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_weight +���������*& +name + +" +input_359_axes_0� + +layer_norm axes  -input_359_axes_0 +input_359_axes_0 +epsilon + +var_7P +betaH +F +Dmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_bias x - input_357' - input_359 -  + input_357S +gammaJ +H +Fmodel_mobilevit_encoder_layer_4_1_transformer_2_layernorm_after_weight& + input_359 +   -� -�� +� +`� linearS biasK I @@ -12341,7 +12565,7 @@ Imodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_weight    � -�C +�C silu x @@ -12350,7 +12574,7 @@ Imodel_mobilevit_encoder_layer_4_1_transformer_2_intermediate_dense_weight    � -�� +�� linearM biasE C @@ -12360,65 +12584,65 @@ Amodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_bias input_363Q weightG E -Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight' - input_365 -  +Cmodel_mobilevit_encoder_layer_4_1_transformer_2_output_dense_weight& + input_365 +   -� -�V +� +`U add x input_365 y - input_357' - input_367 -  + input_357& + input_367 +   -� -�| +� +`| const! patches_37_axes_0  *' +name + +" +patches_37_axes_0*' val     -���������*' -name - -" -patches_37_axes_0� +���������� layer_norm< beta4 2 -0model_mobilevit_encoder_layer_4_1_layernorm_bias? -gamma6 -4 -2model_mobilevit_encoder_layer_4_1_layernorm_weight +0model_mobilevit_encoder_layer_4_1_layernorm_bias axes  -patches_37_axes_0 +patches_37_axes_0? +gamma6 +4 +2model_mobilevit_encoder_layer_4_1_layernorm_weight x input_367 epsilon  -var_7( +var_7' -patches_37 -  +patches_37 +   -� -�m +� +`m const -var_1319 +var_1318  *+ @@ -12431,31 +12655,31 @@ patches_37 name  " -op_1319b -reshape -shape - - -var_1319 +op_1318a +reshape x -patches_37+ -patches -  +patches_37 +shape + + +var_1318* +patches +    -� -�x +� +`x const" -features_27_perm_0 +features_29_perm_0  *( name   " -features_27_perm_0*! +features_29_perm_0*! val  @@ -12464,114 +12688,115 @@ patches_37+  d const -var_1330 +var_1329  -*" +* +name + + " +op_1329*" val    -�* -name - - " -op_1330n - transpose -x - -patches +� m + transpose perm  -features_27_perm_0/ - transpose_1 -  - -� +features_29_perm_0 +x + +patches. + transpose_1 +  + +` � f -reshape +reshape +x + + transpose_1 shape -var_1330 -x - - transpose_1. - features_29 +var_1329. + features_31   -� +�    x const" -features_31_perm_0 +features_33_perm_0  -*! +*( +name  + +" +features_33_perm_0*! val    -*( -name  - -" -features_31_perm_0d +c const -var_1333 +var_1332  * name  " -op_1333*" -val +op_1332*! +val  - -  -� q + + + +` q transpose perm  -features_31_perm_0 +features_33_perm_0 x - features_29. + features_31. transpose_0   -� +�    -d -reshape -x - - transpose_0 +c +reshape shape -var_1333, - input_369 -  - -� +var_1332 +x + + transpose_0+ + input_369 +  + +`    a const -var_1338 +var_1337  * name  " -op_1338* +op_1337* val  @@ -12579,7 +12804,7 @@ patches_37+  a const -var_1340 +var_1339  * @@ -12592,107 +12817,109 @@ patches_37+ name  " -op_1340n +op_1339n const input_371_pad_type_0 -** -name" - -" -input_371_pad_type_0* +* val  " -customr +custom** +name" + +" +input_371_pad_type_0r const input_371_pad_0  -*% -name - -" -input_371_pad_0*! +*! val    -� -const6 -input_373_weight_0 -  -� -� - -*P -valI -  -� -� +*% +name + +" +input_371_pad_0� +const4 +input_373_weight_0 +  +P +`  -*% -@model_path/weights/weight.bin���*( +*( name   " -input_373_weight_0� -const! -input_373_bias_0 -   -�*= -val6 -   -�*% -@model_path/weights/weight.bin���*& +input_373_weight_0*N +valG +  +P +` + +*% +@model_path/weights/weight.bin���� +const +input_373_bias_0 + +  +P*< +val5 + +  +P*% +@model_path/weights/weight.bin���*& name  " -input_373_bias_0� -conv +input_373_bias_0� +conv +weight + +input_373_weight_0 groups  -var_10 -bias - -input_373_bias_0 -pad - -input_371_pad_0 -weight - -input_373_weight_0 - dilations +var_10 +strides -var_1340$ +var_1337 +bias + +input_373_bias_0 +x + + input_369$ pad_type  -input_371_pad_type_0 -strides +input_371_pad_type_0 + dilations -var_1338 -x - - input_369, - input_373 -  - -� +var_1339 +pad + +input_371_pad_0+ + input_373 +  + +P   - G + F silu x - input_373+ -features -  - -� + input_373* +features +  + +P    m const @@ -12706,11 +12933,7 @@ patches_37+  " input_375_interleave_0� -concat -axis - - -var_10# +concat# values input_293 @@ -12720,15 +12943,19 @@ patches_37+ interleave  -input_375_interleave_0, +input_375_interleave_0 +axis + + +var_10, input_375    -� +�    a const -var_1354 +var_1353  * @@ -12741,154 +12968,156 @@ interleave name  " -op_1354a +op_1353a const -var_1356 +var_1355  -* -name - - " -op_1356* +* val    -n +* +name + + " +op_1355n const input_377_pad_type_0 -** -name" - -" -input_377_pad_type_0* +* val  " -customr +custom** +name" + +" +input_377_pad_type_0r const input_377_pad_0  -*! +*% +name + +" +input_377_pad_0*! val    -*% -name - -" -input_377_pad_0� -const6 -input_379_weight_0 -  -� -� - -*P -valI -  -� -� +� +const5 +input_379_weight_0 +  +P +�  -*% -@model_path/weights/weight.bin���*( +*( name   " -input_379_weight_0� -const! -input_379_bias_0 -   -�*& +input_379_weight_0*O +valH +  +P +� + +*% +@model_path/weights/weight.bin���� +const +input_379_bias_0 + +  +P*< +val5 + +  +P*% +@model_path/weights/weight.bin���*& name  " -input_379_bias_0*= -val6 -   -�*% -@model_path/weights/weight.bin��� � -conv$ -pad_type - -input_377_pad_type_0 -pad - -input_377_pad_0 -groups - - -var_10 +input_379_bias_0� +conv weight  input_379_weight_0 bias  -input_379_bias_0 - dilations +input_379_bias_0 +pad + +input_377_pad_0 +strides -var_1356 -strides +var_1353 + dilations + +var_1355 +groups -var_1354 + +var_10 x - input_375, - input_379 -  - -� + input_375$ +pad_type + +input_377_pad_type_0+ + input_379 +  + +P   - H + G silu x - input_379, - input_381 -  - -� + input_379+ + input_381 +  + +P    a const -var_1381 +var_1380  -* -name - - " -op_1381* +* val    -a -const -var_1383 - - -* +* name  " -op_1383* +op_1380a +const +var_1382 + + +* val    -n +* +name + + " +op_1382n const input_383_pad_type_0 ** @@ -12905,32 +13134,32 @@ interleave input_383_pad_0  -*! +*% +name + +" +input_383_pad_0*! val    -*% -name - -" -input_383_pad_0� -const6 -input_385_weight_0 -  -� -� +� +const5 +input_385_weight_0 +  +� +P  -*P -valI -  -� -� +*O +valH +  +� +P  *% -@model_path/weights/weight.bin��� *( +@model_path/weights/weight.bin���*( name   " @@ -12942,7 +13171,7 @@ interleave val6   �*% -@model_path/weights/weight.bin��� *& +@model_path/weights/weight.bin���*& name  " @@ -12950,31 +13179,31 @@ interleave conv$ pad_type  -input_383_pad_type_0 -bias - -input_385_bias_0 +input_383_pad_type_0 pad  input_383_pad_0 x - input_381 -strides + input_381 + dilations -var_1381 +var_1382 +bias + +input_385_bias_0 groups  var_10 weight  -input_385_weight_0 - dilations +input_385_weight_0 +strides -var_1383, +var_1380, input_385    @@ -12985,21 +13214,21 @@ interleave x input_385+ -var_1391 +var_1390    �    a const -var_1395 +var_1394  * name  " -op_1395* +op_1394* val  @@ -13007,66 +13236,66 @@ interleave  a const -var_1397 +var_1396  -* +* +name + + " +op_1396* val    -* -name - - " -op_1397n +n const input_387_pad_type_0 -* +** +name" + +" +input_387_pad_type_0* val  " -custom** -name" - -" -input_387_pad_type_0r +customr const input_387_pad_0  -*% -name - -" -input_387_pad_0*! +*! val    -� -const6 -input_389_weight_0 -  -� -� +*% +name + +" +input_387_pad_0� +const5 +input_389_weight_0 +  +� +P  *( name   " -input_389_weight_0*P -valI -  -� -� +input_389_weight_0*O +valH +  +� +P  *% -@model_path/weights/weight.bin��� � +@model_path/weights/weight.bin���� const! input_389_bias_0   @@ -13074,40 +13303,39 @@ interleave val6   �*% -@model_path/weights/weight.bin��� -*& +@model_path/weights/weight.bin�͕*& name  " input_389_bias_0� -conv - dilations - - -var_1397$ -pad_type - -input_387_pad_type_0 +conv weight  input_389_weight_0 strides -var_1395 -groups - - -var_10 -bias - -input_389_bias_0 +var_1394$ +pad_type + +input_387_pad_type_0 x - input_381 + input_381 +bias + +input_389_bias_0 pad  -input_387_pad_0, +input_387_pad_0 +groups + + +var_10 + dilations + + +var_1396, input_389    @@ -13118,14 +13346,14 @@ interleave x input_389+ -var_1405 +var_1404    �    a const -var_1409 +var_1408  * @@ -13138,16 +13366,16 @@ interleave name  " -op_1409a +op_1408a const -var_1411 +var_1410  * name  " -op_1411* +op_1410* val  @@ -13181,67 +13409,65 @@ interleave name  " -input_391_pad_0� -const6 -input_393_weight_0 -  -� -� - -*P -valI -  -� -� +input_391_pad_0� +const5 +input_393_weight_0 +  +� +P  -*% -@model_path/weights/weight.bin��� -*( +*( name   " -input_393_weight_0� +input_393_weight_0*O +valH +  +� +P + +*% +@model_path/weights/weight.bin�֕� const! input_393_bias_0   -�*= -val6 -   -�*% -@model_path/weights/weight.bin��� -*& +�*& name  " -input_393_bias_0� -conv +input_393_bias_0*= +val6 +   +�*% +@model_path/weights/weight.bin���� +conv +x + + input_381 strides -var_1409 +var_1408 weight  -input_393_weight_0 - dilations - - -var_1411 +input_393_weight_0 +pad + +input_391_pad_0 groups  -var_10 -pad - -input_391_pad_0 -x - - input_381$ -pad_type - -input_391_pad_type_0 +var_10 bias  -input_393_bias_0, +input_393_bias_0$ +pad_type + +input_391_pad_type_0 + dilations + + +var_1410, input_393    @@ -13252,21 +13478,21 @@ interleave x input_393+ -var_1419 +var_1418    �    a const -var_1423 +var_1422  * name  " -op_1423* +op_1422* val  @@ -13274,107 +13500,106 @@ interleave  a const -var_1425 +var_1424  -* -name - - " -op_1425* +* val    -n +* +name + + " +op_1424n const input_395_pad_type_0 -** -name" - -" -input_395_pad_type_0* +* val  " -customr +custom** +name" + +" +input_395_pad_type_0r const input_395_pad_0  -*% -name - -" -input_395_pad_0*! +*! val    -� -const6 -input_397_weight_0 -  -� -� +*% +name + +" +input_395_pad_0� +const5 +input_397_weight_0 +  +� +P  *( name   " -input_397_weight_0*P -valI -  -� -� +input_397_weight_0*O +valH +  +� +P  *% -@model_path/weights/weight.bin��� -� +@model_path/weights/weight.bin���� const! input_397_bias_0   -�*= -val6 -   -�*% -@model_path/weights/weight.bin��� *& +�*& name  " -input_397_bias_0� -conv$ -pad_type - -input_395_pad_type_0 -weight - -input_397_weight_0 +input_397_bias_0*= +val6 +   +�*% +@model_path/weights/weight.bin���� +conv pad  -input_395_pad_0 +input_395_pad_0 +bias + +input_397_bias_0 + dilations + + +var_1424 strides -var_1423 -bias - -input_397_bias_0 +var_1422$ +pad_type + +input_395_pad_type_0 groups  var_10 x - input_381 - dilations - - -var_1425, + input_381 +weight + +input_397_weight_0, input_397    @@ -13385,7 +13610,7 @@ interleave x input_397+ -var_1433 +var_1432    � @@ -13408,73 +13633,73 @@ interleave input_399_axesg const input_399_keep_dims -* -val - - -*) +*) name!  " -input_399_keep_dims� - reduce_mean$ - keep_dims - -input_399_keep_dims -x - - input_381 +input_399_keep_dims* +val + + +� + reduce_mean axes  -input_399_axes, - input_399 -  - -� +input_399_axes +x + + input_381$ + keep_dims + +input_399_keep_dims+ + input_399 +  + +P  a const -var_1443 +var_1442  -* +* +name + + " +op_1442* val    -* -name - - " -op_1443a +a const -var_1445 +var_1444  -* -name - - " -op_1445* +* val    -n +* +name + + " +op_1444n const input_401_pad_type_0 -* +** +name" + +" +input_401_pad_type_0* val  " -custom** -name" - -" -input_401_pad_type_0r +customr const input_401_pad_0 @@ -13490,65 +13715,65 @@ interleave name  " -input_401_pad_0� -const6 -input_403_weight_0 -  -� -� +input_401_pad_0� +const5 +input_403_weight_0 +  +� +P  *( name   " -input_403_weight_0*P -valI -  -� -� +input_403_weight_0*O +valH +  +� +P  *% -@model_path/weights/weight.bin��� � +@model_path/weights/weight.bin���� const! input_403_bias_0   -�*= -val6 -   -�*% -@model_path/weights/weight.bin��� *& +�*& name  " -input_403_bias_0� -conv -x - - input_399 +input_403_bias_0*= +val6 +   +�*% +@model_path/weights/weight.bin���� +conv groups  -var_10 +var_10 +strides + + +var_1442 pad  -input_401_pad_0$ -pad_type - -input_401_pad_type_0 +input_401_pad_0 weight  -input_403_weight_0 - dilations - - -var_1445 +input_403_weight_0$ +pad_type + +input_401_pad_type_0 +x + + input_399 bias  -input_403_bias_0 -strides +input_403_bias_0 + dilations -var_1443, +var_1444, input_403    @@ -13566,12 +13791,12 @@ interleave   const& -var_1455_scale_factor_height_0 +var_1454_scale_factor_height_0  *3 name+ # !" -op_1455_scale_factor_height_0* +op_1454_scale_factor_height_0* val   @@ -13579,43 +13804,43 @@ interleave  B} const% -var_1455_scale_factor_width_0 - *2 -name* -" - " -op_1455_scale_factor_width_0* +var_1454_scale_factor_width_0 + * val     -Bp +B*2 +name* +" + " +op_1454_scale_factor_width_0p const -var_1455_align_corners_0 +var_1454_align_corners_0 *- name%  " -op_1455_align_corners_0* +op_1454_align_corners_0* val   � -upsample_bilinear- - align_corners - -var_1455_align_corners_0 +upsample_bilinear x - input_4057 -scale_factor_width! - -var_1455_scale_factor_width_09 + input_4059 scale_factor_height" -var_1455_scale_factor_height_0+ -var_1455 +var_1454_scale_factor_height_07 +scale_factor_width! + +var_1454_scale_factor_width_0- + align_corners + +var_1454_align_corners_0+ +var_1454    � @@ -13623,40 +13848,40 @@ interleave  m const input_407_interleave_0 -* -val - - -*, +*, name$  " -input_407_interleave_0� -concat -axis - - -var_10( +input_407_interleave_0* +val + + +� +concat( interleave  -input_407_interleave_0F +input_407_interleave_0 +axis + + +var_10F values< -var_1391 +var_1390 -var_1405 +var_1404 -var_1419 +var_1418 -var_1433 +var_1432 -var_1455, +var_1454, input_407    @@ -13665,22 +13890,22 @@ interleave    a const -var_1461 +var_1460  -* -name - - " -op_1461* +* val    -a +* +name + + " +op_1460a const -var_1463 +var_1462  * @@ -13693,19 +13918,19 @@ interleave name  " -op_1463n +op_1462n const input_409_pad_type_0 -** -name" - -" -input_409_pad_type_0* +* val  " -customr +custom** +name" + +" +input_409_pad_type_0r const input_409_pad_0 @@ -13729,11 +13954,7 @@ interleave �   -*( -name  - -" -input_411_weight_0*P +*P valI   � @@ -13741,7 +13962,11 @@ interleave   *% -@model_path/weights/weight.bin��� � +@model_path/weights/weight.bin���*( +name  + +" +input_411_weight_0� const! input_411_bias_0   @@ -13753,32 +13978,32 @@ interleave val6   �*% -@model_path/weights/weight.bin��� � -conv$ -pad_type - -input_409_pad_type_0 +@model_path/weights/weight.bin���� +conv +x + + input_407 +weight + +input_411_weight_0 groups  -var_10 -weight - -input_411_weight_0 -x - - input_407 -pad - -input_409_pad_0 +var_10 dilations -var_1463 +var_1462 strides -var_1461 +var_1460 +pad + +input_409_pad_0$ +pad_type + +input_409_pad_type_0 bias  input_411_bias_0, @@ -13799,35 +14024,35 @@ interleave    a const -var_1478 +var_1477  -* -name - - " -op_1478* +* val    -a -const -var_1480 - - -* +* name  " -op_1480* +op_1477a +const +var_1479 + + +* val    -f +* +name + + " +op_1479f const input_pad_type_0 *& @@ -13845,21 +14070,35 @@ interleave  *! +name + +" + input_pad_0*! val    -*! -name - -" - input_pad_0� +� conv x - input_413 + input_413 +pad + + input_pad_0 + dilations + + +var_1479 +strides + + +var_1477> +weight4 +2 +0model_seg_head_classifier_classifier_conv_weight pad_type  input_pad_type_0 @@ -13869,21 +14108,7 @@ interleave var_10: bias2 0 -.model_seg_head_classifier_classifier_conv_bias -pad - - input_pad_0 -strides - - -var_1478 - dilations - - -var_1480> -weight4 -2 -0model_seg_head_classifier_classifier_conv_weight' +.model_seg_head_classifier_classifier_conv_bias' input    @@ -13892,30 +14117,30 @@ interleave  r const x_scale_factor_height_0 - *- -name% - -" -x_scale_factor_height_0* + * val     -�Ap +�A*- +name% + +" +x_scale_factor_height_0p const x_scale_factor_width_0 - *, -name$ - -" -x_scale_factor_width_0* + * val     -�Ac +�A*, +name$ + +" +x_scale_factor_width_0c const x_align_corners_0 *' @@ -13927,19 +14152,19 @@ interleave   � -upsample_bilinear2 -scale_factor_height - -x_scale_factor_height_0 +upsample_bilinear& + align_corners + +x_align_corners_0 x  -input0 +input2 +scale_factor_height + +x_scale_factor_height_00 scale_factor_width  -x_scale_factor_width_0& - align_corners - -x_align_corners_0% +x_scale_factor_width_0% x    @@ -13947,18 +14172,18 @@ interleave � �P const -var_1495 -* -name - - " -op_1495* +var_1494 +* val   -P +* +name + + " +op_1494P const -var_1496 +var_1495 * val  @@ -13967,20 +14192,20 @@ interleave name  " -op_1496w - reduce_argmax +op_1495w + reduce_argmax + +x + +x keep_dims -var_1496 +var_1495 axis -var_1495 - -x - -x) +var_1494) classLabels