| <?xml version="1.0"?>
|
| <net name="model_2" version="11">
|
| <layers>
|
| <layer id="0" name="input_3" type="Parameter" version="opset1">
|
| <data shape="?,48,48,3" element_type="f32" />
|
| <rt_info>
|
| <attribute name="old_api_map_element_type" version="0" value="f16" />
|
| </rt_info>
|
| <output>
|
| <port id="0" precision="FP32" names="input_3">
|
| <dim>-1</dim>
|
| <dim>48</dim>
|
| <dim>48</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="1" name="Constant_23" type="Const" version="opset1">
|
| <data element_type="i64" shape="4" offset="0" size="32" />
|
| <output>
|
| <port id="0" precision="I64">
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="2" name="Transpose9" type="Transpose" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>48</dim>
|
| <dim>48</dim>
|
| <dim>3</dim>
|
| </port>
|
| <port id="1" precision="I64">
|
| <dim>4</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="adjusted_input3">
|
| <dim>-1</dim>
|
| <dim>3</dim>
|
| <dim>48</dim>
|
| <dim>48</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="3" name="conv2d_8/kernel:0_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="32, 3, 3, 3" offset="32" size="1728" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="4" name="conv2d_8/kernel:0" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="conv2d_8/kernel:0">
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="5" name="conv2d_8/WithoutBiases" type="Convolution" version="opset1">
|
| <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="valid" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>3</dim>
|
| <dim>48</dim>
|
| <dim>48</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>46</dim>
|
| <dim>46</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="6" name="Reshape_36_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 32, 1, 1" offset="1760" size="64" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="7" name="Reshape_36" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="8" name="conv2d_8" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>46</dim>
|
| <dim>46</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="convolution_output3">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>46</dim>
|
| <dim>46</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="9" name="p_re_lu_7_initializer_22_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 32, 1, 1" offset="1824" size="64" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="10" name="p_re_lu_7_initializer_22" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_7_initializer_22">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="11" name="p_re_lu_7" type="PReLU" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>46</dim>
|
| <dim>46</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>32</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="p_re_lu_7/add:0">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>46</dim>
|
| <dim>46</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="12" name="p_re_lu_7/add:0_pooling" type="MaxPool" version="opset8">
|
| <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" kernel="3, 3" rounding_type="floor" auto_pad="same_upper" index_element_type="i64" axis="0" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>46</dim>
|
| <dim>46</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_7/add:0_pooling0">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>23</dim>
|
| <dim>23</dim>
|
| </port>
|
| <port id="2" precision="I64">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>23</dim>
|
| <dim>23</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="13" name="conv2d_9/kernel:0_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="64, 32, 3, 3" offset="1888" size="36864" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>64</dim>
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="14" name="conv2d_9/kernel:0" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>64</dim>
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="conv2d_9/kernel:0">
|
| <dim>64</dim>
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="15" name="conv2d_9/WithoutBiases" type="Convolution" version="opset1">
|
| <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="valid" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>32</dim>
|
| <dim>23</dim>
|
| <dim>23</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>64</dim>
|
| <dim>32</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>21</dim>
|
| <dim>21</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="16" name="Reshape_51_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 64, 1, 1" offset="38752" size="128" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="17" name="Reshape_51" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="18" name="conv2d_9" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>21</dim>
|
| <dim>21</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="convolution_output2">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>21</dim>
|
| <dim>21</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="19" name="p_re_lu_8_initializer_20_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 64, 1, 1" offset="38880" size="128" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="20" name="p_re_lu_8_initializer_20" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_8_initializer_20">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="21" name="p_re_lu_8" type="PReLU" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>21</dim>
|
| <dim>21</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="p_re_lu_8/add:0">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>21</dim>
|
| <dim>21</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="22" name="p_re_lu_8/add:0_pooling" type="MaxPool" version="opset8">
|
| <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" kernel="3, 3" rounding_type="floor" auto_pad="valid" index_element_type="i64" axis="0" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>21</dim>
|
| <dim>21</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_8/add:0_pooling0">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>10</dim>
|
| <dim>10</dim>
|
| </port>
|
| <port id="2" precision="I64">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>10</dim>
|
| <dim>10</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="23" name="conv2d_10/kernel:0_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="64, 64, 3, 3" offset="39008" size="73728" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>64</dim>
|
| <dim>64</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="24" name="conv2d_10/kernel:0" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>64</dim>
|
| <dim>64</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="conv2d_10/kernel:0">
|
| <dim>64</dim>
|
| <dim>64</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="25" name="conv2d_10/WithoutBiases" type="Convolution" version="opset1">
|
| <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="valid" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>10</dim>
|
| <dim>10</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>64</dim>
|
| <dim>64</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>8</dim>
|
| <dim>8</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="26" name="Reshape_66_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 64, 1, 1" offset="112736" size="128" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="27" name="Reshape_66" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="28" name="conv2d_10" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>8</dim>
|
| <dim>8</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="convolution_output1">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>8</dim>
|
| <dim>8</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="29" name="p_re_lu_9_initializer_18_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 64, 1, 1" offset="112864" size="128" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="30" name="p_re_lu_9_initializer_18" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_9_initializer_18">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="31" name="p_re_lu_9" type="PReLU" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>8</dim>
|
| <dim>8</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>64</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="p_re_lu_9/add:0">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>8</dim>
|
| <dim>8</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="32" name="p_re_lu_9/add:0_pooling" type="MaxPool" version="opset8">
|
| <data strides="2, 2" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" kernel="2, 2" rounding_type="floor" auto_pad="same_upper" index_element_type="i64" axis="0" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>8</dim>
|
| <dim>8</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_9/add:0_pooling0">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>4</dim>
|
| <dim>4</dim>
|
| </port>
|
| <port id="2" precision="I64">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>4</dim>
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="33" name="conv2d_11/kernel:0_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="128, 64, 2, 2" offset="112992" size="65536" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>128</dim>
|
| <dim>64</dim>
|
| <dim>2</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="34" name="conv2d_11/kernel:0" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>128</dim>
|
| <dim>64</dim>
|
| <dim>2</dim>
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="conv2d_11/kernel:0">
|
| <dim>128</dim>
|
| <dim>64</dim>
|
| <dim>2</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="35" name="conv2d_11/WithoutBiases" type="Convolution" version="opset1">
|
| <data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="valid" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>64</dim>
|
| <dim>4</dim>
|
| <dim>4</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>128</dim>
|
| <dim>64</dim>
|
| <dim>2</dim>
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>128</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="36" name="Reshape_81_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 128, 1, 1" offset="178528" size="256" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="37" name="Reshape_81" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="38" name="conv2d_11" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>128</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="convolution_output">
|
| <dim>-1</dim>
|
| <dim>128</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="39" name="p_re_lu_10_initializer_16_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 128, 1, 1" offset="178784" size="256" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="40" name="p_re_lu_10_initializer_16" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="p_re_lu_10_initializer_16">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="41" name="p_re_lu_10" type="PReLU" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>128</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>128</dim>
|
| <dim>1</dim>
|
| <dim>1</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="p_re_lu_10/add:0">
|
| <dim>-1</dim>
|
| <dim>128</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="42" name="Constant_84" type="Const" version="opset1">
|
| <data element_type="i64" shape="4" offset="179040" size="32" />
|
| <output>
|
| <port id="0" precision="I64">
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="43" name="PushTranspose_17" type="Transpose" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>128</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| </port>
|
| <port id="1" precision="I64">
|
| <dim>4</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="push_transpose_out_17">
|
| <dim>-1</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| <dim>128</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="44" name="Constant_86" type="Const" version="opset1">
|
| <data element_type="i64" shape="2" offset="179072" size="16" />
|
| <rt_info>
|
| <attribute name="precise" version="0" />
|
| </rt_info>
|
| <output>
|
| <port id="0" precision="I64">
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="45" name="flatten_1" type="Reshape" version="opset1">
|
| <data special_zero="true" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>3</dim>
|
| <dim>3</dim>
|
| <dim>128</dim>
|
| </port>
|
| <port id="1" precision="I64">
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="flatten_1/Reshape:0">
|
| <dim>-1</dim>
|
| <dim>1152</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="46" name="Transpose_1219_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="256, 1152" offset="179088" size="589824" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>256</dim>
|
| <dim>1152</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="47" name="Transpose_1219" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>256</dim>
|
| <dim>1152</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>256</dim>
|
| <dim>1152</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="48" name="dense_3" type="MatMul" version="opset1">
|
| <data transpose_a="false" transpose_b="true" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>1152</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>256</dim>
|
| <dim>1152</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="dense_30">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="49" name="Constant_1243_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 256" offset="768912" size="512" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="50" name="Constant_1243" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="51" name="Add3" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="biased_tensor_name3">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="52" name="slope_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="256" offset="769424" size="512" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="53" name="slope" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="slope">
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="54" name="p_re_lu_11" type="PReLU" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="p_re_lu_11/add:0">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="55" name="Transpose_1222_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="2, 256" offset="769936" size="1024" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>2</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="56" name="Transpose_1222" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>2</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>2</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="57" name="dense_4" type="MatMul" version="opset1">
|
| <data transpose_a="false" transpose_b="true" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>2</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="dense_40">
|
| <dim>-1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="58" name="Constant_1244_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 2" offset="770960" size="4" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="59" name="Constant_1244" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="60" name="Add2" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>2</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="biased_tensor_name2">
|
| <dim>-1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="61" name="softmax_2" type="SoftMax" version="opset8">
|
| <data axis="1" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32" names="softmax_2">
|
| <dim>-1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="63" name="Transpose_1225_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="10, 256" offset="770964" size="5120" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>10</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="64" name="Transpose_1225" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>10</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>10</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="65" name="dense_6" type="MatMul" version="opset1">
|
| <data transpose_a="false" transpose_b="true" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>10</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="dense_60">
|
| <dim>-1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="66" name="Constant_1245_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 10" offset="776084" size="20" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="67" name="Constant_1245" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="68" name="dense_60" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>10</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="dense_6">
|
| <dim>-1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="70" name="Transpose_1228_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="4, 256" offset="776104" size="2048" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>4</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="71" name="Transpose_1228" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>4</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>4</dim>
|
| <dim>256</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="72" name="dense_5" type="MatMul" version="opset1">
|
| <data transpose_a="false" transpose_b="true" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>256</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>4</dim>
|
| <dim>256</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="dense_50">
|
| <dim>-1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="73" name="Constant_1246_compressed" type="Const" version="opset1">
|
| <data element_type="f16" shape="1, 4" offset="778152" size="8" />
|
| <output>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="74" name="Constant_1246" type="Convert" version="opset1">
|
| <data destination_type="f32" />
|
| <rt_info>
|
| <attribute name="decompression" version="0" />
|
| </rt_info>
|
| <input>
|
| <port id="0" precision="FP16">
|
| <dim>1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="75" name="dense_50" type="Add" version="opset1">
|
| <data auto_broadcast="numpy" />
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>4</dim>
|
| </port>
|
| <port id="1" precision="FP32">
|
| <dim>1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </input>
|
| <output>
|
| <port id="2" precision="FP32" names="dense_5">
|
| <dim>-1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </output>
|
| </layer>
|
| <layer id="76" name="dense_5/sink_port_0" type="Result" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>4</dim>
|
| </port>
|
| </input>
|
| </layer>
|
| <layer id="69" name="dense_6/sink_port_0" type="Result" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>10</dim>
|
| </port>
|
| </input>
|
| </layer>
|
| <layer id="62" name="softmax_2/sink_port_0" type="Result" version="opset1">
|
| <input>
|
| <port id="0" precision="FP32">
|
| <dim>-1</dim>
|
| <dim>2</dim>
|
| </port>
|
| </input>
|
| </layer>
|
| </layers>
|
| <edges>
|
| <edge from-layer="0" from-port="0" to-layer="2" to-port="0" />
|
| <edge from-layer="1" from-port="0" to-layer="2" to-port="1" />
|
| <edge from-layer="2" from-port="2" to-layer="5" to-port="0" />
|
| <edge from-layer="3" from-port="0" to-layer="4" to-port="0" />
|
| <edge from-layer="4" from-port="1" to-layer="5" to-port="1" />
|
| <edge from-layer="5" from-port="2" to-layer="8" to-port="0" />
|
| <edge from-layer="6" from-port="0" to-layer="7" to-port="0" />
|
| <edge from-layer="7" from-port="1" to-layer="8" to-port="1" />
|
| <edge from-layer="8" from-port="2" to-layer="11" to-port="0" />
|
| <edge from-layer="9" from-port="0" to-layer="10" to-port="0" />
|
| <edge from-layer="10" from-port="1" to-layer="11" to-port="1" />
|
| <edge from-layer="11" from-port="2" to-layer="12" to-port="0" />
|
| <edge from-layer="12" from-port="1" to-layer="15" to-port="0" />
|
| <edge from-layer="13" from-port="0" to-layer="14" to-port="0" />
|
| <edge from-layer="14" from-port="1" to-layer="15" to-port="1" />
|
| <edge from-layer="15" from-port="2" to-layer="18" to-port="0" />
|
| <edge from-layer="16" from-port="0" to-layer="17" to-port="0" />
|
| <edge from-layer="17" from-port="1" to-layer="18" to-port="1" />
|
| <edge from-layer="18" from-port="2" to-layer="21" to-port="0" />
|
| <edge from-layer="19" from-port="0" to-layer="20" to-port="0" />
|
| <edge from-layer="20" from-port="1" to-layer="21" to-port="1" />
|
| <edge from-layer="21" from-port="2" to-layer="22" to-port="0" />
|
| <edge from-layer="22" from-port="1" to-layer="25" to-port="0" />
|
| <edge from-layer="23" from-port="0" to-layer="24" to-port="0" />
|
| <edge from-layer="24" from-port="1" to-layer="25" to-port="1" />
|
| <edge from-layer="25" from-port="2" to-layer="28" to-port="0" />
|
| <edge from-layer="26" from-port="0" to-layer="27" to-port="0" />
|
| <edge from-layer="27" from-port="1" to-layer="28" to-port="1" />
|
| <edge from-layer="28" from-port="2" to-layer="31" to-port="0" />
|
| <edge from-layer="29" from-port="0" to-layer="30" to-port="0" />
|
| <edge from-layer="30" from-port="1" to-layer="31" to-port="1" />
|
| <edge from-layer="31" from-port="2" to-layer="32" to-port="0" />
|
| <edge from-layer="32" from-port="1" to-layer="35" to-port="0" />
|
| <edge from-layer="33" from-port="0" to-layer="34" to-port="0" />
|
| <edge from-layer="34" from-port="1" to-layer="35" to-port="1" />
|
| <edge from-layer="35" from-port="2" to-layer="38" to-port="0" />
|
| <edge from-layer="36" from-port="0" to-layer="37" to-port="0" />
|
| <edge from-layer="37" from-port="1" to-layer="38" to-port="1" />
|
| <edge from-layer="38" from-port="2" to-layer="41" to-port="0" />
|
| <edge from-layer="39" from-port="0" to-layer="40" to-port="0" />
|
| <edge from-layer="40" from-port="1" to-layer="41" to-port="1" />
|
| <edge from-layer="41" from-port="2" to-layer="43" to-port="0" />
|
| <edge from-layer="42" from-port="0" to-layer="43" to-port="1" />
|
| <edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
|
| <edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
|
| <edge from-layer="45" from-port="2" to-layer="48" to-port="0" />
|
| <edge from-layer="46" from-port="0" to-layer="47" to-port="0" />
|
| <edge from-layer="47" from-port="1" to-layer="48" to-port="1" />
|
| <edge from-layer="48" from-port="2" to-layer="51" to-port="0" />
|
| <edge from-layer="49" from-port="0" to-layer="50" to-port="0" />
|
| <edge from-layer="50" from-port="1" to-layer="51" to-port="1" />
|
| <edge from-layer="51" from-port="2" to-layer="54" to-port="0" />
|
| <edge from-layer="52" from-port="0" to-layer="53" to-port="0" />
|
| <edge from-layer="53" from-port="1" to-layer="54" to-port="1" />
|
| <edge from-layer="54" from-port="2" to-layer="57" to-port="0" />
|
| <edge from-layer="54" from-port="2" to-layer="65" to-port="0" />
|
| <edge from-layer="54" from-port="2" to-layer="72" to-port="0" />
|
| <edge from-layer="55" from-port="0" to-layer="56" to-port="0" />
|
| <edge from-layer="56" from-port="1" to-layer="57" to-port="1" />
|
| <edge from-layer="57" from-port="2" to-layer="60" to-port="0" />
|
| <edge from-layer="58" from-port="0" to-layer="59" to-port="0" />
|
| <edge from-layer="59" from-port="1" to-layer="60" to-port="1" />
|
| <edge from-layer="60" from-port="2" to-layer="61" to-port="0" />
|
| <edge from-layer="61" from-port="1" to-layer="62" to-port="0" />
|
| <edge from-layer="63" from-port="0" to-layer="64" to-port="0" />
|
| <edge from-layer="64" from-port="1" to-layer="65" to-port="1" />
|
| <edge from-layer="65" from-port="2" to-layer="68" to-port="0" />
|
| <edge from-layer="66" from-port="0" to-layer="67" to-port="0" />
|
| <edge from-layer="67" from-port="1" to-layer="68" to-port="1" />
|
| <edge from-layer="68" from-port="2" to-layer="69" to-port="0" />
|
| <edge from-layer="70" from-port="0" to-layer="71" to-port="0" />
|
| <edge from-layer="71" from-port="1" to-layer="72" to-port="1" />
|
| <edge from-layer="72" from-port="2" to-layer="75" to-port="0" />
|
| <edge from-layer="73" from-port="0" to-layer="74" to-port="0" />
|
| <edge from-layer="74" from-port="1" to-layer="75" to-port="1" />
|
| <edge from-layer="75" from-port="2" to-layer="76" to-port="0" />
|
| </edges>
|
| <rt_info>
|
| <MO_version value="2023.2.0-13089-cfd42bd2cb0-HEAD" />
|
| <Runtime_version value="2023.2.0-000--" />
|
| <conversion_parameters>
|
| <input_model value="DIR\onet.onnx" />
|
| <is_python_api_used value="False" />
|
| <output_dir value="DIR" />
|
| </conversion_parameters>
|
| <legacy_frontend value="False" />
|
| </rt_info>
|
| </net>
|
|
|