<?xml version="1.0"?>
|
<net name="Model0" version="11">
|
<layers>
|
<layer id="0" name="x" type="Parameter" version="opset1">
|
<data shape="1,3,640,640" element_type="f32" />
|
<output>
|
<port id="0" precision="FP32" names="x">
|
<dim>1</dim>
|
<dim>3</dim>
|
<dim>640</dim>
|
<dim>640</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="1" name="self.model.0.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="48, 3, 3, 3" offset="0" size="5184" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.0.conv.weight">
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="2" name="__module.model.0.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>3</dim>
|
<dim>640</dim>
|
<dim>640</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>320</dim>
|
<dim>320</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="3" name="__module.model.0.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 48, 1, 1" offset="5184" size="192" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="4" name="__module.model.0.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>320</dim>
|
<dim>320</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="48_1">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>320</dim>
|
<dim>320</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="5" name="__module.model.9.conv.act/aten::silu_/Swish" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>320</dim>
|
<dim>320</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="48,input.1">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>320</dim>
|
<dim>320</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="6" name="self.model.1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 48, 3, 3" offset="5376" size="165888" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.1.conv.weight">
|
<dim>96</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="7" name="__module.model.1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>320</dim>
|
<dim>320</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="8" name="__module.model.1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="171264" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="9" name="__module.model.1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="62_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="10" name="__module.model.9.conv.act/aten::silu_/Swish_1" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="62,input.5">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="11" name="self.model.2.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 1, 1" offset="171648" size="36864" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.2.cv1.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="12" name="__module.model.2.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="13" name="__module.model.2.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="208512" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="14" name="__module.model.2.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="82_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="15" name="__module.model.9.conv.act/aten::silu_/Swish_2" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="82,input.9">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="16" name="68" type="Const" version="opset1">
|
<data element_type="i64" shape="" offset="208896" size="8" />
|
<output>
|
<port id="0" precision="I64" names="68" />
|
</output>
|
</layer>
|
<layer id="17" name="Constant_187" type="Const" version="opset1">
|
<data element_type="i64" shape="2" offset="208904" size="16" />
|
<output>
|
<port id="0" precision="I64" names="84">
|
<dim>2</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="18" name="__module.model.2/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="I64" />
|
<port id="2" precision="I64">
|
<dim>2</dim>
|
</port>
|
</input>
|
<output>
|
<port id="3" precision="FP32" names="86">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="4" precision="FP32" names="87,input.11">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="19" name="self.model.2.m.0.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="48, 48, 3, 3" offset="208920" size="82944" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.2.m.0.cv1.conv.weight">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="20" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="21" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 48, 1, 1" offset="291864" size="192" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="22" name="__module.model.2.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="97_1">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="23" name="__module.model.9.conv.act/aten::silu_/Swish_3" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="97,input.13">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="24" name="self.model.2.m.0.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="48, 48, 3, 3" offset="292056" size="82944" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.2.m.0.cv2.conv.weight">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="25" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="26" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 48, 1, 1" offset="375000" size="192" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="27" name="__module.model.2.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="106_1">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="28" name="__module.model.9.conv.act/aten::silu_/Swish_4" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="106,input.17">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="29" name="__module.model.2.m.0/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="108,input.19">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="30" name="self.model.2.m.1.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="48, 48, 3, 3" offset="375192" size="82944" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.2.m.1.cv1.conv.weight">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="31" name="__module.model.2.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="32" name="__module.model.2.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 48, 1, 1" offset="458136" size="192" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="33" name="__module.model.2.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="118_1">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="34" name="__module.model.9.conv.act/aten::silu_/Swish_5" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="118,input.21">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="35" name="self.model.2.m.1.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="48, 48, 3, 3" offset="458328" size="82944" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.2.m.1.cv2.conv.weight">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="36" name="__module.model.2.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>48</dim>
|
<dim>48</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="37" name="__module.model.2.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 48, 1, 1" offset="541272" size="192" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="38" name="__module.model.2.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="127_1">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="39" name="__module.model.9.conv.act/aten::silu_/Swish_6" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="127,input.25">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="40" name="__module.model.2.m.1/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="129">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="41" name="__module.model.2/aten::cat/Concat" type="Concat" version="opset1">
|
<data axis="1" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="3" precision="FP32">
|
<dim>1</dim>
|
<dim>48</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="4" precision="FP32" names="131,input.27">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="42" name="self.model.2.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 192, 1, 1" offset="541464" size="73728" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.2.cv2.conv.weight">
|
<dim>96</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="43" name="__module.model.2.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="44" name="__module.model.2.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="615192" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="45" name="__module.model.2.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="139_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="46" name="__module.model.9.conv.act/aten::silu_/Swish_7" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="139,input.29">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="47" name="self.model.3.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 96, 3, 3" offset="615576" size="663552" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.3.conv.weight">
|
<dim>192</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="48" name="__module.model.3.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>160</dim>
|
<dim>160</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="49" name="__module.model.3.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="1279128" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="50" name="__module.model.3.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="153_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="51" name="__module.model.9.conv.act/aten::silu_/Swish_8" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="153,input.33">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="52" name="self.model.4.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 1, 1" offset="1279896" size="147456" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.cv1.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="53" name="__module.model.4.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="54" name="__module.model.4.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="1427352" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="55" name="__module.model.4.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="177_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="56" name="__module.model.9.conv.act/aten::silu_/Swish_9" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="177,input.37">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="57" name="159" type="Const" version="opset1">
|
<data element_type="i64" shape="" offset="208896" size="8" />
|
<output>
|
<port id="0" precision="I64" names="159" />
|
</output>
|
</layer>
|
<layer id="58" name="Constant_549" type="Const" version="opset1">
|
<data element_type="i64" shape="2" offset="1428120" size="16" />
|
<output>
|
<port id="0" precision="I64" names="179">
|
<dim>2</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="59" name="__module.model.4/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="I64" />
|
<port id="2" precision="I64">
|
<dim>2</dim>
|
</port>
|
</input>
|
<output>
|
<port id="3" precision="FP32" names="181">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="4" precision="FP32" names="182,input.39">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="60" name="self.model.4.m.0.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="1428136" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.0.cv1.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="61" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="62" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="1759912" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="63" name="__module.model.4.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="192_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="64" name="__module.model.9.conv.act/aten::silu_/Swish_10" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="192,input.41">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="65" name="self.model.4.m.0.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="1760296" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.0.cv2.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="66" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="67" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="2092072" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="68" name="__module.model.4.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="201_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="69" name="__module.model.9.conv.act/aten::silu_/Swish_11" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="201,input.45">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="70" name="__module.model.4.m.0/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="203,input.47">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="71" name="self.model.4.m.1.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="2092456" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.1.cv1.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="72" name="__module.model.4.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="73" name="__module.model.4.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="2424232" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="74" name="__module.model.4.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="213_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="75" name="__module.model.9.conv.act/aten::silu_/Swish_12" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="213,input.49">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="76" name="self.model.4.m.1.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="2424616" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.1.cv2.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="77" name="__module.model.4.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="78" name="__module.model.4.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="2756392" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="79" name="__module.model.4.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="222_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="80" name="__module.model.9.conv.act/aten::silu_/Swish_13" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="222,input.53">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="81" name="__module.model.4.m.1/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="224,input.55">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="82" name="self.model.4.m.2.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="2756776" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.2.cv1.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="83" name="__module.model.4.m.2.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="84" name="__module.model.4.m.2.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="3088552" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="85" name="__module.model.4.m.2.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="234_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="86" name="__module.model.9.conv.act/aten::silu_/Swish_14" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="234,input.57">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="87" name="self.model.4.m.2.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="3088936" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.2.cv2.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="88" name="__module.model.4.m.2.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="89" name="__module.model.4.m.2.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="3420712" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="90" name="__module.model.4.m.2.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="243_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="91" name="__module.model.9.conv.act/aten::silu_/Swish_15" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="243,input.61">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="92" name="__module.model.4.m.2/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="245,input.63">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="93" name="self.model.4.m.3.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="3421096" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.3.cv1.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="94" name="__module.model.4.m.3.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="95" name="__module.model.4.m.3.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="3752872" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="96" name="__module.model.4.m.3.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="255_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="97" name="__module.model.9.conv.act/aten::silu_/Swish_16" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="255,input.65">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="98" name="self.model.4.m.3.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="96, 96, 3, 3" offset="3753256" size="331776" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.m.3.cv2.conv.weight">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="99" name="__module.model.4.m.3.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>96</dim>
|
<dim>96</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="100" name="__module.model.4.m.3.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 96, 1, 1" offset="4085032" size="384" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="101" name="__module.model.4.m.3.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="264_1">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="102" name="__module.model.9.conv.act/aten::silu_/Swish_17" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="264,input.69">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="103" name="__module.model.4.m.3/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="266">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="104" name="__module.model.4/aten::cat/Concat" type="Concat" version="opset1">
|
<data axis="1" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="3" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="4" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="5" precision="FP32">
|
<dim>1</dim>
|
<dim>96</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="6" precision="FP32" names="268,input.71">
|
<dim>1</dim>
|
<dim>576</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="105" name="self.model.4.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 576, 1, 1" offset="4085416" size="442368" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.4.cv2.conv.weight">
|
<dim>192</dim>
|
<dim>576</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="106" name="__module.model.4.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>576</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>576</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="107" name="__module.model.4.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="4527784" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="108" name="__module.model.4.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="276_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="109" name="__module.model.9.conv.act/aten::silu_/Swish_18" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="276,input.73">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="110" name="self.model.5.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 192, 3, 3" offset="4528552" size="2654208" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.5.conv.weight">
|
<dim>384</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="111" name="__module.model.5.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>80</dim>
|
<dim>80</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="112" name="__module.model.5.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="7182760" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="113" name="__module.model.5.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="290_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="114" name="__module.model.9.conv.act/aten::silu_/Swish_19" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="290,input.77">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="115" name="self.model.6.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 384, 1, 1" offset="7184296" size="589824" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.cv1.conv.weight">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="116" name="__module.model.6.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="117" name="__module.model.6.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="7774120" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="118" name="__module.model.6.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="314_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="119" name="__module.model.9.conv.act/aten::silu_/Swish_20" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="314,input.81">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="120" name="296" type="Const" version="opset1">
|
<data element_type="i64" shape="" offset="208896" size="8" />
|
<output>
|
<port id="0" precision="I64" names="296" />
|
</output>
|
</layer>
|
<layer id="121" name="Constant_1107" type="Const" version="opset1">
|
<data element_type="i64" shape="2" offset="7775656" size="16" />
|
<output>
|
<port id="0" precision="I64" names="316">
|
<dim>2</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="122" name="__module.model.6/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="I64" />
|
<port id="2" precision="I64">
|
<dim>2</dim>
|
</port>
|
</input>
|
<output>
|
<port id="3" precision="FP32" names="318">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="4" precision="FP32" names="319,input.83">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="123" name="self.model.6.m.0.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="7775672" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.0.cv1.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="124" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="125" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="9102776" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="126" name="__module.model.6.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="329_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="127" name="__module.model.9.conv.act/aten::silu_/Swish_21" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="329,input.85">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="128" name="self.model.6.m.0.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="9103544" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.0.cv2.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="129" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="130" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="10430648" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="131" name="__module.model.6.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="338_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="132" name="__module.model.9.conv.act/aten::silu_/Swish_22" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="338,input.89">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="133" name="__module.model.6.m.0/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="340,input.91">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="134" name="self.model.6.m.1.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="10431416" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.1.cv1.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="135" name="__module.model.6.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="136" name="__module.model.6.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="11758520" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="137" name="__module.model.6.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="350_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="138" name="__module.model.9.conv.act/aten::silu_/Swish_23" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="350,input.93">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="139" name="self.model.6.m.1.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="11759288" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.1.cv2.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="140" name="__module.model.6.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="141" name="__module.model.6.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="13086392" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="142" name="__module.model.6.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="359_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="143" name="__module.model.9.conv.act/aten::silu_/Swish_24" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="359,input.97">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="144" name="__module.model.6.m.1/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="361,input.99">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="145" name="self.model.6.m.2.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="13087160" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.2.cv1.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="146" name="__module.model.6.m.2.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="147" name="__module.model.6.m.2.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="14414264" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="148" name="__module.model.6.m.2.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="371_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="149" name="__module.model.9.conv.act/aten::silu_/Swish_25" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="371,input.101">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="150" name="self.model.6.m.2.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="14415032" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.2.cv2.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="151" name="__module.model.6.m.2.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="152" name="__module.model.6.m.2.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="15742136" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="153" name="__module.model.6.m.2.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="380_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="154" name="__module.model.9.conv.act/aten::silu_/Swish_26" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="380,input.105">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="155" name="__module.model.6.m.2/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="382,input.107">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="156" name="self.model.6.m.3.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="15742904" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.3.cv1.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="157" name="__module.model.6.m.3.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="158" name="__module.model.6.m.3.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="17070008" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="159" name="__module.model.6.m.3.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="392_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="160" name="__module.model.9.conv.act/aten::silu_/Swish_27" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="392,input.109">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="161" name="self.model.6.m.3.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="192, 192, 3, 3" offset="17070776" size="1327104" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.m.3.cv2.conv.weight">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="162" name="__module.model.6.m.3.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>192</dim>
|
<dim>192</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="163" name="__module.model.6.m.3.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 192, 1, 1" offset="18397880" size="768" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="164" name="__module.model.6.m.3.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="401_1">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="165" name="__module.model.9.conv.act/aten::silu_/Swish_28" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="401,input.113">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="166" name="__module.model.6.m.3/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="403">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="167" name="__module.model.6/aten::cat/Concat" type="Concat" version="opset1">
|
<data axis="1" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="3" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="4" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="5" precision="FP32">
|
<dim>1</dim>
|
<dim>192</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="6" precision="FP32" names="405,input.115">
|
<dim>1</dim>
|
<dim>1152</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="168" name="self.model.6.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 1152, 1, 1" offset="18398648" size="1769472" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.6.cv2.conv.weight">
|
<dim>384</dim>
|
<dim>1152</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="169" name="__module.model.6.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1152</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>1152</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="170" name="__module.model.6.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="20168120" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="171" name="__module.model.6.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="413_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="172" name="__module.model.9.conv.act/aten::silu_/Swish_29" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="413,input.117">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="173" name="self.model.7.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="768, 384, 3, 3" offset="20169656" size="10616832" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.7.conv.weight">
|
<dim>768</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="174" name="__module.model.7.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="2, 2" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>40</dim>
|
<dim>40</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>768</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="175" name="__module.model.7.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 768, 1, 1" offset="30786488" size="3072" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="176" name="__module.model.7.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="427_1">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="177" name="__module.model.9.conv.act/aten::silu_/Swish_30" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="427,input.121">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="178" name="self.model.8.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="768, 768, 1, 1" offset="30789560" size="2359296" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.8.cv1.conv.weight">
|
<dim>768</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="179" name="__module.model.8.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>768</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="180" name="__module.model.8.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 768, 1, 1" offset="33148856" size="3072" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="181" name="__module.model.8.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="447_1">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="182" name="__module.model.9.conv.act/aten::silu_/Swish_31" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="447,input.125">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="183" name="433" type="Const" version="opset1">
|
<data element_type="i64" shape="" offset="208896" size="8" />
|
<output>
|
<port id="0" precision="I64" names="433" />
|
</output>
|
</layer>
|
<layer id="184" name="Constant_1661" type="Const" version="opset1">
|
<data element_type="i64" shape="2" offset="33151928" size="16" />
|
<output>
|
<port id="0" precision="I64" names="449">
|
<dim>2</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="185" name="__module.model.8/prim::ListUnpack" type="VariadicSplit" version="opset1">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="I64" />
|
<port id="2" precision="I64">
|
<dim>2</dim>
|
</port>
|
</input>
|
<output>
|
<port id="3" precision="FP32" names="451">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="4" precision="FP32" names="452,input.127">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="186" name="self.model.8.m.0.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 384, 3, 3" offset="33151944" size="5308416" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.8.m.0.cv1.conv.weight">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="187" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="188" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="38460360" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="189" name="__module.model.8.m.0.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="462_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="190" name="__module.model.9.conv.act/aten::silu_/Swish_32" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="462,input.129">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="191" name="self.model.8.m.0.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 384, 3, 3" offset="38461896" size="5308416" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.8.m.0.cv2.conv.weight">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="192" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="193" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="43770312" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="194" name="__module.model.8.m.0.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="471_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="195" name="__module.model.9.conv.act/aten::silu_/Swish_33" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="471,input.133">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="196" name="__module.model.8.m.0/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="473,input.135">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="197" name="self.model.8.m.1.cv1.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 384, 3, 3" offset="43771848" size="5308416" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.8.m.1.cv1.conv.weight">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="198" name="__module.model.8.m.1.cv1.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="199" name="__module.model.8.m.1.cv1.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="49080264" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="200" name="__module.model.8.m.1.cv1.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="483_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="201" name="__module.model.9.conv.act/aten::silu_/Swish_34" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="483,input.137">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="202" name="self.model.8.m.1.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="384, 384, 3, 3" offset="49081800" size="5308416" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.8.m.1.cv2.conv.weight">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="203" name="__module.model.8.m.1.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="1, 1" pads_end="1, 1" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>384</dim>
|
<dim>384</dim>
|
<dim>3</dim>
|
<dim>3</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="204" name="__module.model.8.m.1.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 384, 1, 1" offset="54390216" size="1536" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="205" name="__module.model.8.m.1.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="492_1">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="206" name="__module.model.9.conv.act/aten::silu_/Swish_35" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="492,input.141">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="207" name="__module.model.8.m.1/aten::add/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="494">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="208" name="__module.model.8/aten::cat/Concat" type="Concat" version="opset1">
|
<data axis="1" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="3" precision="FP32">
|
<dim>1</dim>
|
<dim>384</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="4" precision="FP32" names="496,input.143">
|
<dim>1</dim>
|
<dim>1536</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="209" name="self.model.8.cv2.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="768, 1536, 1, 1" offset="54391752" size="4718592" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.8.cv2.conv.weight">
|
<dim>768</dim>
|
<dim>1536</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="210" name="__module.model.8.cv2.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1536</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>768</dim>
|
<dim>1536</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="211" name="__module.model.8.cv2.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 768, 1, 1" offset="59110344" size="3072" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="212" name="__module.model.8.cv2.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="504_1">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="213" name="__module.model.9.conv.act/aten::silu_/Swish_36" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="504,input.145">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="214" name="self.model.9.conv.conv.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="1280, 768, 1, 1" offset="59113416" size="3932160" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.9.conv.conv.weight">
|
<dim>1280</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="215" name="__module.model.9.conv.conv/aten::_convolution/Convolution" type="Convolution" version="opset1">
|
<data strides="1, 1" dilations="1, 1" pads_begin="0, 0" pads_end="0, 0" auto_pad="explicit" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>768</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1280</dim>
|
<dim>768</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="216" name="__module.model.9.conv.conv/aten::_convolution/Reshape" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 1280, 1, 1" offset="63045576" size="5120" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="217" name="__module.model.9.conv.conv/aten::_convolution/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="525_1">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="218" name="__module.model.9.conv.act/aten::silu_/Swish_37" type="Swish" version="opset4">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32" names="525,input.149">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="219" name="Constant_5408" type="Const" version="opset1">
|
<data element_type="i64" shape="2" offset="63050696" size="16" />
|
<output>
|
<port id="0" precision="I64">
|
<dim>2</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="220" name="__module.model.9.pool/aten::adaptive_avg_pool2d/AdaptiveAvgPool" type="ReduceMean" version="opset1">
|
<data keep_dims="true" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>20</dim>
|
<dim>20</dim>
|
</port>
|
<port id="1" precision="I64">
|
<dim>2</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="528">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="221" name="Concat_3692" type="Const" version="opset1">
|
<data element_type="i64" shape="2" offset="63050712" size="16" />
|
<output>
|
<port id="0" precision="I64">
|
<dim>2</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="222" name="__module.model.9/aten::flatten/Reshape" type="Reshape" version="opset1">
|
<data special_zero="true" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
<dim>1</dim>
|
<dim>1</dim>
|
</port>
|
<port id="1" precision="I64">
|
<dim>2</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="529,input.153">
|
<dim>1</dim>
|
<dim>1280</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="223" name="self.model.9.linear.weight" type="Const" version="opset1">
|
<data element_type="f32" shape="656, 1280" offset="63050728" size="3358720" />
|
<output>
|
<port id="0" precision="FP32" names="self.model.9.linear.weight">
|
<dim>656</dim>
|
<dim>1280</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="224" name="__module.model.9.linear/aten::linear/MatMul" type="MatMul" version="opset1">
|
<data transpose_a="false" transpose_b="true" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>1280</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>656</dim>
|
<dim>1280</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="225" name="Constant_5426" type="Const" version="opset1">
|
<data element_type="f32" shape="1, 656" offset="66409448" size="2624" />
|
<output>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="226" name="__module.model.9.linear/aten::linear/Add" type="Add" version="opset1">
|
<data auto_broadcast="numpy" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</input>
|
<output>
|
<port id="2" precision="FP32" names="533,x_1">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="227" name="__module.model.9/aten::softmax/Softmax" type="SoftMax" version="opset8">
|
<data axis="1" />
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</input>
|
<output>
|
<port id="1" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</output>
|
</layer>
|
<layer id="228" name="Result_2055" type="Result" version="opset1">
|
<input>
|
<port id="0" precision="FP32">
|
<dim>1</dim>
|
<dim>656</dim>
|
</port>
|
</input>
|
</layer>
|
</layers>
|
<edges>
|
<edge from-layer="0" from-port="0" to-layer="2" to-port="0" />
|
<edge from-layer="1" from-port="0" to-layer="2" to-port="1" />
|
<edge from-layer="2" from-port="2" to-layer="4" to-port="0" />
|
<edge from-layer="3" from-port="0" to-layer="4" to-port="1" />
|
<edge from-layer="4" from-port="2" to-layer="5" to-port="0" />
|
<edge from-layer="5" from-port="1" to-layer="7" to-port="0" />
|
<edge from-layer="6" from-port="0" to-layer="7" to-port="1" />
|
<edge from-layer="7" from-port="2" to-layer="9" to-port="0" />
|
<edge from-layer="8" from-port="0" to-layer="9" to-port="1" />
|
<edge from-layer="9" from-port="2" to-layer="10" to-port="0" />
|
<edge from-layer="10" from-port="1" to-layer="12" to-port="0" />
|
<edge from-layer="11" from-port="0" to-layer="12" to-port="1" />
|
<edge from-layer="12" from-port="2" to-layer="14" to-port="0" />
|
<edge from-layer="13" from-port="0" to-layer="14" to-port="1" />
|
<edge from-layer="14" from-port="2" to-layer="15" to-port="0" />
|
<edge from-layer="15" from-port="1" to-layer="18" to-port="0" />
|
<edge from-layer="16" from-port="0" to-layer="18" to-port="1" />
|
<edge from-layer="17" from-port="0" to-layer="18" to-port="2" />
|
<edge from-layer="18" from-port="4" to-layer="20" to-port="0" />
|
<edge from-layer="18" from-port="4" to-layer="29" to-port="0" />
|
<edge from-layer="18" from-port="3" to-layer="41" to-port="0" />
|
<edge from-layer="18" from-port="4" to-layer="41" to-port="1" />
|
<edge from-layer="19" from-port="0" to-layer="20" to-port="1" />
|
<edge from-layer="20" from-port="2" to-layer="22" to-port="0" />
|
<edge from-layer="21" from-port="0" to-layer="22" to-port="1" />
|
<edge from-layer="22" from-port="2" to-layer="23" to-port="0" />
|
<edge from-layer="23" from-port="1" to-layer="25" to-port="0" />
|
<edge from-layer="24" from-port="0" to-layer="25" to-port="1" />
|
<edge from-layer="25" from-port="2" to-layer="27" to-port="0" />
|
<edge from-layer="26" from-port="0" to-layer="27" to-port="1" />
|
<edge from-layer="27" from-port="2" to-layer="28" to-port="0" />
|
<edge from-layer="28" from-port="1" to-layer="29" to-port="1" />
|
<edge from-layer="29" from-port="2" to-layer="40" to-port="0" />
|
<edge from-layer="29" from-port="2" to-layer="31" to-port="0" />
|
<edge from-layer="29" from-port="2" to-layer="41" to-port="2" />
|
<edge from-layer="30" from-port="0" to-layer="31" to-port="1" />
|
<edge from-layer="31" from-port="2" to-layer="33" to-port="0" />
|
<edge from-layer="32" from-port="0" to-layer="33" to-port="1" />
|
<edge from-layer="33" from-port="2" to-layer="34" to-port="0" />
|
<edge from-layer="34" from-port="1" to-layer="36" to-port="0" />
|
<edge from-layer="35" from-port="0" to-layer="36" to-port="1" />
|
<edge from-layer="36" from-port="2" to-layer="38" to-port="0" />
|
<edge from-layer="37" from-port="0" to-layer="38" to-port="1" />
|
<edge from-layer="38" from-port="2" to-layer="39" to-port="0" />
|
<edge from-layer="39" from-port="1" to-layer="40" to-port="1" />
|
<edge from-layer="40" from-port="2" to-layer="41" to-port="3" />
|
<edge from-layer="41" from-port="4" to-layer="43" to-port="0" />
|
<edge from-layer="42" from-port="0" to-layer="43" to-port="1" />
|
<edge from-layer="43" from-port="2" to-layer="45" to-port="0" />
|
<edge from-layer="44" from-port="0" to-layer="45" to-port="1" />
|
<edge from-layer="45" from-port="2" to-layer="46" to-port="0" />
|
<edge from-layer="46" from-port="1" to-layer="48" to-port="0" />
|
<edge from-layer="47" from-port="0" to-layer="48" to-port="1" />
|
<edge from-layer="48" from-port="2" to-layer="50" to-port="0" />
|
<edge from-layer="49" from-port="0" to-layer="50" to-port="1" />
|
<edge from-layer="50" from-port="2" to-layer="51" to-port="0" />
|
<edge from-layer="51" from-port="1" to-layer="53" to-port="0" />
|
<edge from-layer="52" from-port="0" to-layer="53" to-port="1" />
|
<edge from-layer="53" from-port="2" to-layer="55" to-port="0" />
|
<edge from-layer="54" from-port="0" to-layer="55" to-port="1" />
|
<edge from-layer="55" from-port="2" to-layer="56" to-port="0" />
|
<edge from-layer="56" from-port="1" to-layer="59" to-port="0" />
|
<edge from-layer="57" from-port="0" to-layer="59" to-port="1" />
|
<edge from-layer="58" from-port="0" to-layer="59" to-port="2" />
|
<edge from-layer="59" from-port="4" to-layer="61" to-port="0" />
|
<edge from-layer="59" from-port="4" to-layer="70" to-port="0" />
|
<edge from-layer="59" from-port="3" to-layer="104" to-port="0" />
|
<edge from-layer="59" from-port="4" to-layer="104" to-port="1" />
|
<edge from-layer="60" from-port="0" to-layer="61" to-port="1" />
|
<edge from-layer="61" from-port="2" to-layer="63" to-port="0" />
|
<edge from-layer="62" from-port="0" to-layer="63" to-port="1" />
|
<edge from-layer="63" from-port="2" to-layer="64" to-port="0" />
|
<edge from-layer="64" from-port="1" to-layer="66" to-port="0" />
|
<edge from-layer="65" from-port="0" to-layer="66" to-port="1" />
|
<edge from-layer="66" from-port="2" to-layer="68" to-port="0" />
|
<edge from-layer="67" from-port="0" to-layer="68" to-port="1" />
|
<edge from-layer="68" from-port="2" to-layer="69" to-port="0" />
|
<edge from-layer="69" from-port="1" to-layer="70" to-port="1" />
|
<edge from-layer="70" from-port="2" to-layer="72" to-port="0" />
|
<edge from-layer="70" from-port="2" to-layer="81" to-port="0" />
|
<edge from-layer="70" from-port="2" to-layer="104" to-port="2" />
|
<edge from-layer="71" from-port="0" to-layer="72" to-port="1" />
|
<edge from-layer="72" from-port="2" to-layer="74" to-port="0" />
|
<edge from-layer="73" from-port="0" to-layer="74" to-port="1" />
|
<edge from-layer="74" from-port="2" to-layer="75" to-port="0" />
|
<edge from-layer="75" from-port="1" to-layer="77" to-port="0" />
|
<edge from-layer="76" from-port="0" to-layer="77" to-port="1" />
|
<edge from-layer="77" from-port="2" to-layer="79" to-port="0" />
|
<edge from-layer="78" from-port="0" to-layer="79" to-port="1" />
|
<edge from-layer="79" from-port="2" to-layer="80" to-port="0" />
|
<edge from-layer="80" from-port="1" to-layer="81" to-port="1" />
|
<edge from-layer="81" from-port="2" to-layer="83" to-port="0" />
|
<edge from-layer="81" from-port="2" to-layer="92" to-port="0" />
|
<edge from-layer="81" from-port="2" to-layer="104" to-port="3" />
|
<edge from-layer="82" from-port="0" to-layer="83" to-port="1" />
|
<edge from-layer="83" from-port="2" to-layer="85" to-port="0" />
|
<edge from-layer="84" from-port="0" to-layer="85" to-port="1" />
|
<edge from-layer="85" from-port="2" to-layer="86" to-port="0" />
|
<edge from-layer="86" from-port="1" to-layer="88" to-port="0" />
|
<edge from-layer="87" from-port="0" to-layer="88" to-port="1" />
|
<edge from-layer="88" from-port="2" to-layer="90" to-port="0" />
|
<edge from-layer="89" from-port="0" to-layer="90" to-port="1" />
|
<edge from-layer="90" from-port="2" to-layer="91" to-port="0" />
|
<edge from-layer="91" from-port="1" to-layer="92" to-port="1" />
|
<edge from-layer="92" from-port="2" to-layer="94" to-port="0" />
|
<edge from-layer="92" from-port="2" to-layer="103" to-port="0" />
|
<edge from-layer="92" from-port="2" to-layer="104" to-port="4" />
|
<edge from-layer="93" from-port="0" to-layer="94" to-port="1" />
|
<edge from-layer="94" from-port="2" to-layer="96" to-port="0" />
|
<edge from-layer="95" from-port="0" to-layer="96" to-port="1" />
|
<edge from-layer="96" from-port="2" to-layer="97" to-port="0" />
|
<edge from-layer="97" from-port="1" to-layer="99" to-port="0" />
|
<edge from-layer="98" from-port="0" to-layer="99" to-port="1" />
|
<edge from-layer="99" from-port="2" to-layer="101" to-port="0" />
|
<edge from-layer="100" from-port="0" to-layer="101" to-port="1" />
|
<edge from-layer="101" from-port="2" to-layer="102" to-port="0" />
|
<edge from-layer="102" from-port="1" to-layer="103" to-port="1" />
|
<edge from-layer="103" from-port="2" to-layer="104" to-port="5" />
|
<edge from-layer="104" from-port="6" to-layer="106" to-port="0" />
|
<edge from-layer="105" from-port="0" to-layer="106" to-port="1" />
|
<edge from-layer="106" from-port="2" to-layer="108" to-port="0" />
|
<edge from-layer="107" from-port="0" to-layer="108" to-port="1" />
|
<edge from-layer="108" from-port="2" to-layer="109" to-port="0" />
|
<edge from-layer="109" from-port="1" to-layer="111" to-port="0" />
|
<edge from-layer="110" from-port="0" to-layer="111" to-port="1" />
|
<edge from-layer="111" from-port="2" to-layer="113" to-port="0" />
|
<edge from-layer="112" from-port="0" to-layer="113" to-port="1" />
|
<edge from-layer="113" from-port="2" to-layer="114" to-port="0" />
|
<edge from-layer="114" from-port="1" to-layer="116" to-port="0" />
|
<edge from-layer="115" from-port="0" to-layer="116" to-port="1" />
|
<edge from-layer="116" from-port="2" to-layer="118" to-port="0" />
|
<edge from-layer="117" from-port="0" to-layer="118" to-port="1" />
|
<edge from-layer="118" from-port="2" to-layer="119" to-port="0" />
|
<edge from-layer="119" from-port="1" to-layer="122" to-port="0" />
|
<edge from-layer="120" from-port="0" to-layer="122" to-port="1" />
|
<edge from-layer="121" from-port="0" to-layer="122" to-port="2" />
|
<edge from-layer="122" from-port="4" to-layer="124" to-port="0" />
|
<edge from-layer="122" from-port="4" to-layer="133" to-port="0" />
|
<edge from-layer="122" from-port="3" to-layer="167" to-port="0" />
|
<edge from-layer="122" from-port="4" to-layer="167" to-port="1" />
|
<edge from-layer="123" from-port="0" to-layer="124" to-port="1" />
|
<edge from-layer="124" from-port="2" to-layer="126" to-port="0" />
|
<edge from-layer="125" from-port="0" to-layer="126" to-port="1" />
|
<edge from-layer="126" from-port="2" to-layer="127" to-port="0" />
|
<edge from-layer="127" from-port="1" to-layer="129" to-port="0" />
|
<edge from-layer="128" from-port="0" to-layer="129" to-port="1" />
|
<edge from-layer="129" from-port="2" to-layer="131" to-port="0" />
|
<edge from-layer="130" from-port="0" to-layer="131" to-port="1" />
|
<edge from-layer="131" from-port="2" to-layer="132" to-port="0" />
|
<edge from-layer="132" from-port="1" to-layer="133" to-port="1" />
|
<edge from-layer="133" from-port="2" to-layer="144" to-port="0" />
|
<edge from-layer="133" from-port="2" to-layer="135" to-port="0" />
|
<edge from-layer="133" from-port="2" to-layer="167" to-port="2" />
|
<edge from-layer="134" from-port="0" to-layer="135" to-port="1" />
|
<edge from-layer="135" from-port="2" to-layer="137" to-port="0" />
|
<edge from-layer="136" from-port="0" to-layer="137" to-port="1" />
|
<edge from-layer="137" from-port="2" to-layer="138" to-port="0" />
|
<edge from-layer="138" from-port="1" to-layer="140" to-port="0" />
|
<edge from-layer="139" from-port="0" to-layer="140" to-port="1" />
|
<edge from-layer="140" from-port="2" to-layer="142" to-port="0" />
|
<edge from-layer="141" from-port="0" to-layer="142" to-port="1" />
|
<edge from-layer="142" from-port="2" to-layer="143" to-port="0" />
|
<edge from-layer="143" from-port="1" to-layer="144" to-port="1" />
|
<edge from-layer="144" from-port="2" to-layer="155" to-port="0" />
|
<edge from-layer="144" from-port="2" to-layer="146" to-port="0" />
|
<edge from-layer="144" from-port="2" to-layer="167" to-port="3" />
|
<edge from-layer="145" from-port="0" to-layer="146" to-port="1" />
|
<edge from-layer="146" from-port="2" to-layer="148" to-port="0" />
|
<edge from-layer="147" from-port="0" to-layer="148" to-port="1" />
|
<edge from-layer="148" from-port="2" to-layer="149" to-port="0" />
|
<edge from-layer="149" from-port="1" to-layer="151" to-port="0" />
|
<edge from-layer="150" from-port="0" to-layer="151" to-port="1" />
|
<edge from-layer="151" from-port="2" to-layer="153" to-port="0" />
|
<edge from-layer="152" from-port="0" to-layer="153" to-port="1" />
|
<edge from-layer="153" from-port="2" to-layer="154" to-port="0" />
|
<edge from-layer="154" from-port="1" to-layer="155" to-port="1" />
|
<edge from-layer="155" from-port="2" to-layer="157" to-port="0" />
|
<edge from-layer="155" from-port="2" to-layer="166" to-port="0" />
|
<edge from-layer="155" from-port="2" to-layer="167" to-port="4" />
|
<edge from-layer="156" from-port="0" to-layer="157" to-port="1" />
|
<edge from-layer="157" from-port="2" to-layer="159" to-port="0" />
|
<edge from-layer="158" from-port="0" to-layer="159" to-port="1" />
|
<edge from-layer="159" from-port="2" to-layer="160" to-port="0" />
|
<edge from-layer="160" from-port="1" to-layer="162" to-port="0" />
|
<edge from-layer="161" from-port="0" to-layer="162" to-port="1" />
|
<edge from-layer="162" from-port="2" to-layer="164" to-port="0" />
|
<edge from-layer="163" from-port="0" to-layer="164" to-port="1" />
|
<edge from-layer="164" from-port="2" to-layer="165" to-port="0" />
|
<edge from-layer="165" from-port="1" to-layer="166" to-port="1" />
|
<edge from-layer="166" from-port="2" to-layer="167" to-port="5" />
|
<edge from-layer="167" from-port="6" to-layer="169" to-port="0" />
|
<edge from-layer="168" from-port="0" to-layer="169" to-port="1" />
|
<edge from-layer="169" from-port="2" to-layer="171" to-port="0" />
|
<edge from-layer="170" from-port="0" to-layer="171" to-port="1" />
|
<edge from-layer="171" from-port="2" to-layer="172" to-port="0" />
|
<edge from-layer="172" from-port="1" to-layer="174" to-port="0" />
|
<edge from-layer="173" from-port="0" to-layer="174" to-port="1" />
|
<edge from-layer="174" from-port="2" to-layer="176" to-port="0" />
|
<edge from-layer="175" from-port="0" to-layer="176" to-port="1" />
|
<edge from-layer="176" from-port="2" to-layer="177" to-port="0" />
|
<edge from-layer="177" from-port="1" to-layer="179" to-port="0" />
|
<edge from-layer="178" from-port="0" to-layer="179" to-port="1" />
|
<edge from-layer="179" from-port="2" to-layer="181" to-port="0" />
|
<edge from-layer="180" from-port="0" to-layer="181" to-port="1" />
|
<edge from-layer="181" from-port="2" to-layer="182" to-port="0" />
|
<edge from-layer="182" from-port="1" to-layer="185" to-port="0" />
|
<edge from-layer="183" from-port="0" to-layer="185" to-port="1" />
|
<edge from-layer="184" from-port="0" to-layer="185" to-port="2" />
|
<edge from-layer="185" from-port="4" to-layer="187" to-port="0" />
|
<edge from-layer="185" from-port="4" to-layer="196" to-port="0" />
|
<edge from-layer="185" from-port="3" to-layer="208" to-port="0" />
|
<edge from-layer="185" from-port="4" to-layer="208" to-port="1" />
|
<edge from-layer="186" from-port="0" to-layer="187" to-port="1" />
|
<edge from-layer="187" from-port="2" to-layer="189" to-port="0" />
|
<edge from-layer="188" from-port="0" to-layer="189" to-port="1" />
|
<edge from-layer="189" from-port="2" to-layer="190" to-port="0" />
|
<edge from-layer="190" from-port="1" to-layer="192" to-port="0" />
|
<edge from-layer="191" from-port="0" to-layer="192" to-port="1" />
|
<edge from-layer="192" from-port="2" to-layer="194" to-port="0" />
|
<edge from-layer="193" from-port="0" to-layer="194" to-port="1" />
|
<edge from-layer="194" from-port="2" to-layer="195" to-port="0" />
|
<edge from-layer="195" from-port="1" to-layer="196" to-port="1" />
|
<edge from-layer="196" from-port="2" to-layer="198" to-port="0" />
|
<edge from-layer="196" from-port="2" to-layer="207" to-port="0" />
|
<edge from-layer="196" from-port="2" to-layer="208" to-port="2" />
|
<edge from-layer="197" from-port="0" to-layer="198" to-port="1" />
|
<edge from-layer="198" from-port="2" to-layer="200" to-port="0" />
|
<edge from-layer="199" from-port="0" to-layer="200" to-port="1" />
|
<edge from-layer="200" from-port="2" to-layer="201" to-port="0" />
|
<edge from-layer="201" from-port="1" to-layer="203" to-port="0" />
|
<edge from-layer="202" from-port="0" to-layer="203" to-port="1" />
|
<edge from-layer="203" from-port="2" to-layer="205" to-port="0" />
|
<edge from-layer="204" from-port="0" to-layer="205" to-port="1" />
|
<edge from-layer="205" from-port="2" to-layer="206" to-port="0" />
|
<edge from-layer="206" from-port="1" to-layer="207" to-port="1" />
|
<edge from-layer="207" from-port="2" to-layer="208" to-port="3" />
|
<edge from-layer="208" from-port="4" to-layer="210" to-port="0" />
|
<edge from-layer="209" from-port="0" to-layer="210" to-port="1" />
|
<edge from-layer="210" from-port="2" to-layer="212" to-port="0" />
|
<edge from-layer="211" from-port="0" to-layer="212" to-port="1" />
|
<edge from-layer="212" from-port="2" to-layer="213" to-port="0" />
|
<edge from-layer="213" from-port="1" to-layer="215" to-port="0" />
|
<edge from-layer="214" from-port="0" to-layer="215" to-port="1" />
|
<edge from-layer="215" from-port="2" to-layer="217" to-port="0" />
|
<edge from-layer="216" from-port="0" to-layer="217" to-port="1" />
|
<edge from-layer="217" from-port="2" to-layer="218" to-port="0" />
|
<edge from-layer="218" from-port="1" to-layer="220" to-port="0" />
|
<edge from-layer="219" from-port="0" to-layer="220" to-port="1" />
|
<edge from-layer="220" from-port="2" to-layer="222" to-port="0" />
|
<edge from-layer="221" from-port="0" to-layer="222" to-port="1" />
|
<edge from-layer="222" from-port="2" to-layer="224" to-port="0" />
|
<edge from-layer="223" from-port="0" to-layer="224" to-port="1" />
|
<edge from-layer="224" from-port="2" to-layer="226" to-port="0" />
|
<edge from-layer="225" from-port="0" to-layer="226" to-port="1" />
|
<edge from-layer="226" from-port="2" to-layer="227" to-port="0" />
|
<edge from-layer="227" from-port="1" to-layer="228" to-port="0" />
|
</edges>
|
<rt_info>
|
<Runtime_version value="2025.1.0-18503-6fec06580ab-releases/2025/1" />
|
<conversion_parameters>
|
<framework value="pytorch" />
|
<is_python_object value="True" />
|
</conversion_parameters>
|
<model_info>
|
<iou_threshold value="0.7" />
|
<labels value="Choulingdancao Dayezizhu aidicha aiye anxixiang awei badou baibeiyegen baibiandou baibu baifan baifuzi baiguo baihe baihelingzhi baihuadan baihuasheshecao baiji baijiangcao bailian baimagu baimaogen baiqian baishao baishihua baishiying baishouwu baitouweng baiwei baixianpi baiying baizhi baizhu bajiaofeng bajiaohuixiang bajiaolian bajitian banbianlian banfenghe banmao banxia banzhilian baomadingxiang baqia beidougen beiliujinu beishashen bianxu biba(bo) bichengqie biejia bimazi binglang bohe bohuohui buguzhi buzhaye cangerzi cangzhu caodoukou caoguo cebaiye chaihu changchunhua changshan chanhua chansha chansu chantui chenpi chenxiang cheqiancao cheqianzi chishao chishizhi chixiaodou chonglou chongweizi chongyayao chouwutong chuanbeimu chuanlianzi chuanmutong chuanposhi chuanshanjia chuanshanlong chuanwu chuanxinlian chuanxiong chuipencao chunpi chushizi cishi ciweipi ciwujia cixiancai congzi cuiyuncao cujiangcao dabogu dadouhuangjuan dafeiyangcao dafengai dafengzi dafupi dahuang daji dandouchi danfan danggui dangshen dangyao dannanxing danshen danzhuye daodou daokoucao daoya daqingye dasuan daxueteng dayeanye dazao dengxincao dengzhanxixin diburong diercao difengpi difuzi digupi dihuang dijincao dinggongteng dingxiang direngen dongchongxiacao dongfengju dongguapi dongkuizi donglingcao doukou duanxueliu duhuo duzhong duzhongye ebushicao eshen eshu fanbaicao fangfeng fangji fangjieshi fanshiliuye fanxieye feilongzhangxue feizi fenge fengfang fenghuangyi fengweicao fengxiangzhi fengyanguo foshou fuling fulonggan fupenzi fuping fushen fuxiaomai fuzi gancao gangbangui gangmeigen gangrengen ganjiang ganqi gansong gansui gaoben gaoliangjiang gegen gejie gonglaomu gouguye gouji gouqi goushen gouteng gualou guangdonghaitongpi guangdongliujinu guangdongwangbuliuxing guanghuoxiang guangzao guanhuangbo guanyejinsitao guanzhong guazijin guihua guijia guijianyu guipi guizhencao guizhi gujingcao gusuibu haigeke hailong haima haipiaoxiao haishen haitongpi haizao hancai hantaoye hehuamhua hehuanpi heidou heshi heshouwu hetaoren heye hezi hongdji hongdoukou honghua hongjingtian hongqi hongqu hongshen hongtiankui houpu houpuhua huaihua huaijiao huajiao huajuhong huangjing huangjingzi huanglian huangpihe huangqi huangqin huangshukui huangteng huangyaozi huashi hubeibeimu huercao hujiao hujisheng hulucha huomaren huotanmu hutuiziye huzhang jiangcan jianghuang jiangxiang jianhua jiaogulan jicai jidanhua jiegeng jigucao jiguxiang jili jindenglong jineijin jinfeicao jingjie jingtiansanqi jingucao jinguolan jinqiaomai jinsicao jintiesuo jinyingzi jinyinhua jishiteng jiubiying jiucengta jiujiechangpu jiulixiang jiulongteng jiuxiangchong jixuecao jixueteng jiyancao juanbai juemingzi juhe juhua juqu juqu2 ketengzi kuandonghua kuanjinteng kudiding kudingcha kugua kulianpi kumaicai kumu kunbu kushen kushilian kuxingren lajiao laliao lanbuzheng langdu langyupi laoguancao leigongteng leiwan lianfang liangmianzhen liangtoujian lianqiancao lianqiao lianzi lianzixin liaogewang lilu lingxiangcao lingxiaohua lingzhi liushenqu liuyuehan lizhihe longchi longdan longgu longkui longliye longyanrou loulu luganshi lugen luobumaye luofumu luohanguo luole luoshiteng lurong luxiancao luying lvcao lvdou mabiancao mabo machixian madouling mahuang mahuanggen maidong maiya manjingzi manshanhong maozhaocao maqianzi maweilian meiguihua meihua mengchong mianbixie mianbixiepian mimenghua mohanlian molihua mubiezi mudanpi mudingxiang muer mufangji mugua muhao muhudie mujingye mujinhua mujinpi muli mumianhua muxiang muzei nanbanlangen nanguazi nanshanzha nanshashen nanwuweizi naosha naoyanghua niubangzi niudali niuduteng niuhuang niuxi nuodaogen nvzhenzi oujie pangdahai peilan pianjianghuang pijiuhua pipaye pugongying qiancao qianghuo qianhu qianjinba qianjinzi qianliguang qiannianjian qianniuzi qianrihong qianshi qingfengteng qingguo qinghao qingmuxiang qingniudan qingpi qingtiankui qingyangshe qingyedan qinjiao qinpi qiyelian qiyeyizhihua quanshen qumai rendongteng renshen renshenye ricaogen rongshuye roucongrong roudoukou ruiren sanbaicao sanbaicaogen sanfensan sangbaipi sangshen sangye sangzhipian sankezhen sanleng sanqi sanqipian sanyaku shaji shancigu shandayan shandougen shanglu shannai shanyao shanzha shanzhanye shanzhima shanzhuyu sharen shegan sheliugu shemei shencha shengjiangpi shengma shenjincao shetui shicancao shichangpu shidi shidiaolan shihu shihu2 shijueming shijunzi shiliupi shinanteng shinanye shishangbai shisuan shiwei shouwuteng shuangshen shudihuang shuifeiji shuizhi shuqicao sigualuo sijiqing siyeshen songmu songxiang suanzaoren sumu suoluozi suoyang taibaimi taizishen tanxiang taoren taozhi tiandiding tiandong tianguadi tianguazi tianhuafen tianhusui tiankuizi tianma tiannanxing tianxianteng tianzhuhuang tiebaojin tiexian tongcao tougucao tubeimu tubiechong tujingpi tuniuexi walengzi wangbuliuxing wasong weilingxian wenjing wubeizi wugenteng wuhuaguo wuhuaguo2 wuhuanzi wujiapi wujiuzi wulingzhi wumei wushaoshe wutongzi wuyao wuzhimaotao wuzhuyu xiakucao xiangfu xiangru xiangsizi xiangyuan xianhecao xianmao xiaobogu xiaobopi xiaohuixiang xiaoji xiaotongcao xiatianwu xiebai xiecao xiguapi xiheliu xihuangcao xinyi xionghuang xishuguo xixiancao xixin xiyangshen xuancaogen xuanshen xuchangqing xuduan xuefengteng xueshangyizhihao xueyutan xungufeng yadanzi yanduzhong yangqishi yangticao yanhusuo yanwo yazhangmupi yazhicao yejuhua yexiazhu yimucao yinchaihu yinchen yiner yingchunhua yingsuke yingtaohe yinxingye yinyanghuo yiyiren yizhi yizhihuanghua yizhijian yousongjie yuanbaocao yuanhua yuansuizi yuanzhi yuejihua yuganzi yujin yuliren yumixu yuxingcao yuyejinhua yuyuliang yuzhizi yuzhouloulu yuzhu zaojiao zaojiao1 zaojiaoci zaojiaoci1 zelan zeqi zexie zhangmu zhebeimu zhenzhumu zhigancao zhihuangqi zhijuzi zhimu zhiqiao zhishanzhuyu zhishi zhizhuxiang zhizi zhongjiefeng zhongrushi zhujieshen zhuling zhumagen zhuru zhushagen zhuyazao zhuzishen zibeichi zicao zicaorong ziheche zishaohua zisugeng zisuye zisuzi ziyuan zizhuye zonglv zoumajian zoumatai" />
|
<model_type value="YOLO" />
|
<pad_value value="114" />
|
<reverse_input_channels value="YES" />
|
<scale_values value="255" />
|
</model_info>
|
</rt_info>
|
</net>
|