Last active
April 22, 2024 18:52
-
-
Save AmosLewis/1cb839df5dd34fd2f6dc701392aa6bc6 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
module { | |
func.func @torch_jit(%arg0: !torch.vtensor<[1,3,640,640],f32>) -> (!torch.vtensor<[1,84,8400],f32>, !torch.vtensor<[1,144,80,80],f32>, !torch.vtensor<[1,144,40,40],f32>, !torch.vtensor<[1,144,20,20],f32>) attributes {torch.onnx_meta.ir_version = 8 : si64, torch.onnx_meta.opset_version = 17 : si64, torch.onnx_meta.producer_name = "pytorch", torch.onnx_meta.producer_version = "1.13.1"} { | |
%0 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x3x3x3xf32>) : !torch.vtensor<[16,3,3,3],f32> | |
%1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
%2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x16x3x3xf32>) : !torch.vtensor<[32,16,3,3],f32> | |
%3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%4 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x1x1xf32>) : !torch.vtensor<[32,32,1,1],f32> | |
%5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%6 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x16x3x3xf32>) : !torch.vtensor<[16,16,3,3],f32> | |
%7 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
%8 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x16x3x3xf32>) : !torch.vtensor<[16,16,3,3],f32> | |
%9 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
%10 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x48x1x1xf32>) : !torch.vtensor<[32,48,1,1],f32> | |
%11 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%12 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x32x3x3xf32>) : !torch.vtensor<[64,32,3,3],f32> | |
%13 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%14 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
%15 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%16 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
%17 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%18 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
%19 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%20 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
%21 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%22 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
%23 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%24 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x128x1x1xf32>) : !torch.vtensor<[64,128,1,1],f32> | |
%25 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%26 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x64x3x3xf32>) : !torch.vtensor<[128,64,3,3],f32> | |
%27 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%28 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x1x1xf32>) : !torch.vtensor<[128,128,1,1],f32> | |
%29 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%30 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%31 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%32 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%33 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%34 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%35 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%36 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%37 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%38 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x256x1x1xf32>) : !torch.vtensor<[128,256,1,1],f32> | |
%39 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%40 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x128x3x3xf32>) : !torch.vtensor<[256,128,3,3],f32> | |
%41 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
%42 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x256x1x1xf32>) : !torch.vtensor<[256,256,1,1],f32> | |
%43 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
%44 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
%45 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%46 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
%47 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%48 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x384x1x1xf32>) : !torch.vtensor<[256,384,1,1],f32> | |
%49 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
%50 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x256x1x1xf32>) : !torch.vtensor<[128,256,1,1],f32> | |
%51 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%52 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x512x1x1xf32>) : !torch.vtensor<[256,512,1,1],f32> | |
%53 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
%54 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x384x1x1xf32>) : !torch.vtensor<[128,384,1,1],f32> | |
%55 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%56 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%57 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%58 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%59 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%60 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x192x1x1xf32>) : !torch.vtensor<[128,192,1,1],f32> | |
%61 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%62 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x192x1x1xf32>) : !torch.vtensor<[64,192,1,1],f32> | |
%63 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%64 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
%65 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%66 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
%67 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
%68 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x96x1x1xf32>) : !torch.vtensor<[64,96,1,1],f32> | |
%69 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%70 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%71 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%72 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x192x1x1xf32>) : !torch.vtensor<[128,192,1,1],f32> | |
%73 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%74 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%75 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%76 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%77 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%78 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x192x1x1xf32>) : !torch.vtensor<[128,192,1,1],f32> | |
%79 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%80 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
%81 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%82 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x384x1x1xf32>) : !torch.vtensor<[256,384,1,1],f32> | |
%83 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
%84 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
%85 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%86 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
%87 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
%88 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x384x1x1xf32>) : !torch.vtensor<[256,384,1,1],f32> | |
%89 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
%90 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%91 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%92 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%93 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%94 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
%95 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%96 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x64x3x3xf32>) : !torch.vtensor<[80,64,3,3],f32> | |
%97 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%98 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x3x3xf32>) : !torch.vtensor<[80,80,3,3],f32> | |
%99 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%100 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x1x1xf32>) : !torch.vtensor<[80,80,1,1],f32> | |
%101 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%102 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x128x3x3xf32>) : !torch.vtensor<[64,128,3,3],f32> | |
%103 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%104 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%105 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%106 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
%107 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%108 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x128x3x3xf32>) : !torch.vtensor<[80,128,3,3],f32> | |
%109 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%110 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x3x3xf32>) : !torch.vtensor<[80,80,3,3],f32> | |
%111 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%112 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x1x1xf32>) : !torch.vtensor<[80,80,1,1],f32> | |
%113 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%114 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x256x3x3xf32>) : !torch.vtensor<[64,256,3,3],f32> | |
%115 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%116 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
%117 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%118 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
%119 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
%120 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x256x3x3xf32>) : !torch.vtensor<[80,256,3,3],f32> | |
%121 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%122 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x3x3xf32>) : !torch.vtensor<[80,80,3,3],f32> | |
%123 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%124 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x1x1xf32>) : !torch.vtensor<[80,80,1,1],f32> | |
%125 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
%126 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x16x1x1xf32>) : !torch.vtensor<[1,16,1,1],f32> | |
%none = torch.constant.none | |
%127 = torch.vtensor.literal(dense<2.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%128 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%129 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12 = torch.constant.int 12 | |
%130 = torch.aten.item %128 : !torch.vtensor<[],f32> -> !torch.float | |
%131 = torch.aten.item %129 : !torch.vtensor<[],si8> -> !torch.int | |
%132 = torch.aten.quantize_per_tensor %127, %130, %131, %int12 : !torch.vtensor<[],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[],!torch.qint8> | |
%133 = torch.aten.int_repr %132 : !torch.vtensor<[],!torch.qint8> -> !torch.vtensor<[],si8> | |
%134 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%135 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%136 = torch.aten.item %134 : !torch.vtensor<[],f32> -> !torch.float | |
%137 = torch.aten.item %135 : !torch.vtensor<[],si8> -> !torch.int | |
%138 = torch.aten._make_per_tensor_quantized_tensor %133, %136, %137 : !torch.vtensor<[],si8>, !torch.float, !torch.int -> !torch.vtensor<[],!torch.qint8> | |
%139 = torch.aten.dequantize.self %138 : !torch.vtensor<[],!torch.qint8> -> !torch.vtensor<[],f32> | |
%140 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x2x8400xf32>) : !torch.vtensor<[1,2,8400],f32> | |
%141 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%142 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_0 = torch.constant.int 12 | |
%143 = torch.aten.item %141 : !torch.vtensor<[],f32> -> !torch.float | |
%144 = torch.aten.item %142 : !torch.vtensor<[],si8> -> !torch.int | |
%145 = torch.aten.quantize_per_tensor %140, %143, %144, %int12_0 : !torch.vtensor<[1,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,2,8400],!torch.qint8> | |
%146 = torch.aten.int_repr %145 : !torch.vtensor<[1,2,8400],!torch.qint8> -> !torch.vtensor<[1,2,8400],si8> | |
%147 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%148 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%149 = torch.aten.item %147 : !torch.vtensor<[],f32> -> !torch.float | |
%150 = torch.aten.item %148 : !torch.vtensor<[],si8> -> !torch.int | |
%151 = torch.aten._make_per_tensor_quantized_tensor %146, %149, %150 : !torch.vtensor<[1,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,2,8400],!torch.qint8> | |
%152 = torch.aten.dequantize.self %151 : !torch.vtensor<[1,2,8400],!torch.qint8> -> !torch.vtensor<[1,2,8400],f32> | |
%153 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x8400xf32>) : !torch.vtensor<[1,8400],f32> | |
%154 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%155 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1 = torch.constant.int 12 | |
%156 = torch.aten.item %154 : !torch.vtensor<[],f32> -> !torch.float | |
%157 = torch.aten.item %155 : !torch.vtensor<[],si8> -> !torch.int | |
%158 = torch.aten.quantize_per_tensor %153, %156, %157, %int12_1 : !torch.vtensor<[1,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,8400],!torch.qint8> | |
%159 = torch.aten.int_repr %158 : !torch.vtensor<[1,8400],!torch.qint8> -> !torch.vtensor<[1,8400],si8> | |
%160 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%161 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%162 = torch.aten.item %160 : !torch.vtensor<[],f32> -> !torch.float | |
%163 = torch.aten.item %161 : !torch.vtensor<[],si8> -> !torch.int | |
%164 = torch.aten._make_per_tensor_quantized_tensor %159, %162, %163 : !torch.vtensor<[1,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,8400],!torch.qint8> | |
%165 = torch.aten.dequantize.self %164 : !torch.vtensor<[1,8400],!torch.qint8> -> !torch.vtensor<[1,8400],f32> | |
%166 = torch.vtensor.literal(dense<2.44140625E-4> : tensor<f32>) : !torch.vtensor<[],f32> | |
%167 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_2 = torch.constant.int 12 | |
%168 = torch.aten.item %166 : !torch.vtensor<[],f32> -> !torch.float | |
%169 = torch.aten.item %167 : !torch.vtensor<[],si8> -> !torch.int | |
%170 = torch.aten.quantize_per_tensor %arg0, %168, %169, %int12_2 : !torch.vtensor<[1,3,640,640],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,3,640,640],!torch.qint8> | |
%171 = torch.aten.int_repr %170 : !torch.vtensor<[1,3,640,640],!torch.qint8> -> !torch.vtensor<[1,3,640,640],si8> | |
%172 = torch.vtensor.literal(dense<2.44140625E-4> : tensor<f32>) : !torch.vtensor<[],f32> | |
%173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%174 = torch.aten.item %172 : !torch.vtensor<[],f32> -> !torch.float | |
%175 = torch.aten.item %173 : !torch.vtensor<[],si8> -> !torch.int | |
%176 = torch.aten._make_per_tensor_quantized_tensor %171, %174, %175 : !torch.vtensor<[1,3,640,640],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,3,640,640],!torch.qint8> | |
%177 = torch.aten.dequantize.self %176 : !torch.vtensor<[1,3,640,640],!torch.qint8> -> !torch.vtensor<[1,3,640,640],f32> | |
%178 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_3 = torch.constant.int 12 | |
%180 = torch.aten.item %178 : !torch.vtensor<[],f32> -> !torch.float | |
%181 = torch.aten.item %179 : !torch.vtensor<[],si8> -> !torch.int | |
%182 = torch.aten.quantize_per_tensor %0, %180, %181, %int12_3 : !torch.vtensor<[16,3,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,3,3,3],!torch.qint8> | |
%183 = torch.aten.int_repr %182 : !torch.vtensor<[16,3,3,3],!torch.qint8> -> !torch.vtensor<[16,3,3,3],si8> | |
%184 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%185 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%186 = torch.aten.item %184 : !torch.vtensor<[],f32> -> !torch.float | |
%187 = torch.aten.item %185 : !torch.vtensor<[],si8> -> !torch.int | |
%188 = torch.aten._make_per_tensor_quantized_tensor %183, %186, %187 : !torch.vtensor<[16,3,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,3,3,3],!torch.qint8> | |
%189 = torch.aten.dequantize.self %188 : !torch.vtensor<[16,3,3,3],!torch.qint8> -> !torch.vtensor<[16,3,3,3],f32> | |
%190 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%191 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_4 = torch.constant.int 12 | |
%192 = torch.aten.item %190 : !torch.vtensor<[],f32> -> !torch.float | |
%193 = torch.aten.item %191 : !torch.vtensor<[],si8> -> !torch.int | |
%194 = torch.aten.quantize_per_tensor %1, %192, %193, %int12_4 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
%195 = torch.aten.int_repr %194 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
%196 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%197 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%198 = torch.aten.item %196 : !torch.vtensor<[],f32> -> !torch.float | |
%199 = torch.aten.item %197 : !torch.vtensor<[],si8> -> !torch.int | |
%200 = torch.aten._make_per_tensor_quantized_tensor %195, %198, %199 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
%201 = torch.aten.dequantize.self %200 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
%int1 = torch.constant.int 1 | |
%int1_5 = torch.constant.int 1 | |
%int1_6 = torch.constant.int 1 | |
%int1_7 = torch.constant.int 1 | |
%int2 = torch.constant.int 2 | |
%int2_8 = torch.constant.int 2 | |
%int0 = torch.constant.int 0 | |
%202 = torch.prim.ListConstruct %int1, %int1_5 : (!torch.int, !torch.int) -> !torch.list<int> | |
%203 = torch.prim.ListConstruct %int1_6, %int1_7 : (!torch.int, !torch.int) -> !torch.list<int> | |
%204 = torch.prim.ListConstruct %int2, %int2_8 : (!torch.int, !torch.int) -> !torch.list<int> | |
%205 = torch.prim.ListConstruct %int0, %int0 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false = torch.constant.bool false | |
%int1_9 = torch.constant.int 1 | |
%206 = torch.aten.convolution %177, %189, %201, %204, %202, %203, %false, %205, %int1_9 : !torch.vtensor<[1,3,640,640],f32>, !torch.vtensor<[16,3,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,320,320],f32> | |
%207 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%208 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_10 = torch.constant.int 12 | |
%209 = torch.aten.item %207 : !torch.vtensor<[],f32> -> !torch.float | |
%210 = torch.aten.item %208 : !torch.vtensor<[],si8> -> !torch.int | |
%211 = torch.aten.quantize_per_tensor %206, %209, %210, %int12_10 : !torch.vtensor<[1,16,320,320],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
%212 = torch.aten.int_repr %211 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],si8> | |
%213 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%214 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%215 = torch.aten.item %213 : !torch.vtensor<[],f32> -> !torch.float | |
%216 = torch.aten.item %214 : !torch.vtensor<[],si8> -> !torch.int | |
%217 = torch.aten._make_per_tensor_quantized_tensor %212, %215, %216 : !torch.vtensor<[1,16,320,320],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
%218 = torch.aten.dequantize.self %217 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],f32> | |
%219 = torch.aten.sigmoid %218 : !torch.vtensor<[1,16,320,320],f32> -> !torch.vtensor<[1,16,320,320],f32> | |
%220 = torch.aten.mul.Tensor %218, %219 : !torch.vtensor<[1,16,320,320],f32>, !torch.vtensor<[1,16,320,320],f32> -> !torch.vtensor<[1,16,320,320],f32> | |
%221 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_11 = torch.constant.int 12 | |
%223 = torch.aten.item %221 : !torch.vtensor<[],f32> -> !torch.float | |
%224 = torch.aten.item %222 : !torch.vtensor<[],si8> -> !torch.int | |
%225 = torch.aten.quantize_per_tensor %220, %223, %224, %int12_11 : !torch.vtensor<[1,16,320,320],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
%226 = torch.aten.int_repr %225 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],si8> | |
%227 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%228 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%229 = torch.aten.item %227 : !torch.vtensor<[],f32> -> !torch.float | |
%230 = torch.aten.item %228 : !torch.vtensor<[],si8> -> !torch.int | |
%231 = torch.aten._make_per_tensor_quantized_tensor %226, %229, %230 : !torch.vtensor<[1,16,320,320],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
%232 = torch.aten.dequantize.self %231 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],f32> | |
%233 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%234 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_12 = torch.constant.int 12 | |
%235 = torch.aten.item %233 : !torch.vtensor<[],f32> -> !torch.float | |
%236 = torch.aten.item %234 : !torch.vtensor<[],si8> -> !torch.int | |
%237 = torch.aten.quantize_per_tensor %2, %235, %236, %int12_12 : !torch.vtensor<[32,16,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,16,3,3],!torch.qint8> | |
%238 = torch.aten.int_repr %237 : !torch.vtensor<[32,16,3,3],!torch.qint8> -> !torch.vtensor<[32,16,3,3],si8> | |
%239 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%240 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%241 = torch.aten.item %239 : !torch.vtensor<[],f32> -> !torch.float | |
%242 = torch.aten.item %240 : !torch.vtensor<[],si8> -> !torch.int | |
%243 = torch.aten._make_per_tensor_quantized_tensor %238, %241, %242 : !torch.vtensor<[32,16,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,16,3,3],!torch.qint8> | |
%244 = torch.aten.dequantize.self %243 : !torch.vtensor<[32,16,3,3],!torch.qint8> -> !torch.vtensor<[32,16,3,3],f32> | |
%245 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%246 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_13 = torch.constant.int 12 | |
%247 = torch.aten.item %245 : !torch.vtensor<[],f32> -> !torch.float | |
%248 = torch.aten.item %246 : !torch.vtensor<[],si8> -> !torch.int | |
%249 = torch.aten.quantize_per_tensor %3, %247, %248, %int12_13 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%250 = torch.aten.int_repr %249 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%251 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%252 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%253 = torch.aten.item %251 : !torch.vtensor<[],f32> -> !torch.float | |
%254 = torch.aten.item %252 : !torch.vtensor<[],si8> -> !torch.int | |
%255 = torch.aten._make_per_tensor_quantized_tensor %250, %253, %254 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%256 = torch.aten.dequantize.self %255 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_14 = torch.constant.int 1 | |
%int1_15 = torch.constant.int 1 | |
%int1_16 = torch.constant.int 1 | |
%int1_17 = torch.constant.int 1 | |
%int2_18 = torch.constant.int 2 | |
%int2_19 = torch.constant.int 2 | |
%int0_20 = torch.constant.int 0 | |
%257 = torch.prim.ListConstruct %int1_14, %int1_15 : (!torch.int, !torch.int) -> !torch.list<int> | |
%258 = torch.prim.ListConstruct %int1_16, %int1_17 : (!torch.int, !torch.int) -> !torch.list<int> | |
%259 = torch.prim.ListConstruct %int2_18, %int2_19 : (!torch.int, !torch.int) -> !torch.list<int> | |
%260 = torch.prim.ListConstruct %int0_20, %int0_20 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_21 = torch.constant.bool false | |
%int1_22 = torch.constant.int 1 | |
%261 = torch.aten.convolution %232, %244, %256, %259, %257, %258, %false_21, %260, %int1_22 : !torch.vtensor<[1,16,320,320],f32>, !torch.vtensor<[32,16,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,160,160],f32> | |
%262 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%263 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_23 = torch.constant.int 12 | |
%264 = torch.aten.item %262 : !torch.vtensor<[],f32> -> !torch.float | |
%265 = torch.aten.item %263 : !torch.vtensor<[],si8> -> !torch.int | |
%266 = torch.aten.quantize_per_tensor %261, %264, %265, %int12_23 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%267 = torch.aten.int_repr %266 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
%268 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%269 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%270 = torch.aten.item %268 : !torch.vtensor<[],f32> -> !torch.float | |
%271 = torch.aten.item %269 : !torch.vtensor<[],si8> -> !torch.int | |
%272 = torch.aten._make_per_tensor_quantized_tensor %267, %270, %271 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%273 = torch.aten.dequantize.self %272 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
%274 = torch.aten.sigmoid %273 : !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
%275 = torch.aten.mul.Tensor %273, %274 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
%276 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%277 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_24 = torch.constant.int 12 | |
%278 = torch.aten.item %276 : !torch.vtensor<[],f32> -> !torch.float | |
%279 = torch.aten.item %277 : !torch.vtensor<[],si8> -> !torch.int | |
%280 = torch.aten.quantize_per_tensor %275, %278, %279, %int12_24 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%281 = torch.aten.int_repr %280 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
%282 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%283 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%284 = torch.aten.item %282 : !torch.vtensor<[],f32> -> !torch.float | |
%285 = torch.aten.item %283 : !torch.vtensor<[],si8> -> !torch.int | |
%286 = torch.aten._make_per_tensor_quantized_tensor %281, %284, %285 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%287 = torch.aten.dequantize.self %286 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
%288 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%289 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_25 = torch.constant.int 12 | |
%290 = torch.aten.item %288 : !torch.vtensor<[],f32> -> !torch.float | |
%291 = torch.aten.item %289 : !torch.vtensor<[],si8> -> !torch.int | |
%292 = torch.aten.quantize_per_tensor %4, %290, %291, %int12_25 : !torch.vtensor<[32,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,1,1],!torch.qint8> | |
%293 = torch.aten.int_repr %292 : !torch.vtensor<[32,32,1,1],!torch.qint8> -> !torch.vtensor<[32,32,1,1],si8> | |
%294 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%295 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%296 = torch.aten.item %294 : !torch.vtensor<[],f32> -> !torch.float | |
%297 = torch.aten.item %295 : !torch.vtensor<[],si8> -> !torch.int | |
%298 = torch.aten._make_per_tensor_quantized_tensor %293, %296, %297 : !torch.vtensor<[32,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,1,1],!torch.qint8> | |
%299 = torch.aten.dequantize.self %298 : !torch.vtensor<[32,32,1,1],!torch.qint8> -> !torch.vtensor<[32,32,1,1],f32> | |
%300 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%301 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_26 = torch.constant.int 12 | |
%302 = torch.aten.item %300 : !torch.vtensor<[],f32> -> !torch.float | |
%303 = torch.aten.item %301 : !torch.vtensor<[],si8> -> !torch.int | |
%304 = torch.aten.quantize_per_tensor %5, %302, %303, %int12_26 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%305 = torch.aten.int_repr %304 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%306 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%307 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%308 = torch.aten.item %306 : !torch.vtensor<[],f32> -> !torch.float | |
%309 = torch.aten.item %307 : !torch.vtensor<[],si8> -> !torch.int | |
%310 = torch.aten._make_per_tensor_quantized_tensor %305, %308, %309 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%311 = torch.aten.dequantize.self %310 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int0_27 = torch.constant.int 0 | |
%int0_28 = torch.constant.int 0 | |
%int1_29 = torch.constant.int 1 | |
%int1_30 = torch.constant.int 1 | |
%int1_31 = torch.constant.int 1 | |
%int1_32 = torch.constant.int 1 | |
%int0_33 = torch.constant.int 0 | |
%312 = torch.prim.ListConstruct %int0_27, %int0_28 : (!torch.int, !torch.int) -> !torch.list<int> | |
%313 = torch.prim.ListConstruct %int1_29, %int1_30 : (!torch.int, !torch.int) -> !torch.list<int> | |
%314 = torch.prim.ListConstruct %int1_31, %int1_32 : (!torch.int, !torch.int) -> !torch.list<int> | |
%315 = torch.prim.ListConstruct %int0_33, %int0_33 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_34 = torch.constant.bool false | |
%int1_35 = torch.constant.int 1 | |
%316 = torch.aten.convolution %287, %299, %311, %314, %312, %313, %false_34, %315, %int1_35 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[32,32,1,1],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,160,160],f32> | |
%317 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%318 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_36 = torch.constant.int 12 | |
%319 = torch.aten.item %317 : !torch.vtensor<[],f32> -> !torch.float | |
%320 = torch.aten.item %318 : !torch.vtensor<[],si8> -> !torch.int | |
%321 = torch.aten.quantize_per_tensor %316, %319, %320, %int12_36 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%322 = torch.aten.int_repr %321 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
%323 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%324 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%325 = torch.aten.item %323 : !torch.vtensor<[],f32> -> !torch.float | |
%326 = torch.aten.item %324 : !torch.vtensor<[],si8> -> !torch.int | |
%327 = torch.aten._make_per_tensor_quantized_tensor %322, %325, %326 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%328 = torch.aten.dequantize.self %327 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
%329 = torch.aten.sigmoid %328 : !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
%330 = torch.aten.mul.Tensor %328, %329 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
%331 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%332 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_37 = torch.constant.int 12 | |
%333 = torch.aten.item %331 : !torch.vtensor<[],f32> -> !torch.float | |
%334 = torch.aten.item %332 : !torch.vtensor<[],si8> -> !torch.int | |
%335 = torch.aten.quantize_per_tensor %330, %333, %334, %int12_37 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%336 = torch.aten.int_repr %335 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
%337 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%338 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%339 = torch.aten.item %337 : !torch.vtensor<[],f32> -> !torch.float | |
%340 = torch.aten.item %338 : !torch.vtensor<[],si8> -> !torch.int | |
%341 = torch.aten._make_per_tensor_quantized_tensor %336, %339, %340 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%342 = torch.aten.dequantize.self %341 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
%343 = torch.vtensor.literal(dense<16> : tensor<si64>) : !torch.vtensor<[],si64> | |
%344 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%345 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%346 = torch.vtensor.literal(dense<16> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%347 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_38 = torch.constant.int 0 | |
%int0_39 = torch.constant.int 0 | |
%348 = torch.prim.NumToTensor.Scalar %int0_39 : !torch.int -> !torch.vtensor<[1],si64> | |
%349 = torch.aten.index_select %345, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%350 = torch.aten.item %349 : !torch.vtensor<[1],si64> -> !torch.int | |
%351 = torch.aten.index_select %346, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%352 = torch.aten.item %351 : !torch.vtensor<[1],si64> -> !torch.int | |
%353 = torch.aten.index_select %344, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%354 = torch.aten.item %353 : !torch.vtensor<[1],si64> -> !torch.int | |
%355 = torch.aten.index_select %347, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%356 = torch.aten.item %355 : !torch.vtensor<[1],si64> -> !torch.int | |
%357 = torch.aten.slice.Tensor %342, %354, %350, %352, %356 : !torch.vtensor<[1,32,160,160],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
%358 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%359 = torch.vtensor.literal(dense<16> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%360 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%361 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_40 = torch.constant.int 0 | |
%int0_41 = torch.constant.int 0 | |
%362 = torch.prim.NumToTensor.Scalar %int0_41 : !torch.int -> !torch.vtensor<[1],si64> | |
%363 = torch.aten.index_select %359, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%364 = torch.aten.item %363 : !torch.vtensor<[1],si64> -> !torch.int | |
%365 = torch.aten.index_select %360, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%366 = torch.aten.item %365 : !torch.vtensor<[1],si64> -> !torch.int | |
%367 = torch.aten.index_select %358, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%368 = torch.aten.item %367 : !torch.vtensor<[1],si64> -> !torch.int | |
%369 = torch.aten.index_select %361, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%370 = torch.aten.item %369 : !torch.vtensor<[1],si64> -> !torch.int | |
%371 = torch.aten.slice.Tensor %342, %368, %364, %366, %370 : !torch.vtensor<[1,32,160,160],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
%372 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%373 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_42 = torch.constant.int 12 | |
%374 = torch.aten.item %372 : !torch.vtensor<[],f32> -> !torch.float | |
%375 = torch.aten.item %373 : !torch.vtensor<[],si8> -> !torch.int | |
%376 = torch.aten.quantize_per_tensor %6, %374, %375, %int12_42 : !torch.vtensor<[16,16,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
%377 = torch.aten.int_repr %376 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],si8> | |
%378 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%379 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%380 = torch.aten.item %378 : !torch.vtensor<[],f32> -> !torch.float | |
%381 = torch.aten.item %379 : !torch.vtensor<[],si8> -> !torch.int | |
%382 = torch.aten._make_per_tensor_quantized_tensor %377, %380, %381 : !torch.vtensor<[16,16,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
%383 = torch.aten.dequantize.self %382 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],f32> | |
%384 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%385 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_43 = torch.constant.int 12 | |
%386 = torch.aten.item %384 : !torch.vtensor<[],f32> -> !torch.float | |
%387 = torch.aten.item %385 : !torch.vtensor<[],si8> -> !torch.int | |
%388 = torch.aten.quantize_per_tensor %7, %386, %387, %int12_43 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
%389 = torch.aten.int_repr %388 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
%390 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%391 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%392 = torch.aten.item %390 : !torch.vtensor<[],f32> -> !torch.float | |
%393 = torch.aten.item %391 : !torch.vtensor<[],si8> -> !torch.int | |
%394 = torch.aten._make_per_tensor_quantized_tensor %389, %392, %393 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
%395 = torch.aten.dequantize.self %394 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
%int1_44 = torch.constant.int 1 | |
%int1_45 = torch.constant.int 1 | |
%int1_46 = torch.constant.int 1 | |
%int1_47 = torch.constant.int 1 | |
%int1_48 = torch.constant.int 1 | |
%int1_49 = torch.constant.int 1 | |
%int0_50 = torch.constant.int 0 | |
%396 = torch.prim.ListConstruct %int1_44, %int1_45 : (!torch.int, !torch.int) -> !torch.list<int> | |
%397 = torch.prim.ListConstruct %int1_46, %int1_47 : (!torch.int, !torch.int) -> !torch.list<int> | |
%398 = torch.prim.ListConstruct %int1_48, %int1_49 : (!torch.int, !torch.int) -> !torch.list<int> | |
%399 = torch.prim.ListConstruct %int0_50, %int0_50 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_51 = torch.constant.bool false | |
%int1_52 = torch.constant.int 1 | |
%400 = torch.aten.convolution %371, %383, %395, %398, %396, %397, %false_51, %399, %int1_52 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[16,16,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
%401 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%402 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_53 = torch.constant.int 12 | |
%403 = torch.aten.item %401 : !torch.vtensor<[],f32> -> !torch.float | |
%404 = torch.aten.item %402 : !torch.vtensor<[],si8> -> !torch.int | |
%405 = torch.aten.quantize_per_tensor %400, %403, %404, %int12_53 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%406 = torch.aten.int_repr %405 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
%407 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%408 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%409 = torch.aten.item %407 : !torch.vtensor<[],f32> -> !torch.float | |
%410 = torch.aten.item %408 : !torch.vtensor<[],si8> -> !torch.int | |
%411 = torch.aten._make_per_tensor_quantized_tensor %406, %409, %410 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%412 = torch.aten.dequantize.self %411 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
%413 = torch.aten.sigmoid %412 : !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
%414 = torch.aten.mul.Tensor %412, %413 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
%415 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%416 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_54 = torch.constant.int 12 | |
%417 = torch.aten.item %415 : !torch.vtensor<[],f32> -> !torch.float | |
%418 = torch.aten.item %416 : !torch.vtensor<[],si8> -> !torch.int | |
%419 = torch.aten.quantize_per_tensor %414, %417, %418, %int12_54 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%420 = torch.aten.int_repr %419 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
%421 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%422 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%423 = torch.aten.item %421 : !torch.vtensor<[],f32> -> !torch.float | |
%424 = torch.aten.item %422 : !torch.vtensor<[],si8> -> !torch.int | |
%425 = torch.aten._make_per_tensor_quantized_tensor %420, %423, %424 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%426 = torch.aten.dequantize.self %425 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
%427 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%428 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_55 = torch.constant.int 12 | |
%429 = torch.aten.item %427 : !torch.vtensor<[],f32> -> !torch.float | |
%430 = torch.aten.item %428 : !torch.vtensor<[],si8> -> !torch.int | |
%431 = torch.aten.quantize_per_tensor %8, %429, %430, %int12_55 : !torch.vtensor<[16,16,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
%432 = torch.aten.int_repr %431 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],si8> | |
%433 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%434 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%435 = torch.aten.item %433 : !torch.vtensor<[],f32> -> !torch.float | |
%436 = torch.aten.item %434 : !torch.vtensor<[],si8> -> !torch.int | |
%437 = torch.aten._make_per_tensor_quantized_tensor %432, %435, %436 : !torch.vtensor<[16,16,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
%438 = torch.aten.dequantize.self %437 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],f32> | |
%439 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%440 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_56 = torch.constant.int 12 | |
%441 = torch.aten.item %439 : !torch.vtensor<[],f32> -> !torch.float | |
%442 = torch.aten.item %440 : !torch.vtensor<[],si8> -> !torch.int | |
%443 = torch.aten.quantize_per_tensor %9, %441, %442, %int12_56 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
%444 = torch.aten.int_repr %443 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
%445 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%446 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%447 = torch.aten.item %445 : !torch.vtensor<[],f32> -> !torch.float | |
%448 = torch.aten.item %446 : !torch.vtensor<[],si8> -> !torch.int | |
%449 = torch.aten._make_per_tensor_quantized_tensor %444, %447, %448 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
%450 = torch.aten.dequantize.self %449 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
%int1_57 = torch.constant.int 1 | |
%int1_58 = torch.constant.int 1 | |
%int1_59 = torch.constant.int 1 | |
%int1_60 = torch.constant.int 1 | |
%int1_61 = torch.constant.int 1 | |
%int1_62 = torch.constant.int 1 | |
%int0_63 = torch.constant.int 0 | |
%451 = torch.prim.ListConstruct %int1_57, %int1_58 : (!torch.int, !torch.int) -> !torch.list<int> | |
%452 = torch.prim.ListConstruct %int1_59, %int1_60 : (!torch.int, !torch.int) -> !torch.list<int> | |
%453 = torch.prim.ListConstruct %int1_61, %int1_62 : (!torch.int, !torch.int) -> !torch.list<int> | |
%454 = torch.prim.ListConstruct %int0_63, %int0_63 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_64 = torch.constant.bool false | |
%int1_65 = torch.constant.int 1 | |
%455 = torch.aten.convolution %426, %438, %450, %453, %451, %452, %false_64, %454, %int1_65 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[16,16,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
%456 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%457 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_66 = torch.constant.int 12 | |
%458 = torch.aten.item %456 : !torch.vtensor<[],f32> -> !torch.float | |
%459 = torch.aten.item %457 : !torch.vtensor<[],si8> -> !torch.int | |
%460 = torch.aten.quantize_per_tensor %455, %458, %459, %int12_66 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%461 = torch.aten.int_repr %460 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
%462 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%463 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%464 = torch.aten.item %462 : !torch.vtensor<[],f32> -> !torch.float | |
%465 = torch.aten.item %463 : !torch.vtensor<[],si8> -> !torch.int | |
%466 = torch.aten._make_per_tensor_quantized_tensor %461, %464, %465 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%467 = torch.aten.dequantize.self %466 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
%468 = torch.aten.sigmoid %467 : !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
%469 = torch.aten.mul.Tensor %467, %468 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
%470 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%471 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_67 = torch.constant.int 12 | |
%472 = torch.aten.item %470 : !torch.vtensor<[],f32> -> !torch.float | |
%473 = torch.aten.item %471 : !torch.vtensor<[],si8> -> !torch.int | |
%474 = torch.aten.quantize_per_tensor %469, %472, %473, %int12_67 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%475 = torch.aten.int_repr %474 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
%476 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%477 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%478 = torch.aten.item %476 : !torch.vtensor<[],f32> -> !torch.float | |
%479 = torch.aten.item %477 : !torch.vtensor<[],si8> -> !torch.int | |
%480 = torch.aten._make_per_tensor_quantized_tensor %475, %478, %479 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
%481 = torch.aten.dequantize.self %480 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
%int1_68 = torch.constant.int 1 | |
%482 = torch.aten.add.Tensor %371, %481, %int1_68 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32>, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
%483 = torch.prim.ListConstruct %357, %371, %482 : (!torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32>) -> !torch.list<vtensor> | |
%int1_69 = torch.constant.int 1 | |
%484 = torch.aten.cat %483, %int1_69 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,48,160,160],f32> | |
%485 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%486 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_70 = torch.constant.int 12 | |
%487 = torch.aten.item %485 : !torch.vtensor<[],f32> -> !torch.float | |
%488 = torch.aten.item %486 : !torch.vtensor<[],si8> -> !torch.int | |
%489 = torch.aten.quantize_per_tensor %484, %487, %488, %int12_70 : !torch.vtensor<[1,48,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,48,160,160],!torch.qint8> | |
%490 = torch.aten.int_repr %489 : !torch.vtensor<[1,48,160,160],!torch.qint8> -> !torch.vtensor<[1,48,160,160],si8> | |
%491 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%492 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%493 = torch.aten.item %491 : !torch.vtensor<[],f32> -> !torch.float | |
%494 = torch.aten.item %492 : !torch.vtensor<[],si8> -> !torch.int | |
%495 = torch.aten._make_per_tensor_quantized_tensor %490, %493, %494 : !torch.vtensor<[1,48,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,48,160,160],!torch.qint8> | |
%496 = torch.aten.dequantize.self %495 : !torch.vtensor<[1,48,160,160],!torch.qint8> -> !torch.vtensor<[1,48,160,160],f32> | |
%497 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%498 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_71 = torch.constant.int 12 | |
%499 = torch.aten.item %497 : !torch.vtensor<[],f32> -> !torch.float | |
%500 = torch.aten.item %498 : !torch.vtensor<[],si8> -> !torch.int | |
%501 = torch.aten.quantize_per_tensor %10, %499, %500, %int12_71 : !torch.vtensor<[32,48,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,48,1,1],!torch.qint8> | |
%502 = torch.aten.int_repr %501 : !torch.vtensor<[32,48,1,1],!torch.qint8> -> !torch.vtensor<[32,48,1,1],si8> | |
%503 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%504 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%505 = torch.aten.item %503 : !torch.vtensor<[],f32> -> !torch.float | |
%506 = torch.aten.item %504 : !torch.vtensor<[],si8> -> !torch.int | |
%507 = torch.aten._make_per_tensor_quantized_tensor %502, %505, %506 : !torch.vtensor<[32,48,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,48,1,1],!torch.qint8> | |
%508 = torch.aten.dequantize.self %507 : !torch.vtensor<[32,48,1,1],!torch.qint8> -> !torch.vtensor<[32,48,1,1],f32> | |
%509 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%510 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_72 = torch.constant.int 12 | |
%511 = torch.aten.item %509 : !torch.vtensor<[],f32> -> !torch.float | |
%512 = torch.aten.item %510 : !torch.vtensor<[],si8> -> !torch.int | |
%513 = torch.aten.quantize_per_tensor %11, %511, %512, %int12_72 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%514 = torch.aten.int_repr %513 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%515 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%516 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%517 = torch.aten.item %515 : !torch.vtensor<[],f32> -> !torch.float | |
%518 = torch.aten.item %516 : !torch.vtensor<[],si8> -> !torch.int | |
%519 = torch.aten._make_per_tensor_quantized_tensor %514, %517, %518 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%520 = torch.aten.dequantize.self %519 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int0_73 = torch.constant.int 0 | |
%int0_74 = torch.constant.int 0 | |
%int1_75 = torch.constant.int 1 | |
%int1_76 = torch.constant.int 1 | |
%int1_77 = torch.constant.int 1 | |
%int1_78 = torch.constant.int 1 | |
%int0_79 = torch.constant.int 0 | |
%521 = torch.prim.ListConstruct %int0_73, %int0_74 : (!torch.int, !torch.int) -> !torch.list<int> | |
%522 = torch.prim.ListConstruct %int1_75, %int1_76 : (!torch.int, !torch.int) -> !torch.list<int> | |
%523 = torch.prim.ListConstruct %int1_77, %int1_78 : (!torch.int, !torch.int) -> !torch.list<int> | |
%524 = torch.prim.ListConstruct %int0_79, %int0_79 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_80 = torch.constant.bool false | |
%int1_81 = torch.constant.int 1 | |
%525 = torch.aten.convolution %496, %508, %520, %523, %521, %522, %false_80, %524, %int1_81 : !torch.vtensor<[1,48,160,160],f32>, !torch.vtensor<[32,48,1,1],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,160,160],f32> | |
%526 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%527 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_82 = torch.constant.int 12 | |
%528 = torch.aten.item %526 : !torch.vtensor<[],f32> -> !torch.float | |
%529 = torch.aten.item %527 : !torch.vtensor<[],si8> -> !torch.int | |
%530 = torch.aten.quantize_per_tensor %525, %528, %529, %int12_82 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%531 = torch.aten.int_repr %530 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
%532 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%533 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%534 = torch.aten.item %532 : !torch.vtensor<[],f32> -> !torch.float | |
%535 = torch.aten.item %533 : !torch.vtensor<[],si8> -> !torch.int | |
%536 = torch.aten._make_per_tensor_quantized_tensor %531, %534, %535 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%537 = torch.aten.dequantize.self %536 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
%538 = torch.aten.sigmoid %537 : !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
%539 = torch.aten.mul.Tensor %537, %538 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
%540 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%541 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_83 = torch.constant.int 12 | |
%542 = torch.aten.item %540 : !torch.vtensor<[],f32> -> !torch.float | |
%543 = torch.aten.item %541 : !torch.vtensor<[],si8> -> !torch.int | |
%544 = torch.aten.quantize_per_tensor %539, %542, %543, %int12_83 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%545 = torch.aten.int_repr %544 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
%546 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%547 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%548 = torch.aten.item %546 : !torch.vtensor<[],f32> -> !torch.float | |
%549 = torch.aten.item %547 : !torch.vtensor<[],si8> -> !torch.int | |
%550 = torch.aten._make_per_tensor_quantized_tensor %545, %548, %549 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
%551 = torch.aten.dequantize.self %550 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
%552 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%553 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_84 = torch.constant.int 12 | |
%554 = torch.aten.item %552 : !torch.vtensor<[],f32> -> !torch.float | |
%555 = torch.aten.item %553 : !torch.vtensor<[],si8> -> !torch.int | |
%556 = torch.aten.quantize_per_tensor %12, %554, %555, %int12_84 : !torch.vtensor<[64,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,32,3,3],!torch.qint8> | |
%557 = torch.aten.int_repr %556 : !torch.vtensor<[64,32,3,3],!torch.qint8> -> !torch.vtensor<[64,32,3,3],si8> | |
%558 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%559 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%560 = torch.aten.item %558 : !torch.vtensor<[],f32> -> !torch.float | |
%561 = torch.aten.item %559 : !torch.vtensor<[],si8> -> !torch.int | |
%562 = torch.aten._make_per_tensor_quantized_tensor %557, %560, %561 : !torch.vtensor<[64,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,32,3,3],!torch.qint8> | |
%563 = torch.aten.dequantize.self %562 : !torch.vtensor<[64,32,3,3],!torch.qint8> -> !torch.vtensor<[64,32,3,3],f32> | |
%564 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%565 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_85 = torch.constant.int 12 | |
%566 = torch.aten.item %564 : !torch.vtensor<[],f32> -> !torch.float | |
%567 = torch.aten.item %565 : !torch.vtensor<[],si8> -> !torch.int | |
%568 = torch.aten.quantize_per_tensor %13, %566, %567, %int12_85 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%569 = torch.aten.int_repr %568 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%570 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%571 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%572 = torch.aten.item %570 : !torch.vtensor<[],f32> -> !torch.float | |
%573 = torch.aten.item %571 : !torch.vtensor<[],si8> -> !torch.int | |
%574 = torch.aten._make_per_tensor_quantized_tensor %569, %572, %573 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%575 = torch.aten.dequantize.self %574 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_86 = torch.constant.int 1 | |
%int1_87 = torch.constant.int 1 | |
%int1_88 = torch.constant.int 1 | |
%int1_89 = torch.constant.int 1 | |
%int2_90 = torch.constant.int 2 | |
%int2_91 = torch.constant.int 2 | |
%int0_92 = torch.constant.int 0 | |
%576 = torch.prim.ListConstruct %int1_86, %int1_87 : (!torch.int, !torch.int) -> !torch.list<int> | |
%577 = torch.prim.ListConstruct %int1_88, %int1_89 : (!torch.int, !torch.int) -> !torch.list<int> | |
%578 = torch.prim.ListConstruct %int2_90, %int2_91 : (!torch.int, !torch.int) -> !torch.list<int> | |
%579 = torch.prim.ListConstruct %int0_92, %int0_92 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_93 = torch.constant.bool false | |
%int1_94 = torch.constant.int 1 | |
%580 = torch.aten.convolution %551, %563, %575, %578, %576, %577, %false_93, %579, %int1_94 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[64,32,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%581 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%582 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_95 = torch.constant.int 12 | |
%583 = torch.aten.item %581 : !torch.vtensor<[],f32> -> !torch.float | |
%584 = torch.aten.item %582 : !torch.vtensor<[],si8> -> !torch.int | |
%585 = torch.aten.quantize_per_tensor %580, %583, %584, %int12_95 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%586 = torch.aten.int_repr %585 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%587 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%588 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%589 = torch.aten.item %587 : !torch.vtensor<[],f32> -> !torch.float | |
%590 = torch.aten.item %588 : !torch.vtensor<[],si8> -> !torch.int | |
%591 = torch.aten._make_per_tensor_quantized_tensor %586, %589, %590 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%592 = torch.aten.dequantize.self %591 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%593 = torch.aten.sigmoid %592 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%594 = torch.aten.mul.Tensor %592, %593 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%595 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%596 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_96 = torch.constant.int 12 | |
%597 = torch.aten.item %595 : !torch.vtensor<[],f32> -> !torch.float | |
%598 = torch.aten.item %596 : !torch.vtensor<[],si8> -> !torch.int | |
%599 = torch.aten.quantize_per_tensor %594, %597, %598, %int12_96 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%600 = torch.aten.int_repr %599 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%601 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%602 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%603 = torch.aten.item %601 : !torch.vtensor<[],f32> -> !torch.float | |
%604 = torch.aten.item %602 : !torch.vtensor<[],si8> -> !torch.int | |
%605 = torch.aten._make_per_tensor_quantized_tensor %600, %603, %604 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%606 = torch.aten.dequantize.self %605 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%607 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%608 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_97 = torch.constant.int 12 | |
%609 = torch.aten.item %607 : !torch.vtensor<[],f32> -> !torch.float | |
%610 = torch.aten.item %608 : !torch.vtensor<[],si8> -> !torch.int | |
%611 = torch.aten.quantize_per_tensor %14, %609, %610, %int12_97 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%612 = torch.aten.int_repr %611 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
%613 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%614 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%615 = torch.aten.item %613 : !torch.vtensor<[],f32> -> !torch.float | |
%616 = torch.aten.item %614 : !torch.vtensor<[],si8> -> !torch.int | |
%617 = torch.aten._make_per_tensor_quantized_tensor %612, %615, %616 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%618 = torch.aten.dequantize.self %617 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
%619 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%620 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_98 = torch.constant.int 12 | |
%621 = torch.aten.item %619 : !torch.vtensor<[],f32> -> !torch.float | |
%622 = torch.aten.item %620 : !torch.vtensor<[],si8> -> !torch.int | |
%623 = torch.aten.quantize_per_tensor %15, %621, %622, %int12_98 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%624 = torch.aten.int_repr %623 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%625 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%626 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%627 = torch.aten.item %625 : !torch.vtensor<[],f32> -> !torch.float | |
%628 = torch.aten.item %626 : !torch.vtensor<[],si8> -> !torch.int | |
%629 = torch.aten._make_per_tensor_quantized_tensor %624, %627, %628 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%630 = torch.aten.dequantize.self %629 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_99 = torch.constant.int 0 | |
%int0_100 = torch.constant.int 0 | |
%int1_101 = torch.constant.int 1 | |
%int1_102 = torch.constant.int 1 | |
%int1_103 = torch.constant.int 1 | |
%int1_104 = torch.constant.int 1 | |
%int0_105 = torch.constant.int 0 | |
%631 = torch.prim.ListConstruct %int0_99, %int0_100 : (!torch.int, !torch.int) -> !torch.list<int> | |
%632 = torch.prim.ListConstruct %int1_101, %int1_102 : (!torch.int, !torch.int) -> !torch.list<int> | |
%633 = torch.prim.ListConstruct %int1_103, %int1_104 : (!torch.int, !torch.int) -> !torch.list<int> | |
%634 = torch.prim.ListConstruct %int0_105, %int0_105 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_106 = torch.constant.bool false | |
%int1_107 = torch.constant.int 1 | |
%635 = torch.aten.convolution %606, %618, %630, %633, %631, %632, %false_106, %634, %int1_107 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%636 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%637 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_108 = torch.constant.int 12 | |
%638 = torch.aten.item %636 : !torch.vtensor<[],f32> -> !torch.float | |
%639 = torch.aten.item %637 : !torch.vtensor<[],si8> -> !torch.int | |
%640 = torch.aten.quantize_per_tensor %635, %638, %639, %int12_108 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%641 = torch.aten.int_repr %640 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%642 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%643 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%644 = torch.aten.item %642 : !torch.vtensor<[],f32> -> !torch.float | |
%645 = torch.aten.item %643 : !torch.vtensor<[],si8> -> !torch.int | |
%646 = torch.aten._make_per_tensor_quantized_tensor %641, %644, %645 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%647 = torch.aten.dequantize.self %646 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%648 = torch.aten.sigmoid %647 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%649 = torch.aten.mul.Tensor %647, %648 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%650 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%651 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_109 = torch.constant.int 12 | |
%652 = torch.aten.item %650 : !torch.vtensor<[],f32> -> !torch.float | |
%653 = torch.aten.item %651 : !torch.vtensor<[],si8> -> !torch.int | |
%654 = torch.aten.quantize_per_tensor %649, %652, %653, %int12_109 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%655 = torch.aten.int_repr %654 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%656 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%657 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%658 = torch.aten.item %656 : !torch.vtensor<[],f32> -> !torch.float | |
%659 = torch.aten.item %657 : !torch.vtensor<[],si8> -> !torch.int | |
%660 = torch.aten._make_per_tensor_quantized_tensor %655, %658, %659 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%661 = torch.aten.dequantize.self %660 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%662 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%663 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%664 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%665 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_110 = torch.constant.int 0 | |
%int0_111 = torch.constant.int 0 | |
%666 = torch.prim.NumToTensor.Scalar %int0_111 : !torch.int -> !torch.vtensor<[1],si64> | |
%667 = torch.aten.index_select %663, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%668 = torch.aten.item %667 : !torch.vtensor<[1],si64> -> !torch.int | |
%669 = torch.aten.index_select %664, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%670 = torch.aten.item %669 : !torch.vtensor<[1],si64> -> !torch.int | |
%671 = torch.aten.index_select %662, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%672 = torch.aten.item %671 : !torch.vtensor<[1],si64> -> !torch.int | |
%673 = torch.aten.index_select %665, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%674 = torch.aten.item %673 : !torch.vtensor<[1],si64> -> !torch.int | |
%675 = torch.aten.slice.Tensor %661, %672, %668, %670, %674 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%676 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%677 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%678 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%679 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_112 = torch.constant.int 0 | |
%int0_113 = torch.constant.int 0 | |
%680 = torch.prim.NumToTensor.Scalar %int0_113 : !torch.int -> !torch.vtensor<[1],si64> | |
%681 = torch.aten.index_select %677, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%682 = torch.aten.item %681 : !torch.vtensor<[1],si64> -> !torch.int | |
%683 = torch.aten.index_select %678, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%684 = torch.aten.item %683 : !torch.vtensor<[1],si64> -> !torch.int | |
%685 = torch.aten.index_select %676, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%686 = torch.aten.item %685 : !torch.vtensor<[1],si64> -> !torch.int | |
%687 = torch.aten.index_select %679, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%688 = torch.aten.item %687 : !torch.vtensor<[1],si64> -> !torch.int | |
%689 = torch.aten.slice.Tensor %661, %686, %682, %684, %688 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%690 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%691 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_114 = torch.constant.int 12 | |
%692 = torch.aten.item %690 : !torch.vtensor<[],f32> -> !torch.float | |
%693 = torch.aten.item %691 : !torch.vtensor<[],si8> -> !torch.int | |
%694 = torch.aten.quantize_per_tensor %16, %692, %693, %int12_114 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%695 = torch.aten.int_repr %694 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
%696 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%697 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%698 = torch.aten.item %696 : !torch.vtensor<[],f32> -> !torch.float | |
%699 = torch.aten.item %697 : !torch.vtensor<[],si8> -> !torch.int | |
%700 = torch.aten._make_per_tensor_quantized_tensor %695, %698, %699 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%701 = torch.aten.dequantize.self %700 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
%702 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%703 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_115 = torch.constant.int 12 | |
%704 = torch.aten.item %702 : !torch.vtensor<[],f32> -> !torch.float | |
%705 = torch.aten.item %703 : !torch.vtensor<[],si8> -> !torch.int | |
%706 = torch.aten.quantize_per_tensor %17, %704, %705, %int12_115 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%707 = torch.aten.int_repr %706 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%708 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%709 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%710 = torch.aten.item %708 : !torch.vtensor<[],f32> -> !torch.float | |
%711 = torch.aten.item %709 : !torch.vtensor<[],si8> -> !torch.int | |
%712 = torch.aten._make_per_tensor_quantized_tensor %707, %710, %711 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%713 = torch.aten.dequantize.self %712 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_116 = torch.constant.int 1 | |
%int1_117 = torch.constant.int 1 | |
%int1_118 = torch.constant.int 1 | |
%int1_119 = torch.constant.int 1 | |
%int1_120 = torch.constant.int 1 | |
%int1_121 = torch.constant.int 1 | |
%int0_122 = torch.constant.int 0 | |
%714 = torch.prim.ListConstruct %int1_116, %int1_117 : (!torch.int, !torch.int) -> !torch.list<int> | |
%715 = torch.prim.ListConstruct %int1_118, %int1_119 : (!torch.int, !torch.int) -> !torch.list<int> | |
%716 = torch.prim.ListConstruct %int1_120, %int1_121 : (!torch.int, !torch.int) -> !torch.list<int> | |
%717 = torch.prim.ListConstruct %int0_122, %int0_122 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_123 = torch.constant.bool false | |
%int1_124 = torch.constant.int 1 | |
%718 = torch.aten.convolution %689, %701, %713, %716, %714, %715, %false_123, %717, %int1_124 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%719 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%720 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_125 = torch.constant.int 12 | |
%721 = torch.aten.item %719 : !torch.vtensor<[],f32> -> !torch.float | |
%722 = torch.aten.item %720 : !torch.vtensor<[],si8> -> !torch.int | |
%723 = torch.aten.quantize_per_tensor %718, %721, %722, %int12_125 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%724 = torch.aten.int_repr %723 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%725 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%726 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%727 = torch.aten.item %725 : !torch.vtensor<[],f32> -> !torch.float | |
%728 = torch.aten.item %726 : !torch.vtensor<[],si8> -> !torch.int | |
%729 = torch.aten._make_per_tensor_quantized_tensor %724, %727, %728 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%730 = torch.aten.dequantize.self %729 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%731 = torch.aten.sigmoid %730 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%732 = torch.aten.mul.Tensor %730, %731 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%733 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%734 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_126 = torch.constant.int 12 | |
%735 = torch.aten.item %733 : !torch.vtensor<[],f32> -> !torch.float | |
%736 = torch.aten.item %734 : !torch.vtensor<[],si8> -> !torch.int | |
%737 = torch.aten.quantize_per_tensor %732, %735, %736, %int12_126 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%738 = torch.aten.int_repr %737 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%739 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%740 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%741 = torch.aten.item %739 : !torch.vtensor<[],f32> -> !torch.float | |
%742 = torch.aten.item %740 : !torch.vtensor<[],si8> -> !torch.int | |
%743 = torch.aten._make_per_tensor_quantized_tensor %738, %741, %742 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%744 = torch.aten.dequantize.self %743 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%745 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%746 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_127 = torch.constant.int 12 | |
%747 = torch.aten.item %745 : !torch.vtensor<[],f32> -> !torch.float | |
%748 = torch.aten.item %746 : !torch.vtensor<[],si8> -> !torch.int | |
%749 = torch.aten.quantize_per_tensor %18, %747, %748, %int12_127 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%750 = torch.aten.int_repr %749 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
%751 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%752 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%753 = torch.aten.item %751 : !torch.vtensor<[],f32> -> !torch.float | |
%754 = torch.aten.item %752 : !torch.vtensor<[],si8> -> !torch.int | |
%755 = torch.aten._make_per_tensor_quantized_tensor %750, %753, %754 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%756 = torch.aten.dequantize.self %755 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
%757 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%758 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_128 = torch.constant.int 12 | |
%759 = torch.aten.item %757 : !torch.vtensor<[],f32> -> !torch.float | |
%760 = torch.aten.item %758 : !torch.vtensor<[],si8> -> !torch.int | |
%761 = torch.aten.quantize_per_tensor %19, %759, %760, %int12_128 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%762 = torch.aten.int_repr %761 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%763 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%764 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%765 = torch.aten.item %763 : !torch.vtensor<[],f32> -> !torch.float | |
%766 = torch.aten.item %764 : !torch.vtensor<[],si8> -> !torch.int | |
%767 = torch.aten._make_per_tensor_quantized_tensor %762, %765, %766 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%768 = torch.aten.dequantize.self %767 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_129 = torch.constant.int 1 | |
%int1_130 = torch.constant.int 1 | |
%int1_131 = torch.constant.int 1 | |
%int1_132 = torch.constant.int 1 | |
%int1_133 = torch.constant.int 1 | |
%int1_134 = torch.constant.int 1 | |
%int0_135 = torch.constant.int 0 | |
%769 = torch.prim.ListConstruct %int1_129, %int1_130 : (!torch.int, !torch.int) -> !torch.list<int> | |
%770 = torch.prim.ListConstruct %int1_131, %int1_132 : (!torch.int, !torch.int) -> !torch.list<int> | |
%771 = torch.prim.ListConstruct %int1_133, %int1_134 : (!torch.int, !torch.int) -> !torch.list<int> | |
%772 = torch.prim.ListConstruct %int0_135, %int0_135 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_136 = torch.constant.bool false | |
%int1_137 = torch.constant.int 1 | |
%773 = torch.aten.convolution %744, %756, %768, %771, %769, %770, %false_136, %772, %int1_137 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%774 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%775 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_138 = torch.constant.int 12 | |
%776 = torch.aten.item %774 : !torch.vtensor<[],f32> -> !torch.float | |
%777 = torch.aten.item %775 : !torch.vtensor<[],si8> -> !torch.int | |
%778 = torch.aten.quantize_per_tensor %773, %776, %777, %int12_138 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%779 = torch.aten.int_repr %778 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%780 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%781 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%782 = torch.aten.item %780 : !torch.vtensor<[],f32> -> !torch.float | |
%783 = torch.aten.item %781 : !torch.vtensor<[],si8> -> !torch.int | |
%784 = torch.aten._make_per_tensor_quantized_tensor %779, %782, %783 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%785 = torch.aten.dequantize.self %784 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%786 = torch.aten.sigmoid %785 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%787 = torch.aten.mul.Tensor %785, %786 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%788 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%789 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_139 = torch.constant.int 12 | |
%790 = torch.aten.item %788 : !torch.vtensor<[],f32> -> !torch.float | |
%791 = torch.aten.item %789 : !torch.vtensor<[],si8> -> !torch.int | |
%792 = torch.aten.quantize_per_tensor %787, %790, %791, %int12_139 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%793 = torch.aten.int_repr %792 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%794 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%795 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%796 = torch.aten.item %794 : !torch.vtensor<[],f32> -> !torch.float | |
%797 = torch.aten.item %795 : !torch.vtensor<[],si8> -> !torch.int | |
%798 = torch.aten._make_per_tensor_quantized_tensor %793, %796, %797 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%799 = torch.aten.dequantize.self %798 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%int1_140 = torch.constant.int 1 | |
%800 = torch.aten.add.Tensor %689, %799, %int1_140 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%801 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%802 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_141 = torch.constant.int 12 | |
%803 = torch.aten.item %801 : !torch.vtensor<[],f32> -> !torch.float | |
%804 = torch.aten.item %802 : !torch.vtensor<[],si8> -> !torch.int | |
%805 = torch.aten.quantize_per_tensor %800, %803, %804, %int12_141 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%806 = torch.aten.int_repr %805 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%807 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%808 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%809 = torch.aten.item %807 : !torch.vtensor<[],f32> -> !torch.float | |
%810 = torch.aten.item %808 : !torch.vtensor<[],si8> -> !torch.int | |
%811 = torch.aten._make_per_tensor_quantized_tensor %806, %809, %810 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%812 = torch.aten.dequantize.self %811 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%813 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%814 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_142 = torch.constant.int 12 | |
%815 = torch.aten.item %813 : !torch.vtensor<[],f32> -> !torch.float | |
%816 = torch.aten.item %814 : !torch.vtensor<[],si8> -> !torch.int | |
%817 = torch.aten.quantize_per_tensor %20, %815, %816, %int12_142 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%818 = torch.aten.int_repr %817 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
%819 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%820 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%821 = torch.aten.item %819 : !torch.vtensor<[],f32> -> !torch.float | |
%822 = torch.aten.item %820 : !torch.vtensor<[],si8> -> !torch.int | |
%823 = torch.aten._make_per_tensor_quantized_tensor %818, %821, %822 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%824 = torch.aten.dequantize.self %823 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
%825 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%826 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_143 = torch.constant.int 12 | |
%827 = torch.aten.item %825 : !torch.vtensor<[],f32> -> !torch.float | |
%828 = torch.aten.item %826 : !torch.vtensor<[],si8> -> !torch.int | |
%829 = torch.aten.quantize_per_tensor %21, %827, %828, %int12_143 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%830 = torch.aten.int_repr %829 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%831 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%832 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%833 = torch.aten.item %831 : !torch.vtensor<[],f32> -> !torch.float | |
%834 = torch.aten.item %832 : !torch.vtensor<[],si8> -> !torch.int | |
%835 = torch.aten._make_per_tensor_quantized_tensor %830, %833, %834 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%836 = torch.aten.dequantize.self %835 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_144 = torch.constant.int 1 | |
%int1_145 = torch.constant.int 1 | |
%int1_146 = torch.constant.int 1 | |
%int1_147 = torch.constant.int 1 | |
%int1_148 = torch.constant.int 1 | |
%int1_149 = torch.constant.int 1 | |
%int0_150 = torch.constant.int 0 | |
%837 = torch.prim.ListConstruct %int1_144, %int1_145 : (!torch.int, !torch.int) -> !torch.list<int> | |
%838 = torch.prim.ListConstruct %int1_146, %int1_147 : (!torch.int, !torch.int) -> !torch.list<int> | |
%839 = torch.prim.ListConstruct %int1_148, %int1_149 : (!torch.int, !torch.int) -> !torch.list<int> | |
%840 = torch.prim.ListConstruct %int0_150, %int0_150 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_151 = torch.constant.bool false | |
%int1_152 = torch.constant.int 1 | |
%841 = torch.aten.convolution %812, %824, %836, %839, %837, %838, %false_151, %840, %int1_152 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%842 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%843 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_153 = torch.constant.int 12 | |
%844 = torch.aten.item %842 : !torch.vtensor<[],f32> -> !torch.float | |
%845 = torch.aten.item %843 : !torch.vtensor<[],si8> -> !torch.int | |
%846 = torch.aten.quantize_per_tensor %841, %844, %845, %int12_153 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%847 = torch.aten.int_repr %846 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%848 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%849 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%850 = torch.aten.item %848 : !torch.vtensor<[],f32> -> !torch.float | |
%851 = torch.aten.item %849 : !torch.vtensor<[],si8> -> !torch.int | |
%852 = torch.aten._make_per_tensor_quantized_tensor %847, %850, %851 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%853 = torch.aten.dequantize.self %852 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%854 = torch.aten.sigmoid %853 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%855 = torch.aten.mul.Tensor %853, %854 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%856 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%857 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_154 = torch.constant.int 12 | |
%858 = torch.aten.item %856 : !torch.vtensor<[],f32> -> !torch.float | |
%859 = torch.aten.item %857 : !torch.vtensor<[],si8> -> !torch.int | |
%860 = torch.aten.quantize_per_tensor %855, %858, %859, %int12_154 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%861 = torch.aten.int_repr %860 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%862 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%863 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%864 = torch.aten.item %862 : !torch.vtensor<[],f32> -> !torch.float | |
%865 = torch.aten.item %863 : !torch.vtensor<[],si8> -> !torch.int | |
%866 = torch.aten._make_per_tensor_quantized_tensor %861, %864, %865 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%867 = torch.aten.dequantize.self %866 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%868 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%869 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_155 = torch.constant.int 12 | |
%870 = torch.aten.item %868 : !torch.vtensor<[],f32> -> !torch.float | |
%871 = torch.aten.item %869 : !torch.vtensor<[],si8> -> !torch.int | |
%872 = torch.aten.quantize_per_tensor %22, %870, %871, %int12_155 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%873 = torch.aten.int_repr %872 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
%874 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%875 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%876 = torch.aten.item %874 : !torch.vtensor<[],f32> -> !torch.float | |
%877 = torch.aten.item %875 : !torch.vtensor<[],si8> -> !torch.int | |
%878 = torch.aten._make_per_tensor_quantized_tensor %873, %876, %877 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%879 = torch.aten.dequantize.self %878 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
%880 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%881 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_156 = torch.constant.int 12 | |
%882 = torch.aten.item %880 : !torch.vtensor<[],f32> -> !torch.float | |
%883 = torch.aten.item %881 : !torch.vtensor<[],si8> -> !torch.int | |
%884 = torch.aten.quantize_per_tensor %23, %882, %883, %int12_156 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%885 = torch.aten.int_repr %884 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%886 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%887 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%888 = torch.aten.item %886 : !torch.vtensor<[],f32> -> !torch.float | |
%889 = torch.aten.item %887 : !torch.vtensor<[],si8> -> !torch.int | |
%890 = torch.aten._make_per_tensor_quantized_tensor %885, %888, %889 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%891 = torch.aten.dequantize.self %890 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_157 = torch.constant.int 1 | |
%int1_158 = torch.constant.int 1 | |
%int1_159 = torch.constant.int 1 | |
%int1_160 = torch.constant.int 1 | |
%int1_161 = torch.constant.int 1 | |
%int1_162 = torch.constant.int 1 | |
%int0_163 = torch.constant.int 0 | |
%892 = torch.prim.ListConstruct %int1_157, %int1_158 : (!torch.int, !torch.int) -> !torch.list<int> | |
%893 = torch.prim.ListConstruct %int1_159, %int1_160 : (!torch.int, !torch.int) -> !torch.list<int> | |
%894 = torch.prim.ListConstruct %int1_161, %int1_162 : (!torch.int, !torch.int) -> !torch.list<int> | |
%895 = torch.prim.ListConstruct %int0_163, %int0_163 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_164 = torch.constant.bool false | |
%int1_165 = torch.constant.int 1 | |
%896 = torch.aten.convolution %867, %879, %891, %894, %892, %893, %false_164, %895, %int1_165 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%897 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%898 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_166 = torch.constant.int 12 | |
%899 = torch.aten.item %897 : !torch.vtensor<[],f32> -> !torch.float | |
%900 = torch.aten.item %898 : !torch.vtensor<[],si8> -> !torch.int | |
%901 = torch.aten.quantize_per_tensor %896, %899, %900, %int12_166 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%902 = torch.aten.int_repr %901 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%903 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%904 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%905 = torch.aten.item %903 : !torch.vtensor<[],f32> -> !torch.float | |
%906 = torch.aten.item %904 : !torch.vtensor<[],si8> -> !torch.int | |
%907 = torch.aten._make_per_tensor_quantized_tensor %902, %905, %906 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%908 = torch.aten.dequantize.self %907 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%909 = torch.aten.sigmoid %908 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%910 = torch.aten.mul.Tensor %908, %909 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%911 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%912 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_167 = torch.constant.int 12 | |
%913 = torch.aten.item %911 : !torch.vtensor<[],f32> -> !torch.float | |
%914 = torch.aten.item %912 : !torch.vtensor<[],si8> -> !torch.int | |
%915 = torch.aten.quantize_per_tensor %910, %913, %914, %int12_167 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%916 = torch.aten.int_repr %915 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%917 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%918 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%919 = torch.aten.item %917 : !torch.vtensor<[],f32> -> !torch.float | |
%920 = torch.aten.item %918 : !torch.vtensor<[],si8> -> !torch.int | |
%921 = torch.aten._make_per_tensor_quantized_tensor %916, %919, %920 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%922 = torch.aten.dequantize.self %921 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%int1_168 = torch.constant.int 1 | |
%923 = torch.aten.add.Tensor %812, %922, %int1_168 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%924 = torch.prim.ListConstruct %675, %689, %812, %923 : (!torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>) -> !torch.list<vtensor> | |
%int1_169 = torch.constant.int 1 | |
%925 = torch.aten.cat %924, %int1_169 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,128,80,80],f32> | |
%926 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%927 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_170 = torch.constant.int 12 | |
%928 = torch.aten.item %926 : !torch.vtensor<[],f32> -> !torch.float | |
%929 = torch.aten.item %927 : !torch.vtensor<[],si8> -> !torch.int | |
%930 = torch.aten.quantize_per_tensor %925, %928, %929, %int12_170 : !torch.vtensor<[1,128,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,80,80],!torch.qint8> | |
%931 = torch.aten.int_repr %930 : !torch.vtensor<[1,128,80,80],!torch.qint8> -> !torch.vtensor<[1,128,80,80],si8> | |
%932 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%933 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%934 = torch.aten.item %932 : !torch.vtensor<[],f32> -> !torch.float | |
%935 = torch.aten.item %933 : !torch.vtensor<[],si8> -> !torch.int | |
%936 = torch.aten._make_per_tensor_quantized_tensor %931, %934, %935 : !torch.vtensor<[1,128,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,80,80],!torch.qint8> | |
%937 = torch.aten.dequantize.self %936 : !torch.vtensor<[1,128,80,80],!torch.qint8> -> !torch.vtensor<[1,128,80,80],f32> | |
%938 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%939 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_171 = torch.constant.int 12 | |
%940 = torch.aten.item %938 : !torch.vtensor<[],f32> -> !torch.float | |
%941 = torch.aten.item %939 : !torch.vtensor<[],si8> -> !torch.int | |
%942 = torch.aten.quantize_per_tensor %24, %940, %941, %int12_171 : !torch.vtensor<[64,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,128,1,1],!torch.qint8> | |
%943 = torch.aten.int_repr %942 : !torch.vtensor<[64,128,1,1],!torch.qint8> -> !torch.vtensor<[64,128,1,1],si8> | |
%944 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%945 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%946 = torch.aten.item %944 : !torch.vtensor<[],f32> -> !torch.float | |
%947 = torch.aten.item %945 : !torch.vtensor<[],si8> -> !torch.int | |
%948 = torch.aten._make_per_tensor_quantized_tensor %943, %946, %947 : !torch.vtensor<[64,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,128,1,1],!torch.qint8> | |
%949 = torch.aten.dequantize.self %948 : !torch.vtensor<[64,128,1,1],!torch.qint8> -> !torch.vtensor<[64,128,1,1],f32> | |
%950 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%951 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_172 = torch.constant.int 12 | |
%952 = torch.aten.item %950 : !torch.vtensor<[],f32> -> !torch.float | |
%953 = torch.aten.item %951 : !torch.vtensor<[],si8> -> !torch.int | |
%954 = torch.aten.quantize_per_tensor %25, %952, %953, %int12_172 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%955 = torch.aten.int_repr %954 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%956 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%957 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%958 = torch.aten.item %956 : !torch.vtensor<[],f32> -> !torch.float | |
%959 = torch.aten.item %957 : !torch.vtensor<[],si8> -> !torch.int | |
%960 = torch.aten._make_per_tensor_quantized_tensor %955, %958, %959 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%961 = torch.aten.dequantize.self %960 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_173 = torch.constant.int 0 | |
%int0_174 = torch.constant.int 0 | |
%int1_175 = torch.constant.int 1 | |
%int1_176 = torch.constant.int 1 | |
%int1_177 = torch.constant.int 1 | |
%int1_178 = torch.constant.int 1 | |
%int0_179 = torch.constant.int 0 | |
%962 = torch.prim.ListConstruct %int0_173, %int0_174 : (!torch.int, !torch.int) -> !torch.list<int> | |
%963 = torch.prim.ListConstruct %int1_175, %int1_176 : (!torch.int, !torch.int) -> !torch.list<int> | |
%964 = torch.prim.ListConstruct %int1_177, %int1_178 : (!torch.int, !torch.int) -> !torch.list<int> | |
%965 = torch.prim.ListConstruct %int0_179, %int0_179 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_180 = torch.constant.bool false | |
%int1_181 = torch.constant.int 1 | |
%966 = torch.aten.convolution %937, %949, %961, %964, %962, %963, %false_180, %965, %int1_181 : !torch.vtensor<[1,128,80,80],f32>, !torch.vtensor<[64,128,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%967 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%968 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_182 = torch.constant.int 12 | |
%969 = torch.aten.item %967 : !torch.vtensor<[],f32> -> !torch.float | |
%970 = torch.aten.item %968 : !torch.vtensor<[],si8> -> !torch.int | |
%971 = torch.aten.quantize_per_tensor %966, %969, %970, %int12_182 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%972 = torch.aten.int_repr %971 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%973 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%974 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%975 = torch.aten.item %973 : !torch.vtensor<[],f32> -> !torch.float | |
%976 = torch.aten.item %974 : !torch.vtensor<[],si8> -> !torch.int | |
%977 = torch.aten._make_per_tensor_quantized_tensor %972, %975, %976 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%978 = torch.aten.dequantize.self %977 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%979 = torch.aten.sigmoid %978 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%980 = torch.aten.mul.Tensor %978, %979 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%981 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%982 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_183 = torch.constant.int 12 | |
%983 = torch.aten.item %981 : !torch.vtensor<[],f32> -> !torch.float | |
%984 = torch.aten.item %982 : !torch.vtensor<[],si8> -> !torch.int | |
%985 = torch.aten.quantize_per_tensor %980, %983, %984, %int12_183 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%986 = torch.aten.int_repr %985 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%987 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%988 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%989 = torch.aten.item %987 : !torch.vtensor<[],f32> -> !torch.float | |
%990 = torch.aten.item %988 : !torch.vtensor<[],si8> -> !torch.int | |
%991 = torch.aten._make_per_tensor_quantized_tensor %986, %989, %990 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%992 = torch.aten.dequantize.self %991 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%993 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%994 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_184 = torch.constant.int 12 | |
%995 = torch.aten.item %993 : !torch.vtensor<[],f32> -> !torch.float | |
%996 = torch.aten.item %994 : !torch.vtensor<[],si8> -> !torch.int | |
%997 = torch.aten.quantize_per_tensor %26, %995, %996, %int12_184 : !torch.vtensor<[128,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,64,3,3],!torch.qint8> | |
%998 = torch.aten.int_repr %997 : !torch.vtensor<[128,64,3,3],!torch.qint8> -> !torch.vtensor<[128,64,3,3],si8> | |
%999 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1000 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1001 = torch.aten.item %999 : !torch.vtensor<[],f32> -> !torch.float | |
%1002 = torch.aten.item %1000 : !torch.vtensor<[],si8> -> !torch.int | |
%1003 = torch.aten._make_per_tensor_quantized_tensor %998, %1001, %1002 : !torch.vtensor<[128,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,64,3,3],!torch.qint8> | |
%1004 = torch.aten.dequantize.self %1003 : !torch.vtensor<[128,64,3,3],!torch.qint8> -> !torch.vtensor<[128,64,3,3],f32> | |
%1005 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1006 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_185 = torch.constant.int 12 | |
%1007 = torch.aten.item %1005 : !torch.vtensor<[],f32> -> !torch.float | |
%1008 = torch.aten.item %1006 : !torch.vtensor<[],si8> -> !torch.int | |
%1009 = torch.aten.quantize_per_tensor %27, %1007, %1008, %int12_185 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1010 = torch.aten.int_repr %1009 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1011 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1012 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1013 = torch.aten.item %1011 : !torch.vtensor<[],f32> -> !torch.float | |
%1014 = torch.aten.item %1012 : !torch.vtensor<[],si8> -> !torch.int | |
%1015 = torch.aten._make_per_tensor_quantized_tensor %1010, %1013, %1014 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1016 = torch.aten.dequantize.self %1015 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int1_186 = torch.constant.int 1 | |
%int1_187 = torch.constant.int 1 | |
%int1_188 = torch.constant.int 1 | |
%int1_189 = torch.constant.int 1 | |
%int2_190 = torch.constant.int 2 | |
%int2_191 = torch.constant.int 2 | |
%int0_192 = torch.constant.int 0 | |
%1017 = torch.prim.ListConstruct %int1_186, %int1_187 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1018 = torch.prim.ListConstruct %int1_188, %int1_189 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1019 = torch.prim.ListConstruct %int2_190, %int2_191 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1020 = torch.prim.ListConstruct %int0_192, %int0_192 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_193 = torch.constant.bool false | |
%int1_194 = torch.constant.int 1 | |
%1021 = torch.aten.convolution %992, %1004, %1016, %1019, %1017, %1018, %false_193, %1020, %int1_194 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[128,64,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%1022 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1023 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_195 = torch.constant.int 12 | |
%1024 = torch.aten.item %1022 : !torch.vtensor<[],f32> -> !torch.float | |
%1025 = torch.aten.item %1023 : !torch.vtensor<[],si8> -> !torch.int | |
%1026 = torch.aten.quantize_per_tensor %1021, %1024, %1025, %int12_195 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1027 = torch.aten.int_repr %1026 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1028 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1029 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1030 = torch.aten.item %1028 : !torch.vtensor<[],f32> -> !torch.float | |
%1031 = torch.aten.item %1029 : !torch.vtensor<[],si8> -> !torch.int | |
%1032 = torch.aten._make_per_tensor_quantized_tensor %1027, %1030, %1031 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1033 = torch.aten.dequantize.self %1032 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1034 = torch.aten.sigmoid %1033 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1035 = torch.aten.mul.Tensor %1033, %1034 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1036 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1037 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_196 = torch.constant.int 12 | |
%1038 = torch.aten.item %1036 : !torch.vtensor<[],f32> -> !torch.float | |
%1039 = torch.aten.item %1037 : !torch.vtensor<[],si8> -> !torch.int | |
%1040 = torch.aten.quantize_per_tensor %1035, %1038, %1039, %int12_196 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1041 = torch.aten.int_repr %1040 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1042 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1043 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1044 = torch.aten.item %1042 : !torch.vtensor<[],f32> -> !torch.float | |
%1045 = torch.aten.item %1043 : !torch.vtensor<[],si8> -> !torch.int | |
%1046 = torch.aten._make_per_tensor_quantized_tensor %1041, %1044, %1045 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1047 = torch.aten.dequantize.self %1046 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1048 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1049 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_197 = torch.constant.int 12 | |
%1050 = torch.aten.item %1048 : !torch.vtensor<[],f32> -> !torch.float | |
%1051 = torch.aten.item %1049 : !torch.vtensor<[],si8> -> !torch.int | |
%1052 = torch.aten.quantize_per_tensor %28, %1050, %1051, %int12_197 : !torch.vtensor<[128,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,1,1],!torch.qint8> | |
%1053 = torch.aten.int_repr %1052 : !torch.vtensor<[128,128,1,1],!torch.qint8> -> !torch.vtensor<[128,128,1,1],si8> | |
%1054 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1055 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1056 = torch.aten.item %1054 : !torch.vtensor<[],f32> -> !torch.float | |
%1057 = torch.aten.item %1055 : !torch.vtensor<[],si8> -> !torch.int | |
%1058 = torch.aten._make_per_tensor_quantized_tensor %1053, %1056, %1057 : !torch.vtensor<[128,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,1,1],!torch.qint8> | |
%1059 = torch.aten.dequantize.self %1058 : !torch.vtensor<[128,128,1,1],!torch.qint8> -> !torch.vtensor<[128,128,1,1],f32> | |
%1060 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1061 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_198 = torch.constant.int 12 | |
%1062 = torch.aten.item %1060 : !torch.vtensor<[],f32> -> !torch.float | |
%1063 = torch.aten.item %1061 : !torch.vtensor<[],si8> -> !torch.int | |
%1064 = torch.aten.quantize_per_tensor %29, %1062, %1063, %int12_198 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1065 = torch.aten.int_repr %1064 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1066 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1067 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1068 = torch.aten.item %1066 : !torch.vtensor<[],f32> -> !torch.float | |
%1069 = torch.aten.item %1067 : !torch.vtensor<[],si8> -> !torch.int | |
%1070 = torch.aten._make_per_tensor_quantized_tensor %1065, %1068, %1069 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1071 = torch.aten.dequantize.self %1070 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_199 = torch.constant.int 0 | |
%int0_200 = torch.constant.int 0 | |
%int1_201 = torch.constant.int 1 | |
%int1_202 = torch.constant.int 1 | |
%int1_203 = torch.constant.int 1 | |
%int1_204 = torch.constant.int 1 | |
%int0_205 = torch.constant.int 0 | |
%1072 = torch.prim.ListConstruct %int0_199, %int0_200 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1073 = torch.prim.ListConstruct %int1_201, %int1_202 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1074 = torch.prim.ListConstruct %int1_203, %int1_204 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1075 = torch.prim.ListConstruct %int0_205, %int0_205 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_206 = torch.constant.bool false | |
%int1_207 = torch.constant.int 1 | |
%1076 = torch.aten.convolution %1047, %1059, %1071, %1074, %1072, %1073, %false_206, %1075, %int1_207 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[128,128,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%1077 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1078 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_208 = torch.constant.int 12 | |
%1079 = torch.aten.item %1077 : !torch.vtensor<[],f32> -> !torch.float | |
%1080 = torch.aten.item %1078 : !torch.vtensor<[],si8> -> !torch.int | |
%1081 = torch.aten.quantize_per_tensor %1076, %1079, %1080, %int12_208 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1082 = torch.aten.int_repr %1081 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1083 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1084 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1085 = torch.aten.item %1083 : !torch.vtensor<[],f32> -> !torch.float | |
%1086 = torch.aten.item %1084 : !torch.vtensor<[],si8> -> !torch.int | |
%1087 = torch.aten._make_per_tensor_quantized_tensor %1082, %1085, %1086 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1088 = torch.aten.dequantize.self %1087 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1089 = torch.aten.sigmoid %1088 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1090 = torch.aten.mul.Tensor %1088, %1089 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1091 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1092 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_209 = torch.constant.int 12 | |
%1093 = torch.aten.item %1091 : !torch.vtensor<[],f32> -> !torch.float | |
%1094 = torch.aten.item %1092 : !torch.vtensor<[],si8> -> !torch.int | |
%1095 = torch.aten.quantize_per_tensor %1090, %1093, %1094, %int12_209 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1096 = torch.aten.int_repr %1095 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1097 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1098 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1099 = torch.aten.item %1097 : !torch.vtensor<[],f32> -> !torch.float | |
%1100 = torch.aten.item %1098 : !torch.vtensor<[],si8> -> !torch.int | |
%1101 = torch.aten._make_per_tensor_quantized_tensor %1096, %1099, %1100 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1102 = torch.aten.dequantize.self %1101 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1103 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1104 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1105 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1106 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_210 = torch.constant.int 0 | |
%int0_211 = torch.constant.int 0 | |
%1107 = torch.prim.NumToTensor.Scalar %int0_211 : !torch.int -> !torch.vtensor<[1],si64> | |
%1108 = torch.aten.index_select %1104, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1109 = torch.aten.item %1108 : !torch.vtensor<[1],si64> -> !torch.int | |
%1110 = torch.aten.index_select %1105, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1111 = torch.aten.item %1110 : !torch.vtensor<[1],si64> -> !torch.int | |
%1112 = torch.aten.index_select %1103, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1113 = torch.aten.item %1112 : !torch.vtensor<[1],si64> -> !torch.int | |
%1114 = torch.aten.index_select %1106, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1115 = torch.aten.item %1114 : !torch.vtensor<[1],si64> -> !torch.int | |
%1116 = torch.aten.slice.Tensor %1102, %1113, %1109, %1111, %1115 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1117 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1118 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1119 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1120 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_212 = torch.constant.int 0 | |
%int0_213 = torch.constant.int 0 | |
%1121 = torch.prim.NumToTensor.Scalar %int0_213 : !torch.int -> !torch.vtensor<[1],si64> | |
%1122 = torch.aten.index_select %1118, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1123 = torch.aten.item %1122 : !torch.vtensor<[1],si64> -> !torch.int | |
%1124 = torch.aten.index_select %1119, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1125 = torch.aten.item %1124 : !torch.vtensor<[1],si64> -> !torch.int | |
%1126 = torch.aten.index_select %1117, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1127 = torch.aten.item %1126 : !torch.vtensor<[1],si64> -> !torch.int | |
%1128 = torch.aten.index_select %1120, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1129 = torch.aten.item %1128 : !torch.vtensor<[1],si64> -> !torch.int | |
%1130 = torch.aten.slice.Tensor %1102, %1127, %1123, %1125, %1129 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1131 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1132 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_214 = torch.constant.int 12 | |
%1133 = torch.aten.item %1131 : !torch.vtensor<[],f32> -> !torch.float | |
%1134 = torch.aten.item %1132 : !torch.vtensor<[],si8> -> !torch.int | |
%1135 = torch.aten.quantize_per_tensor %30, %1133, %1134, %int12_214 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1136 = torch.aten.int_repr %1135 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%1137 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1138 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1139 = torch.aten.item %1137 : !torch.vtensor<[],f32> -> !torch.float | |
%1140 = torch.aten.item %1138 : !torch.vtensor<[],si8> -> !torch.int | |
%1141 = torch.aten._make_per_tensor_quantized_tensor %1136, %1139, %1140 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1142 = torch.aten.dequantize.self %1141 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%1143 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1144 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_215 = torch.constant.int 12 | |
%1145 = torch.aten.item %1143 : !torch.vtensor<[],f32> -> !torch.float | |
%1146 = torch.aten.item %1144 : !torch.vtensor<[],si8> -> !torch.int | |
%1147 = torch.aten.quantize_per_tensor %31, %1145, %1146, %int12_215 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1148 = torch.aten.int_repr %1147 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%1149 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1150 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1151 = torch.aten.item %1149 : !torch.vtensor<[],f32> -> !torch.float | |
%1152 = torch.aten.item %1150 : !torch.vtensor<[],si8> -> !torch.int | |
%1153 = torch.aten._make_per_tensor_quantized_tensor %1148, %1151, %1152 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1154 = torch.aten.dequantize.self %1153 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_216 = torch.constant.int 1 | |
%int1_217 = torch.constant.int 1 | |
%int1_218 = torch.constant.int 1 | |
%int1_219 = torch.constant.int 1 | |
%int1_220 = torch.constant.int 1 | |
%int1_221 = torch.constant.int 1 | |
%int0_222 = torch.constant.int 0 | |
%1155 = torch.prim.ListConstruct %int1_216, %int1_217 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1156 = torch.prim.ListConstruct %int1_218, %int1_219 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1157 = torch.prim.ListConstruct %int1_220, %int1_221 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1158 = torch.prim.ListConstruct %int0_222, %int0_222 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_223 = torch.constant.bool false | |
%int1_224 = torch.constant.int 1 | |
%1159 = torch.aten.convolution %1130, %1142, %1154, %1157, %1155, %1156, %false_223, %1158, %int1_224 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1160 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1161 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_225 = torch.constant.int 12 | |
%1162 = torch.aten.item %1160 : !torch.vtensor<[],f32> -> !torch.float | |
%1163 = torch.aten.item %1161 : !torch.vtensor<[],si8> -> !torch.int | |
%1164 = torch.aten.quantize_per_tensor %1159, %1162, %1163, %int12_225 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1165 = torch.aten.int_repr %1164 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1166 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1167 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1168 = torch.aten.item %1166 : !torch.vtensor<[],f32> -> !torch.float | |
%1169 = torch.aten.item %1167 : !torch.vtensor<[],si8> -> !torch.int | |
%1170 = torch.aten._make_per_tensor_quantized_tensor %1165, %1168, %1169 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1171 = torch.aten.dequantize.self %1170 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1172 = torch.aten.sigmoid %1171 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1173 = torch.aten.mul.Tensor %1171, %1172 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1174 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1175 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_226 = torch.constant.int 12 | |
%1176 = torch.aten.item %1174 : !torch.vtensor<[],f32> -> !torch.float | |
%1177 = torch.aten.item %1175 : !torch.vtensor<[],si8> -> !torch.int | |
%1178 = torch.aten.quantize_per_tensor %1173, %1176, %1177, %int12_226 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1179 = torch.aten.int_repr %1178 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1180 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1181 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1182 = torch.aten.item %1180 : !torch.vtensor<[],f32> -> !torch.float | |
%1183 = torch.aten.item %1181 : !torch.vtensor<[],si8> -> !torch.int | |
%1184 = torch.aten._make_per_tensor_quantized_tensor %1179, %1182, %1183 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1185 = torch.aten.dequantize.self %1184 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1186 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1187 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_227 = torch.constant.int 12 | |
%1188 = torch.aten.item %1186 : !torch.vtensor<[],f32> -> !torch.float | |
%1189 = torch.aten.item %1187 : !torch.vtensor<[],si8> -> !torch.int | |
%1190 = torch.aten.quantize_per_tensor %32, %1188, %1189, %int12_227 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1191 = torch.aten.int_repr %1190 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%1192 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1193 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1194 = torch.aten.item %1192 : !torch.vtensor<[],f32> -> !torch.float | |
%1195 = torch.aten.item %1193 : !torch.vtensor<[],si8> -> !torch.int | |
%1196 = torch.aten._make_per_tensor_quantized_tensor %1191, %1194, %1195 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1197 = torch.aten.dequantize.self %1196 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%1198 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1199 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_228 = torch.constant.int 12 | |
%1200 = torch.aten.item %1198 : !torch.vtensor<[],f32> -> !torch.float | |
%1201 = torch.aten.item %1199 : !torch.vtensor<[],si8> -> !torch.int | |
%1202 = torch.aten.quantize_per_tensor %33, %1200, %1201, %int12_228 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1203 = torch.aten.int_repr %1202 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%1204 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1205 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1206 = torch.aten.item %1204 : !torch.vtensor<[],f32> -> !torch.float | |
%1207 = torch.aten.item %1205 : !torch.vtensor<[],si8> -> !torch.int | |
%1208 = torch.aten._make_per_tensor_quantized_tensor %1203, %1206, %1207 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1209 = torch.aten.dequantize.self %1208 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_229 = torch.constant.int 1 | |
%int1_230 = torch.constant.int 1 | |
%int1_231 = torch.constant.int 1 | |
%int1_232 = torch.constant.int 1 | |
%int1_233 = torch.constant.int 1 | |
%int1_234 = torch.constant.int 1 | |
%int0_235 = torch.constant.int 0 | |
%1210 = torch.prim.ListConstruct %int1_229, %int1_230 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1211 = torch.prim.ListConstruct %int1_231, %int1_232 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1212 = torch.prim.ListConstruct %int1_233, %int1_234 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1213 = torch.prim.ListConstruct %int0_235, %int0_235 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_236 = torch.constant.bool false | |
%int1_237 = torch.constant.int 1 | |
%1214 = torch.aten.convolution %1185, %1197, %1209, %1212, %1210, %1211, %false_236, %1213, %int1_237 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1215 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1216 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_238 = torch.constant.int 12 | |
%1217 = torch.aten.item %1215 : !torch.vtensor<[],f32> -> !torch.float | |
%1218 = torch.aten.item %1216 : !torch.vtensor<[],si8> -> !torch.int | |
%1219 = torch.aten.quantize_per_tensor %1214, %1217, %1218, %int12_238 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1220 = torch.aten.int_repr %1219 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1221 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1223 = torch.aten.item %1221 : !torch.vtensor<[],f32> -> !torch.float | |
%1224 = torch.aten.item %1222 : !torch.vtensor<[],si8> -> !torch.int | |
%1225 = torch.aten._make_per_tensor_quantized_tensor %1220, %1223, %1224 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1226 = torch.aten.dequantize.self %1225 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1227 = torch.aten.sigmoid %1226 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1228 = torch.aten.mul.Tensor %1226, %1227 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1229 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1230 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_239 = torch.constant.int 12 | |
%1231 = torch.aten.item %1229 : !torch.vtensor<[],f32> -> !torch.float | |
%1232 = torch.aten.item %1230 : !torch.vtensor<[],si8> -> !torch.int | |
%1233 = torch.aten.quantize_per_tensor %1228, %1231, %1232, %int12_239 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1234 = torch.aten.int_repr %1233 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1235 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1236 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1237 = torch.aten.item %1235 : !torch.vtensor<[],f32> -> !torch.float | |
%1238 = torch.aten.item %1236 : !torch.vtensor<[],si8> -> !torch.int | |
%1239 = torch.aten._make_per_tensor_quantized_tensor %1234, %1237, %1238 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1240 = torch.aten.dequantize.self %1239 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%int1_240 = torch.constant.int 1 | |
%1241 = torch.aten.add.Tensor %1130, %1240, %int1_240 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1242 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1243 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_241 = torch.constant.int 12 | |
%1244 = torch.aten.item %1242 : !torch.vtensor<[],f32> -> !torch.float | |
%1245 = torch.aten.item %1243 : !torch.vtensor<[],si8> -> !torch.int | |
%1246 = torch.aten.quantize_per_tensor %1241, %1244, %1245, %int12_241 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1247 = torch.aten.int_repr %1246 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1248 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1249 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1250 = torch.aten.item %1248 : !torch.vtensor<[],f32> -> !torch.float | |
%1251 = torch.aten.item %1249 : !torch.vtensor<[],si8> -> !torch.int | |
%1252 = torch.aten._make_per_tensor_quantized_tensor %1247, %1250, %1251 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1253 = torch.aten.dequantize.self %1252 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1254 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1255 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_242 = torch.constant.int 12 | |
%1256 = torch.aten.item %1254 : !torch.vtensor<[],f32> -> !torch.float | |
%1257 = torch.aten.item %1255 : !torch.vtensor<[],si8> -> !torch.int | |
%1258 = torch.aten.quantize_per_tensor %34, %1256, %1257, %int12_242 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1259 = torch.aten.int_repr %1258 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%1260 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1261 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1262 = torch.aten.item %1260 : !torch.vtensor<[],f32> -> !torch.float | |
%1263 = torch.aten.item %1261 : !torch.vtensor<[],si8> -> !torch.int | |
%1264 = torch.aten._make_per_tensor_quantized_tensor %1259, %1262, %1263 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1265 = torch.aten.dequantize.self %1264 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%1266 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1267 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_243 = torch.constant.int 12 | |
%1268 = torch.aten.item %1266 : !torch.vtensor<[],f32> -> !torch.float | |
%1269 = torch.aten.item %1267 : !torch.vtensor<[],si8> -> !torch.int | |
%1270 = torch.aten.quantize_per_tensor %35, %1268, %1269, %int12_243 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1271 = torch.aten.int_repr %1270 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%1272 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1273 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1274 = torch.aten.item %1272 : !torch.vtensor<[],f32> -> !torch.float | |
%1275 = torch.aten.item %1273 : !torch.vtensor<[],si8> -> !torch.int | |
%1276 = torch.aten._make_per_tensor_quantized_tensor %1271, %1274, %1275 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1277 = torch.aten.dequantize.self %1276 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_244 = torch.constant.int 1 | |
%int1_245 = torch.constant.int 1 | |
%int1_246 = torch.constant.int 1 | |
%int1_247 = torch.constant.int 1 | |
%int1_248 = torch.constant.int 1 | |
%int1_249 = torch.constant.int 1 | |
%int0_250 = torch.constant.int 0 | |
%1278 = torch.prim.ListConstruct %int1_244, %int1_245 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1279 = torch.prim.ListConstruct %int1_246, %int1_247 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1280 = torch.prim.ListConstruct %int1_248, %int1_249 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1281 = torch.prim.ListConstruct %int0_250, %int0_250 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_251 = torch.constant.bool false | |
%int1_252 = torch.constant.int 1 | |
%1282 = torch.aten.convolution %1253, %1265, %1277, %1280, %1278, %1279, %false_251, %1281, %int1_252 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1283 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1284 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_253 = torch.constant.int 12 | |
%1285 = torch.aten.item %1283 : !torch.vtensor<[],f32> -> !torch.float | |
%1286 = torch.aten.item %1284 : !torch.vtensor<[],si8> -> !torch.int | |
%1287 = torch.aten.quantize_per_tensor %1282, %1285, %1286, %int12_253 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1288 = torch.aten.int_repr %1287 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1289 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1290 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1291 = torch.aten.item %1289 : !torch.vtensor<[],f32> -> !torch.float | |
%1292 = torch.aten.item %1290 : !torch.vtensor<[],si8> -> !torch.int | |
%1293 = torch.aten._make_per_tensor_quantized_tensor %1288, %1291, %1292 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1294 = torch.aten.dequantize.self %1293 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1295 = torch.aten.sigmoid %1294 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1296 = torch.aten.mul.Tensor %1294, %1295 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1297 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1298 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_254 = torch.constant.int 12 | |
%1299 = torch.aten.item %1297 : !torch.vtensor<[],f32> -> !torch.float | |
%1300 = torch.aten.item %1298 : !torch.vtensor<[],si8> -> !torch.int | |
%1301 = torch.aten.quantize_per_tensor %1296, %1299, %1300, %int12_254 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1302 = torch.aten.int_repr %1301 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1303 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1304 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1305 = torch.aten.item %1303 : !torch.vtensor<[],f32> -> !torch.float | |
%1306 = torch.aten.item %1304 : !torch.vtensor<[],si8> -> !torch.int | |
%1307 = torch.aten._make_per_tensor_quantized_tensor %1302, %1305, %1306 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1308 = torch.aten.dequantize.self %1307 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1309 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1310 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_255 = torch.constant.int 12 | |
%1311 = torch.aten.item %1309 : !torch.vtensor<[],f32> -> !torch.float | |
%1312 = torch.aten.item %1310 : !torch.vtensor<[],si8> -> !torch.int | |
%1313 = torch.aten.quantize_per_tensor %36, %1311, %1312, %int12_255 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1314 = torch.aten.int_repr %1313 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%1315 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1316 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1317 = torch.aten.item %1315 : !torch.vtensor<[],f32> -> !torch.float | |
%1318 = torch.aten.item %1316 : !torch.vtensor<[],si8> -> !torch.int | |
%1319 = torch.aten._make_per_tensor_quantized_tensor %1314, %1317, %1318 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%1320 = torch.aten.dequantize.self %1319 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%1321 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1322 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_256 = torch.constant.int 12 | |
%1323 = torch.aten.item %1321 : !torch.vtensor<[],f32> -> !torch.float | |
%1324 = torch.aten.item %1322 : !torch.vtensor<[],si8> -> !torch.int | |
%1325 = torch.aten.quantize_per_tensor %37, %1323, %1324, %int12_256 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1326 = torch.aten.int_repr %1325 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%1327 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1328 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1329 = torch.aten.item %1327 : !torch.vtensor<[],f32> -> !torch.float | |
%1330 = torch.aten.item %1328 : !torch.vtensor<[],si8> -> !torch.int | |
%1331 = torch.aten._make_per_tensor_quantized_tensor %1326, %1329, %1330 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%1332 = torch.aten.dequantize.self %1331 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_257 = torch.constant.int 1 | |
%int1_258 = torch.constant.int 1 | |
%int1_259 = torch.constant.int 1 | |
%int1_260 = torch.constant.int 1 | |
%int1_261 = torch.constant.int 1 | |
%int1_262 = torch.constant.int 1 | |
%int0_263 = torch.constant.int 0 | |
%1333 = torch.prim.ListConstruct %int1_257, %int1_258 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1334 = torch.prim.ListConstruct %int1_259, %int1_260 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1335 = torch.prim.ListConstruct %int1_261, %int1_262 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1336 = torch.prim.ListConstruct %int0_263, %int0_263 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_264 = torch.constant.bool false | |
%int1_265 = torch.constant.int 1 | |
%1337 = torch.aten.convolution %1308, %1320, %1332, %1335, %1333, %1334, %false_264, %1336, %int1_265 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1338 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1339 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_266 = torch.constant.int 12 | |
%1340 = torch.aten.item %1338 : !torch.vtensor<[],f32> -> !torch.float | |
%1341 = torch.aten.item %1339 : !torch.vtensor<[],si8> -> !torch.int | |
%1342 = torch.aten.quantize_per_tensor %1337, %1340, %1341, %int12_266 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1343 = torch.aten.int_repr %1342 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1344 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1345 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1346 = torch.aten.item %1344 : !torch.vtensor<[],f32> -> !torch.float | |
%1347 = torch.aten.item %1345 : !torch.vtensor<[],si8> -> !torch.int | |
%1348 = torch.aten._make_per_tensor_quantized_tensor %1343, %1346, %1347 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1349 = torch.aten.dequantize.self %1348 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%1350 = torch.aten.sigmoid %1349 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1351 = torch.aten.mul.Tensor %1349, %1350 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%1352 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1353 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_267 = torch.constant.int 12 | |
%1354 = torch.aten.item %1352 : !torch.vtensor<[],f32> -> !torch.float | |
%1355 = torch.aten.item %1353 : !torch.vtensor<[],si8> -> !torch.int | |
%1356 = torch.aten.quantize_per_tensor %1351, %1354, %1355, %int12_267 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1357 = torch.aten.int_repr %1356 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%1358 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1359 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1360 = torch.aten.item %1358 : !torch.vtensor<[],f32> -> !torch.float | |
%1361 = torch.aten.item %1359 : !torch.vtensor<[],si8> -> !torch.int | |
%1362 = torch.aten._make_per_tensor_quantized_tensor %1357, %1360, %1361 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%1363 = torch.aten.dequantize.self %1362 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%int1_268 = torch.constant.int 1 | |
%1364 = torch.aten.add.Tensor %1253, %1363, %int1_268 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%1365 = torch.prim.ListConstruct %1116, %1130, %1253, %1364 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>) -> !torch.list<vtensor> | |
%int1_269 = torch.constant.int 1 | |
%1366 = torch.aten.cat %1365, %int1_269 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,256,40,40],f32> | |
%1367 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1368 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_270 = torch.constant.int 12 | |
%1369 = torch.aten.item %1367 : !torch.vtensor<[],f32> -> !torch.float | |
%1370 = torch.aten.item %1368 : !torch.vtensor<[],si8> -> !torch.int | |
%1371 = torch.aten.quantize_per_tensor %1366, %1369, %1370, %int12_270 : !torch.vtensor<[1,256,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,40,40],!torch.qint8> | |
%1372 = torch.aten.int_repr %1371 : !torch.vtensor<[1,256,40,40],!torch.qint8> -> !torch.vtensor<[1,256,40,40],si8> | |
%1373 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1374 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1375 = torch.aten.item %1373 : !torch.vtensor<[],f32> -> !torch.float | |
%1376 = torch.aten.item %1374 : !torch.vtensor<[],si8> -> !torch.int | |
%1377 = torch.aten._make_per_tensor_quantized_tensor %1372, %1375, %1376 : !torch.vtensor<[1,256,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,40,40],!torch.qint8> | |
%1378 = torch.aten.dequantize.self %1377 : !torch.vtensor<[1,256,40,40],!torch.qint8> -> !torch.vtensor<[1,256,40,40],f32> | |
%1379 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1380 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_271 = torch.constant.int 12 | |
%1381 = torch.aten.item %1379 : !torch.vtensor<[],f32> -> !torch.float | |
%1382 = torch.aten.item %1380 : !torch.vtensor<[],si8> -> !torch.int | |
%1383 = torch.aten.quantize_per_tensor %38, %1381, %1382, %int12_271 : !torch.vtensor<[128,256,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
%1384 = torch.aten.int_repr %1383 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],si8> | |
%1385 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1386 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1387 = torch.aten.item %1385 : !torch.vtensor<[],f32> -> !torch.float | |
%1388 = torch.aten.item %1386 : !torch.vtensor<[],si8> -> !torch.int | |
%1389 = torch.aten._make_per_tensor_quantized_tensor %1384, %1387, %1388 : !torch.vtensor<[128,256,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
%1390 = torch.aten.dequantize.self %1389 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],f32> | |
%1391 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1392 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_272 = torch.constant.int 12 | |
%1393 = torch.aten.item %1391 : !torch.vtensor<[],f32> -> !torch.float | |
%1394 = torch.aten.item %1392 : !torch.vtensor<[],si8> -> !torch.int | |
%1395 = torch.aten.quantize_per_tensor %39, %1393, %1394, %int12_272 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1396 = torch.aten.int_repr %1395 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1397 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1398 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1399 = torch.aten.item %1397 : !torch.vtensor<[],f32> -> !torch.float | |
%1400 = torch.aten.item %1398 : !torch.vtensor<[],si8> -> !torch.int | |
%1401 = torch.aten._make_per_tensor_quantized_tensor %1396, %1399, %1400 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1402 = torch.aten.dequantize.self %1401 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_273 = torch.constant.int 0 | |
%int0_274 = torch.constant.int 0 | |
%int1_275 = torch.constant.int 1 | |
%int1_276 = torch.constant.int 1 | |
%int1_277 = torch.constant.int 1 | |
%int1_278 = torch.constant.int 1 | |
%int0_279 = torch.constant.int 0 | |
%1403 = torch.prim.ListConstruct %int0_273, %int0_274 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1404 = torch.prim.ListConstruct %int1_275, %int1_276 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1405 = torch.prim.ListConstruct %int1_277, %int1_278 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1406 = torch.prim.ListConstruct %int0_279, %int0_279 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_280 = torch.constant.bool false | |
%int1_281 = torch.constant.int 1 | |
%1407 = torch.aten.convolution %1378, %1390, %1402, %1405, %1403, %1404, %false_280, %1406, %int1_281 : !torch.vtensor<[1,256,40,40],f32>, !torch.vtensor<[128,256,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%1408 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1409 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_282 = torch.constant.int 12 | |
%1410 = torch.aten.item %1408 : !torch.vtensor<[],f32> -> !torch.float | |
%1411 = torch.aten.item %1409 : !torch.vtensor<[],si8> -> !torch.int | |
%1412 = torch.aten.quantize_per_tensor %1407, %1410, %1411, %int12_282 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1413 = torch.aten.int_repr %1412 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1414 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1415 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1416 = torch.aten.item %1414 : !torch.vtensor<[],f32> -> !torch.float | |
%1417 = torch.aten.item %1415 : !torch.vtensor<[],si8> -> !torch.int | |
%1418 = torch.aten._make_per_tensor_quantized_tensor %1413, %1416, %1417 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1419 = torch.aten.dequantize.self %1418 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1420 = torch.aten.sigmoid %1419 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1421 = torch.aten.mul.Tensor %1419, %1420 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1422 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1423 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_283 = torch.constant.int 12 | |
%1424 = torch.aten.item %1422 : !torch.vtensor<[],f32> -> !torch.float | |
%1425 = torch.aten.item %1423 : !torch.vtensor<[],si8> -> !torch.int | |
%1426 = torch.aten.quantize_per_tensor %1421, %1424, %1425, %int12_283 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1427 = torch.aten.int_repr %1426 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1428 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1429 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1430 = torch.aten.item %1428 : !torch.vtensor<[],f32> -> !torch.float | |
%1431 = torch.aten.item %1429 : !torch.vtensor<[],si8> -> !torch.int | |
%1432 = torch.aten._make_per_tensor_quantized_tensor %1427, %1430, %1431 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1433 = torch.aten.dequantize.self %1432 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1434 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1435 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_284 = torch.constant.int 12 | |
%1436 = torch.aten.item %1434 : !torch.vtensor<[],f32> -> !torch.float | |
%1437 = torch.aten.item %1435 : !torch.vtensor<[],si8> -> !torch.int | |
%1438 = torch.aten.quantize_per_tensor %40, %1436, %1437, %int12_284 : !torch.vtensor<[256,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,128,3,3],!torch.qint8> | |
%1439 = torch.aten.int_repr %1438 : !torch.vtensor<[256,128,3,3],!torch.qint8> -> !torch.vtensor<[256,128,3,3],si8> | |
%1440 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1441 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1442 = torch.aten.item %1440 : !torch.vtensor<[],f32> -> !torch.float | |
%1443 = torch.aten.item %1441 : !torch.vtensor<[],si8> -> !torch.int | |
%1444 = torch.aten._make_per_tensor_quantized_tensor %1439, %1442, %1443 : !torch.vtensor<[256,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,128,3,3],!torch.qint8> | |
%1445 = torch.aten.dequantize.self %1444 : !torch.vtensor<[256,128,3,3],!torch.qint8> -> !torch.vtensor<[256,128,3,3],f32> | |
%1446 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1447 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_285 = torch.constant.int 12 | |
%1448 = torch.aten.item %1446 : !torch.vtensor<[],f32> -> !torch.float | |
%1449 = torch.aten.item %1447 : !torch.vtensor<[],si8> -> !torch.int | |
%1450 = torch.aten.quantize_per_tensor %41, %1448, %1449, %int12_285 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1451 = torch.aten.int_repr %1450 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
%1452 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1453 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1454 = torch.aten.item %1452 : !torch.vtensor<[],f32> -> !torch.float | |
%1455 = torch.aten.item %1453 : !torch.vtensor<[],si8> -> !torch.int | |
%1456 = torch.aten._make_per_tensor_quantized_tensor %1451, %1454, %1455 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1457 = torch.aten.dequantize.self %1456 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
%int1_286 = torch.constant.int 1 | |
%int1_287 = torch.constant.int 1 | |
%int1_288 = torch.constant.int 1 | |
%int1_289 = torch.constant.int 1 | |
%int2_290 = torch.constant.int 2 | |
%int2_291 = torch.constant.int 2 | |
%int0_292 = torch.constant.int 0 | |
%1458 = torch.prim.ListConstruct %int1_286, %int1_287 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1459 = torch.prim.ListConstruct %int1_288, %int1_289 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1460 = torch.prim.ListConstruct %int2_290, %int2_291 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1461 = torch.prim.ListConstruct %int0_292, %int0_292 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_293 = torch.constant.bool false | |
%int1_294 = torch.constant.int 1 | |
%1462 = torch.aten.convolution %1433, %1445, %1457, %1460, %1458, %1459, %false_293, %1461, %int1_294 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[256,128,3,3],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
%1463 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1464 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_295 = torch.constant.int 12 | |
%1465 = torch.aten.item %1463 : !torch.vtensor<[],f32> -> !torch.float | |
%1466 = torch.aten.item %1464 : !torch.vtensor<[],si8> -> !torch.int | |
%1467 = torch.aten.quantize_per_tensor %1462, %1465, %1466, %int12_295 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1468 = torch.aten.int_repr %1467 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1469 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1470 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1471 = torch.aten.item %1469 : !torch.vtensor<[],f32> -> !torch.float | |
%1472 = torch.aten.item %1470 : !torch.vtensor<[],si8> -> !torch.int | |
%1473 = torch.aten._make_per_tensor_quantized_tensor %1468, %1471, %1472 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1474 = torch.aten.dequantize.self %1473 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1475 = torch.aten.sigmoid %1474 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1476 = torch.aten.mul.Tensor %1474, %1475 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1477 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1478 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_296 = torch.constant.int 12 | |
%1479 = torch.aten.item %1477 : !torch.vtensor<[],f32> -> !torch.float | |
%1480 = torch.aten.item %1478 : !torch.vtensor<[],si8> -> !torch.int | |
%1481 = torch.aten.quantize_per_tensor %1476, %1479, %1480, %int12_296 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1482 = torch.aten.int_repr %1481 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1483 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1484 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1485 = torch.aten.item %1483 : !torch.vtensor<[],f32> -> !torch.float | |
%1486 = torch.aten.item %1484 : !torch.vtensor<[],si8> -> !torch.int | |
%1487 = torch.aten._make_per_tensor_quantized_tensor %1482, %1485, %1486 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1488 = torch.aten.dequantize.self %1487 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1489 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1490 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_297 = torch.constant.int 12 | |
%1491 = torch.aten.item %1489 : !torch.vtensor<[],f32> -> !torch.float | |
%1492 = torch.aten.item %1490 : !torch.vtensor<[],si8> -> !torch.int | |
%1493 = torch.aten.quantize_per_tensor %42, %1491, %1492, %int12_297 : !torch.vtensor<[256,256,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,256,1,1],!torch.qint8> | |
%1494 = torch.aten.int_repr %1493 : !torch.vtensor<[256,256,1,1],!torch.qint8> -> !torch.vtensor<[256,256,1,1],si8> | |
%1495 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1496 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1497 = torch.aten.item %1495 : !torch.vtensor<[],f32> -> !torch.float | |
%1498 = torch.aten.item %1496 : !torch.vtensor<[],si8> -> !torch.int | |
%1499 = torch.aten._make_per_tensor_quantized_tensor %1494, %1497, %1498 : !torch.vtensor<[256,256,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,256,1,1],!torch.qint8> | |
%1500 = torch.aten.dequantize.self %1499 : !torch.vtensor<[256,256,1,1],!torch.qint8> -> !torch.vtensor<[256,256,1,1],f32> | |
%1501 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1502 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_298 = torch.constant.int 12 | |
%1503 = torch.aten.item %1501 : !torch.vtensor<[],f32> -> !torch.float | |
%1504 = torch.aten.item %1502 : !torch.vtensor<[],si8> -> !torch.int | |
%1505 = torch.aten.quantize_per_tensor %43, %1503, %1504, %int12_298 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1506 = torch.aten.int_repr %1505 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
%1507 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1508 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1509 = torch.aten.item %1507 : !torch.vtensor<[],f32> -> !torch.float | |
%1510 = torch.aten.item %1508 : !torch.vtensor<[],si8> -> !torch.int | |
%1511 = torch.aten._make_per_tensor_quantized_tensor %1506, %1509, %1510 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1512 = torch.aten.dequantize.self %1511 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
%int0_299 = torch.constant.int 0 | |
%int0_300 = torch.constant.int 0 | |
%int1_301 = torch.constant.int 1 | |
%int1_302 = torch.constant.int 1 | |
%int1_303 = torch.constant.int 1 | |
%int1_304 = torch.constant.int 1 | |
%int0_305 = torch.constant.int 0 | |
%1513 = torch.prim.ListConstruct %int0_299, %int0_300 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1514 = torch.prim.ListConstruct %int1_301, %int1_302 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1515 = torch.prim.ListConstruct %int1_303, %int1_304 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1516 = torch.prim.ListConstruct %int0_305, %int0_305 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_306 = torch.constant.bool false | |
%int1_307 = torch.constant.int 1 | |
%1517 = torch.aten.convolution %1488, %1500, %1512, %1515, %1513, %1514, %false_306, %1516, %int1_307 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[256,256,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
%1518 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1519 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_308 = torch.constant.int 12 | |
%1520 = torch.aten.item %1518 : !torch.vtensor<[],f32> -> !torch.float | |
%1521 = torch.aten.item %1519 : !torch.vtensor<[],si8> -> !torch.int | |
%1522 = torch.aten.quantize_per_tensor %1517, %1520, %1521, %int12_308 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1523 = torch.aten.int_repr %1522 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1524 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1525 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1526 = torch.aten.item %1524 : !torch.vtensor<[],f32> -> !torch.float | |
%1527 = torch.aten.item %1525 : !torch.vtensor<[],si8> -> !torch.int | |
%1528 = torch.aten._make_per_tensor_quantized_tensor %1523, %1526, %1527 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1529 = torch.aten.dequantize.self %1528 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1530 = torch.aten.sigmoid %1529 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1531 = torch.aten.mul.Tensor %1529, %1530 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1532 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1533 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_309 = torch.constant.int 12 | |
%1534 = torch.aten.item %1532 : !torch.vtensor<[],f32> -> !torch.float | |
%1535 = torch.aten.item %1533 : !torch.vtensor<[],si8> -> !torch.int | |
%1536 = torch.aten.quantize_per_tensor %1531, %1534, %1535, %int12_309 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1537 = torch.aten.int_repr %1536 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1538 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1539 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1540 = torch.aten.item %1538 : !torch.vtensor<[],f32> -> !torch.float | |
%1541 = torch.aten.item %1539 : !torch.vtensor<[],si8> -> !torch.int | |
%1542 = torch.aten._make_per_tensor_quantized_tensor %1537, %1540, %1541 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1543 = torch.aten.dequantize.self %1542 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1544 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1545 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1546 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1547 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_310 = torch.constant.int 0 | |
%int0_311 = torch.constant.int 0 | |
%1548 = torch.prim.NumToTensor.Scalar %int0_311 : !torch.int -> !torch.vtensor<[1],si64> | |
%1549 = torch.aten.index_select %1545, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1550 = torch.aten.item %1549 : !torch.vtensor<[1],si64> -> !torch.int | |
%1551 = torch.aten.index_select %1546, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1552 = torch.aten.item %1551 : !torch.vtensor<[1],si64> -> !torch.int | |
%1553 = torch.aten.index_select %1544, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1554 = torch.aten.item %1553 : !torch.vtensor<[1],si64> -> !torch.int | |
%1555 = torch.aten.index_select %1547, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1556 = torch.aten.item %1555 : !torch.vtensor<[1],si64> -> !torch.int | |
%1557 = torch.aten.slice.Tensor %1543, %1554, %1550, %1552, %1556 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%1558 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1559 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1560 = torch.vtensor.literal(dense<256> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1561 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_312 = torch.constant.int 0 | |
%int0_313 = torch.constant.int 0 | |
%1562 = torch.prim.NumToTensor.Scalar %int0_313 : !torch.int -> !torch.vtensor<[1],si64> | |
%1563 = torch.aten.index_select %1559, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1564 = torch.aten.item %1563 : !torch.vtensor<[1],si64> -> !torch.int | |
%1565 = torch.aten.index_select %1560, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1566 = torch.aten.item %1565 : !torch.vtensor<[1],si64> -> !torch.int | |
%1567 = torch.aten.index_select %1558, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1568 = torch.aten.item %1567 : !torch.vtensor<[1],si64> -> !torch.int | |
%1569 = torch.aten.index_select %1561, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1570 = torch.aten.item %1569 : !torch.vtensor<[1],si64> -> !torch.int | |
%1571 = torch.aten.slice.Tensor %1543, %1568, %1564, %1566, %1570 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%1572 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1573 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_314 = torch.constant.int 12 | |
%1574 = torch.aten.item %1572 : !torch.vtensor<[],f32> -> !torch.float | |
%1575 = torch.aten.item %1573 : !torch.vtensor<[],si8> -> !torch.int | |
%1576 = torch.aten.quantize_per_tensor %44, %1574, %1575, %int12_314 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%1577 = torch.aten.int_repr %1576 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
%1578 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1579 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1580 = torch.aten.item %1578 : !torch.vtensor<[],f32> -> !torch.float | |
%1581 = torch.aten.item %1579 : !torch.vtensor<[],si8> -> !torch.int | |
%1582 = torch.aten._make_per_tensor_quantized_tensor %1577, %1580, %1581 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%1583 = torch.aten.dequantize.self %1582 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
%1584 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1585 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_315 = torch.constant.int 12 | |
%1586 = torch.aten.item %1584 : !torch.vtensor<[],f32> -> !torch.float | |
%1587 = torch.aten.item %1585 : !torch.vtensor<[],si8> -> !torch.int | |
%1588 = torch.aten.quantize_per_tensor %45, %1586, %1587, %int12_315 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1589 = torch.aten.int_repr %1588 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1590 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1591 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1592 = torch.aten.item %1590 : !torch.vtensor<[],f32> -> !torch.float | |
%1593 = torch.aten.item %1591 : !torch.vtensor<[],si8> -> !torch.int | |
%1594 = torch.aten._make_per_tensor_quantized_tensor %1589, %1592, %1593 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1595 = torch.aten.dequantize.self %1594 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int1_316 = torch.constant.int 1 | |
%int1_317 = torch.constant.int 1 | |
%int1_318 = torch.constant.int 1 | |
%int1_319 = torch.constant.int 1 | |
%int1_320 = torch.constant.int 1 | |
%int1_321 = torch.constant.int 1 | |
%int0_322 = torch.constant.int 0 | |
%1596 = torch.prim.ListConstruct %int1_316, %int1_317 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1597 = torch.prim.ListConstruct %int1_318, %int1_319 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1598 = torch.prim.ListConstruct %int1_320, %int1_321 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1599 = torch.prim.ListConstruct %int0_322, %int0_322 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_323 = torch.constant.bool false | |
%int1_324 = torch.constant.int 1 | |
%1600 = torch.aten.convolution %1571, %1583, %1595, %1598, %1596, %1597, %false_323, %1599, %int1_324 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%1601 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1602 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_325 = torch.constant.int 12 | |
%1603 = torch.aten.item %1601 : !torch.vtensor<[],f32> -> !torch.float | |
%1604 = torch.aten.item %1602 : !torch.vtensor<[],si8> -> !torch.int | |
%1605 = torch.aten.quantize_per_tensor %1600, %1603, %1604, %int12_325 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1606 = torch.aten.int_repr %1605 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1607 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1608 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1609 = torch.aten.item %1607 : !torch.vtensor<[],f32> -> !torch.float | |
%1610 = torch.aten.item %1608 : !torch.vtensor<[],si8> -> !torch.int | |
%1611 = torch.aten._make_per_tensor_quantized_tensor %1606, %1609, %1610 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1612 = torch.aten.dequantize.self %1611 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%1613 = torch.aten.sigmoid %1612 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%1614 = torch.aten.mul.Tensor %1612, %1613 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%1615 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1616 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_326 = torch.constant.int 12 | |
%1617 = torch.aten.item %1615 : !torch.vtensor<[],f32> -> !torch.float | |
%1618 = torch.aten.item %1616 : !torch.vtensor<[],si8> -> !torch.int | |
%1619 = torch.aten.quantize_per_tensor %1614, %1617, %1618, %int12_326 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1620 = torch.aten.int_repr %1619 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1621 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1622 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1623 = torch.aten.item %1621 : !torch.vtensor<[],f32> -> !torch.float | |
%1624 = torch.aten.item %1622 : !torch.vtensor<[],si8> -> !torch.int | |
%1625 = torch.aten._make_per_tensor_quantized_tensor %1620, %1623, %1624 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1626 = torch.aten.dequantize.self %1625 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%1627 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1628 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_327 = torch.constant.int 12 | |
%1629 = torch.aten.item %1627 : !torch.vtensor<[],f32> -> !torch.float | |
%1630 = torch.aten.item %1628 : !torch.vtensor<[],si8> -> !torch.int | |
%1631 = torch.aten.quantize_per_tensor %46, %1629, %1630, %int12_327 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%1632 = torch.aten.int_repr %1631 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
%1633 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1634 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1635 = torch.aten.item %1633 : !torch.vtensor<[],f32> -> !torch.float | |
%1636 = torch.aten.item %1634 : !torch.vtensor<[],si8> -> !torch.int | |
%1637 = torch.aten._make_per_tensor_quantized_tensor %1632, %1635, %1636 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%1638 = torch.aten.dequantize.self %1637 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
%1639 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1640 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_328 = torch.constant.int 12 | |
%1641 = torch.aten.item %1639 : !torch.vtensor<[],f32> -> !torch.float | |
%1642 = torch.aten.item %1640 : !torch.vtensor<[],si8> -> !torch.int | |
%1643 = torch.aten.quantize_per_tensor %47, %1641, %1642, %int12_328 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1644 = torch.aten.int_repr %1643 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1645 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1646 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1647 = torch.aten.item %1645 : !torch.vtensor<[],f32> -> !torch.float | |
%1648 = torch.aten.item %1646 : !torch.vtensor<[],si8> -> !torch.int | |
%1649 = torch.aten._make_per_tensor_quantized_tensor %1644, %1647, %1648 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1650 = torch.aten.dequantize.self %1649 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int1_329 = torch.constant.int 1 | |
%int1_330 = torch.constant.int 1 | |
%int1_331 = torch.constant.int 1 | |
%int1_332 = torch.constant.int 1 | |
%int1_333 = torch.constant.int 1 | |
%int1_334 = torch.constant.int 1 | |
%int0_335 = torch.constant.int 0 | |
%1651 = torch.prim.ListConstruct %int1_329, %int1_330 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1652 = torch.prim.ListConstruct %int1_331, %int1_332 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1653 = torch.prim.ListConstruct %int1_333, %int1_334 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1654 = torch.prim.ListConstruct %int0_335, %int0_335 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_336 = torch.constant.bool false | |
%int1_337 = torch.constant.int 1 | |
%1655 = torch.aten.convolution %1626, %1638, %1650, %1653, %1651, %1652, %false_336, %1654, %int1_337 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%1656 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1657 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_338 = torch.constant.int 12 | |
%1658 = torch.aten.item %1656 : !torch.vtensor<[],f32> -> !torch.float | |
%1659 = torch.aten.item %1657 : !torch.vtensor<[],si8> -> !torch.int | |
%1660 = torch.aten.quantize_per_tensor %1655, %1658, %1659, %int12_338 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1661 = torch.aten.int_repr %1660 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1662 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1663 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1664 = torch.aten.item %1662 : !torch.vtensor<[],f32> -> !torch.float | |
%1665 = torch.aten.item %1663 : !torch.vtensor<[],si8> -> !torch.int | |
%1666 = torch.aten._make_per_tensor_quantized_tensor %1661, %1664, %1665 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1667 = torch.aten.dequantize.self %1666 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%1668 = torch.aten.sigmoid %1667 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%1669 = torch.aten.mul.Tensor %1667, %1668 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%1670 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1671 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_339 = torch.constant.int 12 | |
%1672 = torch.aten.item %1670 : !torch.vtensor<[],f32> -> !torch.float | |
%1673 = torch.aten.item %1671 : !torch.vtensor<[],si8> -> !torch.int | |
%1674 = torch.aten.quantize_per_tensor %1669, %1672, %1673, %int12_339 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1675 = torch.aten.int_repr %1674 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1676 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1677 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1678 = torch.aten.item %1676 : !torch.vtensor<[],f32> -> !torch.float | |
%1679 = torch.aten.item %1677 : !torch.vtensor<[],si8> -> !torch.int | |
%1680 = torch.aten._make_per_tensor_quantized_tensor %1675, %1678, %1679 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1681 = torch.aten.dequantize.self %1680 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%int1_340 = torch.constant.int 1 | |
%1682 = torch.aten.add.Tensor %1571, %1681, %int1_340 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%1683 = torch.prim.ListConstruct %1557, %1571, %1682 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>) -> !torch.list<vtensor> | |
%int1_341 = torch.constant.int 1 | |
%1684 = torch.aten.cat %1683, %int1_341 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,384,20,20],f32> | |
%1685 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1686 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_342 = torch.constant.int 12 | |
%1687 = torch.aten.item %1685 : !torch.vtensor<[],f32> -> !torch.float | |
%1688 = torch.aten.item %1686 : !torch.vtensor<[],si8> -> !torch.int | |
%1689 = torch.aten.quantize_per_tensor %1684, %1687, %1688, %int12_342 : !torch.vtensor<[1,384,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
%1690 = torch.aten.int_repr %1689 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],si8> | |
%1691 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1692 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1693 = torch.aten.item %1691 : !torch.vtensor<[],f32> -> !torch.float | |
%1694 = torch.aten.item %1692 : !torch.vtensor<[],si8> -> !torch.int | |
%1695 = torch.aten._make_per_tensor_quantized_tensor %1690, %1693, %1694 : !torch.vtensor<[1,384,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
%1696 = torch.aten.dequantize.self %1695 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],f32> | |
%1697 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1698 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_343 = torch.constant.int 12 | |
%1699 = torch.aten.item %1697 : !torch.vtensor<[],f32> -> !torch.float | |
%1700 = torch.aten.item %1698 : !torch.vtensor<[],si8> -> !torch.int | |
%1701 = torch.aten.quantize_per_tensor %48, %1699, %1700, %int12_343 : !torch.vtensor<[256,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
%1702 = torch.aten.int_repr %1701 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],si8> | |
%1703 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1704 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1705 = torch.aten.item %1703 : !torch.vtensor<[],f32> -> !torch.float | |
%1706 = torch.aten.item %1704 : !torch.vtensor<[],si8> -> !torch.int | |
%1707 = torch.aten._make_per_tensor_quantized_tensor %1702, %1705, %1706 : !torch.vtensor<[256,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
%1708 = torch.aten.dequantize.self %1707 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],f32> | |
%1709 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1710 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_344 = torch.constant.int 12 | |
%1711 = torch.aten.item %1709 : !torch.vtensor<[],f32> -> !torch.float | |
%1712 = torch.aten.item %1710 : !torch.vtensor<[],si8> -> !torch.int | |
%1713 = torch.aten.quantize_per_tensor %49, %1711, %1712, %int12_344 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1714 = torch.aten.int_repr %1713 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
%1715 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1716 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1717 = torch.aten.item %1715 : !torch.vtensor<[],f32> -> !torch.float | |
%1718 = torch.aten.item %1716 : !torch.vtensor<[],si8> -> !torch.int | |
%1719 = torch.aten._make_per_tensor_quantized_tensor %1714, %1717, %1718 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1720 = torch.aten.dequantize.self %1719 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
%int0_345 = torch.constant.int 0 | |
%int0_346 = torch.constant.int 0 | |
%int1_347 = torch.constant.int 1 | |
%int1_348 = torch.constant.int 1 | |
%int1_349 = torch.constant.int 1 | |
%int1_350 = torch.constant.int 1 | |
%int0_351 = torch.constant.int 0 | |
%1721 = torch.prim.ListConstruct %int0_345, %int0_346 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1722 = torch.prim.ListConstruct %int1_347, %int1_348 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1723 = torch.prim.ListConstruct %int1_349, %int1_350 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1724 = torch.prim.ListConstruct %int0_351, %int0_351 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_352 = torch.constant.bool false | |
%int1_353 = torch.constant.int 1 | |
%1725 = torch.aten.convolution %1696, %1708, %1720, %1723, %1721, %1722, %false_352, %1724, %int1_353 : !torch.vtensor<[1,384,20,20],f32>, !torch.vtensor<[256,384,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
%1726 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1727 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_354 = torch.constant.int 12 | |
%1728 = torch.aten.item %1726 : !torch.vtensor<[],f32> -> !torch.float | |
%1729 = torch.aten.item %1727 : !torch.vtensor<[],si8> -> !torch.int | |
%1730 = torch.aten.quantize_per_tensor %1725, %1728, %1729, %int12_354 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1731 = torch.aten.int_repr %1730 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1732 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1733 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1734 = torch.aten.item %1732 : !torch.vtensor<[],f32> -> !torch.float | |
%1735 = torch.aten.item %1733 : !torch.vtensor<[],si8> -> !torch.int | |
%1736 = torch.aten._make_per_tensor_quantized_tensor %1731, %1734, %1735 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1737 = torch.aten.dequantize.self %1736 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1738 = torch.aten.sigmoid %1737 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1739 = torch.aten.mul.Tensor %1737, %1738 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1740 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1741 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_355 = torch.constant.int 12 | |
%1742 = torch.aten.item %1740 : !torch.vtensor<[],f32> -> !torch.float | |
%1743 = torch.aten.item %1741 : !torch.vtensor<[],si8> -> !torch.int | |
%1744 = torch.aten.quantize_per_tensor %1739, %1742, %1743, %int12_355 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1745 = torch.aten.int_repr %1744 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1746 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1747 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1748 = torch.aten.item %1746 : !torch.vtensor<[],f32> -> !torch.float | |
%1749 = torch.aten.item %1747 : !torch.vtensor<[],si8> -> !torch.int | |
%1750 = torch.aten._make_per_tensor_quantized_tensor %1745, %1748, %1749 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1751 = torch.aten.dequantize.self %1750 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1752 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1753 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_356 = torch.constant.int 12 | |
%1754 = torch.aten.item %1752 : !torch.vtensor<[],f32> -> !torch.float | |
%1755 = torch.aten.item %1753 : !torch.vtensor<[],si8> -> !torch.int | |
%1756 = torch.aten.quantize_per_tensor %50, %1754, %1755, %int12_356 : !torch.vtensor<[128,256,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
%1757 = torch.aten.int_repr %1756 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],si8> | |
%1758 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1759 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1760 = torch.aten.item %1758 : !torch.vtensor<[],f32> -> !torch.float | |
%1761 = torch.aten.item %1759 : !torch.vtensor<[],si8> -> !torch.int | |
%1762 = torch.aten._make_per_tensor_quantized_tensor %1757, %1760, %1761 : !torch.vtensor<[128,256,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
%1763 = torch.aten.dequantize.self %1762 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],f32> | |
%1764 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1765 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_357 = torch.constant.int 12 | |
%1766 = torch.aten.item %1764 : !torch.vtensor<[],f32> -> !torch.float | |
%1767 = torch.aten.item %1765 : !torch.vtensor<[],si8> -> !torch.int | |
%1768 = torch.aten.quantize_per_tensor %51, %1766, %1767, %int12_357 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1769 = torch.aten.int_repr %1768 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1770 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1771 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1772 = torch.aten.item %1770 : !torch.vtensor<[],f32> -> !torch.float | |
%1773 = torch.aten.item %1771 : !torch.vtensor<[],si8> -> !torch.int | |
%1774 = torch.aten._make_per_tensor_quantized_tensor %1769, %1772, %1773 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1775 = torch.aten.dequantize.self %1774 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_358 = torch.constant.int 0 | |
%int0_359 = torch.constant.int 0 | |
%int1_360 = torch.constant.int 1 | |
%int1_361 = torch.constant.int 1 | |
%int1_362 = torch.constant.int 1 | |
%int1_363 = torch.constant.int 1 | |
%int0_364 = torch.constant.int 0 | |
%1776 = torch.prim.ListConstruct %int0_358, %int0_359 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1777 = torch.prim.ListConstruct %int1_360, %int1_361 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1778 = torch.prim.ListConstruct %int1_362, %int1_363 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1779 = torch.prim.ListConstruct %int0_364, %int0_364 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_365 = torch.constant.bool false | |
%int1_366 = torch.constant.int 1 | |
%1780 = torch.aten.convolution %1751, %1763, %1775, %1778, %1776, %1777, %false_365, %1779, %int1_366 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[128,256,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%1781 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1782 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_367 = torch.constant.int 12 | |
%1783 = torch.aten.item %1781 : !torch.vtensor<[],f32> -> !torch.float | |
%1784 = torch.aten.item %1782 : !torch.vtensor<[],si8> -> !torch.int | |
%1785 = torch.aten.quantize_per_tensor %1780, %1783, %1784, %int12_367 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1786 = torch.aten.int_repr %1785 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1787 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1788 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1789 = torch.aten.item %1787 : !torch.vtensor<[],f32> -> !torch.float | |
%1790 = torch.aten.item %1788 : !torch.vtensor<[],si8> -> !torch.int | |
%1791 = torch.aten._make_per_tensor_quantized_tensor %1786, %1789, %1790 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1792 = torch.aten.dequantize.self %1791 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%1793 = torch.aten.sigmoid %1792 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%1794 = torch.aten.mul.Tensor %1792, %1793 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%1795 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1796 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_368 = torch.constant.int 12 | |
%1797 = torch.aten.item %1795 : !torch.vtensor<[],f32> -> !torch.float | |
%1798 = torch.aten.item %1796 : !torch.vtensor<[],si8> -> !torch.int | |
%1799 = torch.aten.quantize_per_tensor %1794, %1797, %1798, %int12_368 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1800 = torch.aten.int_repr %1799 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1801 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1802 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1803 = torch.aten.item %1801 : !torch.vtensor<[],f32> -> !torch.float | |
%1804 = torch.aten.item %1802 : !torch.vtensor<[],si8> -> !torch.int | |
%1805 = torch.aten._make_per_tensor_quantized_tensor %1800, %1803, %1804 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1806 = torch.aten.dequantize.self %1805 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%int5 = torch.constant.int 5 | |
%int5_369 = torch.constant.int 5 | |
%1807 = torch.prim.ListConstruct %int5, %int5_369 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int2_370 = torch.constant.int 2 | |
%int2_371 = torch.constant.int 2 | |
%1808 = torch.prim.ListConstruct %int2_370, %int2_371 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int1_372 = torch.constant.int 1 | |
%int1_373 = torch.constant.int 1 | |
%1809 = torch.prim.ListConstruct %int1_372, %int1_373 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int1_374 = torch.constant.int 1 | |
%int1_375 = torch.constant.int 1 | |
%1810 = torch.prim.ListConstruct %int1_374, %int1_375 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_376 = torch.constant.bool false | |
%1811 = torch.aten.max_pool2d %1806, %1807, %1809, %1808, %1810, %false_376 : !torch.vtensor<[1,128,20,20],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[1,128,20,20],f32> | |
%1812 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1813 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_377 = torch.constant.int 12 | |
%1814 = torch.aten.item %1812 : !torch.vtensor<[],f32> -> !torch.float | |
%1815 = torch.aten.item %1813 : !torch.vtensor<[],si8> -> !torch.int | |
%1816 = torch.aten.quantize_per_tensor %1811, %1814, %1815, %int12_377 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1817 = torch.aten.int_repr %1816 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1818 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1819 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1820 = torch.aten.item %1818 : !torch.vtensor<[],f32> -> !torch.float | |
%1821 = torch.aten.item %1819 : !torch.vtensor<[],si8> -> !torch.int | |
%1822 = torch.aten._make_per_tensor_quantized_tensor %1817, %1820, %1821 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1823 = torch.aten.dequantize.self %1822 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%int5_378 = torch.constant.int 5 | |
%int5_379 = torch.constant.int 5 | |
%1824 = torch.prim.ListConstruct %int5_378, %int5_379 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int2_380 = torch.constant.int 2 | |
%int2_381 = torch.constant.int 2 | |
%1825 = torch.prim.ListConstruct %int2_380, %int2_381 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int1_382 = torch.constant.int 1 | |
%int1_383 = torch.constant.int 1 | |
%1826 = torch.prim.ListConstruct %int1_382, %int1_383 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int1_384 = torch.constant.int 1 | |
%int1_385 = torch.constant.int 1 | |
%1827 = torch.prim.ListConstruct %int1_384, %int1_385 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_386 = torch.constant.bool false | |
%1828 = torch.aten.max_pool2d %1823, %1824, %1826, %1825, %1827, %false_386 : !torch.vtensor<[1,128,20,20],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[1,128,20,20],f32> | |
%1829 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1830 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_387 = torch.constant.int 12 | |
%1831 = torch.aten.item %1829 : !torch.vtensor<[],f32> -> !torch.float | |
%1832 = torch.aten.item %1830 : !torch.vtensor<[],si8> -> !torch.int | |
%1833 = torch.aten.quantize_per_tensor %1828, %1831, %1832, %int12_387 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1834 = torch.aten.int_repr %1833 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%1835 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1836 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1837 = torch.aten.item %1835 : !torch.vtensor<[],f32> -> !torch.float | |
%1838 = torch.aten.item %1836 : !torch.vtensor<[],si8> -> !torch.int | |
%1839 = torch.aten._make_per_tensor_quantized_tensor %1834, %1837, %1838 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%1840 = torch.aten.dequantize.self %1839 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%int5_388 = torch.constant.int 5 | |
%int5_389 = torch.constant.int 5 | |
%1841 = torch.prim.ListConstruct %int5_388, %int5_389 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int2_390 = torch.constant.int 2 | |
%int2_391 = torch.constant.int 2 | |
%1842 = torch.prim.ListConstruct %int2_390, %int2_391 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int1_392 = torch.constant.int 1 | |
%int1_393 = torch.constant.int 1 | |
%1843 = torch.prim.ListConstruct %int1_392, %int1_393 : (!torch.int, !torch.int) -> !torch.list<int> | |
%int1_394 = torch.constant.int 1 | |
%int1_395 = torch.constant.int 1 | |
%1844 = torch.prim.ListConstruct %int1_394, %int1_395 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_396 = torch.constant.bool false | |
%1845 = torch.aten.max_pool2d %1840, %1841, %1843, %1842, %1844, %false_396 : !torch.vtensor<[1,128,20,20],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[1,128,20,20],f32> | |
%1846 = torch.prim.ListConstruct %1806, %1823, %1840, %1845 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>) -> !torch.list<vtensor> | |
%int1_397 = torch.constant.int 1 | |
%1847 = torch.aten.cat %1846, %int1_397 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,512,20,20],f32> | |
%1848 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1849 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_398 = torch.constant.int 12 | |
%1850 = torch.aten.item %1848 : !torch.vtensor<[],f32> -> !torch.float | |
%1851 = torch.aten.item %1849 : !torch.vtensor<[],si8> -> !torch.int | |
%1852 = torch.aten.quantize_per_tensor %1847, %1850, %1851, %int12_398 : !torch.vtensor<[1,512,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,512,20,20],!torch.qint8> | |
%1853 = torch.aten.int_repr %1852 : !torch.vtensor<[1,512,20,20],!torch.qint8> -> !torch.vtensor<[1,512,20,20],si8> | |
%1854 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1855 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1856 = torch.aten.item %1854 : !torch.vtensor<[],f32> -> !torch.float | |
%1857 = torch.aten.item %1855 : !torch.vtensor<[],si8> -> !torch.int | |
%1858 = torch.aten._make_per_tensor_quantized_tensor %1853, %1856, %1857 : !torch.vtensor<[1,512,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,512,20,20],!torch.qint8> | |
%1859 = torch.aten.dequantize.self %1858 : !torch.vtensor<[1,512,20,20],!torch.qint8> -> !torch.vtensor<[1,512,20,20],f32> | |
%1860 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1861 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_399 = torch.constant.int 12 | |
%1862 = torch.aten.item %1860 : !torch.vtensor<[],f32> -> !torch.float | |
%1863 = torch.aten.item %1861 : !torch.vtensor<[],si8> -> !torch.int | |
%1864 = torch.aten.quantize_per_tensor %52, %1862, %1863, %int12_399 : !torch.vtensor<[256,512,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,512,1,1],!torch.qint8> | |
%1865 = torch.aten.int_repr %1864 : !torch.vtensor<[256,512,1,1],!torch.qint8> -> !torch.vtensor<[256,512,1,1],si8> | |
%1866 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1867 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1868 = torch.aten.item %1866 : !torch.vtensor<[],f32> -> !torch.float | |
%1869 = torch.aten.item %1867 : !torch.vtensor<[],si8> -> !torch.int | |
%1870 = torch.aten._make_per_tensor_quantized_tensor %1865, %1868, %1869 : !torch.vtensor<[256,512,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,512,1,1],!torch.qint8> | |
%1871 = torch.aten.dequantize.self %1870 : !torch.vtensor<[256,512,1,1],!torch.qint8> -> !torch.vtensor<[256,512,1,1],f32> | |
%1872 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1873 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_400 = torch.constant.int 12 | |
%1874 = torch.aten.item %1872 : !torch.vtensor<[],f32> -> !torch.float | |
%1875 = torch.aten.item %1873 : !torch.vtensor<[],si8> -> !torch.int | |
%1876 = torch.aten.quantize_per_tensor %53, %1874, %1875, %int12_400 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1877 = torch.aten.int_repr %1876 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
%1878 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1879 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1880 = torch.aten.item %1878 : !torch.vtensor<[],f32> -> !torch.float | |
%1881 = torch.aten.item %1879 : !torch.vtensor<[],si8> -> !torch.int | |
%1882 = torch.aten._make_per_tensor_quantized_tensor %1877, %1880, %1881 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%1883 = torch.aten.dequantize.self %1882 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
%int0_401 = torch.constant.int 0 | |
%int0_402 = torch.constant.int 0 | |
%int1_403 = torch.constant.int 1 | |
%int1_404 = torch.constant.int 1 | |
%int1_405 = torch.constant.int 1 | |
%int1_406 = torch.constant.int 1 | |
%int0_407 = torch.constant.int 0 | |
%1884 = torch.prim.ListConstruct %int0_401, %int0_402 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1885 = torch.prim.ListConstruct %int1_403, %int1_404 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1886 = torch.prim.ListConstruct %int1_405, %int1_406 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1887 = torch.prim.ListConstruct %int0_407, %int0_407 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_408 = torch.constant.bool false | |
%int1_409 = torch.constant.int 1 | |
%1888 = torch.aten.convolution %1859, %1871, %1883, %1886, %1884, %1885, %false_408, %1887, %int1_409 : !torch.vtensor<[1,512,20,20],f32>, !torch.vtensor<[256,512,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
%1889 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1890 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_410 = torch.constant.int 12 | |
%1891 = torch.aten.item %1889 : !torch.vtensor<[],f32> -> !torch.float | |
%1892 = torch.aten.item %1890 : !torch.vtensor<[],si8> -> !torch.int | |
%1893 = torch.aten.quantize_per_tensor %1888, %1891, %1892, %int12_410 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1894 = torch.aten.int_repr %1893 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1895 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1896 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1897 = torch.aten.item %1895 : !torch.vtensor<[],f32> -> !torch.float | |
%1898 = torch.aten.item %1896 : !torch.vtensor<[],si8> -> !torch.int | |
%1899 = torch.aten._make_per_tensor_quantized_tensor %1894, %1897, %1898 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1900 = torch.aten.dequantize.self %1899 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1901 = torch.aten.sigmoid %1900 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1902 = torch.aten.mul.Tensor %1900, %1901 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%1903 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1904 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_411 = torch.constant.int 12 | |
%1905 = torch.aten.item %1903 : !torch.vtensor<[],f32> -> !torch.float | |
%1906 = torch.aten.item %1904 : !torch.vtensor<[],si8> -> !torch.int | |
%1907 = torch.aten.quantize_per_tensor %1902, %1905, %1906, %int12_411 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1908 = torch.aten.int_repr %1907 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%1909 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1910 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1911 = torch.aten.item %1909 : !torch.vtensor<[],f32> -> !torch.float | |
%1912 = torch.aten.item %1910 : !torch.vtensor<[],si8> -> !torch.int | |
%1913 = torch.aten._make_per_tensor_quantized_tensor %1908, %1911, %1912 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%1914 = torch.aten.dequantize.self %1913 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%1915 = torch.vtensor.literal(dense<1.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
%1916 = torch.vtensor.literal(dense<2.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
%1917 = torch.prim.ListConstruct %1915, %1916 : (!torch.vtensor<[2],f32>, !torch.vtensor<[2],f32>) -> !torch.list<vtensor> | |
%int0_412 = torch.constant.int 0 | |
%1918 = torch.aten.cat %1917, %int0_412 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],f32> | |
%1919 = torch.operator "onnx.Resize"(%1914, %none, %1918) {torch.onnx.coordinate_transformation_mode = "asymmetric", torch.onnx.cubic_coeff_a = -7.500000e-01 : f32, torch.onnx.mode = "nearest", torch.onnx.nearest_mode = "floor"} : (!torch.vtensor<[1,256,20,20],f32>, !torch.none, !torch.vtensor<[4],f32>) -> !torch.vtensor<[?,?,?,?],f32> | |
%1920 = torch.prim.ListConstruct %1919, %1433 : (!torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[1,128,40,40],f32>) -> !torch.list<vtensor> | |
%int1_413 = torch.constant.int 1 | |
%1921 = torch.aten.cat %1920, %int1_413 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,?,40,40],f32> | |
%1922 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1923 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_414 = torch.constant.int 12 | |
%1924 = torch.aten.item %1922 : !torch.vtensor<[],f32> -> !torch.float | |
%1925 = torch.aten.item %1923 : !torch.vtensor<[],si8> -> !torch.int | |
%1926 = torch.aten.quantize_per_tensor %1921, %1924, %1925, %int12_414 : !torch.vtensor<[1,?,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,?,40,40],!torch.qint8> | |
%1927 = torch.aten.int_repr %1926 : !torch.vtensor<[1,?,40,40],!torch.qint8> -> !torch.vtensor<[1,?,40,40],si8> | |
%1928 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1929 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1930 = torch.aten.item %1928 : !torch.vtensor<[],f32> -> !torch.float | |
%1931 = torch.aten.item %1929 : !torch.vtensor<[],si8> -> !torch.int | |
%1932 = torch.aten._make_per_tensor_quantized_tensor %1927, %1930, %1931 : !torch.vtensor<[1,?,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,?,40,40],!torch.qint8> | |
%1933 = torch.aten.dequantize.self %1932 : !torch.vtensor<[1,?,40,40],!torch.qint8> -> !torch.vtensor<[1,?,40,40],f32> | |
%1934 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1935 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_415 = torch.constant.int 12 | |
%1936 = torch.aten.item %1934 : !torch.vtensor<[],f32> -> !torch.float | |
%1937 = torch.aten.item %1935 : !torch.vtensor<[],si8> -> !torch.int | |
%1938 = torch.aten.quantize_per_tensor %54, %1936, %1937, %int12_415 : !torch.vtensor<[128,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,384,1,1],!torch.qint8> | |
%1939 = torch.aten.int_repr %1938 : !torch.vtensor<[128,384,1,1],!torch.qint8> -> !torch.vtensor<[128,384,1,1],si8> | |
%1940 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1941 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1942 = torch.aten.item %1940 : !torch.vtensor<[],f32> -> !torch.float | |
%1943 = torch.aten.item %1941 : !torch.vtensor<[],si8> -> !torch.int | |
%1944 = torch.aten._make_per_tensor_quantized_tensor %1939, %1942, %1943 : !torch.vtensor<[128,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,384,1,1],!torch.qint8> | |
%1945 = torch.aten.dequantize.self %1944 : !torch.vtensor<[128,384,1,1],!torch.qint8> -> !torch.vtensor<[128,384,1,1],f32> | |
%1946 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1947 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_416 = torch.constant.int 12 | |
%1948 = torch.aten.item %1946 : !torch.vtensor<[],f32> -> !torch.float | |
%1949 = torch.aten.item %1947 : !torch.vtensor<[],si8> -> !torch.int | |
%1950 = torch.aten.quantize_per_tensor %55, %1948, %1949, %int12_416 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1951 = torch.aten.int_repr %1950 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%1952 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1953 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1954 = torch.aten.item %1952 : !torch.vtensor<[],f32> -> !torch.float | |
%1955 = torch.aten.item %1953 : !torch.vtensor<[],si8> -> !torch.int | |
%1956 = torch.aten._make_per_tensor_quantized_tensor %1951, %1954, %1955 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%1957 = torch.aten.dequantize.self %1956 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_417 = torch.constant.int 0 | |
%int0_418 = torch.constant.int 0 | |
%int1_419 = torch.constant.int 1 | |
%int1_420 = torch.constant.int 1 | |
%int1_421 = torch.constant.int 1 | |
%int1_422 = torch.constant.int 1 | |
%int0_423 = torch.constant.int 0 | |
%1958 = torch.prim.ListConstruct %int0_417, %int0_418 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1959 = torch.prim.ListConstruct %int1_419, %int1_420 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1960 = torch.prim.ListConstruct %int1_421, %int1_422 : (!torch.int, !torch.int) -> !torch.list<int> | |
%1961 = torch.prim.ListConstruct %int0_423, %int0_423 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_424 = torch.constant.bool false | |
%int1_425 = torch.constant.int 1 | |
%1962 = torch.aten.convolution %1933, %1945, %1957, %1960, %1958, %1959, %false_424, %1961, %int1_425 : !torch.vtensor<[1,?,40,40],f32>, !torch.vtensor<[128,384,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%1963 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1964 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_426 = torch.constant.int 12 | |
%1965 = torch.aten.item %1963 : !torch.vtensor<[],f32> -> !torch.float | |
%1966 = torch.aten.item %1964 : !torch.vtensor<[],si8> -> !torch.int | |
%1967 = torch.aten.quantize_per_tensor %1962, %1965, %1966, %int12_426 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1968 = torch.aten.int_repr %1967 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1969 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1970 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1971 = torch.aten.item %1969 : !torch.vtensor<[],f32> -> !torch.float | |
%1972 = torch.aten.item %1970 : !torch.vtensor<[],si8> -> !torch.int | |
%1973 = torch.aten._make_per_tensor_quantized_tensor %1968, %1971, %1972 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1974 = torch.aten.dequantize.self %1973 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1975 = torch.aten.sigmoid %1974 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1976 = torch.aten.mul.Tensor %1974, %1975 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%1977 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1978 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_427 = torch.constant.int 12 | |
%1979 = torch.aten.item %1977 : !torch.vtensor<[],f32> -> !torch.float | |
%1980 = torch.aten.item %1978 : !torch.vtensor<[],si8> -> !torch.int | |
%1981 = torch.aten.quantize_per_tensor %1976, %1979, %1980, %int12_427 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1982 = torch.aten.int_repr %1981 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%1983 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%1984 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%1985 = torch.aten.item %1983 : !torch.vtensor<[],f32> -> !torch.float | |
%1986 = torch.aten.item %1984 : !torch.vtensor<[],si8> -> !torch.int | |
%1987 = torch.aten._make_per_tensor_quantized_tensor %1982, %1985, %1986 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%1988 = torch.aten.dequantize.self %1987 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%1989 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1990 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1991 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%1992 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_428 = torch.constant.int 0 | |
%int0_429 = torch.constant.int 0 | |
%1993 = torch.prim.NumToTensor.Scalar %int0_429 : !torch.int -> !torch.vtensor<[1],si64> | |
%1994 = torch.aten.index_select %1990, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1995 = torch.aten.item %1994 : !torch.vtensor<[1],si64> -> !torch.int | |
%1996 = torch.aten.index_select %1991, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1997 = torch.aten.item %1996 : !torch.vtensor<[1],si64> -> !torch.int | |
%1998 = torch.aten.index_select %1989, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%1999 = torch.aten.item %1998 : !torch.vtensor<[1],si64> -> !torch.int | |
%2000 = torch.aten.index_select %1992, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2001 = torch.aten.item %2000 : !torch.vtensor<[1],si64> -> !torch.int | |
%2002 = torch.aten.slice.Tensor %1988, %1999, %1995, %1997, %2001 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2003 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2004 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2005 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2006 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_430 = torch.constant.int 0 | |
%int0_431 = torch.constant.int 0 | |
%2007 = torch.prim.NumToTensor.Scalar %int0_431 : !torch.int -> !torch.vtensor<[1],si64> | |
%2008 = torch.aten.index_select %2004, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2009 = torch.aten.item %2008 : !torch.vtensor<[1],si64> -> !torch.int | |
%2010 = torch.aten.index_select %2005, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2011 = torch.aten.item %2010 : !torch.vtensor<[1],si64> -> !torch.int | |
%2012 = torch.aten.index_select %2003, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2013 = torch.aten.item %2012 : !torch.vtensor<[1],si64> -> !torch.int | |
%2014 = torch.aten.index_select %2006, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2015 = torch.aten.item %2014 : !torch.vtensor<[1],si64> -> !torch.int | |
%2016 = torch.aten.slice.Tensor %1988, %2013, %2009, %2011, %2015 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2017 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2018 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_432 = torch.constant.int 12 | |
%2019 = torch.aten.item %2017 : !torch.vtensor<[],f32> -> !torch.float | |
%2020 = torch.aten.item %2018 : !torch.vtensor<[],si8> -> !torch.int | |
%2021 = torch.aten.quantize_per_tensor %56, %2019, %2020, %int12_432 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2022 = torch.aten.int_repr %2021 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%2023 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2024 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2025 = torch.aten.item %2023 : !torch.vtensor<[],f32> -> !torch.float | |
%2026 = torch.aten.item %2024 : !torch.vtensor<[],si8> -> !torch.int | |
%2027 = torch.aten._make_per_tensor_quantized_tensor %2022, %2025, %2026 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2028 = torch.aten.dequantize.self %2027 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%2029 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2030 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_433 = torch.constant.int 12 | |
%2031 = torch.aten.item %2029 : !torch.vtensor<[],f32> -> !torch.float | |
%2032 = torch.aten.item %2030 : !torch.vtensor<[],si8> -> !torch.int | |
%2033 = torch.aten.quantize_per_tensor %57, %2031, %2032, %int12_433 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2034 = torch.aten.int_repr %2033 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2035 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2036 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2037 = torch.aten.item %2035 : !torch.vtensor<[],f32> -> !torch.float | |
%2038 = torch.aten.item %2036 : !torch.vtensor<[],si8> -> !torch.int | |
%2039 = torch.aten._make_per_tensor_quantized_tensor %2034, %2037, %2038 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2040 = torch.aten.dequantize.self %2039 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_434 = torch.constant.int 1 | |
%int1_435 = torch.constant.int 1 | |
%int1_436 = torch.constant.int 1 | |
%int1_437 = torch.constant.int 1 | |
%int1_438 = torch.constant.int 1 | |
%int1_439 = torch.constant.int 1 | |
%int0_440 = torch.constant.int 0 | |
%2041 = torch.prim.ListConstruct %int1_434, %int1_435 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2042 = torch.prim.ListConstruct %int1_436, %int1_437 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2043 = torch.prim.ListConstruct %int1_438, %int1_439 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2044 = torch.prim.ListConstruct %int0_440, %int0_440 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_441 = torch.constant.bool false | |
%int1_442 = torch.constant.int 1 | |
%2045 = torch.aten.convolution %2016, %2028, %2040, %2043, %2041, %2042, %false_441, %2044, %int1_442 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2046 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2047 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_443 = torch.constant.int 12 | |
%2048 = torch.aten.item %2046 : !torch.vtensor<[],f32> -> !torch.float | |
%2049 = torch.aten.item %2047 : !torch.vtensor<[],si8> -> !torch.int | |
%2050 = torch.aten.quantize_per_tensor %2045, %2048, %2049, %int12_443 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2051 = torch.aten.int_repr %2050 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2052 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2053 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2054 = torch.aten.item %2052 : !torch.vtensor<[],f32> -> !torch.float | |
%2055 = torch.aten.item %2053 : !torch.vtensor<[],si8> -> !torch.int | |
%2056 = torch.aten._make_per_tensor_quantized_tensor %2051, %2054, %2055 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2057 = torch.aten.dequantize.self %2056 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2058 = torch.aten.sigmoid %2057 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2059 = torch.aten.mul.Tensor %2057, %2058 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2060 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2061 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_444 = torch.constant.int 12 | |
%2062 = torch.aten.item %2060 : !torch.vtensor<[],f32> -> !torch.float | |
%2063 = torch.aten.item %2061 : !torch.vtensor<[],si8> -> !torch.int | |
%2064 = torch.aten.quantize_per_tensor %2059, %2062, %2063, %int12_444 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2065 = torch.aten.int_repr %2064 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2066 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2067 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2068 = torch.aten.item %2066 : !torch.vtensor<[],f32> -> !torch.float | |
%2069 = torch.aten.item %2067 : !torch.vtensor<[],si8> -> !torch.int | |
%2070 = torch.aten._make_per_tensor_quantized_tensor %2065, %2068, %2069 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2071 = torch.aten.dequantize.self %2070 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2072 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2073 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_445 = torch.constant.int 12 | |
%2074 = torch.aten.item %2072 : !torch.vtensor<[],f32> -> !torch.float | |
%2075 = torch.aten.item %2073 : !torch.vtensor<[],si8> -> !torch.int | |
%2076 = torch.aten.quantize_per_tensor %58, %2074, %2075, %int12_445 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2077 = torch.aten.int_repr %2076 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%2078 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2079 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2080 = torch.aten.item %2078 : !torch.vtensor<[],f32> -> !torch.float | |
%2081 = torch.aten.item %2079 : !torch.vtensor<[],si8> -> !torch.int | |
%2082 = torch.aten._make_per_tensor_quantized_tensor %2077, %2080, %2081 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2083 = torch.aten.dequantize.self %2082 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%2084 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2085 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_446 = torch.constant.int 12 | |
%2086 = torch.aten.item %2084 : !torch.vtensor<[],f32> -> !torch.float | |
%2087 = torch.aten.item %2085 : !torch.vtensor<[],si8> -> !torch.int | |
%2088 = torch.aten.quantize_per_tensor %59, %2086, %2087, %int12_446 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2089 = torch.aten.int_repr %2088 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2090 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2091 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2092 = torch.aten.item %2090 : !torch.vtensor<[],f32> -> !torch.float | |
%2093 = torch.aten.item %2091 : !torch.vtensor<[],si8> -> !torch.int | |
%2094 = torch.aten._make_per_tensor_quantized_tensor %2089, %2092, %2093 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2095 = torch.aten.dequantize.self %2094 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_447 = torch.constant.int 1 | |
%int1_448 = torch.constant.int 1 | |
%int1_449 = torch.constant.int 1 | |
%int1_450 = torch.constant.int 1 | |
%int1_451 = torch.constant.int 1 | |
%int1_452 = torch.constant.int 1 | |
%int0_453 = torch.constant.int 0 | |
%2096 = torch.prim.ListConstruct %int1_447, %int1_448 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2097 = torch.prim.ListConstruct %int1_449, %int1_450 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2098 = torch.prim.ListConstruct %int1_451, %int1_452 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2099 = torch.prim.ListConstruct %int0_453, %int0_453 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_454 = torch.constant.bool false | |
%int1_455 = torch.constant.int 1 | |
%2100 = torch.aten.convolution %2071, %2083, %2095, %2098, %2096, %2097, %false_454, %2099, %int1_455 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2101 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2102 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_456 = torch.constant.int 12 | |
%2103 = torch.aten.item %2101 : !torch.vtensor<[],f32> -> !torch.float | |
%2104 = torch.aten.item %2102 : !torch.vtensor<[],si8> -> !torch.int | |
%2105 = torch.aten.quantize_per_tensor %2100, %2103, %2104, %int12_456 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2106 = torch.aten.int_repr %2105 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2107 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2108 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2109 = torch.aten.item %2107 : !torch.vtensor<[],f32> -> !torch.float | |
%2110 = torch.aten.item %2108 : !torch.vtensor<[],si8> -> !torch.int | |
%2111 = torch.aten._make_per_tensor_quantized_tensor %2106, %2109, %2110 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2112 = torch.aten.dequantize.self %2111 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2113 = torch.aten.sigmoid %2112 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2114 = torch.aten.mul.Tensor %2112, %2113 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2115 = torch.prim.ListConstruct %2002, %2016, %2114 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>) -> !torch.list<vtensor> | |
%int1_457 = torch.constant.int 1 | |
%2116 = torch.aten.cat %2115, %int1_457 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,192,40,40],f32> | |
%2117 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2118 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_458 = torch.constant.int 12 | |
%2119 = torch.aten.item %2117 : !torch.vtensor<[],f32> -> !torch.float | |
%2120 = torch.aten.item %2118 : !torch.vtensor<[],si8> -> !torch.int | |
%2121 = torch.aten.quantize_per_tensor %2116, %2119, %2120, %int12_458 : !torch.vtensor<[1,192,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
%2122 = torch.aten.int_repr %2121 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],si8> | |
%2123 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2124 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2125 = torch.aten.item %2123 : !torch.vtensor<[],f32> -> !torch.float | |
%2126 = torch.aten.item %2124 : !torch.vtensor<[],si8> -> !torch.int | |
%2127 = torch.aten._make_per_tensor_quantized_tensor %2122, %2125, %2126 : !torch.vtensor<[1,192,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
%2128 = torch.aten.dequantize.self %2127 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],f32> | |
%2129 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2130 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_459 = torch.constant.int 12 | |
%2131 = torch.aten.item %2129 : !torch.vtensor<[],f32> -> !torch.float | |
%2132 = torch.aten.item %2130 : !torch.vtensor<[],si8> -> !torch.int | |
%2133 = torch.aten.quantize_per_tensor %60, %2131, %2132, %int12_459 : !torch.vtensor<[128,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
%2134 = torch.aten.int_repr %2133 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],si8> | |
%2135 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2136 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2137 = torch.aten.item %2135 : !torch.vtensor<[],f32> -> !torch.float | |
%2138 = torch.aten.item %2136 : !torch.vtensor<[],si8> -> !torch.int | |
%2139 = torch.aten._make_per_tensor_quantized_tensor %2134, %2137, %2138 : !torch.vtensor<[128,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
%2140 = torch.aten.dequantize.self %2139 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],f32> | |
%2141 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2142 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_460 = torch.constant.int 12 | |
%2143 = torch.aten.item %2141 : !torch.vtensor<[],f32> -> !torch.float | |
%2144 = torch.aten.item %2142 : !torch.vtensor<[],si8> -> !torch.int | |
%2145 = torch.aten.quantize_per_tensor %61, %2143, %2144, %int12_460 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2146 = torch.aten.int_repr %2145 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%2147 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2148 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2149 = torch.aten.item %2147 : !torch.vtensor<[],f32> -> !torch.float | |
%2150 = torch.aten.item %2148 : !torch.vtensor<[],si8> -> !torch.int | |
%2151 = torch.aten._make_per_tensor_quantized_tensor %2146, %2149, %2150 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2152 = torch.aten.dequantize.self %2151 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_461 = torch.constant.int 0 | |
%int0_462 = torch.constant.int 0 | |
%int1_463 = torch.constant.int 1 | |
%int1_464 = torch.constant.int 1 | |
%int1_465 = torch.constant.int 1 | |
%int1_466 = torch.constant.int 1 | |
%int0_467 = torch.constant.int 0 | |
%2153 = torch.prim.ListConstruct %int0_461, %int0_462 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2154 = torch.prim.ListConstruct %int1_463, %int1_464 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2155 = torch.prim.ListConstruct %int1_465, %int1_466 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2156 = torch.prim.ListConstruct %int0_467, %int0_467 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_468 = torch.constant.bool false | |
%int1_469 = torch.constant.int 1 | |
%2157 = torch.aten.convolution %2128, %2140, %2152, %2155, %2153, %2154, %false_468, %2156, %int1_469 : !torch.vtensor<[1,192,40,40],f32>, !torch.vtensor<[128,192,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%2158 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2159 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_470 = torch.constant.int 12 | |
%2160 = torch.aten.item %2158 : !torch.vtensor<[],f32> -> !torch.float | |
%2161 = torch.aten.item %2159 : !torch.vtensor<[],si8> -> !torch.int | |
%2162 = torch.aten.quantize_per_tensor %2157, %2160, %2161, %int12_470 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2163 = torch.aten.int_repr %2162 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%2164 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2165 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2166 = torch.aten.item %2164 : !torch.vtensor<[],f32> -> !torch.float | |
%2167 = torch.aten.item %2165 : !torch.vtensor<[],si8> -> !torch.int | |
%2168 = torch.aten._make_per_tensor_quantized_tensor %2163, %2166, %2167 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2169 = torch.aten.dequantize.self %2168 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%2170 = torch.aten.sigmoid %2169 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%2171 = torch.aten.mul.Tensor %2169, %2170 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%2172 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_471 = torch.constant.int 12 | |
%2174 = torch.aten.item %2172 : !torch.vtensor<[],f32> -> !torch.float | |
%2175 = torch.aten.item %2173 : !torch.vtensor<[],si8> -> !torch.int | |
%2176 = torch.aten.quantize_per_tensor %2171, %2174, %2175, %int12_471 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2177 = torch.aten.int_repr %2176 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%2178 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2180 = torch.aten.item %2178 : !torch.vtensor<[],f32> -> !torch.float | |
%2181 = torch.aten.item %2179 : !torch.vtensor<[],si8> -> !torch.int | |
%2182 = torch.aten._make_per_tensor_quantized_tensor %2177, %2180, %2181 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2183 = torch.aten.dequantize.self %2182 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%2184 = torch.vtensor.literal(dense<1.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
%2185 = torch.vtensor.literal(dense<2.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
%2186 = torch.prim.ListConstruct %2184, %2185 : (!torch.vtensor<[2],f32>, !torch.vtensor<[2],f32>) -> !torch.list<vtensor> | |
%int0_472 = torch.constant.int 0 | |
%2187 = torch.aten.cat %2186, %int0_472 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],f32> | |
%2188 = torch.operator "onnx.Resize"(%2183, %none, %2187) {torch.onnx.coordinate_transformation_mode = "asymmetric", torch.onnx.cubic_coeff_a = -7.500000e-01 : f32, torch.onnx.mode = "nearest", torch.onnx.nearest_mode = "floor"} : (!torch.vtensor<[1,128,40,40],f32>, !torch.none, !torch.vtensor<[4],f32>) -> !torch.vtensor<[?,?,?,?],f32> | |
%2189 = torch.prim.ListConstruct %2188, %992 : (!torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[1,64,80,80],f32>) -> !torch.list<vtensor> | |
%int1_473 = torch.constant.int 1 | |
%2190 = torch.aten.cat %2189, %int1_473 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,?,80,80],f32> | |
%2191 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2192 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_474 = torch.constant.int 12 | |
%2193 = torch.aten.item %2191 : !torch.vtensor<[],f32> -> !torch.float | |
%2194 = torch.aten.item %2192 : !torch.vtensor<[],si8> -> !torch.int | |
%2195 = torch.aten.quantize_per_tensor %2190, %2193, %2194, %int12_474 : !torch.vtensor<[1,?,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,?,80,80],!torch.qint8> | |
%2196 = torch.aten.int_repr %2195 : !torch.vtensor<[1,?,80,80],!torch.qint8> -> !torch.vtensor<[1,?,80,80],si8> | |
%2197 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2198 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2199 = torch.aten.item %2197 : !torch.vtensor<[],f32> -> !torch.float | |
%2200 = torch.aten.item %2198 : !torch.vtensor<[],si8> -> !torch.int | |
%2201 = torch.aten._make_per_tensor_quantized_tensor %2196, %2199, %2200 : !torch.vtensor<[1,?,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,?,80,80],!torch.qint8> | |
%2202 = torch.aten.dequantize.self %2201 : !torch.vtensor<[1,?,80,80],!torch.qint8> -> !torch.vtensor<[1,?,80,80],f32> | |
%2203 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2204 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_475 = torch.constant.int 12 | |
%2205 = torch.aten.item %2203 : !torch.vtensor<[],f32> -> !torch.float | |
%2206 = torch.aten.item %2204 : !torch.vtensor<[],si8> -> !torch.int | |
%2207 = torch.aten.quantize_per_tensor %62, %2205, %2206, %int12_475 : !torch.vtensor<[64,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,192,1,1],!torch.qint8> | |
%2208 = torch.aten.int_repr %2207 : !torch.vtensor<[64,192,1,1],!torch.qint8> -> !torch.vtensor<[64,192,1,1],si8> | |
%2209 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2210 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2211 = torch.aten.item %2209 : !torch.vtensor<[],f32> -> !torch.float | |
%2212 = torch.aten.item %2210 : !torch.vtensor<[],si8> -> !torch.int | |
%2213 = torch.aten._make_per_tensor_quantized_tensor %2208, %2211, %2212 : !torch.vtensor<[64,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,192,1,1],!torch.qint8> | |
%2214 = torch.aten.dequantize.self %2213 : !torch.vtensor<[64,192,1,1],!torch.qint8> -> !torch.vtensor<[64,192,1,1],f32> | |
%2215 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2216 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_476 = torch.constant.int 12 | |
%2217 = torch.aten.item %2215 : !torch.vtensor<[],f32> -> !torch.float | |
%2218 = torch.aten.item %2216 : !torch.vtensor<[],si8> -> !torch.int | |
%2219 = torch.aten.quantize_per_tensor %63, %2217, %2218, %int12_476 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2220 = torch.aten.int_repr %2219 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2221 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2223 = torch.aten.item %2221 : !torch.vtensor<[],f32> -> !torch.float | |
%2224 = torch.aten.item %2222 : !torch.vtensor<[],si8> -> !torch.int | |
%2225 = torch.aten._make_per_tensor_quantized_tensor %2220, %2223, %2224 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2226 = torch.aten.dequantize.self %2225 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_477 = torch.constant.int 0 | |
%int0_478 = torch.constant.int 0 | |
%int1_479 = torch.constant.int 1 | |
%int1_480 = torch.constant.int 1 | |
%int1_481 = torch.constant.int 1 | |
%int1_482 = torch.constant.int 1 | |
%int0_483 = torch.constant.int 0 | |
%2227 = torch.prim.ListConstruct %int0_477, %int0_478 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2228 = torch.prim.ListConstruct %int1_479, %int1_480 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2229 = torch.prim.ListConstruct %int1_481, %int1_482 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2230 = torch.prim.ListConstruct %int0_483, %int0_483 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_484 = torch.constant.bool false | |
%int1_485 = torch.constant.int 1 | |
%2231 = torch.aten.convolution %2202, %2214, %2226, %2229, %2227, %2228, %false_484, %2230, %int1_485 : !torch.vtensor<[1,?,80,80],f32>, !torch.vtensor<[64,192,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%2232 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2233 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_486 = torch.constant.int 12 | |
%2234 = torch.aten.item %2232 : !torch.vtensor<[],f32> -> !torch.float | |
%2235 = torch.aten.item %2233 : !torch.vtensor<[],si8> -> !torch.int | |
%2236 = torch.aten.quantize_per_tensor %2231, %2234, %2235, %int12_486 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2237 = torch.aten.int_repr %2236 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%2238 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2239 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2240 = torch.aten.item %2238 : !torch.vtensor<[],f32> -> !torch.float | |
%2241 = torch.aten.item %2239 : !torch.vtensor<[],si8> -> !torch.int | |
%2242 = torch.aten._make_per_tensor_quantized_tensor %2237, %2240, %2241 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2243 = torch.aten.dequantize.self %2242 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%2244 = torch.aten.sigmoid %2243 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%2245 = torch.aten.mul.Tensor %2243, %2244 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%2246 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2247 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_487 = torch.constant.int 12 | |
%2248 = torch.aten.item %2246 : !torch.vtensor<[],f32> -> !torch.float | |
%2249 = torch.aten.item %2247 : !torch.vtensor<[],si8> -> !torch.int | |
%2250 = torch.aten.quantize_per_tensor %2245, %2248, %2249, %int12_487 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2251 = torch.aten.int_repr %2250 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%2252 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2253 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2254 = torch.aten.item %2252 : !torch.vtensor<[],f32> -> !torch.float | |
%2255 = torch.aten.item %2253 : !torch.vtensor<[],si8> -> !torch.int | |
%2256 = torch.aten._make_per_tensor_quantized_tensor %2251, %2254, %2255 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2257 = torch.aten.dequantize.self %2256 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%2258 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2259 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2260 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2261 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_488 = torch.constant.int 0 | |
%int0_489 = torch.constant.int 0 | |
%2262 = torch.prim.NumToTensor.Scalar %int0_489 : !torch.int -> !torch.vtensor<[1],si64> | |
%2263 = torch.aten.index_select %2259, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2264 = torch.aten.item %2263 : !torch.vtensor<[1],si64> -> !torch.int | |
%2265 = torch.aten.index_select %2260, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2266 = torch.aten.item %2265 : !torch.vtensor<[1],si64> -> !torch.int | |
%2267 = torch.aten.index_select %2258, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2268 = torch.aten.item %2267 : !torch.vtensor<[1],si64> -> !torch.int | |
%2269 = torch.aten.index_select %2261, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2270 = torch.aten.item %2269 : !torch.vtensor<[1],si64> -> !torch.int | |
%2271 = torch.aten.slice.Tensor %2257, %2268, %2264, %2266, %2270 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%2272 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2273 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2274 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2275 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_490 = torch.constant.int 0 | |
%int0_491 = torch.constant.int 0 | |
%2276 = torch.prim.NumToTensor.Scalar %int0_491 : !torch.int -> !torch.vtensor<[1],si64> | |
%2277 = torch.aten.index_select %2273, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2278 = torch.aten.item %2277 : !torch.vtensor<[1],si64> -> !torch.int | |
%2279 = torch.aten.index_select %2274, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2280 = torch.aten.item %2279 : !torch.vtensor<[1],si64> -> !torch.int | |
%2281 = torch.aten.index_select %2272, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2282 = torch.aten.item %2281 : !torch.vtensor<[1],si64> -> !torch.int | |
%2283 = torch.aten.index_select %2275, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2284 = torch.aten.item %2283 : !torch.vtensor<[1],si64> -> !torch.int | |
%2285 = torch.aten.slice.Tensor %2257, %2282, %2278, %2280, %2284 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%2286 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2287 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_492 = torch.constant.int 12 | |
%2288 = torch.aten.item %2286 : !torch.vtensor<[],f32> -> !torch.float | |
%2289 = torch.aten.item %2287 : !torch.vtensor<[],si8> -> !torch.int | |
%2290 = torch.aten.quantize_per_tensor %64, %2288, %2289, %int12_492 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%2291 = torch.aten.int_repr %2290 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
%2292 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2293 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2294 = torch.aten.item %2292 : !torch.vtensor<[],f32> -> !torch.float | |
%2295 = torch.aten.item %2293 : !torch.vtensor<[],si8> -> !torch.int | |
%2296 = torch.aten._make_per_tensor_quantized_tensor %2291, %2294, %2295 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%2297 = torch.aten.dequantize.self %2296 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
%2298 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2299 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_493 = torch.constant.int 12 | |
%2300 = torch.aten.item %2298 : !torch.vtensor<[],f32> -> !torch.float | |
%2301 = torch.aten.item %2299 : !torch.vtensor<[],si8> -> !torch.int | |
%2302 = torch.aten.quantize_per_tensor %65, %2300, %2301, %int12_493 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%2303 = torch.aten.int_repr %2302 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%2304 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2305 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2306 = torch.aten.item %2304 : !torch.vtensor<[],f32> -> !torch.float | |
%2307 = torch.aten.item %2305 : !torch.vtensor<[],si8> -> !torch.int | |
%2308 = torch.aten._make_per_tensor_quantized_tensor %2303, %2306, %2307 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%2309 = torch.aten.dequantize.self %2308 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_494 = torch.constant.int 1 | |
%int1_495 = torch.constant.int 1 | |
%int1_496 = torch.constant.int 1 | |
%int1_497 = torch.constant.int 1 | |
%int1_498 = torch.constant.int 1 | |
%int1_499 = torch.constant.int 1 | |
%int0_500 = torch.constant.int 0 | |
%2310 = torch.prim.ListConstruct %int1_494, %int1_495 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2311 = torch.prim.ListConstruct %int1_496, %int1_497 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2312 = torch.prim.ListConstruct %int1_498, %int1_499 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2313 = torch.prim.ListConstruct %int0_500, %int0_500 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_501 = torch.constant.bool false | |
%int1_502 = torch.constant.int 1 | |
%2314 = torch.aten.convolution %2285, %2297, %2309, %2312, %2310, %2311, %false_501, %2313, %int1_502 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%2315 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2316 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_503 = torch.constant.int 12 | |
%2317 = torch.aten.item %2315 : !torch.vtensor<[],f32> -> !torch.float | |
%2318 = torch.aten.item %2316 : !torch.vtensor<[],si8> -> !torch.int | |
%2319 = torch.aten.quantize_per_tensor %2314, %2317, %2318, %int12_503 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%2320 = torch.aten.int_repr %2319 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%2321 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2322 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2323 = torch.aten.item %2321 : !torch.vtensor<[],f32> -> !torch.float | |
%2324 = torch.aten.item %2322 : !torch.vtensor<[],si8> -> !torch.int | |
%2325 = torch.aten._make_per_tensor_quantized_tensor %2320, %2323, %2324 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%2326 = torch.aten.dequantize.self %2325 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%2327 = torch.aten.sigmoid %2326 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%2328 = torch.aten.mul.Tensor %2326, %2327 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%2329 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2330 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_504 = torch.constant.int 12 | |
%2331 = torch.aten.item %2329 : !torch.vtensor<[],f32> -> !torch.float | |
%2332 = torch.aten.item %2330 : !torch.vtensor<[],si8> -> !torch.int | |
%2333 = torch.aten.quantize_per_tensor %2328, %2331, %2332, %int12_504 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%2334 = torch.aten.int_repr %2333 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%2335 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2336 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2337 = torch.aten.item %2335 : !torch.vtensor<[],f32> -> !torch.float | |
%2338 = torch.aten.item %2336 : !torch.vtensor<[],si8> -> !torch.int | |
%2339 = torch.aten._make_per_tensor_quantized_tensor %2334, %2337, %2338 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%2340 = torch.aten.dequantize.self %2339 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%2341 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2342 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_505 = torch.constant.int 12 | |
%2343 = torch.aten.item %2341 : !torch.vtensor<[],f32> -> !torch.float | |
%2344 = torch.aten.item %2342 : !torch.vtensor<[],si8> -> !torch.int | |
%2345 = torch.aten.quantize_per_tensor %66, %2343, %2344, %int12_505 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%2346 = torch.aten.int_repr %2345 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
%2347 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2348 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2349 = torch.aten.item %2347 : !torch.vtensor<[],f32> -> !torch.float | |
%2350 = torch.aten.item %2348 : !torch.vtensor<[],si8> -> !torch.int | |
%2351 = torch.aten._make_per_tensor_quantized_tensor %2346, %2349, %2350 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
%2352 = torch.aten.dequantize.self %2351 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
%2353 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2354 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_506 = torch.constant.int 12 | |
%2355 = torch.aten.item %2353 : !torch.vtensor<[],f32> -> !torch.float | |
%2356 = torch.aten.item %2354 : !torch.vtensor<[],si8> -> !torch.int | |
%2357 = torch.aten.quantize_per_tensor %67, %2355, %2356, %int12_506 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%2358 = torch.aten.int_repr %2357 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
%2359 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2360 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2361 = torch.aten.item %2359 : !torch.vtensor<[],f32> -> !torch.float | |
%2362 = torch.aten.item %2360 : !torch.vtensor<[],si8> -> !torch.int | |
%2363 = torch.aten._make_per_tensor_quantized_tensor %2358, %2361, %2362 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
%2364 = torch.aten.dequantize.self %2363 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
%int1_507 = torch.constant.int 1 | |
%int1_508 = torch.constant.int 1 | |
%int1_509 = torch.constant.int 1 | |
%int1_510 = torch.constant.int 1 | |
%int1_511 = torch.constant.int 1 | |
%int1_512 = torch.constant.int 1 | |
%int0_513 = torch.constant.int 0 | |
%2365 = torch.prim.ListConstruct %int1_507, %int1_508 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2366 = torch.prim.ListConstruct %int1_509, %int1_510 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2367 = torch.prim.ListConstruct %int1_511, %int1_512 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2368 = torch.prim.ListConstruct %int0_513, %int0_513 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_514 = torch.constant.bool false | |
%int1_515 = torch.constant.int 1 | |
%2369 = torch.aten.convolution %2340, %2352, %2364, %2367, %2365, %2366, %false_514, %2368, %int1_515 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
%2370 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2371 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_516 = torch.constant.int 12 | |
%2372 = torch.aten.item %2370 : !torch.vtensor<[],f32> -> !torch.float | |
%2373 = torch.aten.item %2371 : !torch.vtensor<[],si8> -> !torch.int | |
%2374 = torch.aten.quantize_per_tensor %2369, %2372, %2373, %int12_516 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%2375 = torch.aten.int_repr %2374 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
%2376 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2377 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2378 = torch.aten.item %2376 : !torch.vtensor<[],f32> -> !torch.float | |
%2379 = torch.aten.item %2377 : !torch.vtensor<[],si8> -> !torch.int | |
%2380 = torch.aten._make_per_tensor_quantized_tensor %2375, %2378, %2379 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
%2381 = torch.aten.dequantize.self %2380 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
%2382 = torch.aten.sigmoid %2381 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%2383 = torch.aten.mul.Tensor %2381, %2382 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
%2384 = torch.prim.ListConstruct %2271, %2285, %2383 : (!torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>) -> !torch.list<vtensor> | |
%int1_517 = torch.constant.int 1 | |
%2385 = torch.aten.cat %2384, %int1_517 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,96,80,80],f32> | |
%2386 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2387 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_518 = torch.constant.int 12 | |
%2388 = torch.aten.item %2386 : !torch.vtensor<[],f32> -> !torch.float | |
%2389 = torch.aten.item %2387 : !torch.vtensor<[],si8> -> !torch.int | |
%2390 = torch.aten.quantize_per_tensor %2385, %2388, %2389, %int12_518 : !torch.vtensor<[1,96,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,96,80,80],!torch.qint8> | |
%2391 = torch.aten.int_repr %2390 : !torch.vtensor<[1,96,80,80],!torch.qint8> -> !torch.vtensor<[1,96,80,80],si8> | |
%2392 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2393 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2394 = torch.aten.item %2392 : !torch.vtensor<[],f32> -> !torch.float | |
%2395 = torch.aten.item %2393 : !torch.vtensor<[],si8> -> !torch.int | |
%2396 = torch.aten._make_per_tensor_quantized_tensor %2391, %2394, %2395 : !torch.vtensor<[1,96,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,96,80,80],!torch.qint8> | |
%2397 = torch.aten.dequantize.self %2396 : !torch.vtensor<[1,96,80,80],!torch.qint8> -> !torch.vtensor<[1,96,80,80],f32> | |
%2398 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2399 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_519 = torch.constant.int 12 | |
%2400 = torch.aten.item %2398 : !torch.vtensor<[],f32> -> !torch.float | |
%2401 = torch.aten.item %2399 : !torch.vtensor<[],si8> -> !torch.int | |
%2402 = torch.aten.quantize_per_tensor %68, %2400, %2401, %int12_519 : !torch.vtensor<[64,96,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,96,1,1],!torch.qint8> | |
%2403 = torch.aten.int_repr %2402 : !torch.vtensor<[64,96,1,1],!torch.qint8> -> !torch.vtensor<[64,96,1,1],si8> | |
%2404 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2405 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2406 = torch.aten.item %2404 : !torch.vtensor<[],f32> -> !torch.float | |
%2407 = torch.aten.item %2405 : !torch.vtensor<[],si8> -> !torch.int | |
%2408 = torch.aten._make_per_tensor_quantized_tensor %2403, %2406, %2407 : !torch.vtensor<[64,96,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,96,1,1],!torch.qint8> | |
%2409 = torch.aten.dequantize.self %2408 : !torch.vtensor<[64,96,1,1],!torch.qint8> -> !torch.vtensor<[64,96,1,1],f32> | |
%2410 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2411 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_520 = torch.constant.int 12 | |
%2412 = torch.aten.item %2410 : !torch.vtensor<[],f32> -> !torch.float | |
%2413 = torch.aten.item %2411 : !torch.vtensor<[],si8> -> !torch.int | |
%2414 = torch.aten.quantize_per_tensor %69, %2412, %2413, %int12_520 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2415 = torch.aten.int_repr %2414 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2416 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2417 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2418 = torch.aten.item %2416 : !torch.vtensor<[],f32> -> !torch.float | |
%2419 = torch.aten.item %2417 : !torch.vtensor<[],si8> -> !torch.int | |
%2420 = torch.aten._make_per_tensor_quantized_tensor %2415, %2418, %2419 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2421 = torch.aten.dequantize.self %2420 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_521 = torch.constant.int 0 | |
%int0_522 = torch.constant.int 0 | |
%int1_523 = torch.constant.int 1 | |
%int1_524 = torch.constant.int 1 | |
%int1_525 = torch.constant.int 1 | |
%int1_526 = torch.constant.int 1 | |
%int0_527 = torch.constant.int 0 | |
%2422 = torch.prim.ListConstruct %int0_521, %int0_522 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2423 = torch.prim.ListConstruct %int1_523, %int1_524 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2424 = torch.prim.ListConstruct %int1_525, %int1_526 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2425 = torch.prim.ListConstruct %int0_527, %int0_527 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_528 = torch.constant.bool false | |
%int1_529 = torch.constant.int 1 | |
%2426 = torch.aten.convolution %2397, %2409, %2421, %2424, %2422, %2423, %false_528, %2425, %int1_529 : !torch.vtensor<[1,96,80,80],f32>, !torch.vtensor<[64,96,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%2427 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2428 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_530 = torch.constant.int 12 | |
%2429 = torch.aten.item %2427 : !torch.vtensor<[],f32> -> !torch.float | |
%2430 = torch.aten.item %2428 : !torch.vtensor<[],si8> -> !torch.int | |
%2431 = torch.aten.quantize_per_tensor %2426, %2429, %2430, %int12_530 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2432 = torch.aten.int_repr %2431 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%2433 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2434 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2435 = torch.aten.item %2433 : !torch.vtensor<[],f32> -> !torch.float | |
%2436 = torch.aten.item %2434 : !torch.vtensor<[],si8> -> !torch.int | |
%2437 = torch.aten._make_per_tensor_quantized_tensor %2432, %2435, %2436 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2438 = torch.aten.dequantize.self %2437 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%2439 = torch.aten.sigmoid %2438 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%2440 = torch.aten.mul.Tensor %2438, %2439 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%2441 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2442 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_531 = torch.constant.int 12 | |
%2443 = torch.aten.item %2441 : !torch.vtensor<[],f32> -> !torch.float | |
%2444 = torch.aten.item %2442 : !torch.vtensor<[],si8> -> !torch.int | |
%2445 = torch.aten.quantize_per_tensor %2440, %2443, %2444, %int12_531 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2446 = torch.aten.int_repr %2445 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%2447 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2448 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2449 = torch.aten.item %2447 : !torch.vtensor<[],f32> -> !torch.float | |
%2450 = torch.aten.item %2448 : !torch.vtensor<[],si8> -> !torch.int | |
%2451 = torch.aten._make_per_tensor_quantized_tensor %2446, %2449, %2450 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%2452 = torch.aten.dequantize.self %2451 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%2453 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2454 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_532 = torch.constant.int 12 | |
%2455 = torch.aten.item %2453 : !torch.vtensor<[],f32> -> !torch.float | |
%2456 = torch.aten.item %2454 : !torch.vtensor<[],si8> -> !torch.int | |
%2457 = torch.aten.quantize_per_tensor %70, %2455, %2456, %int12_532 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2458 = torch.aten.int_repr %2457 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%2459 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2460 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2461 = torch.aten.item %2459 : !torch.vtensor<[],f32> -> !torch.float | |
%2462 = torch.aten.item %2460 : !torch.vtensor<[],si8> -> !torch.int | |
%2463 = torch.aten._make_per_tensor_quantized_tensor %2458, %2461, %2462 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2464 = torch.aten.dequantize.self %2463 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%2465 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2466 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_533 = torch.constant.int 12 | |
%2467 = torch.aten.item %2465 : !torch.vtensor<[],f32> -> !torch.float | |
%2468 = torch.aten.item %2466 : !torch.vtensor<[],si8> -> !torch.int | |
%2469 = torch.aten.quantize_per_tensor %71, %2467, %2468, %int12_533 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2470 = torch.aten.int_repr %2469 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2471 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2472 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2473 = torch.aten.item %2471 : !torch.vtensor<[],f32> -> !torch.float | |
%2474 = torch.aten.item %2472 : !torch.vtensor<[],si8> -> !torch.int | |
%2475 = torch.aten._make_per_tensor_quantized_tensor %2470, %2473, %2474 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2476 = torch.aten.dequantize.self %2475 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_534 = torch.constant.int 1 | |
%int1_535 = torch.constant.int 1 | |
%int1_536 = torch.constant.int 1 | |
%int1_537 = torch.constant.int 1 | |
%int2_538 = torch.constant.int 2 | |
%int2_539 = torch.constant.int 2 | |
%int0_540 = torch.constant.int 0 | |
%2477 = torch.prim.ListConstruct %int1_534, %int1_535 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2478 = torch.prim.ListConstruct %int1_536, %int1_537 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2479 = torch.prim.ListConstruct %int2_538, %int2_539 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2480 = torch.prim.ListConstruct %int0_540, %int0_540 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_541 = torch.constant.bool false | |
%int1_542 = torch.constant.int 1 | |
%2481 = torch.aten.convolution %2452, %2464, %2476, %2479, %2477, %2478, %false_541, %2480, %int1_542 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2482 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2483 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_543 = torch.constant.int 12 | |
%2484 = torch.aten.item %2482 : !torch.vtensor<[],f32> -> !torch.float | |
%2485 = torch.aten.item %2483 : !torch.vtensor<[],si8> -> !torch.int | |
%2486 = torch.aten.quantize_per_tensor %2481, %2484, %2485, %int12_543 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2487 = torch.aten.int_repr %2486 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2488 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2489 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2490 = torch.aten.item %2488 : !torch.vtensor<[],f32> -> !torch.float | |
%2491 = torch.aten.item %2489 : !torch.vtensor<[],si8> -> !torch.int | |
%2492 = torch.aten._make_per_tensor_quantized_tensor %2487, %2490, %2491 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2493 = torch.aten.dequantize.self %2492 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2494 = torch.aten.sigmoid %2493 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2495 = torch.aten.mul.Tensor %2493, %2494 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2496 = torch.prim.ListConstruct %2495, %2183 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,128,40,40],f32>) -> !torch.list<vtensor> | |
%int1_544 = torch.constant.int 1 | |
%2497 = torch.aten.cat %2496, %int1_544 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,192,40,40],f32> | |
%2498 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2499 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_545 = torch.constant.int 12 | |
%2500 = torch.aten.item %2498 : !torch.vtensor<[],f32> -> !torch.float | |
%2501 = torch.aten.item %2499 : !torch.vtensor<[],si8> -> !torch.int | |
%2502 = torch.aten.quantize_per_tensor %2497, %2500, %2501, %int12_545 : !torch.vtensor<[1,192,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
%2503 = torch.aten.int_repr %2502 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],si8> | |
%2504 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2505 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2506 = torch.aten.item %2504 : !torch.vtensor<[],f32> -> !torch.float | |
%2507 = torch.aten.item %2505 : !torch.vtensor<[],si8> -> !torch.int | |
%2508 = torch.aten._make_per_tensor_quantized_tensor %2503, %2506, %2507 : !torch.vtensor<[1,192,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
%2509 = torch.aten.dequantize.self %2508 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],f32> | |
%2510 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2511 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_546 = torch.constant.int 12 | |
%2512 = torch.aten.item %2510 : !torch.vtensor<[],f32> -> !torch.float | |
%2513 = torch.aten.item %2511 : !torch.vtensor<[],si8> -> !torch.int | |
%2514 = torch.aten.quantize_per_tensor %72, %2512, %2513, %int12_546 : !torch.vtensor<[128,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
%2515 = torch.aten.int_repr %2514 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],si8> | |
%2516 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2517 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2518 = torch.aten.item %2516 : !torch.vtensor<[],f32> -> !torch.float | |
%2519 = torch.aten.item %2517 : !torch.vtensor<[],si8> -> !torch.int | |
%2520 = torch.aten._make_per_tensor_quantized_tensor %2515, %2518, %2519 : !torch.vtensor<[128,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
%2521 = torch.aten.dequantize.self %2520 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],f32> | |
%2522 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2523 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_547 = torch.constant.int 12 | |
%2524 = torch.aten.item %2522 : !torch.vtensor<[],f32> -> !torch.float | |
%2525 = torch.aten.item %2523 : !torch.vtensor<[],si8> -> !torch.int | |
%2526 = torch.aten.quantize_per_tensor %73, %2524, %2525, %int12_547 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2527 = torch.aten.int_repr %2526 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%2528 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2529 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2530 = torch.aten.item %2528 : !torch.vtensor<[],f32> -> !torch.float | |
%2531 = torch.aten.item %2529 : !torch.vtensor<[],si8> -> !torch.int | |
%2532 = torch.aten._make_per_tensor_quantized_tensor %2527, %2530, %2531 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2533 = torch.aten.dequantize.self %2532 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_548 = torch.constant.int 0 | |
%int0_549 = torch.constant.int 0 | |
%int1_550 = torch.constant.int 1 | |
%int1_551 = torch.constant.int 1 | |
%int1_552 = torch.constant.int 1 | |
%int1_553 = torch.constant.int 1 | |
%int0_554 = torch.constant.int 0 | |
%2534 = torch.prim.ListConstruct %int0_548, %int0_549 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2535 = torch.prim.ListConstruct %int1_550, %int1_551 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2536 = torch.prim.ListConstruct %int1_552, %int1_553 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2537 = torch.prim.ListConstruct %int0_554, %int0_554 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_555 = torch.constant.bool false | |
%int1_556 = torch.constant.int 1 | |
%2538 = torch.aten.convolution %2509, %2521, %2533, %2536, %2534, %2535, %false_555, %2537, %int1_556 : !torch.vtensor<[1,192,40,40],f32>, !torch.vtensor<[128,192,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%2539 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2540 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_557 = torch.constant.int 12 | |
%2541 = torch.aten.item %2539 : !torch.vtensor<[],f32> -> !torch.float | |
%2542 = torch.aten.item %2540 : !torch.vtensor<[],si8> -> !torch.int | |
%2543 = torch.aten.quantize_per_tensor %2538, %2541, %2542, %int12_557 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2544 = torch.aten.int_repr %2543 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%2545 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2546 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2547 = torch.aten.item %2545 : !torch.vtensor<[],f32> -> !torch.float | |
%2548 = torch.aten.item %2546 : !torch.vtensor<[],si8> -> !torch.int | |
%2549 = torch.aten._make_per_tensor_quantized_tensor %2544, %2547, %2548 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2550 = torch.aten.dequantize.self %2549 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%2551 = torch.aten.sigmoid %2550 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%2552 = torch.aten.mul.Tensor %2550, %2551 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%2553 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2554 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_558 = torch.constant.int 12 | |
%2555 = torch.aten.item %2553 : !torch.vtensor<[],f32> -> !torch.float | |
%2556 = torch.aten.item %2554 : !torch.vtensor<[],si8> -> !torch.int | |
%2557 = torch.aten.quantize_per_tensor %2552, %2555, %2556, %int12_558 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2558 = torch.aten.int_repr %2557 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%2559 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2560 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2561 = torch.aten.item %2559 : !torch.vtensor<[],f32> -> !torch.float | |
%2562 = torch.aten.item %2560 : !torch.vtensor<[],si8> -> !torch.int | |
%2563 = torch.aten._make_per_tensor_quantized_tensor %2558, %2561, %2562 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2564 = torch.aten.dequantize.self %2563 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%2565 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2566 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2567 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2568 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_559 = torch.constant.int 0 | |
%int0_560 = torch.constant.int 0 | |
%2569 = torch.prim.NumToTensor.Scalar %int0_560 : !torch.int -> !torch.vtensor<[1],si64> | |
%2570 = torch.aten.index_select %2566, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2571 = torch.aten.item %2570 : !torch.vtensor<[1],si64> -> !torch.int | |
%2572 = torch.aten.index_select %2567, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2573 = torch.aten.item %2572 : !torch.vtensor<[1],si64> -> !torch.int | |
%2574 = torch.aten.index_select %2565, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2575 = torch.aten.item %2574 : !torch.vtensor<[1],si64> -> !torch.int | |
%2576 = torch.aten.index_select %2568, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2577 = torch.aten.item %2576 : !torch.vtensor<[1],si64> -> !torch.int | |
%2578 = torch.aten.slice.Tensor %2564, %2575, %2571, %2573, %2577 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2579 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2580 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2581 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2582 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_561 = torch.constant.int 0 | |
%int0_562 = torch.constant.int 0 | |
%2583 = torch.prim.NumToTensor.Scalar %int0_562 : !torch.int -> !torch.vtensor<[1],si64> | |
%2584 = torch.aten.index_select %2580, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2585 = torch.aten.item %2584 : !torch.vtensor<[1],si64> -> !torch.int | |
%2586 = torch.aten.index_select %2581, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2587 = torch.aten.item %2586 : !torch.vtensor<[1],si64> -> !torch.int | |
%2588 = torch.aten.index_select %2579, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2589 = torch.aten.item %2588 : !torch.vtensor<[1],si64> -> !torch.int | |
%2590 = torch.aten.index_select %2582, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2591 = torch.aten.item %2590 : !torch.vtensor<[1],si64> -> !torch.int | |
%2592 = torch.aten.slice.Tensor %2564, %2589, %2585, %2587, %2591 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2593 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2594 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_563 = torch.constant.int 12 | |
%2595 = torch.aten.item %2593 : !torch.vtensor<[],f32> -> !torch.float | |
%2596 = torch.aten.item %2594 : !torch.vtensor<[],si8> -> !torch.int | |
%2597 = torch.aten.quantize_per_tensor %74, %2595, %2596, %int12_563 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2598 = torch.aten.int_repr %2597 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%2599 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2600 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2601 = torch.aten.item %2599 : !torch.vtensor<[],f32> -> !torch.float | |
%2602 = torch.aten.item %2600 : !torch.vtensor<[],si8> -> !torch.int | |
%2603 = torch.aten._make_per_tensor_quantized_tensor %2598, %2601, %2602 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2604 = torch.aten.dequantize.self %2603 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%2605 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2606 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_564 = torch.constant.int 12 | |
%2607 = torch.aten.item %2605 : !torch.vtensor<[],f32> -> !torch.float | |
%2608 = torch.aten.item %2606 : !torch.vtensor<[],si8> -> !torch.int | |
%2609 = torch.aten.quantize_per_tensor %75, %2607, %2608, %int12_564 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2610 = torch.aten.int_repr %2609 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2611 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2612 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2613 = torch.aten.item %2611 : !torch.vtensor<[],f32> -> !torch.float | |
%2614 = torch.aten.item %2612 : !torch.vtensor<[],si8> -> !torch.int | |
%2615 = torch.aten._make_per_tensor_quantized_tensor %2610, %2613, %2614 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2616 = torch.aten.dequantize.self %2615 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_565 = torch.constant.int 1 | |
%int1_566 = torch.constant.int 1 | |
%int1_567 = torch.constant.int 1 | |
%int1_568 = torch.constant.int 1 | |
%int1_569 = torch.constant.int 1 | |
%int1_570 = torch.constant.int 1 | |
%int0_571 = torch.constant.int 0 | |
%2617 = torch.prim.ListConstruct %int1_565, %int1_566 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2618 = torch.prim.ListConstruct %int1_567, %int1_568 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2619 = torch.prim.ListConstruct %int1_569, %int1_570 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2620 = torch.prim.ListConstruct %int0_571, %int0_571 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_572 = torch.constant.bool false | |
%int1_573 = torch.constant.int 1 | |
%2621 = torch.aten.convolution %2592, %2604, %2616, %2619, %2617, %2618, %false_572, %2620, %int1_573 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2622 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2623 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_574 = torch.constant.int 12 | |
%2624 = torch.aten.item %2622 : !torch.vtensor<[],f32> -> !torch.float | |
%2625 = torch.aten.item %2623 : !torch.vtensor<[],si8> -> !torch.int | |
%2626 = torch.aten.quantize_per_tensor %2621, %2624, %2625, %int12_574 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2627 = torch.aten.int_repr %2626 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2628 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2629 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2630 = torch.aten.item %2628 : !torch.vtensor<[],f32> -> !torch.float | |
%2631 = torch.aten.item %2629 : !torch.vtensor<[],si8> -> !torch.int | |
%2632 = torch.aten._make_per_tensor_quantized_tensor %2627, %2630, %2631 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2633 = torch.aten.dequantize.self %2632 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2634 = torch.aten.sigmoid %2633 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2635 = torch.aten.mul.Tensor %2633, %2634 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2636 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2637 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_575 = torch.constant.int 12 | |
%2638 = torch.aten.item %2636 : !torch.vtensor<[],f32> -> !torch.float | |
%2639 = torch.aten.item %2637 : !torch.vtensor<[],si8> -> !torch.int | |
%2640 = torch.aten.quantize_per_tensor %2635, %2638, %2639, %int12_575 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2641 = torch.aten.int_repr %2640 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2642 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2643 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2644 = torch.aten.item %2642 : !torch.vtensor<[],f32> -> !torch.float | |
%2645 = torch.aten.item %2643 : !torch.vtensor<[],si8> -> !torch.int | |
%2646 = torch.aten._make_per_tensor_quantized_tensor %2641, %2644, %2645 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2647 = torch.aten.dequantize.self %2646 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2648 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2649 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_576 = torch.constant.int 12 | |
%2650 = torch.aten.item %2648 : !torch.vtensor<[],f32> -> !torch.float | |
%2651 = torch.aten.item %2649 : !torch.vtensor<[],si8> -> !torch.int | |
%2652 = torch.aten.quantize_per_tensor %76, %2650, %2651, %int12_576 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2653 = torch.aten.int_repr %2652 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%2654 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2655 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2656 = torch.aten.item %2654 : !torch.vtensor<[],f32> -> !torch.float | |
%2657 = torch.aten.item %2655 : !torch.vtensor<[],si8> -> !torch.int | |
%2658 = torch.aten._make_per_tensor_quantized_tensor %2653, %2656, %2657 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%2659 = torch.aten.dequantize.self %2658 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%2660 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2661 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_577 = torch.constant.int 12 | |
%2662 = torch.aten.item %2660 : !torch.vtensor<[],f32> -> !torch.float | |
%2663 = torch.aten.item %2661 : !torch.vtensor<[],si8> -> !torch.int | |
%2664 = torch.aten.quantize_per_tensor %77, %2662, %2663, %int12_577 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2665 = torch.aten.int_repr %2664 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%2666 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2667 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2668 = torch.aten.item %2666 : !torch.vtensor<[],f32> -> !torch.float | |
%2669 = torch.aten.item %2667 : !torch.vtensor<[],si8> -> !torch.int | |
%2670 = torch.aten._make_per_tensor_quantized_tensor %2665, %2668, %2669 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%2671 = torch.aten.dequantize.self %2670 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_578 = torch.constant.int 1 | |
%int1_579 = torch.constant.int 1 | |
%int1_580 = torch.constant.int 1 | |
%int1_581 = torch.constant.int 1 | |
%int1_582 = torch.constant.int 1 | |
%int1_583 = torch.constant.int 1 | |
%int0_584 = torch.constant.int 0 | |
%2672 = torch.prim.ListConstruct %int1_578, %int1_579 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2673 = torch.prim.ListConstruct %int1_580, %int1_581 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2674 = torch.prim.ListConstruct %int1_582, %int1_583 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2675 = torch.prim.ListConstruct %int0_584, %int0_584 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_585 = torch.constant.bool false | |
%int1_586 = torch.constant.int 1 | |
%2676 = torch.aten.convolution %2647, %2659, %2671, %2674, %2672, %2673, %false_585, %2675, %int1_586 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%2677 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2678 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_587 = torch.constant.int 12 | |
%2679 = torch.aten.item %2677 : !torch.vtensor<[],f32> -> !torch.float | |
%2680 = torch.aten.item %2678 : !torch.vtensor<[],si8> -> !torch.int | |
%2681 = torch.aten.quantize_per_tensor %2676, %2679, %2680, %int12_587 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2682 = torch.aten.int_repr %2681 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%2683 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2684 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2685 = torch.aten.item %2683 : !torch.vtensor<[],f32> -> !torch.float | |
%2686 = torch.aten.item %2684 : !torch.vtensor<[],si8> -> !torch.int | |
%2687 = torch.aten._make_per_tensor_quantized_tensor %2682, %2685, %2686 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%2688 = torch.aten.dequantize.self %2687 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%2689 = torch.aten.sigmoid %2688 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2690 = torch.aten.mul.Tensor %2688, %2689 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%2691 = torch.prim.ListConstruct %2578, %2592, %2690 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>) -> !torch.list<vtensor> | |
%int1_588 = torch.constant.int 1 | |
%2692 = torch.aten.cat %2691, %int1_588 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,192,40,40],f32> | |
%2693 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2694 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_589 = torch.constant.int 12 | |
%2695 = torch.aten.item %2693 : !torch.vtensor<[],f32> -> !torch.float | |
%2696 = torch.aten.item %2694 : !torch.vtensor<[],si8> -> !torch.int | |
%2697 = torch.aten.quantize_per_tensor %2692, %2695, %2696, %int12_589 : !torch.vtensor<[1,192,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
%2698 = torch.aten.int_repr %2697 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],si8> | |
%2699 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2700 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2701 = torch.aten.item %2699 : !torch.vtensor<[],f32> -> !torch.float | |
%2702 = torch.aten.item %2700 : !torch.vtensor<[],si8> -> !torch.int | |
%2703 = torch.aten._make_per_tensor_quantized_tensor %2698, %2701, %2702 : !torch.vtensor<[1,192,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
%2704 = torch.aten.dequantize.self %2703 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],f32> | |
%2705 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2706 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_590 = torch.constant.int 12 | |
%2707 = torch.aten.item %2705 : !torch.vtensor<[],f32> -> !torch.float | |
%2708 = torch.aten.item %2706 : !torch.vtensor<[],si8> -> !torch.int | |
%2709 = torch.aten.quantize_per_tensor %78, %2707, %2708, %int12_590 : !torch.vtensor<[128,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
%2710 = torch.aten.int_repr %2709 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],si8> | |
%2711 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2712 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2713 = torch.aten.item %2711 : !torch.vtensor<[],f32> -> !torch.float | |
%2714 = torch.aten.item %2712 : !torch.vtensor<[],si8> -> !torch.int | |
%2715 = torch.aten._make_per_tensor_quantized_tensor %2710, %2713, %2714 : !torch.vtensor<[128,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
%2716 = torch.aten.dequantize.self %2715 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],f32> | |
%2717 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2718 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_591 = torch.constant.int 12 | |
%2719 = torch.aten.item %2717 : !torch.vtensor<[],f32> -> !torch.float | |
%2720 = torch.aten.item %2718 : !torch.vtensor<[],si8> -> !torch.int | |
%2721 = torch.aten.quantize_per_tensor %79, %2719, %2720, %int12_591 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2722 = torch.aten.int_repr %2721 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%2723 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2724 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2725 = torch.aten.item %2723 : !torch.vtensor<[],f32> -> !torch.float | |
%2726 = torch.aten.item %2724 : !torch.vtensor<[],si8> -> !torch.int | |
%2727 = torch.aten._make_per_tensor_quantized_tensor %2722, %2725, %2726 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2728 = torch.aten.dequantize.self %2727 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int0_592 = torch.constant.int 0 | |
%int0_593 = torch.constant.int 0 | |
%int1_594 = torch.constant.int 1 | |
%int1_595 = torch.constant.int 1 | |
%int1_596 = torch.constant.int 1 | |
%int1_597 = torch.constant.int 1 | |
%int0_598 = torch.constant.int 0 | |
%2729 = torch.prim.ListConstruct %int0_592, %int0_593 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2730 = torch.prim.ListConstruct %int1_594, %int1_595 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2731 = torch.prim.ListConstruct %int1_596, %int1_597 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2732 = torch.prim.ListConstruct %int0_598, %int0_598 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_599 = torch.constant.bool false | |
%int1_600 = torch.constant.int 1 | |
%2733 = torch.aten.convolution %2704, %2716, %2728, %2731, %2729, %2730, %false_599, %2732, %int1_600 : !torch.vtensor<[1,192,40,40],f32>, !torch.vtensor<[128,192,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
%2734 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2735 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_601 = torch.constant.int 12 | |
%2736 = torch.aten.item %2734 : !torch.vtensor<[],f32> -> !torch.float | |
%2737 = torch.aten.item %2735 : !torch.vtensor<[],si8> -> !torch.int | |
%2738 = torch.aten.quantize_per_tensor %2733, %2736, %2737, %int12_601 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2739 = torch.aten.int_repr %2738 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%2740 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2741 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2742 = torch.aten.item %2740 : !torch.vtensor<[],f32> -> !torch.float | |
%2743 = torch.aten.item %2741 : !torch.vtensor<[],si8> -> !torch.int | |
%2744 = torch.aten._make_per_tensor_quantized_tensor %2739, %2742, %2743 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2745 = torch.aten.dequantize.self %2744 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%2746 = torch.aten.sigmoid %2745 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%2747 = torch.aten.mul.Tensor %2745, %2746 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
%2748 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2749 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_602 = torch.constant.int 12 | |
%2750 = torch.aten.item %2748 : !torch.vtensor<[],f32> -> !torch.float | |
%2751 = torch.aten.item %2749 : !torch.vtensor<[],si8> -> !torch.int | |
%2752 = torch.aten.quantize_per_tensor %2747, %2750, %2751, %int12_602 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2753 = torch.aten.int_repr %2752 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
%2754 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2755 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2756 = torch.aten.item %2754 : !torch.vtensor<[],f32> -> !torch.float | |
%2757 = torch.aten.item %2755 : !torch.vtensor<[],si8> -> !torch.int | |
%2758 = torch.aten._make_per_tensor_quantized_tensor %2753, %2756, %2757 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
%2759 = torch.aten.dequantize.self %2758 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
%2760 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2761 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_603 = torch.constant.int 12 | |
%2762 = torch.aten.item %2760 : !torch.vtensor<[],f32> -> !torch.float | |
%2763 = torch.aten.item %2761 : !torch.vtensor<[],si8> -> !torch.int | |
%2764 = torch.aten.quantize_per_tensor %80, %2762, %2763, %int12_603 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%2765 = torch.aten.int_repr %2764 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
%2766 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2767 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2768 = torch.aten.item %2766 : !torch.vtensor<[],f32> -> !torch.float | |
%2769 = torch.aten.item %2767 : !torch.vtensor<[],si8> -> !torch.int | |
%2770 = torch.aten._make_per_tensor_quantized_tensor %2765, %2768, %2769 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%2771 = torch.aten.dequantize.self %2770 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
%2772 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2773 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_604 = torch.constant.int 12 | |
%2774 = torch.aten.item %2772 : !torch.vtensor<[],f32> -> !torch.float | |
%2775 = torch.aten.item %2773 : !torch.vtensor<[],si8> -> !torch.int | |
%2776 = torch.aten.quantize_per_tensor %81, %2774, %2775, %int12_604 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2777 = torch.aten.int_repr %2776 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%2778 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2779 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2780 = torch.aten.item %2778 : !torch.vtensor<[],f32> -> !torch.float | |
%2781 = torch.aten.item %2779 : !torch.vtensor<[],si8> -> !torch.int | |
%2782 = torch.aten._make_per_tensor_quantized_tensor %2777, %2780, %2781 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2783 = torch.aten.dequantize.self %2782 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int1_605 = torch.constant.int 1 | |
%int1_606 = torch.constant.int 1 | |
%int1_607 = torch.constant.int 1 | |
%int1_608 = torch.constant.int 1 | |
%int2_609 = torch.constant.int 2 | |
%int2_610 = torch.constant.int 2 | |
%int0_611 = torch.constant.int 0 | |
%2784 = torch.prim.ListConstruct %int1_605, %int1_606 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2785 = torch.prim.ListConstruct %int1_607, %int1_608 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2786 = torch.prim.ListConstruct %int2_609, %int2_610 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2787 = torch.prim.ListConstruct %int0_611, %int0_611 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_612 = torch.constant.bool false | |
%int1_613 = torch.constant.int 1 | |
%2788 = torch.aten.convolution %2759, %2771, %2783, %2786, %2784, %2785, %false_612, %2787, %int1_613 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%2789 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2790 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_614 = torch.constant.int 12 | |
%2791 = torch.aten.item %2789 : !torch.vtensor<[],f32> -> !torch.float | |
%2792 = torch.aten.item %2790 : !torch.vtensor<[],si8> -> !torch.int | |
%2793 = torch.aten.quantize_per_tensor %2788, %2791, %2792, %int12_614 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2794 = torch.aten.int_repr %2793 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%2795 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2796 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2797 = torch.aten.item %2795 : !torch.vtensor<[],f32> -> !torch.float | |
%2798 = torch.aten.item %2796 : !torch.vtensor<[],si8> -> !torch.int | |
%2799 = torch.aten._make_per_tensor_quantized_tensor %2794, %2797, %2798 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2800 = torch.aten.dequantize.self %2799 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%2801 = torch.aten.sigmoid %2800 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%2802 = torch.aten.mul.Tensor %2800, %2801 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%2803 = torch.prim.ListConstruct %2802, %1914 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,256,20,20],f32>) -> !torch.list<vtensor> | |
%int1_615 = torch.constant.int 1 | |
%2804 = torch.aten.cat %2803, %int1_615 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,384,20,20],f32> | |
%2805 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2806 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_616 = torch.constant.int 12 | |
%2807 = torch.aten.item %2805 : !torch.vtensor<[],f32> -> !torch.float | |
%2808 = torch.aten.item %2806 : !torch.vtensor<[],si8> -> !torch.int | |
%2809 = torch.aten.quantize_per_tensor %2804, %2807, %2808, %int12_616 : !torch.vtensor<[1,384,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
%2810 = torch.aten.int_repr %2809 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],si8> | |
%2811 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2812 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2813 = torch.aten.item %2811 : !torch.vtensor<[],f32> -> !torch.float | |
%2814 = torch.aten.item %2812 : !torch.vtensor<[],si8> -> !torch.int | |
%2815 = torch.aten._make_per_tensor_quantized_tensor %2810, %2813, %2814 : !torch.vtensor<[1,384,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
%2816 = torch.aten.dequantize.self %2815 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],f32> | |
%2817 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2818 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_617 = torch.constant.int 12 | |
%2819 = torch.aten.item %2817 : !torch.vtensor<[],f32> -> !torch.float | |
%2820 = torch.aten.item %2818 : !torch.vtensor<[],si8> -> !torch.int | |
%2821 = torch.aten.quantize_per_tensor %82, %2819, %2820, %int12_617 : !torch.vtensor<[256,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
%2822 = torch.aten.int_repr %2821 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],si8> | |
%2823 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2824 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2825 = torch.aten.item %2823 : !torch.vtensor<[],f32> -> !torch.float | |
%2826 = torch.aten.item %2824 : !torch.vtensor<[],si8> -> !torch.int | |
%2827 = torch.aten._make_per_tensor_quantized_tensor %2822, %2825, %2826 : !torch.vtensor<[256,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
%2828 = torch.aten.dequantize.self %2827 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],f32> | |
%2829 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2830 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_618 = torch.constant.int 12 | |
%2831 = torch.aten.item %2829 : !torch.vtensor<[],f32> -> !torch.float | |
%2832 = torch.aten.item %2830 : !torch.vtensor<[],si8> -> !torch.int | |
%2833 = torch.aten.quantize_per_tensor %83, %2831, %2832, %int12_618 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%2834 = torch.aten.int_repr %2833 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
%2835 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2836 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2837 = torch.aten.item %2835 : !torch.vtensor<[],f32> -> !torch.float | |
%2838 = torch.aten.item %2836 : !torch.vtensor<[],si8> -> !torch.int | |
%2839 = torch.aten._make_per_tensor_quantized_tensor %2834, %2837, %2838 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%2840 = torch.aten.dequantize.self %2839 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
%int0_619 = torch.constant.int 0 | |
%int0_620 = torch.constant.int 0 | |
%int1_621 = torch.constant.int 1 | |
%int1_622 = torch.constant.int 1 | |
%int1_623 = torch.constant.int 1 | |
%int1_624 = torch.constant.int 1 | |
%int0_625 = torch.constant.int 0 | |
%2841 = torch.prim.ListConstruct %int0_619, %int0_620 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2842 = torch.prim.ListConstruct %int1_621, %int1_622 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2843 = torch.prim.ListConstruct %int1_623, %int1_624 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2844 = torch.prim.ListConstruct %int0_625, %int0_625 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_626 = torch.constant.bool false | |
%int1_627 = torch.constant.int 1 | |
%2845 = torch.aten.convolution %2816, %2828, %2840, %2843, %2841, %2842, %false_626, %2844, %int1_627 : !torch.vtensor<[1,384,20,20],f32>, !torch.vtensor<[256,384,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
%2846 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2847 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_628 = torch.constant.int 12 | |
%2848 = torch.aten.item %2846 : !torch.vtensor<[],f32> -> !torch.float | |
%2849 = torch.aten.item %2847 : !torch.vtensor<[],si8> -> !torch.int | |
%2850 = torch.aten.quantize_per_tensor %2845, %2848, %2849, %int12_628 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%2851 = torch.aten.int_repr %2850 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%2852 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2853 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2854 = torch.aten.item %2852 : !torch.vtensor<[],f32> -> !torch.float | |
%2855 = torch.aten.item %2853 : !torch.vtensor<[],si8> -> !torch.int | |
%2856 = torch.aten._make_per_tensor_quantized_tensor %2851, %2854, %2855 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%2857 = torch.aten.dequantize.self %2856 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%2858 = torch.aten.sigmoid %2857 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%2859 = torch.aten.mul.Tensor %2857, %2858 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%2860 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2861 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_629 = torch.constant.int 12 | |
%2862 = torch.aten.item %2860 : !torch.vtensor<[],f32> -> !torch.float | |
%2863 = torch.aten.item %2861 : !torch.vtensor<[],si8> -> !torch.int | |
%2864 = torch.aten.quantize_per_tensor %2859, %2862, %2863, %int12_629 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%2865 = torch.aten.int_repr %2864 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%2866 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2867 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2868 = torch.aten.item %2866 : !torch.vtensor<[],f32> -> !torch.float | |
%2869 = torch.aten.item %2867 : !torch.vtensor<[],si8> -> !torch.int | |
%2870 = torch.aten._make_per_tensor_quantized_tensor %2865, %2868, %2869 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%2871 = torch.aten.dequantize.self %2870 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%2872 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2873 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2874 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2875 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_630 = torch.constant.int 0 | |
%int0_631 = torch.constant.int 0 | |
%2876 = torch.prim.NumToTensor.Scalar %int0_631 : !torch.int -> !torch.vtensor<[1],si64> | |
%2877 = torch.aten.index_select %2873, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2878 = torch.aten.item %2877 : !torch.vtensor<[1],si64> -> !torch.int | |
%2879 = torch.aten.index_select %2874, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2880 = torch.aten.item %2879 : !torch.vtensor<[1],si64> -> !torch.int | |
%2881 = torch.aten.index_select %2872, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2882 = torch.aten.item %2881 : !torch.vtensor<[1],si64> -> !torch.int | |
%2883 = torch.aten.index_select %2875, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2884 = torch.aten.item %2883 : !torch.vtensor<[1],si64> -> !torch.int | |
%2885 = torch.aten.slice.Tensor %2871, %2882, %2878, %2880, %2884 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%2886 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2887 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2888 = torch.vtensor.literal(dense<256> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%2889 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_632 = torch.constant.int 0 | |
%int0_633 = torch.constant.int 0 | |
%2890 = torch.prim.NumToTensor.Scalar %int0_633 : !torch.int -> !torch.vtensor<[1],si64> | |
%2891 = torch.aten.index_select %2887, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2892 = torch.aten.item %2891 : !torch.vtensor<[1],si64> -> !torch.int | |
%2893 = torch.aten.index_select %2888, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2894 = torch.aten.item %2893 : !torch.vtensor<[1],si64> -> !torch.int | |
%2895 = torch.aten.index_select %2886, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2896 = torch.aten.item %2895 : !torch.vtensor<[1],si64> -> !torch.int | |
%2897 = torch.aten.index_select %2889, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%2898 = torch.aten.item %2897 : !torch.vtensor<[1],si64> -> !torch.int | |
%2899 = torch.aten.slice.Tensor %2871, %2896, %2892, %2894, %2898 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%2900 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2901 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_634 = torch.constant.int 12 | |
%2902 = torch.aten.item %2900 : !torch.vtensor<[],f32> -> !torch.float | |
%2903 = torch.aten.item %2901 : !torch.vtensor<[],si8> -> !torch.int | |
%2904 = torch.aten.quantize_per_tensor %84, %2902, %2903, %int12_634 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%2905 = torch.aten.int_repr %2904 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
%2906 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2907 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2908 = torch.aten.item %2906 : !torch.vtensor<[],f32> -> !torch.float | |
%2909 = torch.aten.item %2907 : !torch.vtensor<[],si8> -> !torch.int | |
%2910 = torch.aten._make_per_tensor_quantized_tensor %2905, %2908, %2909 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%2911 = torch.aten.dequantize.self %2910 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
%2912 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2913 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_635 = torch.constant.int 12 | |
%2914 = torch.aten.item %2912 : !torch.vtensor<[],f32> -> !torch.float | |
%2915 = torch.aten.item %2913 : !torch.vtensor<[],si8> -> !torch.int | |
%2916 = torch.aten.quantize_per_tensor %85, %2914, %2915, %int12_635 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2917 = torch.aten.int_repr %2916 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%2918 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2919 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2920 = torch.aten.item %2918 : !torch.vtensor<[],f32> -> !torch.float | |
%2921 = torch.aten.item %2919 : !torch.vtensor<[],si8> -> !torch.int | |
%2922 = torch.aten._make_per_tensor_quantized_tensor %2917, %2920, %2921 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2923 = torch.aten.dequantize.self %2922 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int1_636 = torch.constant.int 1 | |
%int1_637 = torch.constant.int 1 | |
%int1_638 = torch.constant.int 1 | |
%int1_639 = torch.constant.int 1 | |
%int1_640 = torch.constant.int 1 | |
%int1_641 = torch.constant.int 1 | |
%int0_642 = torch.constant.int 0 | |
%2924 = torch.prim.ListConstruct %int1_636, %int1_637 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2925 = torch.prim.ListConstruct %int1_638, %int1_639 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2926 = torch.prim.ListConstruct %int1_640, %int1_641 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2927 = torch.prim.ListConstruct %int0_642, %int0_642 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_643 = torch.constant.bool false | |
%int1_644 = torch.constant.int 1 | |
%2928 = torch.aten.convolution %2899, %2911, %2923, %2926, %2924, %2925, %false_643, %2927, %int1_644 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%2929 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2930 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_645 = torch.constant.int 12 | |
%2931 = torch.aten.item %2929 : !torch.vtensor<[],f32> -> !torch.float | |
%2932 = torch.aten.item %2930 : !torch.vtensor<[],si8> -> !torch.int | |
%2933 = torch.aten.quantize_per_tensor %2928, %2931, %2932, %int12_645 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2934 = torch.aten.int_repr %2933 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%2935 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2936 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2937 = torch.aten.item %2935 : !torch.vtensor<[],f32> -> !torch.float | |
%2938 = torch.aten.item %2936 : !torch.vtensor<[],si8> -> !torch.int | |
%2939 = torch.aten._make_per_tensor_quantized_tensor %2934, %2937, %2938 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2940 = torch.aten.dequantize.self %2939 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%2941 = torch.aten.sigmoid %2940 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%2942 = torch.aten.mul.Tensor %2940, %2941 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%2943 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2944 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_646 = torch.constant.int 12 | |
%2945 = torch.aten.item %2943 : !torch.vtensor<[],f32> -> !torch.float | |
%2946 = torch.aten.item %2944 : !torch.vtensor<[],si8> -> !torch.int | |
%2947 = torch.aten.quantize_per_tensor %2942, %2945, %2946, %int12_646 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2948 = torch.aten.int_repr %2947 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%2949 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2950 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2951 = torch.aten.item %2949 : !torch.vtensor<[],f32> -> !torch.float | |
%2952 = torch.aten.item %2950 : !torch.vtensor<[],si8> -> !torch.int | |
%2953 = torch.aten._make_per_tensor_quantized_tensor %2948, %2951, %2952 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2954 = torch.aten.dequantize.self %2953 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%2955 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2956 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_647 = torch.constant.int 12 | |
%2957 = torch.aten.item %2955 : !torch.vtensor<[],f32> -> !torch.float | |
%2958 = torch.aten.item %2956 : !torch.vtensor<[],si8> -> !torch.int | |
%2959 = torch.aten.quantize_per_tensor %86, %2957, %2958, %int12_647 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%2960 = torch.aten.int_repr %2959 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
%2961 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2962 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2963 = torch.aten.item %2961 : !torch.vtensor<[],f32> -> !torch.float | |
%2964 = torch.aten.item %2962 : !torch.vtensor<[],si8> -> !torch.int | |
%2965 = torch.aten._make_per_tensor_quantized_tensor %2960, %2963, %2964 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
%2966 = torch.aten.dequantize.self %2965 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
%2967 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2968 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_648 = torch.constant.int 12 | |
%2969 = torch.aten.item %2967 : !torch.vtensor<[],f32> -> !torch.float | |
%2970 = torch.aten.item %2968 : !torch.vtensor<[],si8> -> !torch.int | |
%2971 = torch.aten.quantize_per_tensor %87, %2969, %2970, %int12_648 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2972 = torch.aten.int_repr %2971 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
%2973 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2974 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2975 = torch.aten.item %2973 : !torch.vtensor<[],f32> -> !torch.float | |
%2976 = torch.aten.item %2974 : !torch.vtensor<[],si8> -> !torch.int | |
%2977 = torch.aten._make_per_tensor_quantized_tensor %2972, %2975, %2976 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
%2978 = torch.aten.dequantize.self %2977 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
%int1_649 = torch.constant.int 1 | |
%int1_650 = torch.constant.int 1 | |
%int1_651 = torch.constant.int 1 | |
%int1_652 = torch.constant.int 1 | |
%int1_653 = torch.constant.int 1 | |
%int1_654 = torch.constant.int 1 | |
%int0_655 = torch.constant.int 0 | |
%2979 = torch.prim.ListConstruct %int1_649, %int1_650 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2980 = torch.prim.ListConstruct %int1_651, %int1_652 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2981 = torch.prim.ListConstruct %int1_653, %int1_654 : (!torch.int, !torch.int) -> !torch.list<int> | |
%2982 = torch.prim.ListConstruct %int0_655, %int0_655 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_656 = torch.constant.bool false | |
%int1_657 = torch.constant.int 1 | |
%2983 = torch.aten.convolution %2954, %2966, %2978, %2981, %2979, %2980, %false_656, %2982, %int1_657 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
%2984 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2985 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_658 = torch.constant.int 12 | |
%2986 = torch.aten.item %2984 : !torch.vtensor<[],f32> -> !torch.float | |
%2987 = torch.aten.item %2985 : !torch.vtensor<[],si8> -> !torch.int | |
%2988 = torch.aten.quantize_per_tensor %2983, %2986, %2987, %int12_658 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2989 = torch.aten.int_repr %2988 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
%2990 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%2991 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%2992 = torch.aten.item %2990 : !torch.vtensor<[],f32> -> !torch.float | |
%2993 = torch.aten.item %2991 : !torch.vtensor<[],si8> -> !torch.int | |
%2994 = torch.aten._make_per_tensor_quantized_tensor %2989, %2992, %2993 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
%2995 = torch.aten.dequantize.self %2994 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
%2996 = torch.aten.sigmoid %2995 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%2997 = torch.aten.mul.Tensor %2995, %2996 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
%2998 = torch.prim.ListConstruct %2885, %2899, %2997 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>) -> !torch.list<vtensor> | |
%int1_659 = torch.constant.int 1 | |
%2999 = torch.aten.cat %2998, %int1_659 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,384,20,20],f32> | |
%3000 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3001 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_660 = torch.constant.int 12 | |
%3002 = torch.aten.item %3000 : !torch.vtensor<[],f32> -> !torch.float | |
%3003 = torch.aten.item %3001 : !torch.vtensor<[],si8> -> !torch.int | |
%3004 = torch.aten.quantize_per_tensor %2999, %3002, %3003, %int12_660 : !torch.vtensor<[1,384,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
%3005 = torch.aten.int_repr %3004 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],si8> | |
%3006 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3007 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3008 = torch.aten.item %3006 : !torch.vtensor<[],f32> -> !torch.float | |
%3009 = torch.aten.item %3007 : !torch.vtensor<[],si8> -> !torch.int | |
%3010 = torch.aten._make_per_tensor_quantized_tensor %3005, %3008, %3009 : !torch.vtensor<[1,384,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
%3011 = torch.aten.dequantize.self %3010 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],f32> | |
%3012 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3013 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_661 = torch.constant.int 12 | |
%3014 = torch.aten.item %3012 : !torch.vtensor<[],f32> -> !torch.float | |
%3015 = torch.aten.item %3013 : !torch.vtensor<[],si8> -> !torch.int | |
%3016 = torch.aten.quantize_per_tensor %88, %3014, %3015, %int12_661 : !torch.vtensor<[256,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
%3017 = torch.aten.int_repr %3016 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],si8> | |
%3018 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3019 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3020 = torch.aten.item %3018 : !torch.vtensor<[],f32> -> !torch.float | |
%3021 = torch.aten.item %3019 : !torch.vtensor<[],si8> -> !torch.int | |
%3022 = torch.aten._make_per_tensor_quantized_tensor %3017, %3020, %3021 : !torch.vtensor<[256,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
%3023 = torch.aten.dequantize.self %3022 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],f32> | |
%3024 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3025 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_662 = torch.constant.int 12 | |
%3026 = torch.aten.item %3024 : !torch.vtensor<[],f32> -> !torch.float | |
%3027 = torch.aten.item %3025 : !torch.vtensor<[],si8> -> !torch.int | |
%3028 = torch.aten.quantize_per_tensor %89, %3026, %3027, %int12_662 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%3029 = torch.aten.int_repr %3028 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
%3030 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3031 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3032 = torch.aten.item %3030 : !torch.vtensor<[],f32> -> !torch.float | |
%3033 = torch.aten.item %3031 : !torch.vtensor<[],si8> -> !torch.int | |
%3034 = torch.aten._make_per_tensor_quantized_tensor %3029, %3032, %3033 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
%3035 = torch.aten.dequantize.self %3034 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
%int0_663 = torch.constant.int 0 | |
%int0_664 = torch.constant.int 0 | |
%int1_665 = torch.constant.int 1 | |
%int1_666 = torch.constant.int 1 | |
%int1_667 = torch.constant.int 1 | |
%int1_668 = torch.constant.int 1 | |
%int0_669 = torch.constant.int 0 | |
%3036 = torch.prim.ListConstruct %int0_663, %int0_664 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3037 = torch.prim.ListConstruct %int1_665, %int1_666 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3038 = torch.prim.ListConstruct %int1_667, %int1_668 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3039 = torch.prim.ListConstruct %int0_669, %int0_669 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_670 = torch.constant.bool false | |
%int1_671 = torch.constant.int 1 | |
%3040 = torch.aten.convolution %3011, %3023, %3035, %3038, %3036, %3037, %false_670, %3039, %int1_671 : !torch.vtensor<[1,384,20,20],f32>, !torch.vtensor<[256,384,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
%3041 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3042 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_672 = torch.constant.int 12 | |
%3043 = torch.aten.item %3041 : !torch.vtensor<[],f32> -> !torch.float | |
%3044 = torch.aten.item %3042 : !torch.vtensor<[],si8> -> !torch.int | |
%3045 = torch.aten.quantize_per_tensor %3040, %3043, %3044, %int12_672 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%3046 = torch.aten.int_repr %3045 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%3047 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3048 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3049 = torch.aten.item %3047 : !torch.vtensor<[],f32> -> !torch.float | |
%3050 = torch.aten.item %3048 : !torch.vtensor<[],si8> -> !torch.int | |
%3051 = torch.aten._make_per_tensor_quantized_tensor %3046, %3049, %3050 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%3052 = torch.aten.dequantize.self %3051 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%3053 = torch.aten.sigmoid %3052 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%3054 = torch.aten.mul.Tensor %3052, %3053 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
%3055 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3056 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_673 = torch.constant.int 12 | |
%3057 = torch.aten.item %3055 : !torch.vtensor<[],f32> -> !torch.float | |
%3058 = torch.aten.item %3056 : !torch.vtensor<[],si8> -> !torch.int | |
%3059 = torch.aten.quantize_per_tensor %3054, %3057, %3058, %int12_673 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%3060 = torch.aten.int_repr %3059 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
%3061 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3062 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3063 = torch.aten.item %3061 : !torch.vtensor<[],f32> -> !torch.float | |
%3064 = torch.aten.item %3062 : !torch.vtensor<[],si8> -> !torch.int | |
%3065 = torch.aten._make_per_tensor_quantized_tensor %3060, %3063, %3064 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
%3066 = torch.aten.dequantize.self %3065 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
%3067 = torch.vtensor.literal(dense<1> : tensor<si64>) : !torch.vtensor<[],si64> | |
%3068 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3069 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_674 = torch.constant.int 12 | |
%3070 = torch.aten.item %3068 : !torch.vtensor<[],f32> -> !torch.float | |
%3071 = torch.aten.item %3069 : !torch.vtensor<[],si8> -> !torch.int | |
%3072 = torch.aten.quantize_per_tensor %90, %3070, %3071, %int12_674 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3073 = torch.aten.int_repr %3072 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%3074 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3075 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3076 = torch.aten.item %3074 : !torch.vtensor<[],f32> -> !torch.float | |
%3077 = torch.aten.item %3075 : !torch.vtensor<[],si8> -> !torch.int | |
%3078 = torch.aten._make_per_tensor_quantized_tensor %3073, %3076, %3077 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3079 = torch.aten.dequantize.self %3078 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%3080 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3081 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_675 = torch.constant.int 12 | |
%3082 = torch.aten.item %3080 : !torch.vtensor<[],f32> -> !torch.float | |
%3083 = torch.aten.item %3081 : !torch.vtensor<[],si8> -> !torch.int | |
%3084 = torch.aten.quantize_per_tensor %91, %3082, %3083, %int12_675 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3085 = torch.aten.int_repr %3084 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3086 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3087 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3088 = torch.aten.item %3086 : !torch.vtensor<[],f32> -> !torch.float | |
%3089 = torch.aten.item %3087 : !torch.vtensor<[],si8> -> !torch.int | |
%3090 = torch.aten._make_per_tensor_quantized_tensor %3085, %3088, %3089 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3091 = torch.aten.dequantize.self %3090 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_676 = torch.constant.int 1 | |
%int1_677 = torch.constant.int 1 | |
%int1_678 = torch.constant.int 1 | |
%int1_679 = torch.constant.int 1 | |
%int1_680 = torch.constant.int 1 | |
%int1_681 = torch.constant.int 1 | |
%int0_682 = torch.constant.int 0 | |
%3092 = torch.prim.ListConstruct %int1_676, %int1_677 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3093 = torch.prim.ListConstruct %int1_678, %int1_679 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3094 = torch.prim.ListConstruct %int1_680, %int1_681 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3095 = torch.prim.ListConstruct %int0_682, %int0_682 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_683 = torch.constant.bool false | |
%int1_684 = torch.constant.int 1 | |
%3096 = torch.aten.convolution %2452, %3079, %3091, %3094, %3092, %3093, %false_683, %3095, %int1_684 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%3097 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3098 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_685 = torch.constant.int 12 | |
%3099 = torch.aten.item %3097 : !torch.vtensor<[],f32> -> !torch.float | |
%3100 = torch.aten.item %3098 : !torch.vtensor<[],si8> -> !torch.int | |
%3101 = torch.aten.quantize_per_tensor %3096, %3099, %3100, %int12_685 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3102 = torch.aten.int_repr %3101 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%3103 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3104 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3105 = torch.aten.item %3103 : !torch.vtensor<[],f32> -> !torch.float | |
%3106 = torch.aten.item %3104 : !torch.vtensor<[],si8> -> !torch.int | |
%3107 = torch.aten._make_per_tensor_quantized_tensor %3102, %3105, %3106 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3108 = torch.aten.dequantize.self %3107 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%3109 = torch.aten.sigmoid %3108 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%3110 = torch.aten.mul.Tensor %3108, %3109 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%3111 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3112 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_686 = torch.constant.int 12 | |
%3113 = torch.aten.item %3111 : !torch.vtensor<[],f32> -> !torch.float | |
%3114 = torch.aten.item %3112 : !torch.vtensor<[],si8> -> !torch.int | |
%3115 = torch.aten.quantize_per_tensor %3110, %3113, %3114, %int12_686 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3116 = torch.aten.int_repr %3115 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%3117 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3118 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3119 = torch.aten.item %3117 : !torch.vtensor<[],f32> -> !torch.float | |
%3120 = torch.aten.item %3118 : !torch.vtensor<[],si8> -> !torch.int | |
%3121 = torch.aten._make_per_tensor_quantized_tensor %3116, %3119, %3120 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3122 = torch.aten.dequantize.self %3121 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%3123 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3124 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_687 = torch.constant.int 12 | |
%3125 = torch.aten.item %3123 : !torch.vtensor<[],f32> -> !torch.float | |
%3126 = torch.aten.item %3124 : !torch.vtensor<[],si8> -> !torch.int | |
%3127 = torch.aten.quantize_per_tensor %92, %3125, %3126, %int12_687 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3128 = torch.aten.int_repr %3127 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%3129 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3130 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3131 = torch.aten.item %3129 : !torch.vtensor<[],f32> -> !torch.float | |
%3132 = torch.aten.item %3130 : !torch.vtensor<[],si8> -> !torch.int | |
%3133 = torch.aten._make_per_tensor_quantized_tensor %3128, %3131, %3132 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3134 = torch.aten.dequantize.self %3133 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%3135 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3136 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_688 = torch.constant.int 12 | |
%3137 = torch.aten.item %3135 : !torch.vtensor<[],f32> -> !torch.float | |
%3138 = torch.aten.item %3136 : !torch.vtensor<[],si8> -> !torch.int | |
%3139 = torch.aten.quantize_per_tensor %93, %3137, %3138, %int12_688 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3140 = torch.aten.int_repr %3139 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3141 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3142 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3143 = torch.aten.item %3141 : !torch.vtensor<[],f32> -> !torch.float | |
%3144 = torch.aten.item %3142 : !torch.vtensor<[],si8> -> !torch.int | |
%3145 = torch.aten._make_per_tensor_quantized_tensor %3140, %3143, %3144 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3146 = torch.aten.dequantize.self %3145 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_689 = torch.constant.int 1 | |
%int1_690 = torch.constant.int 1 | |
%int1_691 = torch.constant.int 1 | |
%int1_692 = torch.constant.int 1 | |
%int1_693 = torch.constant.int 1 | |
%int1_694 = torch.constant.int 1 | |
%int0_695 = torch.constant.int 0 | |
%3147 = torch.prim.ListConstruct %int1_689, %int1_690 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3148 = torch.prim.ListConstruct %int1_691, %int1_692 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3149 = torch.prim.ListConstruct %int1_693, %int1_694 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3150 = torch.prim.ListConstruct %int0_695, %int0_695 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_696 = torch.constant.bool false | |
%int1_697 = torch.constant.int 1 | |
%3151 = torch.aten.convolution %3122, %3134, %3146, %3149, %3147, %3148, %false_696, %3150, %int1_697 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%3152 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3153 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_698 = torch.constant.int 12 | |
%3154 = torch.aten.item %3152 : !torch.vtensor<[],f32> -> !torch.float | |
%3155 = torch.aten.item %3153 : !torch.vtensor<[],si8> -> !torch.int | |
%3156 = torch.aten.quantize_per_tensor %3151, %3154, %3155, %int12_698 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3157 = torch.aten.int_repr %3156 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%3158 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3159 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3160 = torch.aten.item %3158 : !torch.vtensor<[],f32> -> !torch.float | |
%3161 = torch.aten.item %3159 : !torch.vtensor<[],si8> -> !torch.int | |
%3162 = torch.aten._make_per_tensor_quantized_tensor %3157, %3160, %3161 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3163 = torch.aten.dequantize.self %3162 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%3164 = torch.aten.sigmoid %3163 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%3165 = torch.aten.mul.Tensor %3163, %3164 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
%3166 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3167 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_699 = torch.constant.int 12 | |
%3168 = torch.aten.item %3166 : !torch.vtensor<[],f32> -> !torch.float | |
%3169 = torch.aten.item %3167 : !torch.vtensor<[],si8> -> !torch.int | |
%3170 = torch.aten.quantize_per_tensor %3165, %3168, %3169, %int12_699 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3171 = torch.aten.int_repr %3170 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
%3172 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3174 = torch.aten.item %3172 : !torch.vtensor<[],f32> -> !torch.float | |
%3175 = torch.aten.item %3173 : !torch.vtensor<[],si8> -> !torch.int | |
%3176 = torch.aten._make_per_tensor_quantized_tensor %3171, %3174, %3175 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
%3177 = torch.aten.dequantize.self %3176 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
%3178 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_700 = torch.constant.int 12 | |
%3180 = torch.aten.item %3178 : !torch.vtensor<[],f32> -> !torch.float | |
%3181 = torch.aten.item %3179 : !torch.vtensor<[],si8> -> !torch.int | |
%3182 = torch.aten.quantize_per_tensor %94, %3180, %3181, %int12_700 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%3183 = torch.aten.int_repr %3182 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
%3184 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3185 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3186 = torch.aten.item %3184 : !torch.vtensor<[],f32> -> !torch.float | |
%3187 = torch.aten.item %3185 : !torch.vtensor<[],si8> -> !torch.int | |
%3188 = torch.aten._make_per_tensor_quantized_tensor %3183, %3186, %3187 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%3189 = torch.aten.dequantize.self %3188 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
%3190 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3191 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_701 = torch.constant.int 12 | |
%3192 = torch.aten.item %3190 : !torch.vtensor<[],f32> -> !torch.float | |
%3193 = torch.aten.item %3191 : !torch.vtensor<[],si8> -> !torch.int | |
%3194 = torch.aten.quantize_per_tensor %95, %3192, %3193, %int12_701 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3195 = torch.aten.int_repr %3194 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3196 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3197 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3198 = torch.aten.item %3196 : !torch.vtensor<[],f32> -> !torch.float | |
%3199 = torch.aten.item %3197 : !torch.vtensor<[],si8> -> !torch.int | |
%3200 = torch.aten._make_per_tensor_quantized_tensor %3195, %3198, %3199 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3201 = torch.aten.dequantize.self %3200 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_702 = torch.constant.int 0 | |
%int0_703 = torch.constant.int 0 | |
%int1_704 = torch.constant.int 1 | |
%int1_705 = torch.constant.int 1 | |
%int1_706 = torch.constant.int 1 | |
%int1_707 = torch.constant.int 1 | |
%int0_708 = torch.constant.int 0 | |
%3202 = torch.prim.ListConstruct %int0_702, %int0_703 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3203 = torch.prim.ListConstruct %int1_704, %int1_705 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3204 = torch.prim.ListConstruct %int1_706, %int1_707 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3205 = torch.prim.ListConstruct %int0_708, %int0_708 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_709 = torch.constant.bool false | |
%int1_710 = torch.constant.int 1 | |
%3206 = torch.aten.convolution %3177, %3189, %3201, %3204, %3202, %3203, %false_709, %3205, %int1_710 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
%3207 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3208 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_711 = torch.constant.int 12 | |
%3209 = torch.aten.item %3207 : !torch.vtensor<[],f32> -> !torch.float | |
%3210 = torch.aten.item %3208 : !torch.vtensor<[],si8> -> !torch.int | |
%3211 = torch.aten.quantize_per_tensor %96, %3209, %3210, %int12_711 : !torch.vtensor<[80,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,64,3,3],!torch.qint8> | |
%3212 = torch.aten.int_repr %3211 : !torch.vtensor<[80,64,3,3],!torch.qint8> -> !torch.vtensor<[80,64,3,3],si8> | |
%3213 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3214 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3215 = torch.aten.item %3213 : !torch.vtensor<[],f32> -> !torch.float | |
%3216 = torch.aten.item %3214 : !torch.vtensor<[],si8> -> !torch.int | |
%3217 = torch.aten._make_per_tensor_quantized_tensor %3212, %3215, %3216 : !torch.vtensor<[80,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,64,3,3],!torch.qint8> | |
%3218 = torch.aten.dequantize.self %3217 : !torch.vtensor<[80,64,3,3],!torch.qint8> -> !torch.vtensor<[80,64,3,3],f32> | |
%3219 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3220 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_712 = torch.constant.int 12 | |
%3221 = torch.aten.item %3219 : !torch.vtensor<[],f32> -> !torch.float | |
%3222 = torch.aten.item %3220 : !torch.vtensor<[],si8> -> !torch.int | |
%3223 = torch.aten.quantize_per_tensor %97, %3221, %3222, %int12_712 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3224 = torch.aten.int_repr %3223 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3225 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3226 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3227 = torch.aten.item %3225 : !torch.vtensor<[],f32> -> !torch.float | |
%3228 = torch.aten.item %3226 : !torch.vtensor<[],si8> -> !torch.int | |
%3229 = torch.aten._make_per_tensor_quantized_tensor %3224, %3227, %3228 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3230 = torch.aten.dequantize.self %3229 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int1_713 = torch.constant.int 1 | |
%int1_714 = torch.constant.int 1 | |
%int1_715 = torch.constant.int 1 | |
%int1_716 = torch.constant.int 1 | |
%int1_717 = torch.constant.int 1 | |
%int1_718 = torch.constant.int 1 | |
%int0_719 = torch.constant.int 0 | |
%3231 = torch.prim.ListConstruct %int1_713, %int1_714 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3232 = torch.prim.ListConstruct %int1_715, %int1_716 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3233 = torch.prim.ListConstruct %int1_717, %int1_718 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3234 = torch.prim.ListConstruct %int0_719, %int0_719 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_720 = torch.constant.bool false | |
%int1_721 = torch.constant.int 1 | |
%3235 = torch.aten.convolution %2452, %3218, %3230, %3233, %3231, %3232, %false_720, %3234, %int1_721 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[80,64,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,80,80],f32> | |
%3236 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3237 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_722 = torch.constant.int 12 | |
%3238 = torch.aten.item %3236 : !torch.vtensor<[],f32> -> !torch.float | |
%3239 = torch.aten.item %3237 : !torch.vtensor<[],si8> -> !torch.int | |
%3240 = torch.aten.quantize_per_tensor %3235, %3238, %3239, %int12_722 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3241 = torch.aten.int_repr %3240 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
%3242 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3243 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3244 = torch.aten.item %3242 : !torch.vtensor<[],f32> -> !torch.float | |
%3245 = torch.aten.item %3243 : !torch.vtensor<[],si8> -> !torch.int | |
%3246 = torch.aten._make_per_tensor_quantized_tensor %3241, %3244, %3245 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3247 = torch.aten.dequantize.self %3246 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
%3248 = torch.aten.sigmoid %3247 : !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
%3249 = torch.aten.mul.Tensor %3247, %3248 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
%3250 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3251 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_723 = torch.constant.int 12 | |
%3252 = torch.aten.item %3250 : !torch.vtensor<[],f32> -> !torch.float | |
%3253 = torch.aten.item %3251 : !torch.vtensor<[],si8> -> !torch.int | |
%3254 = torch.aten.quantize_per_tensor %3249, %3252, %3253, %int12_723 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3255 = torch.aten.int_repr %3254 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
%3256 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3257 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3258 = torch.aten.item %3256 : !torch.vtensor<[],f32> -> !torch.float | |
%3259 = torch.aten.item %3257 : !torch.vtensor<[],si8> -> !torch.int | |
%3260 = torch.aten._make_per_tensor_quantized_tensor %3255, %3258, %3259 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3261 = torch.aten.dequantize.self %3260 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
%3262 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3263 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_724 = torch.constant.int 12 | |
%3264 = torch.aten.item %3262 : !torch.vtensor<[],f32> -> !torch.float | |
%3265 = torch.aten.item %3263 : !torch.vtensor<[],si8> -> !torch.int | |
%3266 = torch.aten.quantize_per_tensor %98, %3264, %3265, %int12_724 : !torch.vtensor<[80,80,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
%3267 = torch.aten.int_repr %3266 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],si8> | |
%3268 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3269 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3270 = torch.aten.item %3268 : !torch.vtensor<[],f32> -> !torch.float | |
%3271 = torch.aten.item %3269 : !torch.vtensor<[],si8> -> !torch.int | |
%3272 = torch.aten._make_per_tensor_quantized_tensor %3267, %3270, %3271 : !torch.vtensor<[80,80,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
%3273 = torch.aten.dequantize.self %3272 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],f32> | |
%3274 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3275 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_725 = torch.constant.int 12 | |
%3276 = torch.aten.item %3274 : !torch.vtensor<[],f32> -> !torch.float | |
%3277 = torch.aten.item %3275 : !torch.vtensor<[],si8> -> !torch.int | |
%3278 = torch.aten.quantize_per_tensor %99, %3276, %3277, %int12_725 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3279 = torch.aten.int_repr %3278 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3280 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3281 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3282 = torch.aten.item %3280 : !torch.vtensor<[],f32> -> !torch.float | |
%3283 = torch.aten.item %3281 : !torch.vtensor<[],si8> -> !torch.int | |
%3284 = torch.aten._make_per_tensor_quantized_tensor %3279, %3282, %3283 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3285 = torch.aten.dequantize.self %3284 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int1_726 = torch.constant.int 1 | |
%int1_727 = torch.constant.int 1 | |
%int1_728 = torch.constant.int 1 | |
%int1_729 = torch.constant.int 1 | |
%int1_730 = torch.constant.int 1 | |
%int1_731 = torch.constant.int 1 | |
%int0_732 = torch.constant.int 0 | |
%3286 = torch.prim.ListConstruct %int1_726, %int1_727 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3287 = torch.prim.ListConstruct %int1_728, %int1_729 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3288 = torch.prim.ListConstruct %int1_730, %int1_731 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3289 = torch.prim.ListConstruct %int0_732, %int0_732 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_733 = torch.constant.bool false | |
%int1_734 = torch.constant.int 1 | |
%3290 = torch.aten.convolution %3261, %3273, %3285, %3288, %3286, %3287, %false_733, %3289, %int1_734 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[80,80,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,80,80],f32> | |
%3291 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3292 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_735 = torch.constant.int 12 | |
%3293 = torch.aten.item %3291 : !torch.vtensor<[],f32> -> !torch.float | |
%3294 = torch.aten.item %3292 : !torch.vtensor<[],si8> -> !torch.int | |
%3295 = torch.aten.quantize_per_tensor %3290, %3293, %3294, %int12_735 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3296 = torch.aten.int_repr %3295 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
%3297 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3298 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3299 = torch.aten.item %3297 : !torch.vtensor<[],f32> -> !torch.float | |
%3300 = torch.aten.item %3298 : !torch.vtensor<[],si8> -> !torch.int | |
%3301 = torch.aten._make_per_tensor_quantized_tensor %3296, %3299, %3300 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3302 = torch.aten.dequantize.self %3301 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
%3303 = torch.aten.sigmoid %3302 : !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
%3304 = torch.aten.mul.Tensor %3302, %3303 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
%3305 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3306 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_736 = torch.constant.int 12 | |
%3307 = torch.aten.item %3305 : !torch.vtensor<[],f32> -> !torch.float | |
%3308 = torch.aten.item %3306 : !torch.vtensor<[],si8> -> !torch.int | |
%3309 = torch.aten.quantize_per_tensor %3304, %3307, %3308, %int12_736 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3310 = torch.aten.int_repr %3309 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
%3311 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3312 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3313 = torch.aten.item %3311 : !torch.vtensor<[],f32> -> !torch.float | |
%3314 = torch.aten.item %3312 : !torch.vtensor<[],si8> -> !torch.int | |
%3315 = torch.aten._make_per_tensor_quantized_tensor %3310, %3313, %3314 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
%3316 = torch.aten.dequantize.self %3315 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
%3317 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3318 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_737 = torch.constant.int 12 | |
%3319 = torch.aten.item %3317 : !torch.vtensor<[],f32> -> !torch.float | |
%3320 = torch.aten.item %3318 : !torch.vtensor<[],si8> -> !torch.int | |
%3321 = torch.aten.quantize_per_tensor %100, %3319, %3320, %int12_737 : !torch.vtensor<[80,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
%3322 = torch.aten.int_repr %3321 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],si8> | |
%3323 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3324 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3325 = torch.aten.item %3323 : !torch.vtensor<[],f32> -> !torch.float | |
%3326 = torch.aten.item %3324 : !torch.vtensor<[],si8> -> !torch.int | |
%3327 = torch.aten._make_per_tensor_quantized_tensor %3322, %3325, %3326 : !torch.vtensor<[80,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
%3328 = torch.aten.dequantize.self %3327 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],f32> | |
%3329 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3330 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_738 = torch.constant.int 12 | |
%3331 = torch.aten.item %3329 : !torch.vtensor<[],f32> -> !torch.float | |
%3332 = torch.aten.item %3330 : !torch.vtensor<[],si8> -> !torch.int | |
%3333 = torch.aten.quantize_per_tensor %101, %3331, %3332, %int12_738 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3334 = torch.aten.int_repr %3333 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3335 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3336 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3337 = torch.aten.item %3335 : !torch.vtensor<[],f32> -> !torch.float | |
%3338 = torch.aten.item %3336 : !torch.vtensor<[],si8> -> !torch.int | |
%3339 = torch.aten._make_per_tensor_quantized_tensor %3334, %3337, %3338 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3340 = torch.aten.dequantize.self %3339 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int0_739 = torch.constant.int 0 | |
%int0_740 = torch.constant.int 0 | |
%int1_741 = torch.constant.int 1 | |
%int1_742 = torch.constant.int 1 | |
%int1_743 = torch.constant.int 1 | |
%int1_744 = torch.constant.int 1 | |
%int0_745 = torch.constant.int 0 | |
%3341 = torch.prim.ListConstruct %int0_739, %int0_740 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3342 = torch.prim.ListConstruct %int1_741, %int1_742 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3343 = torch.prim.ListConstruct %int1_743, %int1_744 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3344 = torch.prim.ListConstruct %int0_745, %int0_745 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_746 = torch.constant.bool false | |
%int1_747 = torch.constant.int 1 | |
%3345 = torch.aten.convolution %3316, %3328, %3340, %3343, %3341, %3342, %false_746, %3344, %int1_747 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[80,80,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,80,80],f32> | |
%3346 = torch.prim.ListConstruct %3206, %3345 : (!torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,80,80,80],f32>) -> !torch.list<vtensor> | |
%int1_748 = torch.constant.int 1 | |
%3347 = torch.aten.cat %3346, %int1_748 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,144,80,80],f32> | |
%3348 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3349 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_749 = torch.constant.int 12 | |
%3350 = torch.aten.item %3348 : !torch.vtensor<[],f32> -> !torch.float | |
%3351 = torch.aten.item %3349 : !torch.vtensor<[],si8> -> !torch.int | |
%3352 = torch.aten.quantize_per_tensor %3347, %3350, %3351, %int12_749 : !torch.vtensor<[1,144,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,144,80,80],!torch.qint8> | |
%3353 = torch.aten.int_repr %3352 : !torch.vtensor<[1,144,80,80],!torch.qint8> -> !torch.vtensor<[1,144,80,80],si8> | |
%3354 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3355 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3356 = torch.aten.item %3354 : !torch.vtensor<[],f32> -> !torch.float | |
%3357 = torch.aten.item %3355 : !torch.vtensor<[],si8> -> !torch.int | |
%3358 = torch.aten._make_per_tensor_quantized_tensor %3353, %3356, %3357 : !torch.vtensor<[1,144,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,144,80,80],!torch.qint8> | |
%3359 = torch.aten.dequantize.self %3358 : !torch.vtensor<[1,144,80,80],!torch.qint8> -> !torch.vtensor<[1,144,80,80],f32> | |
%3360 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3361 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_750 = torch.constant.int 12 | |
%3362 = torch.aten.item %3360 : !torch.vtensor<[],f32> -> !torch.float | |
%3363 = torch.aten.item %3361 : !torch.vtensor<[],si8> -> !torch.int | |
%3364 = torch.aten.quantize_per_tensor %102, %3362, %3363, %int12_750 : !torch.vtensor<[64,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,128,3,3],!torch.qint8> | |
%3365 = torch.aten.int_repr %3364 : !torch.vtensor<[64,128,3,3],!torch.qint8> -> !torch.vtensor<[64,128,3,3],si8> | |
%3366 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3367 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3368 = torch.aten.item %3366 : !torch.vtensor<[],f32> -> !torch.float | |
%3369 = torch.aten.item %3367 : !torch.vtensor<[],si8> -> !torch.int | |
%3370 = torch.aten._make_per_tensor_quantized_tensor %3365, %3368, %3369 : !torch.vtensor<[64,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,128,3,3],!torch.qint8> | |
%3371 = torch.aten.dequantize.self %3370 : !torch.vtensor<[64,128,3,3],!torch.qint8> -> !torch.vtensor<[64,128,3,3],f32> | |
%3372 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3373 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_751 = torch.constant.int 12 | |
%3374 = torch.aten.item %3372 : !torch.vtensor<[],f32> -> !torch.float | |
%3375 = torch.aten.item %3373 : !torch.vtensor<[],si8> -> !torch.int | |
%3376 = torch.aten.quantize_per_tensor %103, %3374, %3375, %int12_751 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3377 = torch.aten.int_repr %3376 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3378 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3379 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3380 = torch.aten.item %3378 : !torch.vtensor<[],f32> -> !torch.float | |
%3381 = torch.aten.item %3379 : !torch.vtensor<[],si8> -> !torch.int | |
%3382 = torch.aten._make_per_tensor_quantized_tensor %3377, %3380, %3381 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3383 = torch.aten.dequantize.self %3382 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_752 = torch.constant.int 1 | |
%int1_753 = torch.constant.int 1 | |
%int1_754 = torch.constant.int 1 | |
%int1_755 = torch.constant.int 1 | |
%int1_756 = torch.constant.int 1 | |
%int1_757 = torch.constant.int 1 | |
%int0_758 = torch.constant.int 0 | |
%3384 = torch.prim.ListConstruct %int1_752, %int1_753 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3385 = torch.prim.ListConstruct %int1_754, %int1_755 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3386 = torch.prim.ListConstruct %int1_756, %int1_757 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3387 = torch.prim.ListConstruct %int0_758, %int0_758 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_759 = torch.constant.bool false | |
%int1_760 = torch.constant.int 1 | |
%3388 = torch.aten.convolution %2759, %3371, %3383, %3386, %3384, %3385, %false_759, %3387, %int1_760 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[64,128,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%3389 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3390 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_761 = torch.constant.int 12 | |
%3391 = torch.aten.item %3389 : !torch.vtensor<[],f32> -> !torch.float | |
%3392 = torch.aten.item %3390 : !torch.vtensor<[],si8> -> !torch.int | |
%3393 = torch.aten.quantize_per_tensor %3388, %3391, %3392, %int12_761 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3394 = torch.aten.int_repr %3393 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%3395 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3396 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3397 = torch.aten.item %3395 : !torch.vtensor<[],f32> -> !torch.float | |
%3398 = torch.aten.item %3396 : !torch.vtensor<[],si8> -> !torch.int | |
%3399 = torch.aten._make_per_tensor_quantized_tensor %3394, %3397, %3398 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3400 = torch.aten.dequantize.self %3399 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%3401 = torch.aten.sigmoid %3400 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%3402 = torch.aten.mul.Tensor %3400, %3401 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%3403 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3404 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_762 = torch.constant.int 12 | |
%3405 = torch.aten.item %3403 : !torch.vtensor<[],f32> -> !torch.float | |
%3406 = torch.aten.item %3404 : !torch.vtensor<[],si8> -> !torch.int | |
%3407 = torch.aten.quantize_per_tensor %3402, %3405, %3406, %int12_762 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3408 = torch.aten.int_repr %3407 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%3409 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3410 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3411 = torch.aten.item %3409 : !torch.vtensor<[],f32> -> !torch.float | |
%3412 = torch.aten.item %3410 : !torch.vtensor<[],si8> -> !torch.int | |
%3413 = torch.aten._make_per_tensor_quantized_tensor %3408, %3411, %3412 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3414 = torch.aten.dequantize.self %3413 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%3415 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3416 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_763 = torch.constant.int 12 | |
%3417 = torch.aten.item %3415 : !torch.vtensor<[],f32> -> !torch.float | |
%3418 = torch.aten.item %3416 : !torch.vtensor<[],si8> -> !torch.int | |
%3419 = torch.aten.quantize_per_tensor %104, %3417, %3418, %int12_763 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3420 = torch.aten.int_repr %3419 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%3421 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3422 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3423 = torch.aten.item %3421 : !torch.vtensor<[],f32> -> !torch.float | |
%3424 = torch.aten.item %3422 : !torch.vtensor<[],si8> -> !torch.int | |
%3425 = torch.aten._make_per_tensor_quantized_tensor %3420, %3423, %3424 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3426 = torch.aten.dequantize.self %3425 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%3427 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3428 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_764 = torch.constant.int 12 | |
%3429 = torch.aten.item %3427 : !torch.vtensor<[],f32> -> !torch.float | |
%3430 = torch.aten.item %3428 : !torch.vtensor<[],si8> -> !torch.int | |
%3431 = torch.aten.quantize_per_tensor %105, %3429, %3430, %int12_764 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3432 = torch.aten.int_repr %3431 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3433 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3434 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3435 = torch.aten.item %3433 : !torch.vtensor<[],f32> -> !torch.float | |
%3436 = torch.aten.item %3434 : !torch.vtensor<[],si8> -> !torch.int | |
%3437 = torch.aten._make_per_tensor_quantized_tensor %3432, %3435, %3436 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3438 = torch.aten.dequantize.self %3437 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_765 = torch.constant.int 1 | |
%int1_766 = torch.constant.int 1 | |
%int1_767 = torch.constant.int 1 | |
%int1_768 = torch.constant.int 1 | |
%int1_769 = torch.constant.int 1 | |
%int1_770 = torch.constant.int 1 | |
%int0_771 = torch.constant.int 0 | |
%3439 = torch.prim.ListConstruct %int1_765, %int1_766 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3440 = torch.prim.ListConstruct %int1_767, %int1_768 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3441 = torch.prim.ListConstruct %int1_769, %int1_770 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3442 = torch.prim.ListConstruct %int0_771, %int0_771 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_772 = torch.constant.bool false | |
%int1_773 = torch.constant.int 1 | |
%3443 = torch.aten.convolution %3414, %3426, %3438, %3441, %3439, %3440, %false_772, %3442, %int1_773 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%3444 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3445 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_774 = torch.constant.int 12 | |
%3446 = torch.aten.item %3444 : !torch.vtensor<[],f32> -> !torch.float | |
%3447 = torch.aten.item %3445 : !torch.vtensor<[],si8> -> !torch.int | |
%3448 = torch.aten.quantize_per_tensor %3443, %3446, %3447, %int12_774 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3449 = torch.aten.int_repr %3448 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%3450 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3451 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3452 = torch.aten.item %3450 : !torch.vtensor<[],f32> -> !torch.float | |
%3453 = torch.aten.item %3451 : !torch.vtensor<[],si8> -> !torch.int | |
%3454 = torch.aten._make_per_tensor_quantized_tensor %3449, %3452, %3453 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3455 = torch.aten.dequantize.self %3454 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%3456 = torch.aten.sigmoid %3455 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%3457 = torch.aten.mul.Tensor %3455, %3456 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
%3458 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3459 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_775 = torch.constant.int 12 | |
%3460 = torch.aten.item %3458 : !torch.vtensor<[],f32> -> !torch.float | |
%3461 = torch.aten.item %3459 : !torch.vtensor<[],si8> -> !torch.int | |
%3462 = torch.aten.quantize_per_tensor %3457, %3460, %3461, %int12_775 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3463 = torch.aten.int_repr %3462 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
%3464 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3465 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3466 = torch.aten.item %3464 : !torch.vtensor<[],f32> -> !torch.float | |
%3467 = torch.aten.item %3465 : !torch.vtensor<[],si8> -> !torch.int | |
%3468 = torch.aten._make_per_tensor_quantized_tensor %3463, %3466, %3467 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
%3469 = torch.aten.dequantize.self %3468 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
%3470 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3471 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_776 = torch.constant.int 12 | |
%3472 = torch.aten.item %3470 : !torch.vtensor<[],f32> -> !torch.float | |
%3473 = torch.aten.item %3471 : !torch.vtensor<[],si8> -> !torch.int | |
%3474 = torch.aten.quantize_per_tensor %106, %3472, %3473, %int12_776 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%3475 = torch.aten.int_repr %3474 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
%3476 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3477 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3478 = torch.aten.item %3476 : !torch.vtensor<[],f32> -> !torch.float | |
%3479 = torch.aten.item %3477 : !torch.vtensor<[],si8> -> !torch.int | |
%3480 = torch.aten._make_per_tensor_quantized_tensor %3475, %3478, %3479 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%3481 = torch.aten.dequantize.self %3480 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
%3482 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3483 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_777 = torch.constant.int 12 | |
%3484 = torch.aten.item %3482 : !torch.vtensor<[],f32> -> !torch.float | |
%3485 = torch.aten.item %3483 : !torch.vtensor<[],si8> -> !torch.int | |
%3486 = torch.aten.quantize_per_tensor %107, %3484, %3485, %int12_777 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3487 = torch.aten.int_repr %3486 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3488 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3489 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3490 = torch.aten.item %3488 : !torch.vtensor<[],f32> -> !torch.float | |
%3491 = torch.aten.item %3489 : !torch.vtensor<[],si8> -> !torch.int | |
%3492 = torch.aten._make_per_tensor_quantized_tensor %3487, %3490, %3491 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3493 = torch.aten.dequantize.self %3492 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_778 = torch.constant.int 0 | |
%int0_779 = torch.constant.int 0 | |
%int1_780 = torch.constant.int 1 | |
%int1_781 = torch.constant.int 1 | |
%int1_782 = torch.constant.int 1 | |
%int1_783 = torch.constant.int 1 | |
%int0_784 = torch.constant.int 0 | |
%3494 = torch.prim.ListConstruct %int0_778, %int0_779 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3495 = torch.prim.ListConstruct %int1_780, %int1_781 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3496 = torch.prim.ListConstruct %int1_782, %int1_783 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3497 = torch.prim.ListConstruct %int0_784, %int0_784 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_785 = torch.constant.bool false | |
%int1_786 = torch.constant.int 1 | |
%3498 = torch.aten.convolution %3469, %3481, %3493, %3496, %3494, %3495, %false_785, %3497, %int1_786 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
%3499 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3500 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_787 = torch.constant.int 12 | |
%3501 = torch.aten.item %3499 : !torch.vtensor<[],f32> -> !torch.float | |
%3502 = torch.aten.item %3500 : !torch.vtensor<[],si8> -> !torch.int | |
%3503 = torch.aten.quantize_per_tensor %108, %3501, %3502, %int12_787 : !torch.vtensor<[80,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,128,3,3],!torch.qint8> | |
%3504 = torch.aten.int_repr %3503 : !torch.vtensor<[80,128,3,3],!torch.qint8> -> !torch.vtensor<[80,128,3,3],si8> | |
%3505 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3506 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3507 = torch.aten.item %3505 : !torch.vtensor<[],f32> -> !torch.float | |
%3508 = torch.aten.item %3506 : !torch.vtensor<[],si8> -> !torch.int | |
%3509 = torch.aten._make_per_tensor_quantized_tensor %3504, %3507, %3508 : !torch.vtensor<[80,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,128,3,3],!torch.qint8> | |
%3510 = torch.aten.dequantize.self %3509 : !torch.vtensor<[80,128,3,3],!torch.qint8> -> !torch.vtensor<[80,128,3,3],f32> | |
%3511 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3512 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_788 = torch.constant.int 12 | |
%3513 = torch.aten.item %3511 : !torch.vtensor<[],f32> -> !torch.float | |
%3514 = torch.aten.item %3512 : !torch.vtensor<[],si8> -> !torch.int | |
%3515 = torch.aten.quantize_per_tensor %109, %3513, %3514, %int12_788 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3516 = torch.aten.int_repr %3515 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3517 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3518 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3519 = torch.aten.item %3517 : !torch.vtensor<[],f32> -> !torch.float | |
%3520 = torch.aten.item %3518 : !torch.vtensor<[],si8> -> !torch.int | |
%3521 = torch.aten._make_per_tensor_quantized_tensor %3516, %3519, %3520 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3522 = torch.aten.dequantize.self %3521 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int1_789 = torch.constant.int 1 | |
%int1_790 = torch.constant.int 1 | |
%int1_791 = torch.constant.int 1 | |
%int1_792 = torch.constant.int 1 | |
%int1_793 = torch.constant.int 1 | |
%int1_794 = torch.constant.int 1 | |
%int0_795 = torch.constant.int 0 | |
%3523 = torch.prim.ListConstruct %int1_789, %int1_790 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3524 = torch.prim.ListConstruct %int1_791, %int1_792 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3525 = torch.prim.ListConstruct %int1_793, %int1_794 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3526 = torch.prim.ListConstruct %int0_795, %int0_795 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_796 = torch.constant.bool false | |
%int1_797 = torch.constant.int 1 | |
%3527 = torch.aten.convolution %2759, %3510, %3522, %3525, %3523, %3524, %false_796, %3526, %int1_797 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[80,128,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,40,40],f32> | |
%3528 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3529 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_798 = torch.constant.int 12 | |
%3530 = torch.aten.item %3528 : !torch.vtensor<[],f32> -> !torch.float | |
%3531 = torch.aten.item %3529 : !torch.vtensor<[],si8> -> !torch.int | |
%3532 = torch.aten.quantize_per_tensor %3527, %3530, %3531, %int12_798 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3533 = torch.aten.int_repr %3532 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
%3534 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3535 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3536 = torch.aten.item %3534 : !torch.vtensor<[],f32> -> !torch.float | |
%3537 = torch.aten.item %3535 : !torch.vtensor<[],si8> -> !torch.int | |
%3538 = torch.aten._make_per_tensor_quantized_tensor %3533, %3536, %3537 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3539 = torch.aten.dequantize.self %3538 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
%3540 = torch.aten.sigmoid %3539 : !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
%3541 = torch.aten.mul.Tensor %3539, %3540 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
%3542 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3543 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_799 = torch.constant.int 12 | |
%3544 = torch.aten.item %3542 : !torch.vtensor<[],f32> -> !torch.float | |
%3545 = torch.aten.item %3543 : !torch.vtensor<[],si8> -> !torch.int | |
%3546 = torch.aten.quantize_per_tensor %3541, %3544, %3545, %int12_799 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3547 = torch.aten.int_repr %3546 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
%3548 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3549 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3550 = torch.aten.item %3548 : !torch.vtensor<[],f32> -> !torch.float | |
%3551 = torch.aten.item %3549 : !torch.vtensor<[],si8> -> !torch.int | |
%3552 = torch.aten._make_per_tensor_quantized_tensor %3547, %3550, %3551 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3553 = torch.aten.dequantize.self %3552 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
%3554 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3555 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_800 = torch.constant.int 12 | |
%3556 = torch.aten.item %3554 : !torch.vtensor<[],f32> -> !torch.float | |
%3557 = torch.aten.item %3555 : !torch.vtensor<[],si8> -> !torch.int | |
%3558 = torch.aten.quantize_per_tensor %110, %3556, %3557, %int12_800 : !torch.vtensor<[80,80,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
%3559 = torch.aten.int_repr %3558 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],si8> | |
%3560 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3561 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3562 = torch.aten.item %3560 : !torch.vtensor<[],f32> -> !torch.float | |
%3563 = torch.aten.item %3561 : !torch.vtensor<[],si8> -> !torch.int | |
%3564 = torch.aten._make_per_tensor_quantized_tensor %3559, %3562, %3563 : !torch.vtensor<[80,80,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
%3565 = torch.aten.dequantize.self %3564 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],f32> | |
%3566 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3567 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_801 = torch.constant.int 12 | |
%3568 = torch.aten.item %3566 : !torch.vtensor<[],f32> -> !torch.float | |
%3569 = torch.aten.item %3567 : !torch.vtensor<[],si8> -> !torch.int | |
%3570 = torch.aten.quantize_per_tensor %111, %3568, %3569, %int12_801 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3571 = torch.aten.int_repr %3570 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3572 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3573 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3574 = torch.aten.item %3572 : !torch.vtensor<[],f32> -> !torch.float | |
%3575 = torch.aten.item %3573 : !torch.vtensor<[],si8> -> !torch.int | |
%3576 = torch.aten._make_per_tensor_quantized_tensor %3571, %3574, %3575 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3577 = torch.aten.dequantize.self %3576 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int1_802 = torch.constant.int 1 | |
%int1_803 = torch.constant.int 1 | |
%int1_804 = torch.constant.int 1 | |
%int1_805 = torch.constant.int 1 | |
%int1_806 = torch.constant.int 1 | |
%int1_807 = torch.constant.int 1 | |
%int0_808 = torch.constant.int 0 | |
%3578 = torch.prim.ListConstruct %int1_802, %int1_803 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3579 = torch.prim.ListConstruct %int1_804, %int1_805 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3580 = torch.prim.ListConstruct %int1_806, %int1_807 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3581 = torch.prim.ListConstruct %int0_808, %int0_808 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_809 = torch.constant.bool false | |
%int1_810 = torch.constant.int 1 | |
%3582 = torch.aten.convolution %3553, %3565, %3577, %3580, %3578, %3579, %false_809, %3581, %int1_810 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[80,80,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,40,40],f32> | |
%3583 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3584 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_811 = torch.constant.int 12 | |
%3585 = torch.aten.item %3583 : !torch.vtensor<[],f32> -> !torch.float | |
%3586 = torch.aten.item %3584 : !torch.vtensor<[],si8> -> !torch.int | |
%3587 = torch.aten.quantize_per_tensor %3582, %3585, %3586, %int12_811 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3588 = torch.aten.int_repr %3587 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
%3589 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3590 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3591 = torch.aten.item %3589 : !torch.vtensor<[],f32> -> !torch.float | |
%3592 = torch.aten.item %3590 : !torch.vtensor<[],si8> -> !torch.int | |
%3593 = torch.aten._make_per_tensor_quantized_tensor %3588, %3591, %3592 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3594 = torch.aten.dequantize.self %3593 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
%3595 = torch.aten.sigmoid %3594 : !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
%3596 = torch.aten.mul.Tensor %3594, %3595 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
%3597 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3598 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_812 = torch.constant.int 12 | |
%3599 = torch.aten.item %3597 : !torch.vtensor<[],f32> -> !torch.float | |
%3600 = torch.aten.item %3598 : !torch.vtensor<[],si8> -> !torch.int | |
%3601 = torch.aten.quantize_per_tensor %3596, %3599, %3600, %int12_812 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3602 = torch.aten.int_repr %3601 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
%3603 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3604 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3605 = torch.aten.item %3603 : !torch.vtensor<[],f32> -> !torch.float | |
%3606 = torch.aten.item %3604 : !torch.vtensor<[],si8> -> !torch.int | |
%3607 = torch.aten._make_per_tensor_quantized_tensor %3602, %3605, %3606 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
%3608 = torch.aten.dequantize.self %3607 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
%3609 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3610 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_813 = torch.constant.int 12 | |
%3611 = torch.aten.item %3609 : !torch.vtensor<[],f32> -> !torch.float | |
%3612 = torch.aten.item %3610 : !torch.vtensor<[],si8> -> !torch.int | |
%3613 = torch.aten.quantize_per_tensor %112, %3611, %3612, %int12_813 : !torch.vtensor<[80,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
%3614 = torch.aten.int_repr %3613 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],si8> | |
%3615 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3616 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3617 = torch.aten.item %3615 : !torch.vtensor<[],f32> -> !torch.float | |
%3618 = torch.aten.item %3616 : !torch.vtensor<[],si8> -> !torch.int | |
%3619 = torch.aten._make_per_tensor_quantized_tensor %3614, %3617, %3618 : !torch.vtensor<[80,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
%3620 = torch.aten.dequantize.self %3619 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],f32> | |
%3621 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3622 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_814 = torch.constant.int 12 | |
%3623 = torch.aten.item %3621 : !torch.vtensor<[],f32> -> !torch.float | |
%3624 = torch.aten.item %3622 : !torch.vtensor<[],si8> -> !torch.int | |
%3625 = torch.aten.quantize_per_tensor %113, %3623, %3624, %int12_814 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3626 = torch.aten.int_repr %3625 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3627 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3628 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3629 = torch.aten.item %3627 : !torch.vtensor<[],f32> -> !torch.float | |
%3630 = torch.aten.item %3628 : !torch.vtensor<[],si8> -> !torch.int | |
%3631 = torch.aten._make_per_tensor_quantized_tensor %3626, %3629, %3630 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3632 = torch.aten.dequantize.self %3631 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int0_815 = torch.constant.int 0 | |
%int0_816 = torch.constant.int 0 | |
%int1_817 = torch.constant.int 1 | |
%int1_818 = torch.constant.int 1 | |
%int1_819 = torch.constant.int 1 | |
%int1_820 = torch.constant.int 1 | |
%int0_821 = torch.constant.int 0 | |
%3633 = torch.prim.ListConstruct %int0_815, %int0_816 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3634 = torch.prim.ListConstruct %int1_817, %int1_818 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3635 = torch.prim.ListConstruct %int1_819, %int1_820 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3636 = torch.prim.ListConstruct %int0_821, %int0_821 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_822 = torch.constant.bool false | |
%int1_823 = torch.constant.int 1 | |
%3637 = torch.aten.convolution %3608, %3620, %3632, %3635, %3633, %3634, %false_822, %3636, %int1_823 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[80,80,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,40,40],f32> | |
%3638 = torch.prim.ListConstruct %3498, %3637 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,80,40,40],f32>) -> !torch.list<vtensor> | |
%int1_824 = torch.constant.int 1 | |
%3639 = torch.aten.cat %3638, %int1_824 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,144,40,40],f32> | |
%3640 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3641 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_825 = torch.constant.int 12 | |
%3642 = torch.aten.item %3640 : !torch.vtensor<[],f32> -> !torch.float | |
%3643 = torch.aten.item %3641 : !torch.vtensor<[],si8> -> !torch.int | |
%3644 = torch.aten.quantize_per_tensor %3639, %3642, %3643, %int12_825 : !torch.vtensor<[1,144,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,144,40,40],!torch.qint8> | |
%3645 = torch.aten.int_repr %3644 : !torch.vtensor<[1,144,40,40],!torch.qint8> -> !torch.vtensor<[1,144,40,40],si8> | |
%3646 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3647 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3648 = torch.aten.item %3646 : !torch.vtensor<[],f32> -> !torch.float | |
%3649 = torch.aten.item %3647 : !torch.vtensor<[],si8> -> !torch.int | |
%3650 = torch.aten._make_per_tensor_quantized_tensor %3645, %3648, %3649 : !torch.vtensor<[1,144,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,144,40,40],!torch.qint8> | |
%3651 = torch.aten.dequantize.self %3650 : !torch.vtensor<[1,144,40,40],!torch.qint8> -> !torch.vtensor<[1,144,40,40],f32> | |
%3652 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3653 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_826 = torch.constant.int 12 | |
%3654 = torch.aten.item %3652 : !torch.vtensor<[],f32> -> !torch.float | |
%3655 = torch.aten.item %3653 : !torch.vtensor<[],si8> -> !torch.int | |
%3656 = torch.aten.quantize_per_tensor %114, %3654, %3655, %int12_826 : !torch.vtensor<[64,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,256,3,3],!torch.qint8> | |
%3657 = torch.aten.int_repr %3656 : !torch.vtensor<[64,256,3,3],!torch.qint8> -> !torch.vtensor<[64,256,3,3],si8> | |
%3658 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3659 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3660 = torch.aten.item %3658 : !torch.vtensor<[],f32> -> !torch.float | |
%3661 = torch.aten.item %3659 : !torch.vtensor<[],si8> -> !torch.int | |
%3662 = torch.aten._make_per_tensor_quantized_tensor %3657, %3660, %3661 : !torch.vtensor<[64,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,256,3,3],!torch.qint8> | |
%3663 = torch.aten.dequantize.self %3662 : !torch.vtensor<[64,256,3,3],!torch.qint8> -> !torch.vtensor<[64,256,3,3],f32> | |
%3664 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3665 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_827 = torch.constant.int 12 | |
%3666 = torch.aten.item %3664 : !torch.vtensor<[],f32> -> !torch.float | |
%3667 = torch.aten.item %3665 : !torch.vtensor<[],si8> -> !torch.int | |
%3668 = torch.aten.quantize_per_tensor %115, %3666, %3667, %int12_827 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3669 = torch.aten.int_repr %3668 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3670 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3671 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3672 = torch.aten.item %3670 : !torch.vtensor<[],f32> -> !torch.float | |
%3673 = torch.aten.item %3671 : !torch.vtensor<[],si8> -> !torch.int | |
%3674 = torch.aten._make_per_tensor_quantized_tensor %3669, %3672, %3673 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3675 = torch.aten.dequantize.self %3674 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_828 = torch.constant.int 1 | |
%int1_829 = torch.constant.int 1 | |
%int1_830 = torch.constant.int 1 | |
%int1_831 = torch.constant.int 1 | |
%int1_832 = torch.constant.int 1 | |
%int1_833 = torch.constant.int 1 | |
%int0_834 = torch.constant.int 0 | |
%3676 = torch.prim.ListConstruct %int1_828, %int1_829 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3677 = torch.prim.ListConstruct %int1_830, %int1_831 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3678 = torch.prim.ListConstruct %int1_832, %int1_833 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3679 = torch.prim.ListConstruct %int0_834, %int0_834 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_835 = torch.constant.bool false | |
%int1_836 = torch.constant.int 1 | |
%3680 = torch.aten.convolution %3066, %3663, %3675, %3678, %3676, %3677, %false_835, %3679, %int1_836 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[64,256,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,20,20],f32> | |
%3681 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3682 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_837 = torch.constant.int 12 | |
%3683 = torch.aten.item %3681 : !torch.vtensor<[],f32> -> !torch.float | |
%3684 = torch.aten.item %3682 : !torch.vtensor<[],si8> -> !torch.int | |
%3685 = torch.aten.quantize_per_tensor %3680, %3683, %3684, %int12_837 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3686 = torch.aten.int_repr %3685 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
%3687 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3688 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3689 = torch.aten.item %3687 : !torch.vtensor<[],f32> -> !torch.float | |
%3690 = torch.aten.item %3688 : !torch.vtensor<[],si8> -> !torch.int | |
%3691 = torch.aten._make_per_tensor_quantized_tensor %3686, %3689, %3690 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3692 = torch.aten.dequantize.self %3691 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
%3693 = torch.aten.sigmoid %3692 : !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
%3694 = torch.aten.mul.Tensor %3692, %3693 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
%3695 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3696 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_838 = torch.constant.int 12 | |
%3697 = torch.aten.item %3695 : !torch.vtensor<[],f32> -> !torch.float | |
%3698 = torch.aten.item %3696 : !torch.vtensor<[],si8> -> !torch.int | |
%3699 = torch.aten.quantize_per_tensor %3694, %3697, %3698, %int12_838 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3700 = torch.aten.int_repr %3699 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
%3701 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3702 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3703 = torch.aten.item %3701 : !torch.vtensor<[],f32> -> !torch.float | |
%3704 = torch.aten.item %3702 : !torch.vtensor<[],si8> -> !torch.int | |
%3705 = torch.aten._make_per_tensor_quantized_tensor %3700, %3703, %3704 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3706 = torch.aten.dequantize.self %3705 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
%3707 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3708 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_839 = torch.constant.int 12 | |
%3709 = torch.aten.item %3707 : !torch.vtensor<[],f32> -> !torch.float | |
%3710 = torch.aten.item %3708 : !torch.vtensor<[],si8> -> !torch.int | |
%3711 = torch.aten.quantize_per_tensor %116, %3709, %3710, %int12_839 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3712 = torch.aten.int_repr %3711 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
%3713 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3714 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3715 = torch.aten.item %3713 : !torch.vtensor<[],f32> -> !torch.float | |
%3716 = torch.aten.item %3714 : !torch.vtensor<[],si8> -> !torch.int | |
%3717 = torch.aten._make_per_tensor_quantized_tensor %3712, %3715, %3716 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
%3718 = torch.aten.dequantize.self %3717 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
%3719 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3720 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_840 = torch.constant.int 12 | |
%3721 = torch.aten.item %3719 : !torch.vtensor<[],f32> -> !torch.float | |
%3722 = torch.aten.item %3720 : !torch.vtensor<[],si8> -> !torch.int | |
%3723 = torch.aten.quantize_per_tensor %117, %3721, %3722, %int12_840 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3724 = torch.aten.int_repr %3723 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3725 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3726 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3727 = torch.aten.item %3725 : !torch.vtensor<[],f32> -> !torch.float | |
%3728 = torch.aten.item %3726 : !torch.vtensor<[],si8> -> !torch.int | |
%3729 = torch.aten._make_per_tensor_quantized_tensor %3724, %3727, %3728 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3730 = torch.aten.dequantize.self %3729 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int1_841 = torch.constant.int 1 | |
%int1_842 = torch.constant.int 1 | |
%int1_843 = torch.constant.int 1 | |
%int1_844 = torch.constant.int 1 | |
%int1_845 = torch.constant.int 1 | |
%int1_846 = torch.constant.int 1 | |
%int0_847 = torch.constant.int 0 | |
%3731 = torch.prim.ListConstruct %int1_841, %int1_842 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3732 = torch.prim.ListConstruct %int1_843, %int1_844 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3733 = torch.prim.ListConstruct %int1_845, %int1_846 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3734 = torch.prim.ListConstruct %int0_847, %int0_847 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_848 = torch.constant.bool false | |
%int1_849 = torch.constant.int 1 | |
%3735 = torch.aten.convolution %3706, %3718, %3730, %3733, %3731, %3732, %false_848, %3734, %int1_849 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,20,20],f32> | |
%3736 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3737 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_850 = torch.constant.int 12 | |
%3738 = torch.aten.item %3736 : !torch.vtensor<[],f32> -> !torch.float | |
%3739 = torch.aten.item %3737 : !torch.vtensor<[],si8> -> !torch.int | |
%3740 = torch.aten.quantize_per_tensor %3735, %3738, %3739, %int12_850 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3741 = torch.aten.int_repr %3740 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
%3742 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3743 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3744 = torch.aten.item %3742 : !torch.vtensor<[],f32> -> !torch.float | |
%3745 = torch.aten.item %3743 : !torch.vtensor<[],si8> -> !torch.int | |
%3746 = torch.aten._make_per_tensor_quantized_tensor %3741, %3744, %3745 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3747 = torch.aten.dequantize.self %3746 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
%3748 = torch.aten.sigmoid %3747 : !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
%3749 = torch.aten.mul.Tensor %3747, %3748 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
%3750 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3751 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_851 = torch.constant.int 12 | |
%3752 = torch.aten.item %3750 : !torch.vtensor<[],f32> -> !torch.float | |
%3753 = torch.aten.item %3751 : !torch.vtensor<[],si8> -> !torch.int | |
%3754 = torch.aten.quantize_per_tensor %3749, %3752, %3753, %int12_851 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3755 = torch.aten.int_repr %3754 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
%3756 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3757 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3758 = torch.aten.item %3756 : !torch.vtensor<[],f32> -> !torch.float | |
%3759 = torch.aten.item %3757 : !torch.vtensor<[],si8> -> !torch.int | |
%3760 = torch.aten._make_per_tensor_quantized_tensor %3755, %3758, %3759 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
%3761 = torch.aten.dequantize.self %3760 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
%3762 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3763 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_852 = torch.constant.int 12 | |
%3764 = torch.aten.item %3762 : !torch.vtensor<[],f32> -> !torch.float | |
%3765 = torch.aten.item %3763 : !torch.vtensor<[],si8> -> !torch.int | |
%3766 = torch.aten.quantize_per_tensor %118, %3764, %3765, %int12_852 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%3767 = torch.aten.int_repr %3766 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
%3768 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3769 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3770 = torch.aten.item %3768 : !torch.vtensor<[],f32> -> !torch.float | |
%3771 = torch.aten.item %3769 : !torch.vtensor<[],si8> -> !torch.int | |
%3772 = torch.aten._make_per_tensor_quantized_tensor %3767, %3770, %3771 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
%3773 = torch.aten.dequantize.self %3772 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
%3774 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3775 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_853 = torch.constant.int 12 | |
%3776 = torch.aten.item %3774 : !torch.vtensor<[],f32> -> !torch.float | |
%3777 = torch.aten.item %3775 : !torch.vtensor<[],si8> -> !torch.int | |
%3778 = torch.aten.quantize_per_tensor %119, %3776, %3777, %int12_853 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3779 = torch.aten.int_repr %3778 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
%3780 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3781 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3782 = torch.aten.item %3780 : !torch.vtensor<[],f32> -> !torch.float | |
%3783 = torch.aten.item %3781 : !torch.vtensor<[],si8> -> !torch.int | |
%3784 = torch.aten._make_per_tensor_quantized_tensor %3779, %3782, %3783 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
%3785 = torch.aten.dequantize.self %3784 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
%int0_854 = torch.constant.int 0 | |
%int0_855 = torch.constant.int 0 | |
%int1_856 = torch.constant.int 1 | |
%int1_857 = torch.constant.int 1 | |
%int1_858 = torch.constant.int 1 | |
%int1_859 = torch.constant.int 1 | |
%int0_860 = torch.constant.int 0 | |
%3786 = torch.prim.ListConstruct %int0_854, %int0_855 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3787 = torch.prim.ListConstruct %int1_856, %int1_857 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3788 = torch.prim.ListConstruct %int1_858, %int1_859 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3789 = torch.prim.ListConstruct %int0_860, %int0_860 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_861 = torch.constant.bool false | |
%int1_862 = torch.constant.int 1 | |
%3790 = torch.aten.convolution %3761, %3773, %3785, %3788, %3786, %3787, %false_861, %3789, %int1_862 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,20,20],f32> | |
%3791 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3792 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_863 = torch.constant.int 12 | |
%3793 = torch.aten.item %3791 : !torch.vtensor<[],f32> -> !torch.float | |
%3794 = torch.aten.item %3792 : !torch.vtensor<[],si8> -> !torch.int | |
%3795 = torch.aten.quantize_per_tensor %120, %3793, %3794, %int12_863 : !torch.vtensor<[80,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,256,3,3],!torch.qint8> | |
%3796 = torch.aten.int_repr %3795 : !torch.vtensor<[80,256,3,3],!torch.qint8> -> !torch.vtensor<[80,256,3,3],si8> | |
%3797 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3798 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3799 = torch.aten.item %3797 : !torch.vtensor<[],f32> -> !torch.float | |
%3800 = torch.aten.item %3798 : !torch.vtensor<[],si8> -> !torch.int | |
%3801 = torch.aten._make_per_tensor_quantized_tensor %3796, %3799, %3800 : !torch.vtensor<[80,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,256,3,3],!torch.qint8> | |
%3802 = torch.aten.dequantize.self %3801 : !torch.vtensor<[80,256,3,3],!torch.qint8> -> !torch.vtensor<[80,256,3,3],f32> | |
%3803 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3804 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_864 = torch.constant.int 12 | |
%3805 = torch.aten.item %3803 : !torch.vtensor<[],f32> -> !torch.float | |
%3806 = torch.aten.item %3804 : !torch.vtensor<[],si8> -> !torch.int | |
%3807 = torch.aten.quantize_per_tensor %121, %3805, %3806, %int12_864 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3808 = torch.aten.int_repr %3807 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3809 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3810 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3811 = torch.aten.item %3809 : !torch.vtensor<[],f32> -> !torch.float | |
%3812 = torch.aten.item %3810 : !torch.vtensor<[],si8> -> !torch.int | |
%3813 = torch.aten._make_per_tensor_quantized_tensor %3808, %3811, %3812 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3814 = torch.aten.dequantize.self %3813 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int1_865 = torch.constant.int 1 | |
%int1_866 = torch.constant.int 1 | |
%int1_867 = torch.constant.int 1 | |
%int1_868 = torch.constant.int 1 | |
%int1_869 = torch.constant.int 1 | |
%int1_870 = torch.constant.int 1 | |
%int0_871 = torch.constant.int 0 | |
%3815 = torch.prim.ListConstruct %int1_865, %int1_866 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3816 = torch.prim.ListConstruct %int1_867, %int1_868 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3817 = torch.prim.ListConstruct %int1_869, %int1_870 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3818 = torch.prim.ListConstruct %int0_871, %int0_871 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_872 = torch.constant.bool false | |
%int1_873 = torch.constant.int 1 | |
%3819 = torch.aten.convolution %3066, %3802, %3814, %3817, %3815, %3816, %false_872, %3818, %int1_873 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[80,256,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,20,20],f32> | |
%3820 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3821 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_874 = torch.constant.int 12 | |
%3822 = torch.aten.item %3820 : !torch.vtensor<[],f32> -> !torch.float | |
%3823 = torch.aten.item %3821 : !torch.vtensor<[],si8> -> !torch.int | |
%3824 = torch.aten.quantize_per_tensor %3819, %3822, %3823, %int12_874 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3825 = torch.aten.int_repr %3824 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
%3826 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3827 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3828 = torch.aten.item %3826 : !torch.vtensor<[],f32> -> !torch.float | |
%3829 = torch.aten.item %3827 : !torch.vtensor<[],si8> -> !torch.int | |
%3830 = torch.aten._make_per_tensor_quantized_tensor %3825, %3828, %3829 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3831 = torch.aten.dequantize.self %3830 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
%3832 = torch.aten.sigmoid %3831 : !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
%3833 = torch.aten.mul.Tensor %3831, %3832 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
%3834 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3835 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_875 = torch.constant.int 12 | |
%3836 = torch.aten.item %3834 : !torch.vtensor<[],f32> -> !torch.float | |
%3837 = torch.aten.item %3835 : !torch.vtensor<[],si8> -> !torch.int | |
%3838 = torch.aten.quantize_per_tensor %3833, %3836, %3837, %int12_875 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3839 = torch.aten.int_repr %3838 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
%3840 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3841 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3842 = torch.aten.item %3840 : !torch.vtensor<[],f32> -> !torch.float | |
%3843 = torch.aten.item %3841 : !torch.vtensor<[],si8> -> !torch.int | |
%3844 = torch.aten._make_per_tensor_quantized_tensor %3839, %3842, %3843 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3845 = torch.aten.dequantize.self %3844 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
%3846 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3847 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_876 = torch.constant.int 12 | |
%3848 = torch.aten.item %3846 : !torch.vtensor<[],f32> -> !torch.float | |
%3849 = torch.aten.item %3847 : !torch.vtensor<[],si8> -> !torch.int | |
%3850 = torch.aten.quantize_per_tensor %122, %3848, %3849, %int12_876 : !torch.vtensor<[80,80,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
%3851 = torch.aten.int_repr %3850 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],si8> | |
%3852 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3853 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3854 = torch.aten.item %3852 : !torch.vtensor<[],f32> -> !torch.float | |
%3855 = torch.aten.item %3853 : !torch.vtensor<[],si8> -> !torch.int | |
%3856 = torch.aten._make_per_tensor_quantized_tensor %3851, %3854, %3855 : !torch.vtensor<[80,80,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
%3857 = torch.aten.dequantize.self %3856 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],f32> | |
%3858 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3859 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_877 = torch.constant.int 12 | |
%3860 = torch.aten.item %3858 : !torch.vtensor<[],f32> -> !torch.float | |
%3861 = torch.aten.item %3859 : !torch.vtensor<[],si8> -> !torch.int | |
%3862 = torch.aten.quantize_per_tensor %123, %3860, %3861, %int12_877 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3863 = torch.aten.int_repr %3862 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3864 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3865 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3866 = torch.aten.item %3864 : !torch.vtensor<[],f32> -> !torch.float | |
%3867 = torch.aten.item %3865 : !torch.vtensor<[],si8> -> !torch.int | |
%3868 = torch.aten._make_per_tensor_quantized_tensor %3863, %3866, %3867 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3869 = torch.aten.dequantize.self %3868 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int1_878 = torch.constant.int 1 | |
%int1_879 = torch.constant.int 1 | |
%int1_880 = torch.constant.int 1 | |
%int1_881 = torch.constant.int 1 | |
%int1_882 = torch.constant.int 1 | |
%int1_883 = torch.constant.int 1 | |
%int0_884 = torch.constant.int 0 | |
%3870 = torch.prim.ListConstruct %int1_878, %int1_879 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3871 = torch.prim.ListConstruct %int1_880, %int1_881 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3872 = torch.prim.ListConstruct %int1_882, %int1_883 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3873 = torch.prim.ListConstruct %int0_884, %int0_884 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_885 = torch.constant.bool false | |
%int1_886 = torch.constant.int 1 | |
%3874 = torch.aten.convolution %3845, %3857, %3869, %3872, %3870, %3871, %false_885, %3873, %int1_886 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[80,80,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,20,20],f32> | |
%3875 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3876 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_887 = torch.constant.int 12 | |
%3877 = torch.aten.item %3875 : !torch.vtensor<[],f32> -> !torch.float | |
%3878 = torch.aten.item %3876 : !torch.vtensor<[],si8> -> !torch.int | |
%3879 = torch.aten.quantize_per_tensor %3874, %3877, %3878, %int12_887 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3880 = torch.aten.int_repr %3879 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
%3881 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3882 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3883 = torch.aten.item %3881 : !torch.vtensor<[],f32> -> !torch.float | |
%3884 = torch.aten.item %3882 : !torch.vtensor<[],si8> -> !torch.int | |
%3885 = torch.aten._make_per_tensor_quantized_tensor %3880, %3883, %3884 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3886 = torch.aten.dequantize.self %3885 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
%3887 = torch.aten.sigmoid %3886 : !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
%3888 = torch.aten.mul.Tensor %3886, %3887 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
%3889 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3890 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_888 = torch.constant.int 12 | |
%3891 = torch.aten.item %3889 : !torch.vtensor<[],f32> -> !torch.float | |
%3892 = torch.aten.item %3890 : !torch.vtensor<[],si8> -> !torch.int | |
%3893 = torch.aten.quantize_per_tensor %3888, %3891, %3892, %int12_888 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3894 = torch.aten.int_repr %3893 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
%3895 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3896 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3897 = torch.aten.item %3895 : !torch.vtensor<[],f32> -> !torch.float | |
%3898 = torch.aten.item %3896 : !torch.vtensor<[],si8> -> !torch.int | |
%3899 = torch.aten._make_per_tensor_quantized_tensor %3894, %3897, %3898 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
%3900 = torch.aten.dequantize.self %3899 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
%3901 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3902 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_889 = torch.constant.int 12 | |
%3903 = torch.aten.item %3901 : !torch.vtensor<[],f32> -> !torch.float | |
%3904 = torch.aten.item %3902 : !torch.vtensor<[],si8> -> !torch.int | |
%3905 = torch.aten.quantize_per_tensor %124, %3903, %3904, %int12_889 : !torch.vtensor<[80,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
%3906 = torch.aten.int_repr %3905 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],si8> | |
%3907 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3908 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3909 = torch.aten.item %3907 : !torch.vtensor<[],f32> -> !torch.float | |
%3910 = torch.aten.item %3908 : !torch.vtensor<[],si8> -> !torch.int | |
%3911 = torch.aten._make_per_tensor_quantized_tensor %3906, %3909, %3910 : !torch.vtensor<[80,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
%3912 = torch.aten.dequantize.self %3911 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],f32> | |
%3913 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3914 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_890 = torch.constant.int 12 | |
%3915 = torch.aten.item %3913 : !torch.vtensor<[],f32> -> !torch.float | |
%3916 = torch.aten.item %3914 : !torch.vtensor<[],si8> -> !torch.int | |
%3917 = torch.aten.quantize_per_tensor %125, %3915, %3916, %int12_890 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3918 = torch.aten.int_repr %3917 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
%3919 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3920 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3921 = torch.aten.item %3919 : !torch.vtensor<[],f32> -> !torch.float | |
%3922 = torch.aten.item %3920 : !torch.vtensor<[],si8> -> !torch.int | |
%3923 = torch.aten._make_per_tensor_quantized_tensor %3918, %3921, %3922 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
%3924 = torch.aten.dequantize.self %3923 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
%int0_891 = torch.constant.int 0 | |
%int0_892 = torch.constant.int 0 | |
%int1_893 = torch.constant.int 1 | |
%int1_894 = torch.constant.int 1 | |
%int1_895 = torch.constant.int 1 | |
%int1_896 = torch.constant.int 1 | |
%int0_897 = torch.constant.int 0 | |
%3925 = torch.prim.ListConstruct %int0_891, %int0_892 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3926 = torch.prim.ListConstruct %int1_893, %int1_894 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3927 = torch.prim.ListConstruct %int1_895, %int1_896 : (!torch.int, !torch.int) -> !torch.list<int> | |
%3928 = torch.prim.ListConstruct %int0_897, %int0_897 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_898 = torch.constant.bool false | |
%int1_899 = torch.constant.int 1 | |
%3929 = torch.aten.convolution %3900, %3912, %3924, %3927, %3925, %3926, %false_898, %3928, %int1_899 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[80,80,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,20,20],f32> | |
%3930 = torch.prim.ListConstruct %3790, %3929 : (!torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[1,80,20,20],f32>) -> !torch.list<vtensor> | |
%int1_900 = torch.constant.int 1 | |
%3931 = torch.aten.cat %3930, %int1_900 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,144,20,20],f32> | |
%3932 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3933 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_901 = torch.constant.int 12 | |
%3934 = torch.aten.item %3932 : !torch.vtensor<[],f32> -> !torch.float | |
%3935 = torch.aten.item %3933 : !torch.vtensor<[],si8> -> !torch.int | |
%3936 = torch.aten.quantize_per_tensor %3931, %3934, %3935, %int12_901 : !torch.vtensor<[1,144,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,144,20,20],!torch.qint8> | |
%3937 = torch.aten.int_repr %3936 : !torch.vtensor<[1,144,20,20],!torch.qint8> -> !torch.vtensor<[1,144,20,20],si8> | |
%3938 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%3939 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%3940 = torch.aten.item %3938 : !torch.vtensor<[],f32> -> !torch.float | |
%3941 = torch.aten.item %3939 : !torch.vtensor<[],si8> -> !torch.int | |
%3942 = torch.aten._make_per_tensor_quantized_tensor %3937, %3940, %3941 : !torch.vtensor<[1,144,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,144,20,20],!torch.qint8> | |
%3943 = torch.aten.dequantize.self %3942 : !torch.vtensor<[1,144,20,20],!torch.qint8> -> !torch.vtensor<[1,144,20,20],f32> | |
%3944 = torch.vtensor.literal(dense<144> : tensor<si64>) : !torch.vtensor<[],si64> | |
%3945 = torch.vtensor.literal(dense<-1> : tensor<si64>) : !torch.vtensor<[],si64> | |
%3946 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_902 = torch.constant.int 0 | |
%int0_903 = torch.constant.int 0 | |
%int0_904 = torch.constant.int 0 | |
%3947 = torch.aten.select.int %3946, %int0_902, %int0_904 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3948 = torch.aten.item %3947 : !torch.vtensor<[1],si64> -> !torch.int | |
%3949 = torch.aten.lt.int %3948, %int0_902 : !torch.int, !torch.int -> !torch.bool | |
%3950 = torch.aten.Int.bool %3949 : !torch.bool -> !torch.int | |
%3951 = torch.aten.mul.int %3950, %int0_903 : !torch.int, !torch.int -> !torch.int | |
%3952 = torch.aten.add.int %3948, %3951 : !torch.int, !torch.int -> !torch.int | |
%3953 = torch.prim.ListConstruct %3952 : (!torch.int) -> !torch.list<int> | |
%false_905 = torch.constant.bool false | |
%none_906 = torch.constant.none | |
%3954 = torch.aten.tensor %3953, %none_906, %none_906, %false_905 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values, %indices = torch.aten.sort %3954, %int0_902, %false_905 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_907 = torch.constant.int 0 | |
%3955 = torch.aten.select.int %values, %int0_902, %int0_907 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3956 = torch.aten.item %3955 : !torch.vtensor<[1],si64> -> !torch.int | |
%3957 = torch.aten.unsqueeze %3067, %3956 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%3958 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_908 = torch.constant.int 0 | |
%int0_909 = torch.constant.int 0 | |
%int0_910 = torch.constant.int 0 | |
%3959 = torch.aten.select.int %3958, %int0_908, %int0_910 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3960 = torch.aten.item %3959 : !torch.vtensor<[1],si64> -> !torch.int | |
%3961 = torch.aten.lt.int %3960, %int0_908 : !torch.int, !torch.int -> !torch.bool | |
%3962 = torch.aten.Int.bool %3961 : !torch.bool -> !torch.int | |
%3963 = torch.aten.mul.int %3962, %int0_909 : !torch.int, !torch.int -> !torch.int | |
%3964 = torch.aten.add.int %3960, %3963 : !torch.int, !torch.int -> !torch.int | |
%3965 = torch.prim.ListConstruct %3964 : (!torch.int) -> !torch.list<int> | |
%false_911 = torch.constant.bool false | |
%none_912 = torch.constant.none | |
%3966 = torch.aten.tensor %3965, %none_912, %none_912, %false_911 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_913, %indices_914 = torch.aten.sort %3966, %int0_908, %false_911 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_915 = torch.constant.int 0 | |
%3967 = torch.aten.select.int %values_913, %int0_908, %int0_915 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3968 = torch.aten.item %3967 : !torch.vtensor<[1],si64> -> !torch.int | |
%3969 = torch.aten.unsqueeze %3944, %3968 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%3970 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_916 = torch.constant.int 0 | |
%int0_917 = torch.constant.int 0 | |
%int0_918 = torch.constant.int 0 | |
%3971 = torch.aten.select.int %3970, %int0_916, %int0_918 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3972 = torch.aten.item %3971 : !torch.vtensor<[1],si64> -> !torch.int | |
%3973 = torch.aten.lt.int %3972, %int0_916 : !torch.int, !torch.int -> !torch.bool | |
%3974 = torch.aten.Int.bool %3973 : !torch.bool -> !torch.int | |
%3975 = torch.aten.mul.int %3974, %int0_917 : !torch.int, !torch.int -> !torch.int | |
%3976 = torch.aten.add.int %3972, %3975 : !torch.int, !torch.int -> !torch.int | |
%3977 = torch.prim.ListConstruct %3976 : (!torch.int) -> !torch.list<int> | |
%false_919 = torch.constant.bool false | |
%none_920 = torch.constant.none | |
%3978 = torch.aten.tensor %3977, %none_920, %none_920, %false_919 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_921, %indices_922 = torch.aten.sort %3978, %int0_916, %false_919 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_923 = torch.constant.int 0 | |
%3979 = torch.aten.select.int %values_921, %int0_916, %int0_923 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3980 = torch.aten.item %3979 : !torch.vtensor<[1],si64> -> !torch.int | |
%3981 = torch.aten.unsqueeze %3945, %3980 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%3982 = torch.prim.ListConstruct %3957, %3969, %3981 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
%int0_924 = torch.constant.int 0 | |
%3983 = torch.aten.cat %3982, %int0_924 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
%3984 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_925 = torch.constant.int 0 | |
%int0_926 = torch.constant.int 0 | |
%int0_927 = torch.constant.int 0 | |
%3985 = torch.aten.select.int %3984, %int0_925, %int0_927 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3986 = torch.aten.item %3985 : !torch.vtensor<[1],si64> -> !torch.int | |
%3987 = torch.aten.lt.int %3986, %int0_925 : !torch.int, !torch.int -> !torch.bool | |
%3988 = torch.aten.Int.bool %3987 : !torch.bool -> !torch.int | |
%3989 = torch.aten.mul.int %3988, %int0_926 : !torch.int, !torch.int -> !torch.int | |
%3990 = torch.aten.add.int %3986, %3989 : !torch.int, !torch.int -> !torch.int | |
%3991 = torch.prim.ListConstruct %3990 : (!torch.int) -> !torch.list<int> | |
%false_928 = torch.constant.bool false | |
%none_929 = torch.constant.none | |
%3992 = torch.aten.tensor %3991, %none_929, %none_929, %false_928 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_930, %indices_931 = torch.aten.sort %3992, %int0_925, %false_928 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_932 = torch.constant.int 0 | |
%3993 = torch.aten.select.int %values_930, %int0_925, %int0_932 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3994 = torch.aten.item %3993 : !torch.vtensor<[1],si64> -> !torch.int | |
%3995 = torch.aten.unsqueeze %3067, %3994 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%3996 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_933 = torch.constant.int 0 | |
%int0_934 = torch.constant.int 0 | |
%int0_935 = torch.constant.int 0 | |
%3997 = torch.aten.select.int %3996, %int0_933, %int0_935 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%3998 = torch.aten.item %3997 : !torch.vtensor<[1],si64> -> !torch.int | |
%3999 = torch.aten.lt.int %3998, %int0_933 : !torch.int, !torch.int -> !torch.bool | |
%4000 = torch.aten.Int.bool %3999 : !torch.bool -> !torch.int | |
%4001 = torch.aten.mul.int %4000, %int0_934 : !torch.int, !torch.int -> !torch.int | |
%4002 = torch.aten.add.int %3998, %4001 : !torch.int, !torch.int -> !torch.int | |
%4003 = torch.prim.ListConstruct %4002 : (!torch.int) -> !torch.list<int> | |
%false_936 = torch.constant.bool false | |
%none_937 = torch.constant.none | |
%4004 = torch.aten.tensor %4003, %none_937, %none_937, %false_936 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_938, %indices_939 = torch.aten.sort %4004, %int0_933, %false_936 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_940 = torch.constant.int 0 | |
%4005 = torch.aten.select.int %values_938, %int0_933, %int0_940 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4006 = torch.aten.item %4005 : !torch.vtensor<[1],si64> -> !torch.int | |
%4007 = torch.aten.unsqueeze %3944, %4006 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4008 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_941 = torch.constant.int 0 | |
%int0_942 = torch.constant.int 0 | |
%int0_943 = torch.constant.int 0 | |
%4009 = torch.aten.select.int %4008, %int0_941, %int0_943 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4010 = torch.aten.item %4009 : !torch.vtensor<[1],si64> -> !torch.int | |
%4011 = torch.aten.lt.int %4010, %int0_941 : !torch.int, !torch.int -> !torch.bool | |
%4012 = torch.aten.Int.bool %4011 : !torch.bool -> !torch.int | |
%4013 = torch.aten.mul.int %4012, %int0_942 : !torch.int, !torch.int -> !torch.int | |
%4014 = torch.aten.add.int %4010, %4013 : !torch.int, !torch.int -> !torch.int | |
%4015 = torch.prim.ListConstruct %4014 : (!torch.int) -> !torch.list<int> | |
%false_944 = torch.constant.bool false | |
%none_945 = torch.constant.none | |
%4016 = torch.aten.tensor %4015, %none_945, %none_945, %false_944 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_946, %indices_947 = torch.aten.sort %4016, %int0_941, %false_944 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_948 = torch.constant.int 0 | |
%4017 = torch.aten.select.int %values_946, %int0_941, %int0_948 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4018 = torch.aten.item %4017 : !torch.vtensor<[1],si64> -> !torch.int | |
%4019 = torch.aten.unsqueeze %3945, %4018 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4020 = torch.prim.ListConstruct %3995, %4007, %4019 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
%int0_949 = torch.constant.int 0 | |
%4021 = torch.aten.cat %4020, %int0_949 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
%4022 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_950 = torch.constant.int 0 | |
%int0_951 = torch.constant.int 0 | |
%int0_952 = torch.constant.int 0 | |
%4023 = torch.aten.select.int %4022, %int0_950, %int0_952 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4024 = torch.aten.item %4023 : !torch.vtensor<[1],si64> -> !torch.int | |
%4025 = torch.aten.lt.int %4024, %int0_950 : !torch.int, !torch.int -> !torch.bool | |
%4026 = torch.aten.Int.bool %4025 : !torch.bool -> !torch.int | |
%4027 = torch.aten.mul.int %4026, %int0_951 : !torch.int, !torch.int -> !torch.int | |
%4028 = torch.aten.add.int %4024, %4027 : !torch.int, !torch.int -> !torch.int | |
%4029 = torch.prim.ListConstruct %4028 : (!torch.int) -> !torch.list<int> | |
%false_953 = torch.constant.bool false | |
%none_954 = torch.constant.none | |
%4030 = torch.aten.tensor %4029, %none_954, %none_954, %false_953 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_955, %indices_956 = torch.aten.sort %4030, %int0_950, %false_953 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_957 = torch.constant.int 0 | |
%4031 = torch.aten.select.int %values_955, %int0_950, %int0_957 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4032 = torch.aten.item %4031 : !torch.vtensor<[1],si64> -> !torch.int | |
%4033 = torch.aten.unsqueeze %3067, %4032 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4034 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_958 = torch.constant.int 0 | |
%int0_959 = torch.constant.int 0 | |
%int0_960 = torch.constant.int 0 | |
%4035 = torch.aten.select.int %4034, %int0_958, %int0_960 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4036 = torch.aten.item %4035 : !torch.vtensor<[1],si64> -> !torch.int | |
%4037 = torch.aten.lt.int %4036, %int0_958 : !torch.int, !torch.int -> !torch.bool | |
%4038 = torch.aten.Int.bool %4037 : !torch.bool -> !torch.int | |
%4039 = torch.aten.mul.int %4038, %int0_959 : !torch.int, !torch.int -> !torch.int | |
%4040 = torch.aten.add.int %4036, %4039 : !torch.int, !torch.int -> !torch.int | |
%4041 = torch.prim.ListConstruct %4040 : (!torch.int) -> !torch.list<int> | |
%false_961 = torch.constant.bool false | |
%none_962 = torch.constant.none | |
%4042 = torch.aten.tensor %4041, %none_962, %none_962, %false_961 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_963, %indices_964 = torch.aten.sort %4042, %int0_958, %false_961 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_965 = torch.constant.int 0 | |
%4043 = torch.aten.select.int %values_963, %int0_958, %int0_965 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4044 = torch.aten.item %4043 : !torch.vtensor<[1],si64> -> !torch.int | |
%4045 = torch.aten.unsqueeze %3944, %4044 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4046 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_966 = torch.constant.int 0 | |
%int0_967 = torch.constant.int 0 | |
%int0_968 = torch.constant.int 0 | |
%4047 = torch.aten.select.int %4046, %int0_966, %int0_968 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4048 = torch.aten.item %4047 : !torch.vtensor<[1],si64> -> !torch.int | |
%4049 = torch.aten.lt.int %4048, %int0_966 : !torch.int, !torch.int -> !torch.bool | |
%4050 = torch.aten.Int.bool %4049 : !torch.bool -> !torch.int | |
%4051 = torch.aten.mul.int %4050, %int0_967 : !torch.int, !torch.int -> !torch.int | |
%4052 = torch.aten.add.int %4048, %4051 : !torch.int, !torch.int -> !torch.int | |
%4053 = torch.prim.ListConstruct %4052 : (!torch.int) -> !torch.list<int> | |
%false_969 = torch.constant.bool false | |
%none_970 = torch.constant.none | |
%4054 = torch.aten.tensor %4053, %none_970, %none_970, %false_969 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_971, %indices_972 = torch.aten.sort %4054, %int0_966, %false_969 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_973 = torch.constant.int 0 | |
%4055 = torch.aten.select.int %values_971, %int0_966, %int0_973 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4056 = torch.aten.item %4055 : !torch.vtensor<[1],si64> -> !torch.int | |
%4057 = torch.aten.unsqueeze %3945, %4056 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4058 = torch.prim.ListConstruct %4033, %4045, %4057 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
%int0_974 = torch.constant.int 0 | |
%4059 = torch.aten.cat %4058, %int0_974 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
%int0_975 = torch.constant.int 0 | |
%int0_976 = torch.constant.int 0 | |
%4060 = torch.aten.select.int %3983, %int0_975, %int0_976 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4061 = torch.aten.item %4060 : !torch.vtensor<[1],si64> -> !torch.int | |
%4062 = torch.aten.eq.int %4061, %int0_975 : !torch.int, !torch.int -> !torch.bool | |
%4063 = torch.aten.Int.bool %4062 : !torch.bool -> !torch.int | |
%int0_977 = torch.constant.int 0 | |
%4064 = torch.aten.size.int %3359, %int0_977 : !torch.vtensor<[1,144,80,80],f32>, !torch.int -> !torch.int | |
%4065 = torch.prim.NumToTensor.Scalar %4063 : !torch.int -> !torch.vtensor<[],i1> | |
%4066 = torch.prim.NumToTensor.Scalar %4064 : !torch.int -> !torch.vtensor<[],si64> | |
%4067 = torch.prim.NumToTensor.Scalar %4061 : !torch.int -> !torch.vtensor<[],si64> | |
%4068 = torch.aten.where.self %4065, %4066, %4067 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4069 = torch.aten.item %4068 : !torch.vtensor<[],si64> -> !torch.int | |
%int1_978 = torch.constant.int 1 | |
%4070 = torch.aten.select.int %3983, %int0_975, %int1_978 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4071 = torch.aten.item %4070 : !torch.vtensor<[1],si64> -> !torch.int | |
%4072 = torch.aten.eq.int %4071, %int0_975 : !torch.int, !torch.int -> !torch.bool | |
%4073 = torch.aten.Int.bool %4072 : !torch.bool -> !torch.int | |
%int1_979 = torch.constant.int 1 | |
%4074 = torch.aten.size.int %3359, %int1_979 : !torch.vtensor<[1,144,80,80],f32>, !torch.int -> !torch.int | |
%4075 = torch.prim.NumToTensor.Scalar %4073 : !torch.int -> !torch.vtensor<[],i1> | |
%4076 = torch.prim.NumToTensor.Scalar %4074 : !torch.int -> !torch.vtensor<[],si64> | |
%4077 = torch.prim.NumToTensor.Scalar %4071 : !torch.int -> !torch.vtensor<[],si64> | |
%4078 = torch.aten.where.self %4075, %4076, %4077 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4079 = torch.aten.item %4078 : !torch.vtensor<[],si64> -> !torch.int | |
%int2_980 = torch.constant.int 2 | |
%4080 = torch.aten.select.int %3983, %int0_975, %int2_980 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4081 = torch.aten.item %4080 : !torch.vtensor<[1],si64> -> !torch.int | |
%4082 = torch.aten.eq.int %4081, %int0_975 : !torch.int, !torch.int -> !torch.bool | |
%4083 = torch.aten.Int.bool %4082 : !torch.bool -> !torch.int | |
%int2_981 = torch.constant.int 2 | |
%4084 = torch.aten.size.int %3359, %int2_981 : !torch.vtensor<[1,144,80,80],f32>, !torch.int -> !torch.int | |
%4085 = torch.prim.NumToTensor.Scalar %4083 : !torch.int -> !torch.vtensor<[],i1> | |
%4086 = torch.prim.NumToTensor.Scalar %4084 : !torch.int -> !torch.vtensor<[],si64> | |
%4087 = torch.prim.NumToTensor.Scalar %4081 : !torch.int -> !torch.vtensor<[],si64> | |
%4088 = torch.aten.where.self %4085, %4086, %4087 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4089 = torch.aten.item %4088 : !torch.vtensor<[],si64> -> !torch.int | |
%4090 = torch.prim.ListConstruct %4069, %4079, %4089 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%4091 = torch.aten.reshape %3359, %4090 : !torch.vtensor<[1,144,80,80],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
%int0_982 = torch.constant.int 0 | |
%int0_983 = torch.constant.int 0 | |
%4092 = torch.aten.select.int %4021, %int0_982, %int0_983 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4093 = torch.aten.item %4092 : !torch.vtensor<[1],si64> -> !torch.int | |
%4094 = torch.aten.eq.int %4093, %int0_982 : !torch.int, !torch.int -> !torch.bool | |
%4095 = torch.aten.Int.bool %4094 : !torch.bool -> !torch.int | |
%int0_984 = torch.constant.int 0 | |
%4096 = torch.aten.size.int %3651, %int0_984 : !torch.vtensor<[1,144,40,40],f32>, !torch.int -> !torch.int | |
%4097 = torch.prim.NumToTensor.Scalar %4095 : !torch.int -> !torch.vtensor<[],i1> | |
%4098 = torch.prim.NumToTensor.Scalar %4096 : !torch.int -> !torch.vtensor<[],si64> | |
%4099 = torch.prim.NumToTensor.Scalar %4093 : !torch.int -> !torch.vtensor<[],si64> | |
%4100 = torch.aten.where.self %4097, %4098, %4099 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4101 = torch.aten.item %4100 : !torch.vtensor<[],si64> -> !torch.int | |
%int1_985 = torch.constant.int 1 | |
%4102 = torch.aten.select.int %4021, %int0_982, %int1_985 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4103 = torch.aten.item %4102 : !torch.vtensor<[1],si64> -> !torch.int | |
%4104 = torch.aten.eq.int %4103, %int0_982 : !torch.int, !torch.int -> !torch.bool | |
%4105 = torch.aten.Int.bool %4104 : !torch.bool -> !torch.int | |
%int1_986 = torch.constant.int 1 | |
%4106 = torch.aten.size.int %3651, %int1_986 : !torch.vtensor<[1,144,40,40],f32>, !torch.int -> !torch.int | |
%4107 = torch.prim.NumToTensor.Scalar %4105 : !torch.int -> !torch.vtensor<[],i1> | |
%4108 = torch.prim.NumToTensor.Scalar %4106 : !torch.int -> !torch.vtensor<[],si64> | |
%4109 = torch.prim.NumToTensor.Scalar %4103 : !torch.int -> !torch.vtensor<[],si64> | |
%4110 = torch.aten.where.self %4107, %4108, %4109 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4111 = torch.aten.item %4110 : !torch.vtensor<[],si64> -> !torch.int | |
%int2_987 = torch.constant.int 2 | |
%4112 = torch.aten.select.int %4021, %int0_982, %int2_987 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4113 = torch.aten.item %4112 : !torch.vtensor<[1],si64> -> !torch.int | |
%4114 = torch.aten.eq.int %4113, %int0_982 : !torch.int, !torch.int -> !torch.bool | |
%4115 = torch.aten.Int.bool %4114 : !torch.bool -> !torch.int | |
%int2_988 = torch.constant.int 2 | |
%4116 = torch.aten.size.int %3651, %int2_988 : !torch.vtensor<[1,144,40,40],f32>, !torch.int -> !torch.int | |
%4117 = torch.prim.NumToTensor.Scalar %4115 : !torch.int -> !torch.vtensor<[],i1> | |
%4118 = torch.prim.NumToTensor.Scalar %4116 : !torch.int -> !torch.vtensor<[],si64> | |
%4119 = torch.prim.NumToTensor.Scalar %4113 : !torch.int -> !torch.vtensor<[],si64> | |
%4120 = torch.aten.where.self %4117, %4118, %4119 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4121 = torch.aten.item %4120 : !torch.vtensor<[],si64> -> !torch.int | |
%4122 = torch.prim.ListConstruct %4101, %4111, %4121 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%4123 = torch.aten.reshape %3651, %4122 : !torch.vtensor<[1,144,40,40],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
%int0_989 = torch.constant.int 0 | |
%int0_990 = torch.constant.int 0 | |
%4124 = torch.aten.select.int %4059, %int0_989, %int0_990 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4125 = torch.aten.item %4124 : !torch.vtensor<[1],si64> -> !torch.int | |
%4126 = torch.aten.eq.int %4125, %int0_989 : !torch.int, !torch.int -> !torch.bool | |
%4127 = torch.aten.Int.bool %4126 : !torch.bool -> !torch.int | |
%int0_991 = torch.constant.int 0 | |
%4128 = torch.aten.size.int %3943, %int0_991 : !torch.vtensor<[1,144,20,20],f32>, !torch.int -> !torch.int | |
%4129 = torch.prim.NumToTensor.Scalar %4127 : !torch.int -> !torch.vtensor<[],i1> | |
%4130 = torch.prim.NumToTensor.Scalar %4128 : !torch.int -> !torch.vtensor<[],si64> | |
%4131 = torch.prim.NumToTensor.Scalar %4125 : !torch.int -> !torch.vtensor<[],si64> | |
%4132 = torch.aten.where.self %4129, %4130, %4131 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4133 = torch.aten.item %4132 : !torch.vtensor<[],si64> -> !torch.int | |
%int1_992 = torch.constant.int 1 | |
%4134 = torch.aten.select.int %4059, %int0_989, %int1_992 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4135 = torch.aten.item %4134 : !torch.vtensor<[1],si64> -> !torch.int | |
%4136 = torch.aten.eq.int %4135, %int0_989 : !torch.int, !torch.int -> !torch.bool | |
%4137 = torch.aten.Int.bool %4136 : !torch.bool -> !torch.int | |
%int1_993 = torch.constant.int 1 | |
%4138 = torch.aten.size.int %3943, %int1_993 : !torch.vtensor<[1,144,20,20],f32>, !torch.int -> !torch.int | |
%4139 = torch.prim.NumToTensor.Scalar %4137 : !torch.int -> !torch.vtensor<[],i1> | |
%4140 = torch.prim.NumToTensor.Scalar %4138 : !torch.int -> !torch.vtensor<[],si64> | |
%4141 = torch.prim.NumToTensor.Scalar %4135 : !torch.int -> !torch.vtensor<[],si64> | |
%4142 = torch.aten.where.self %4139, %4140, %4141 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4143 = torch.aten.item %4142 : !torch.vtensor<[],si64> -> !torch.int | |
%int2_994 = torch.constant.int 2 | |
%4144 = torch.aten.select.int %4059, %int0_989, %int2_994 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4145 = torch.aten.item %4144 : !torch.vtensor<[1],si64> -> !torch.int | |
%4146 = torch.aten.eq.int %4145, %int0_989 : !torch.int, !torch.int -> !torch.bool | |
%4147 = torch.aten.Int.bool %4146 : !torch.bool -> !torch.int | |
%int2_995 = torch.constant.int 2 | |
%4148 = torch.aten.size.int %3943, %int2_995 : !torch.vtensor<[1,144,20,20],f32>, !torch.int -> !torch.int | |
%4149 = torch.prim.NumToTensor.Scalar %4147 : !torch.int -> !torch.vtensor<[],i1> | |
%4150 = torch.prim.NumToTensor.Scalar %4148 : !torch.int -> !torch.vtensor<[],si64> | |
%4151 = torch.prim.NumToTensor.Scalar %4145 : !torch.int -> !torch.vtensor<[],si64> | |
%4152 = torch.aten.where.self %4149, %4150, %4151 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4153 = torch.aten.item %4152 : !torch.vtensor<[],si64> -> !torch.int | |
%4154 = torch.prim.ListConstruct %4133, %4143, %4153 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%4155 = torch.aten.reshape %3943, %4154 : !torch.vtensor<[1,144,20,20],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
%4156 = torch.prim.ListConstruct %4091, %4123, %4155 : (!torch.vtensor<[?,?,?],f32>, !torch.vtensor<[?,?,?],f32>, !torch.vtensor<[?,?,?],f32>) -> !torch.list<vtensor> | |
%int2_996 = torch.constant.int 2 | |
%4157 = torch.aten.cat %4156, %int2_996 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
%4158 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4159 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_997 = torch.constant.int 12 | |
%4160 = torch.aten.item %4158 : !torch.vtensor<[],f32> -> !torch.float | |
%4161 = torch.aten.item %4159 : !torch.vtensor<[],si8> -> !torch.int | |
%4162 = torch.aten.quantize_per_tensor %4157, %4160, %4161, %int12_997 : !torch.vtensor<[?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],!torch.qint8> | |
%4163 = torch.aten.int_repr %4162 : !torch.vtensor<[?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?],si8> | |
%4164 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4165 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4166 = torch.aten.item %4164 : !torch.vtensor<[],f32> -> !torch.float | |
%4167 = torch.aten.item %4165 : !torch.vtensor<[],si8> -> !torch.int | |
%4168 = torch.aten._make_per_tensor_quantized_tensor %4163, %4166, %4167 : !torch.vtensor<[?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?],!torch.qint8> | |
%4169 = torch.aten.dequantize.self %4168 : !torch.vtensor<[?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?],f32> | |
%4170 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4171 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4172 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4173 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_998 = torch.constant.int 0 | |
%int0_999 = torch.constant.int 0 | |
%4174 = torch.prim.NumToTensor.Scalar %int0_999 : !torch.int -> !torch.vtensor<[1],si64> | |
%4175 = torch.aten.index_select %4171, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4176 = torch.aten.item %4175 : !torch.vtensor<[1],si64> -> !torch.int | |
%4177 = torch.aten.index_select %4172, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4178 = torch.aten.item %4177 : !torch.vtensor<[1],si64> -> !torch.int | |
%4179 = torch.aten.index_select %4170, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4180 = torch.aten.item %4179 : !torch.vtensor<[1],si64> -> !torch.int | |
%4181 = torch.aten.index_select %4173, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4182 = torch.aten.item %4181 : !torch.vtensor<[1],si64> -> !torch.int | |
%4183 = torch.aten.slice.Tensor %4169, %4180, %4176, %4178, %4182 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
%4184 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4185 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4186 = torch.vtensor.literal(dense<144> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4187 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1000 = torch.constant.int 0 | |
%int0_1001 = torch.constant.int 0 | |
%4188 = torch.prim.NumToTensor.Scalar %int0_1001 : !torch.int -> !torch.vtensor<[1],si64> | |
%4189 = torch.aten.index_select %4185, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4190 = torch.aten.item %4189 : !torch.vtensor<[1],si64> -> !torch.int | |
%4191 = torch.aten.index_select %4186, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4192 = torch.aten.item %4191 : !torch.vtensor<[1],si64> -> !torch.int | |
%4193 = torch.aten.index_select %4184, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4194 = torch.aten.item %4193 : !torch.vtensor<[1],si64> -> !torch.int | |
%4195 = torch.aten.index_select %4187, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4196 = torch.aten.item %4195 : !torch.vtensor<[1],si64> -> !torch.int | |
%4197 = torch.aten.slice.Tensor %4169, %4194, %4190, %4192, %4196 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
%4198 = torch.vtensor.literal(dense<1> : tensor<si64>) : !torch.vtensor<[],si64> | |
%4199 = torch.vtensor.literal(dense<8400> : tensor<si64>) : !torch.vtensor<[],si64> | |
%4200 = torch.vtensor.literal(dense<4> : tensor<si64>) : !torch.vtensor<[],si64> | |
%4201 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1002 = torch.constant.int 0 | |
%int0_1003 = torch.constant.int 0 | |
%int0_1004 = torch.constant.int 0 | |
%4202 = torch.aten.select.int %4201, %int0_1002, %int0_1004 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4203 = torch.aten.item %4202 : !torch.vtensor<[1],si64> -> !torch.int | |
%4204 = torch.aten.lt.int %4203, %int0_1002 : !torch.int, !torch.int -> !torch.bool | |
%4205 = torch.aten.Int.bool %4204 : !torch.bool -> !torch.int | |
%4206 = torch.aten.mul.int %4205, %int0_1003 : !torch.int, !torch.int -> !torch.int | |
%4207 = torch.aten.add.int %4203, %4206 : !torch.int, !torch.int -> !torch.int | |
%4208 = torch.prim.ListConstruct %4207 : (!torch.int) -> !torch.list<int> | |
%false_1005 = torch.constant.bool false | |
%none_1006 = torch.constant.none | |
%4209 = torch.aten.tensor %4208, %none_1006, %none_1006, %false_1005 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1007, %indices_1008 = torch.aten.sort %4209, %int0_1002, %false_1005 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1009 = torch.constant.int 0 | |
%4210 = torch.aten.select.int %values_1007, %int0_1002, %int0_1009 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4211 = torch.aten.item %4210 : !torch.vtensor<[1],si64> -> !torch.int | |
%4212 = torch.aten.unsqueeze %4198, %4211 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4213 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1010 = torch.constant.int 0 | |
%int0_1011 = torch.constant.int 0 | |
%int0_1012 = torch.constant.int 0 | |
%4214 = torch.aten.select.int %4213, %int0_1010, %int0_1012 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4215 = torch.aten.item %4214 : !torch.vtensor<[1],si64> -> !torch.int | |
%4216 = torch.aten.lt.int %4215, %int0_1010 : !torch.int, !torch.int -> !torch.bool | |
%4217 = torch.aten.Int.bool %4216 : !torch.bool -> !torch.int | |
%4218 = torch.aten.mul.int %4217, %int0_1011 : !torch.int, !torch.int -> !torch.int | |
%4219 = torch.aten.add.int %4215, %4218 : !torch.int, !torch.int -> !torch.int | |
%4220 = torch.prim.ListConstruct %4219 : (!torch.int) -> !torch.list<int> | |
%false_1013 = torch.constant.bool false | |
%none_1014 = torch.constant.none | |
%4221 = torch.aten.tensor %4220, %none_1014, %none_1014, %false_1013 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1015, %indices_1016 = torch.aten.sort %4221, %int0_1010, %false_1013 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1017 = torch.constant.int 0 | |
%4222 = torch.aten.select.int %values_1015, %int0_1010, %int0_1017 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4223 = torch.aten.item %4222 : !torch.vtensor<[1],si64> -> !torch.int | |
%4224 = torch.aten.unsqueeze %4200, %4223 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4225 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1018 = torch.constant.int 0 | |
%int0_1019 = torch.constant.int 0 | |
%int0_1020 = torch.constant.int 0 | |
%4226 = torch.aten.select.int %4225, %int0_1018, %int0_1020 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4227 = torch.aten.item %4226 : !torch.vtensor<[1],si64> -> !torch.int | |
%4228 = torch.aten.lt.int %4227, %int0_1018 : !torch.int, !torch.int -> !torch.bool | |
%4229 = torch.aten.Int.bool %4228 : !torch.bool -> !torch.int | |
%4230 = torch.aten.mul.int %4229, %int0_1019 : !torch.int, !torch.int -> !torch.int | |
%4231 = torch.aten.add.int %4227, %4230 : !torch.int, !torch.int -> !torch.int | |
%4232 = torch.prim.ListConstruct %4231 : (!torch.int) -> !torch.list<int> | |
%false_1021 = torch.constant.bool false | |
%none_1022 = torch.constant.none | |
%4233 = torch.aten.tensor %4232, %none_1022, %none_1022, %false_1021 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1023, %indices_1024 = torch.aten.sort %4233, %int0_1018, %false_1021 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1025 = torch.constant.int 0 | |
%4234 = torch.aten.select.int %values_1023, %int0_1018, %int0_1025 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4235 = torch.aten.item %4234 : !torch.vtensor<[1],si64> -> !torch.int | |
%4236 = torch.aten.unsqueeze %343, %4235 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4237 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1026 = torch.constant.int 0 | |
%int0_1027 = torch.constant.int 0 | |
%int0_1028 = torch.constant.int 0 | |
%4238 = torch.aten.select.int %4237, %int0_1026, %int0_1028 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4239 = torch.aten.item %4238 : !torch.vtensor<[1],si64> -> !torch.int | |
%4240 = torch.aten.lt.int %4239, %int0_1026 : !torch.int, !torch.int -> !torch.bool | |
%4241 = torch.aten.Int.bool %4240 : !torch.bool -> !torch.int | |
%4242 = torch.aten.mul.int %4241, %int0_1027 : !torch.int, !torch.int -> !torch.int | |
%4243 = torch.aten.add.int %4239, %4242 : !torch.int, !torch.int -> !torch.int | |
%4244 = torch.prim.ListConstruct %4243 : (!torch.int) -> !torch.list<int> | |
%false_1029 = torch.constant.bool false | |
%none_1030 = torch.constant.none | |
%4245 = torch.aten.tensor %4244, %none_1030, %none_1030, %false_1029 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1031, %indices_1032 = torch.aten.sort %4245, %int0_1026, %false_1029 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1033 = torch.constant.int 0 | |
%4246 = torch.aten.select.int %values_1031, %int0_1026, %int0_1033 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4247 = torch.aten.item %4246 : !torch.vtensor<[1],si64> -> !torch.int | |
%4248 = torch.aten.unsqueeze %4199, %4247 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4249 = torch.prim.ListConstruct %4212, %4224, %4236, %4248 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
%int0_1034 = torch.constant.int 0 | |
%4250 = torch.aten.cat %4249, %int0_1034 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],si64> | |
%int0_1035 = torch.constant.int 0 | |
%int0_1036 = torch.constant.int 0 | |
%4251 = torch.aten.select.int %4250, %int0_1035, %int0_1036 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4252 = torch.aten.item %4251 : !torch.vtensor<[1],si64> -> !torch.int | |
%4253 = torch.aten.eq.int %4252, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
%4254 = torch.aten.Int.bool %4253 : !torch.bool -> !torch.int | |
%int0_1037 = torch.constant.int 0 | |
%4255 = torch.aten.size.int %4183, %int0_1037 : !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.int | |
%4256 = torch.prim.NumToTensor.Scalar %4254 : !torch.int -> !torch.vtensor<[],i1> | |
%4257 = torch.prim.NumToTensor.Scalar %4255 : !torch.int -> !torch.vtensor<[],si64> | |
%4258 = torch.prim.NumToTensor.Scalar %4252 : !torch.int -> !torch.vtensor<[],si64> | |
%4259 = torch.aten.where.self %4256, %4257, %4258 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4260 = torch.aten.item %4259 : !torch.vtensor<[],si64> -> !torch.int | |
%int1_1038 = torch.constant.int 1 | |
%4261 = torch.aten.select.int %4250, %int0_1035, %int1_1038 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4262 = torch.aten.item %4261 : !torch.vtensor<[1],si64> -> !torch.int | |
%4263 = torch.aten.eq.int %4262, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
%4264 = torch.aten.Int.bool %4263 : !torch.bool -> !torch.int | |
%int1_1039 = torch.constant.int 1 | |
%4265 = torch.aten.size.int %4183, %int1_1039 : !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.int | |
%4266 = torch.prim.NumToTensor.Scalar %4264 : !torch.int -> !torch.vtensor<[],i1> | |
%4267 = torch.prim.NumToTensor.Scalar %4265 : !torch.int -> !torch.vtensor<[],si64> | |
%4268 = torch.prim.NumToTensor.Scalar %4262 : !torch.int -> !torch.vtensor<[],si64> | |
%4269 = torch.aten.where.self %4266, %4267, %4268 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4270 = torch.aten.item %4269 : !torch.vtensor<[],si64> -> !torch.int | |
%int2_1040 = torch.constant.int 2 | |
%4271 = torch.aten.select.int %4250, %int0_1035, %int2_1040 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4272 = torch.aten.item %4271 : !torch.vtensor<[1],si64> -> !torch.int | |
%4273 = torch.aten.eq.int %4272, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
%4274 = torch.aten.Int.bool %4273 : !torch.bool -> !torch.int | |
%int2_1041 = torch.constant.int 2 | |
%4275 = torch.aten.size.int %4183, %int2_1041 : !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.int | |
%4276 = torch.prim.NumToTensor.Scalar %4274 : !torch.int -> !torch.vtensor<[],i1> | |
%4277 = torch.prim.NumToTensor.Scalar %4275 : !torch.int -> !torch.vtensor<[],si64> | |
%4278 = torch.prim.NumToTensor.Scalar %4272 : !torch.int -> !torch.vtensor<[],si64> | |
%4279 = torch.aten.where.self %4276, %4277, %4278 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4280 = torch.aten.item %4279 : !torch.vtensor<[],si64> -> !torch.int | |
%int3 = torch.constant.int 3 | |
%4281 = torch.aten.select.int %4250, %int0_1035, %int3 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4282 = torch.aten.item %4281 : !torch.vtensor<[1],si64> -> !torch.int | |
%4283 = torch.aten.eq.int %4282, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
%4284 = torch.aten.Int.bool %4283 : !torch.bool -> !torch.int | |
%4285 = torch.prim.ListConstruct %4260, %4270, %4280, %4282 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%4286 = torch.aten.reshape %4183, %4285 : !torch.vtensor<[?,?,?],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?,?],f32> | |
%int1_1042 = torch.constant.int 1 | |
%int2_1043 = torch.constant.int 2 | |
%4287 = torch.aten.transpose.int %4286, %int1_1042, %int2_1043 : !torch.vtensor<[?,?,?,?],f32>, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],f32> | |
%4288 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4289 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1044 = torch.constant.int 12 | |
%4290 = torch.aten.item %4288 : !torch.vtensor<[],f32> -> !torch.float | |
%4291 = torch.aten.item %4289 : !torch.vtensor<[],si8> -> !torch.int | |
%4292 = torch.aten.quantize_per_tensor %4287, %4290, %4291, %int12_1044 : !torch.vtensor<[?,?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
%4293 = torch.aten.int_repr %4292 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],si8> | |
%4294 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4295 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4296 = torch.aten.item %4294 : !torch.vtensor<[],f32> -> !torch.float | |
%4297 = torch.aten.item %4295 : !torch.vtensor<[],si8> -> !torch.int | |
%4298 = torch.aten._make_per_tensor_quantized_tensor %4293, %4296, %4297 : !torch.vtensor<[?,?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
%4299 = torch.aten.dequantize.self %4298 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],f32> | |
%int1_1045 = torch.constant.int 1 | |
%none_1046 = torch.constant.none | |
%4300 = torch.aten.softmax.int %4299, %int1_1045, %none_1046 : !torch.vtensor<[?,?,?,?],f32>, !torch.int, !torch.none -> !torch.vtensor<[?,?,?,?],f32> | |
%4301 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4302 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1047 = torch.constant.int 12 | |
%4303 = torch.aten.item %4301 : !torch.vtensor<[],f32> -> !torch.float | |
%4304 = torch.aten.item %4302 : !torch.vtensor<[],si8> -> !torch.int | |
%4305 = torch.aten.quantize_per_tensor %4300, %4303, %4304, %int12_1047 : !torch.vtensor<[?,?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
%4306 = torch.aten.int_repr %4305 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],si8> | |
%4307 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4308 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4309 = torch.aten.item %4307 : !torch.vtensor<[],f32> -> !torch.float | |
%4310 = torch.aten.item %4308 : !torch.vtensor<[],si8> -> !torch.int | |
%4311 = torch.aten._make_per_tensor_quantized_tensor %4306, %4309, %4310 : !torch.vtensor<[?,?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
%4312 = torch.aten.dequantize.self %4311 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],f32> | |
%4313 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4314 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1048 = torch.constant.int 12 | |
%4315 = torch.aten.item %4313 : !torch.vtensor<[],f32> -> !torch.float | |
%4316 = torch.aten.item %4314 : !torch.vtensor<[],si8> -> !torch.int | |
%4317 = torch.aten.quantize_per_tensor %126, %4315, %4316, %int12_1048 : !torch.vtensor<[1,16,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,1,1],!torch.qint8> | |
%4318 = torch.aten.int_repr %4317 : !torch.vtensor<[1,16,1,1],!torch.qint8> -> !torch.vtensor<[1,16,1,1],si8> | |
%4319 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4320 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4321 = torch.aten.item %4319 : !torch.vtensor<[],f32> -> !torch.float | |
%4322 = torch.aten.item %4320 : !torch.vtensor<[],si8> -> !torch.int | |
%4323 = torch.aten._make_per_tensor_quantized_tensor %4318, %4321, %4322 : !torch.vtensor<[1,16,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,1,1],!torch.qint8> | |
%4324 = torch.aten.dequantize.self %4323 : !torch.vtensor<[1,16,1,1],!torch.qint8> -> !torch.vtensor<[1,16,1,1],f32> | |
%int0_1049 = torch.constant.int 0 | |
%int0_1050 = torch.constant.int 0 | |
%int1_1051 = torch.constant.int 1 | |
%int1_1052 = torch.constant.int 1 | |
%int1_1053 = torch.constant.int 1 | |
%int1_1054 = torch.constant.int 1 | |
%int0_1055 = torch.constant.int 0 | |
%4325 = torch.prim.ListConstruct %int0_1049, %int0_1050 : (!torch.int, !torch.int) -> !torch.list<int> | |
%4326 = torch.prim.ListConstruct %int1_1051, %int1_1052 : (!torch.int, !torch.int) -> !torch.list<int> | |
%4327 = torch.prim.ListConstruct %int1_1053, %int1_1054 : (!torch.int, !torch.int) -> !torch.list<int> | |
%4328 = torch.prim.ListConstruct %int0_1055, %int0_1055 : (!torch.int, !torch.int) -> !torch.list<int> | |
%false_1056 = torch.constant.bool false | |
%none_1057 = torch.constant.none | |
%int1_1058 = torch.constant.int 1 | |
%4329 = torch.aten.convolution %4312, %4324, %none_1057, %4327, %4325, %4326, %false_1056, %4328, %int1_1058 : !torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[1,16,1,1],f32>, !torch.none, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[?,1,?,?],f32> | |
%4330 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4331 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1059 = torch.constant.int 12 | |
%4332 = torch.aten.item %4330 : !torch.vtensor<[],f32> -> !torch.float | |
%4333 = torch.aten.item %4331 : !torch.vtensor<[],si8> -> !torch.int | |
%4334 = torch.aten.quantize_per_tensor %4329, %4332, %4333, %int12_1059 : !torch.vtensor<[?,1,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,1,?,?],!torch.qint8> | |
%4335 = torch.aten.int_repr %4334 : !torch.vtensor<[?,1,?,?],!torch.qint8> -> !torch.vtensor<[?,1,?,?],si8> | |
%4336 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4337 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4338 = torch.aten.item %4336 : !torch.vtensor<[],f32> -> !torch.float | |
%4339 = torch.aten.item %4337 : !torch.vtensor<[],si8> -> !torch.int | |
%4340 = torch.aten._make_per_tensor_quantized_tensor %4335, %4338, %4339 : !torch.vtensor<[?,1,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,1,?,?],!torch.qint8> | |
%4341 = torch.aten.dequantize.self %4340 : !torch.vtensor<[?,1,?,?],!torch.qint8> -> !torch.vtensor<[?,1,?,?],f32> | |
%4342 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1060 = torch.constant.int 0 | |
%int0_1061 = torch.constant.int 0 | |
%int0_1062 = torch.constant.int 0 | |
%4343 = torch.aten.select.int %4342, %int0_1060, %int0_1062 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4344 = torch.aten.item %4343 : !torch.vtensor<[1],si64> -> !torch.int | |
%4345 = torch.aten.lt.int %4344, %int0_1060 : !torch.int, !torch.int -> !torch.bool | |
%4346 = torch.aten.Int.bool %4345 : !torch.bool -> !torch.int | |
%4347 = torch.aten.mul.int %4346, %int0_1061 : !torch.int, !torch.int -> !torch.int | |
%4348 = torch.aten.add.int %4344, %4347 : !torch.int, !torch.int -> !torch.int | |
%4349 = torch.prim.ListConstruct %4348 : (!torch.int) -> !torch.list<int> | |
%false_1063 = torch.constant.bool false | |
%none_1064 = torch.constant.none | |
%4350 = torch.aten.tensor %4349, %none_1064, %none_1064, %false_1063 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1065, %indices_1066 = torch.aten.sort %4350, %int0_1060, %false_1063 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1067 = torch.constant.int 0 | |
%4351 = torch.aten.select.int %values_1065, %int0_1060, %int0_1067 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4352 = torch.aten.item %4351 : !torch.vtensor<[1],si64> -> !torch.int | |
%4353 = torch.aten.unsqueeze %4198, %4352 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4354 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1068 = torch.constant.int 0 | |
%int0_1069 = torch.constant.int 0 | |
%int0_1070 = torch.constant.int 0 | |
%4355 = torch.aten.select.int %4354, %int0_1068, %int0_1070 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4356 = torch.aten.item %4355 : !torch.vtensor<[1],si64> -> !torch.int | |
%4357 = torch.aten.lt.int %4356, %int0_1068 : !torch.int, !torch.int -> !torch.bool | |
%4358 = torch.aten.Int.bool %4357 : !torch.bool -> !torch.int | |
%4359 = torch.aten.mul.int %4358, %int0_1069 : !torch.int, !torch.int -> !torch.int | |
%4360 = torch.aten.add.int %4356, %4359 : !torch.int, !torch.int -> !torch.int | |
%4361 = torch.prim.ListConstruct %4360 : (!torch.int) -> !torch.list<int> | |
%false_1071 = torch.constant.bool false | |
%none_1072 = torch.constant.none | |
%4362 = torch.aten.tensor %4361, %none_1072, %none_1072, %false_1071 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1073, %indices_1074 = torch.aten.sort %4362, %int0_1068, %false_1071 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1075 = torch.constant.int 0 | |
%4363 = torch.aten.select.int %values_1073, %int0_1068, %int0_1075 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4364 = torch.aten.item %4363 : !torch.vtensor<[1],si64> -> !torch.int | |
%4365 = torch.aten.unsqueeze %4200, %4364 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4366 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1076 = torch.constant.int 0 | |
%int0_1077 = torch.constant.int 0 | |
%int0_1078 = torch.constant.int 0 | |
%4367 = torch.aten.select.int %4366, %int0_1076, %int0_1078 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4368 = torch.aten.item %4367 : !torch.vtensor<[1],si64> -> !torch.int | |
%4369 = torch.aten.lt.int %4368, %int0_1076 : !torch.int, !torch.int -> !torch.bool | |
%4370 = torch.aten.Int.bool %4369 : !torch.bool -> !torch.int | |
%4371 = torch.aten.mul.int %4370, %int0_1077 : !torch.int, !torch.int -> !torch.int | |
%4372 = torch.aten.add.int %4368, %4371 : !torch.int, !torch.int -> !torch.int | |
%4373 = torch.prim.ListConstruct %4372 : (!torch.int) -> !torch.list<int> | |
%false_1079 = torch.constant.bool false | |
%none_1080 = torch.constant.none | |
%4374 = torch.aten.tensor %4373, %none_1080, %none_1080, %false_1079 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
%values_1081, %indices_1082 = torch.aten.sort %4374, %int0_1076, %false_1079 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
%int0_1083 = torch.constant.int 0 | |
%4375 = torch.aten.select.int %values_1081, %int0_1076, %int0_1083 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4376 = torch.aten.item %4375 : !torch.vtensor<[1],si64> -> !torch.int | |
%4377 = torch.aten.unsqueeze %4199, %4376 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
%4378 = torch.prim.ListConstruct %4353, %4365, %4377 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
%int0_1084 = torch.constant.int 0 | |
%4379 = torch.aten.cat %4378, %int0_1084 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
%int0_1085 = torch.constant.int 0 | |
%int0_1086 = torch.constant.int 0 | |
%4380 = torch.aten.select.int %4379, %int0_1085, %int0_1086 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4381 = torch.aten.item %4380 : !torch.vtensor<[1],si64> -> !torch.int | |
%4382 = torch.aten.eq.int %4381, %int0_1085 : !torch.int, !torch.int -> !torch.bool | |
%4383 = torch.aten.Int.bool %4382 : !torch.bool -> !torch.int | |
%int0_1087 = torch.constant.int 0 | |
%4384 = torch.aten.size.int %4341, %int0_1087 : !torch.vtensor<[?,1,?,?],f32>, !torch.int -> !torch.int | |
%4385 = torch.prim.NumToTensor.Scalar %4383 : !torch.int -> !torch.vtensor<[],i1> | |
%4386 = torch.prim.NumToTensor.Scalar %4384 : !torch.int -> !torch.vtensor<[],si64> | |
%4387 = torch.prim.NumToTensor.Scalar %4381 : !torch.int -> !torch.vtensor<[],si64> | |
%4388 = torch.aten.where.self %4385, %4386, %4387 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4389 = torch.aten.item %4388 : !torch.vtensor<[],si64> -> !torch.int | |
%int1_1088 = torch.constant.int 1 | |
%4390 = torch.aten.select.int %4379, %int0_1085, %int1_1088 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4391 = torch.aten.item %4390 : !torch.vtensor<[1],si64> -> !torch.int | |
%4392 = torch.aten.eq.int %4391, %int0_1085 : !torch.int, !torch.int -> !torch.bool | |
%4393 = torch.aten.Int.bool %4392 : !torch.bool -> !torch.int | |
%int1_1089 = torch.constant.int 1 | |
%4394 = torch.aten.size.int %4341, %int1_1089 : !torch.vtensor<[?,1,?,?],f32>, !torch.int -> !torch.int | |
%4395 = torch.prim.NumToTensor.Scalar %4393 : !torch.int -> !torch.vtensor<[],i1> | |
%4396 = torch.prim.NumToTensor.Scalar %4394 : !torch.int -> !torch.vtensor<[],si64> | |
%4397 = torch.prim.NumToTensor.Scalar %4391 : !torch.int -> !torch.vtensor<[],si64> | |
%4398 = torch.aten.where.self %4395, %4396, %4397 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4399 = torch.aten.item %4398 : !torch.vtensor<[],si64> -> !torch.int | |
%int2_1090 = torch.constant.int 2 | |
%4400 = torch.aten.select.int %4379, %int0_1085, %int2_1090 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
%4401 = torch.aten.item %4400 : !torch.vtensor<[1],si64> -> !torch.int | |
%4402 = torch.aten.eq.int %4401, %int0_1085 : !torch.int, !torch.int -> !torch.bool | |
%4403 = torch.aten.Int.bool %4402 : !torch.bool -> !torch.int | |
%int2_1091 = torch.constant.int 2 | |
%4404 = torch.aten.size.int %4341, %int2_1091 : !torch.vtensor<[?,1,?,?],f32>, !torch.int -> !torch.int | |
%4405 = torch.prim.NumToTensor.Scalar %4403 : !torch.int -> !torch.vtensor<[],i1> | |
%4406 = torch.prim.NumToTensor.Scalar %4404 : !torch.int -> !torch.vtensor<[],si64> | |
%4407 = torch.prim.NumToTensor.Scalar %4401 : !torch.int -> !torch.vtensor<[],si64> | |
%4408 = torch.aten.where.self %4405, %4406, %4407 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
%4409 = torch.aten.item %4408 : !torch.vtensor<[],si64> -> !torch.int | |
%4410 = torch.prim.ListConstruct %4389, %4399, %4409 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%4411 = torch.aten.reshape %4341, %4410 : !torch.vtensor<[?,1,?,?],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
%4412 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4413 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4414 = torch.vtensor.literal(dense<2> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4415 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1092 = torch.constant.int 0 | |
%int0_1093 = torch.constant.int 0 | |
%4416 = torch.prim.NumToTensor.Scalar %int0_1093 : !torch.int -> !torch.vtensor<[1],si64> | |
%4417 = torch.aten.index_select %4413, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4418 = torch.aten.item %4417 : !torch.vtensor<[1],si64> -> !torch.int | |
%4419 = torch.aten.index_select %4414, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4420 = torch.aten.item %4419 : !torch.vtensor<[1],si64> -> !torch.int | |
%4421 = torch.aten.index_select %4412, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4422 = torch.aten.item %4421 : !torch.vtensor<[1],si64> -> !torch.int | |
%4423 = torch.aten.index_select %4415, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4424 = torch.aten.item %4423 : !torch.vtensor<[1],si64> -> !torch.int | |
%4425 = torch.aten.slice.Tensor %4411, %4422, %4418, %4420, %4424 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
%4426 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4427 = torch.vtensor.literal(dense<2> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4428 = torch.vtensor.literal(dense<4> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%4429 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
%int0_1094 = torch.constant.int 0 | |
%int0_1095 = torch.constant.int 0 | |
%4430 = torch.prim.NumToTensor.Scalar %int0_1095 : !torch.int -> !torch.vtensor<[1],si64> | |
%4431 = torch.aten.index_select %4427, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4432 = torch.aten.item %4431 : !torch.vtensor<[1],si64> -> !torch.int | |
%4433 = torch.aten.index_select %4428, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4434 = torch.aten.item %4433 : !torch.vtensor<[1],si64> -> !torch.int | |
%4435 = torch.aten.index_select %4426, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4436 = torch.aten.item %4435 : !torch.vtensor<[1],si64> -> !torch.int | |
%4437 = torch.aten.index_select %4429, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
%4438 = torch.aten.item %4437 : !torch.vtensor<[1],si64> -> !torch.int | |
%4439 = torch.aten.slice.Tensor %4411, %4436, %4432, %4434, %4438 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
%4440 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4441 = torch.aten.mul.Tensor %4425, %4440 : !torch.vtensor<[?,?,?],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[?,?,?],f32> | |
%int1_1096 = torch.constant.int 1 | |
%4442 = torch.aten.sub.Tensor %152, %4441, %int1_1096 : !torch.vtensor<[1,2,8400],f32>, !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
%4443 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4444 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1097 = torch.constant.int 12 | |
%4445 = torch.aten.item %4443 : !torch.vtensor<[],f32> -> !torch.float | |
%4446 = torch.aten.item %4444 : !torch.vtensor<[],si8> -> !torch.int | |
%4447 = torch.aten.quantize_per_tensor %4442, %4445, %4446, %int12_1097 : !torch.vtensor<[?,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
%4448 = torch.aten.int_repr %4447 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],si8> | |
%4449 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4450 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4451 = torch.aten.item %4449 : !torch.vtensor<[],f32> -> !torch.float | |
%4452 = torch.aten.item %4450 : !torch.vtensor<[],si8> -> !torch.int | |
%4453 = torch.aten._make_per_tensor_quantized_tensor %4448, %4451, %4452 : !torch.vtensor<[?,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
%4454 = torch.aten.dequantize.self %4453 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],f32> | |
%int1_1098 = torch.constant.int 1 | |
%4455 = torch.aten.add.Tensor %152, %4439, %int1_1098 : !torch.vtensor<[1,2,8400],f32>, !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
%4456 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4457 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1099 = torch.constant.int 12 | |
%4458 = torch.aten.item %4456 : !torch.vtensor<[],f32> -> !torch.float | |
%4459 = torch.aten.item %4457 : !torch.vtensor<[],si8> -> !torch.int | |
%4460 = torch.aten.quantize_per_tensor %4455, %4458, %4459, %int12_1099 : !torch.vtensor<[?,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
%4461 = torch.aten.int_repr %4460 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],si8> | |
%4462 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4463 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4464 = torch.aten.item %4462 : !torch.vtensor<[],f32> -> !torch.float | |
%4465 = torch.aten.item %4463 : !torch.vtensor<[],si8> -> !torch.int | |
%4466 = torch.aten._make_per_tensor_quantized_tensor %4461, %4464, %4465 : !torch.vtensor<[?,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
%4467 = torch.aten.dequantize.self %4466 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],f32> | |
%int1_1100 = torch.constant.int 1 | |
%4468 = torch.aten.add.Tensor %4454, %4467, %int1_1100 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[?,2,8400],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
%4469 = torch.vtensor.literal(dense<2.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4470 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1101 = torch.constant.int 12 | |
%4471 = torch.aten.item %4469 : !torch.vtensor<[],f32> -> !torch.float | |
%4472 = torch.aten.item %4470 : !torch.vtensor<[],si8> -> !torch.int | |
%4473 = torch.aten.quantize_per_tensor %4468, %4471, %4472, %int12_1101 : !torch.vtensor<[?,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
%4474 = torch.aten.int_repr %4473 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],si8> | |
%4475 = torch.vtensor.literal(dense<2.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4476 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4477 = torch.aten.item %4475 : !torch.vtensor<[],f32> -> !torch.float | |
%4478 = torch.aten.item %4476 : !torch.vtensor<[],si8> -> !torch.int | |
%4479 = torch.aten._make_per_tensor_quantized_tensor %4474, %4477, %4478 : !torch.vtensor<[?,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
%4480 = torch.aten.dequantize.self %4479 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],f32> | |
%4481 = torch.aten.div.Tensor %4480, %139 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[?,2,8400],f32> | |
%4482 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4483 = torch.aten.mul.Tensor %4454, %4482 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[?,2,8400],f32> | |
%int1_1102 = torch.constant.int 1 | |
%4484 = torch.aten.sub.Tensor %4467, %4483, %int1_1102 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[?,2,8400],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
%4485 = torch.prim.ListConstruct %4481, %4484 : (!torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[?,2,8400],f32>) -> !torch.list<vtensor> | |
%int1_1103 = torch.constant.int 1 | |
%4486 = torch.aten.cat %4485, %int1_1103 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[?,4,8400],f32> | |
%4487 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4488 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1104 = torch.constant.int 12 | |
%4489 = torch.aten.item %4487 : !torch.vtensor<[],f32> -> !torch.float | |
%4490 = torch.aten.item %4488 : !torch.vtensor<[],si8> -> !torch.int | |
%4491 = torch.aten.quantize_per_tensor %4486, %4489, %4490, %int12_1104 : !torch.vtensor<[?,4,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,4,8400],!torch.qint8> | |
%4492 = torch.aten.int_repr %4491 : !torch.vtensor<[?,4,8400],!torch.qint8> -> !torch.vtensor<[?,4,8400],si8> | |
%4493 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4494 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4495 = torch.aten.item %4493 : !torch.vtensor<[],f32> -> !torch.float | |
%4496 = torch.aten.item %4494 : !torch.vtensor<[],si8> -> !torch.int | |
%4497 = torch.aten._make_per_tensor_quantized_tensor %4492, %4495, %4496 : !torch.vtensor<[?,4,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,4,8400],!torch.qint8> | |
%4498 = torch.aten.dequantize.self %4497 : !torch.vtensor<[?,4,8400],!torch.qint8> -> !torch.vtensor<[?,4,8400],f32> | |
%4499 = torch.aten.mul.Tensor %4498, %165 : !torch.vtensor<[?,4,8400],f32>, !torch.vtensor<[1,8400],f32> -> !torch.vtensor<[?,4,8400],f32> | |
%4500 = torch.aten.sigmoid %4197 : !torch.vtensor<[?,?,?],f32> -> !torch.vtensor<[?,?,?],f32> | |
%4501 = torch.prim.ListConstruct %4499, %4500 : (!torch.vtensor<[?,4,8400],f32>, !torch.vtensor<[?,?,?],f32>) -> !torch.list<vtensor> | |
%int1_1105 = torch.constant.int 1 | |
%4502 = torch.aten.cat %4501, %int1_1105 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[?,?,8400],f32> | |
%4503 = torch.vtensor.literal(dense<8.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4504 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%int12_1106 = torch.constant.int 12 | |
%4505 = torch.aten.item %4503 : !torch.vtensor<[],f32> -> !torch.float | |
%4506 = torch.aten.item %4504 : !torch.vtensor<[],si8> -> !torch.int | |
%4507 = torch.aten.quantize_per_tensor %4502, %4505, %4506, %int12_1106 : !torch.vtensor<[?,?,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,8400],!torch.qint8> | |
%4508 = torch.aten.int_repr %4507 : !torch.vtensor<[?,?,8400],!torch.qint8> -> !torch.vtensor<[?,?,8400],si8> | |
%4509 = torch.vtensor.literal(dense<8.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
%4510 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
%4511 = torch.aten.item %4509 : !torch.vtensor<[],f32> -> !torch.float | |
%4512 = torch.aten.item %4510 : !torch.vtensor<[],si8> -> !torch.int | |
%4513 = torch.aten._make_per_tensor_quantized_tensor %4508, %4511, %4512 : !torch.vtensor<[?,?,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,84,8400],!torch.qint8> | |
%4514 = torch.aten.dequantize.self %4513 : !torch.vtensor<[1,84,8400],!torch.qint8> -> !torch.vtensor<[1,84,8400],f32> | |
return %4514, %3359, %3651, %3943 : !torch.vtensor<[1,84,8400],f32>, !torch.vtensor<[1,144,80,80],f32>, !torch.vtensor<[1,144,40,40],f32>, !torch.vtensor<[1,144,20,20],f32> | |
} | |
} | |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment