Last active
April 22, 2024 18:52
-
-
Save AmosLewis/1cb839df5dd34fd2f6dc701392aa6bc6 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| module { | |
| func.func @torch_jit(%arg0: !torch.vtensor<[1,3,640,640],f32>) -> (!torch.vtensor<[1,84,8400],f32>, !torch.vtensor<[1,144,80,80],f32>, !torch.vtensor<[1,144,40,40],f32>, !torch.vtensor<[1,144,20,20],f32>) attributes {torch.onnx_meta.ir_version = 8 : si64, torch.onnx_meta.opset_version = 17 : si64, torch.onnx_meta.producer_name = "pytorch", torch.onnx_meta.producer_version = "1.13.1"} { | |
| %0 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x3x3x3xf32>) : !torch.vtensor<[16,3,3,3],f32> | |
| %1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
| %2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x16x3x3xf32>) : !torch.vtensor<[32,16,3,3],f32> | |
| %3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %4 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x1x1xf32>) : !torch.vtensor<[32,32,1,1],f32> | |
| %5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %6 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x16x3x3xf32>) : !torch.vtensor<[16,16,3,3],f32> | |
| %7 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
| %8 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x16x3x3xf32>) : !torch.vtensor<[16,16,3,3],f32> | |
| %9 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
| %10 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x48x1x1xf32>) : !torch.vtensor<[32,48,1,1],f32> | |
| %11 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %12 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x32x3x3xf32>) : !torch.vtensor<[64,32,3,3],f32> | |
| %13 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %14 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
| %15 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %16 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %17 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %18 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %19 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %20 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %21 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %22 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %23 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %24 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x128x1x1xf32>) : !torch.vtensor<[64,128,1,1],f32> | |
| %25 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %26 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x64x3x3xf32>) : !torch.vtensor<[128,64,3,3],f32> | |
| %27 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %28 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x1x1xf32>) : !torch.vtensor<[128,128,1,1],f32> | |
| %29 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %30 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %31 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %32 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %33 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %34 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %35 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %36 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %37 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %38 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x256x1x1xf32>) : !torch.vtensor<[128,256,1,1],f32> | |
| %39 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %40 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x128x3x3xf32>) : !torch.vtensor<[256,128,3,3],f32> | |
| %41 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %42 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x256x1x1xf32>) : !torch.vtensor<[256,256,1,1],f32> | |
| %43 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %44 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %45 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %46 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %47 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %48 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x384x1x1xf32>) : !torch.vtensor<[256,384,1,1],f32> | |
| %49 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %50 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x256x1x1xf32>) : !torch.vtensor<[128,256,1,1],f32> | |
| %51 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %52 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x512x1x1xf32>) : !torch.vtensor<[256,512,1,1],f32> | |
| %53 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %54 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x384x1x1xf32>) : !torch.vtensor<[128,384,1,1],f32> | |
| %55 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %56 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %57 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %58 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %59 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %60 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x192x1x1xf32>) : !torch.vtensor<[128,192,1,1],f32> | |
| %61 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %62 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x192x1x1xf32>) : !torch.vtensor<[64,192,1,1],f32> | |
| %63 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %64 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %65 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %66 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %67 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %68 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x96x1x1xf32>) : !torch.vtensor<[64,96,1,1],f32> | |
| %69 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %70 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %71 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %72 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x192x1x1xf32>) : !torch.vtensor<[128,192,1,1],f32> | |
| %73 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %74 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %75 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %76 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %77 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %78 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x192x1x1xf32>) : !torch.vtensor<[128,192,1,1],f32> | |
| %79 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %80 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %81 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %82 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x384x1x1xf32>) : !torch.vtensor<[256,384,1,1],f32> | |
| %83 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %84 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %85 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %86 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %87 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %88 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x384x1x1xf32>) : !torch.vtensor<[256,384,1,1],f32> | |
| %89 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %90 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %91 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %92 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %93 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %94 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
| %95 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %96 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x64x3x3xf32>) : !torch.vtensor<[80,64,3,3],f32> | |
| %97 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %98 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x3x3xf32>) : !torch.vtensor<[80,80,3,3],f32> | |
| %99 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %100 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x1x1xf32>) : !torch.vtensor<[80,80,1,1],f32> | |
| %101 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %102 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x128x3x3xf32>) : !torch.vtensor<[64,128,3,3],f32> | |
| %103 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %104 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %105 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %106 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
| %107 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %108 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x128x3x3xf32>) : !torch.vtensor<[80,128,3,3],f32> | |
| %109 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %110 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x3x3xf32>) : !torch.vtensor<[80,80,3,3],f32> | |
| %111 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %112 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x1x1xf32>) : !torch.vtensor<[80,80,1,1],f32> | |
| %113 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %114 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x256x3x3xf32>) : !torch.vtensor<[64,256,3,3],f32> | |
| %115 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %116 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %117 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %118 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x1x1xf32>) : !torch.vtensor<[64,64,1,1],f32> | |
| %119 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %120 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x256x3x3xf32>) : !torch.vtensor<[80,256,3,3],f32> | |
| %121 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %122 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x3x3xf32>) : !torch.vtensor<[80,80,3,3],f32> | |
| %123 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %124 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x80x1x1xf32>) : !torch.vtensor<[80,80,1,1],f32> | |
| %125 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %126 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x16x1x1xf32>) : !torch.vtensor<[1,16,1,1],f32> | |
| %none = torch.constant.none | |
| %127 = torch.vtensor.literal(dense<2.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %128 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %129 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12 = torch.constant.int 12 | |
| %130 = torch.aten.item %128 : !torch.vtensor<[],f32> -> !torch.float | |
| %131 = torch.aten.item %129 : !torch.vtensor<[],si8> -> !torch.int | |
| %132 = torch.aten.quantize_per_tensor %127, %130, %131, %int12 : !torch.vtensor<[],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[],!torch.qint8> | |
| %133 = torch.aten.int_repr %132 : !torch.vtensor<[],!torch.qint8> -> !torch.vtensor<[],si8> | |
| %134 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %135 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %136 = torch.aten.item %134 : !torch.vtensor<[],f32> -> !torch.float | |
| %137 = torch.aten.item %135 : !torch.vtensor<[],si8> -> !torch.int | |
| %138 = torch.aten._make_per_tensor_quantized_tensor %133, %136, %137 : !torch.vtensor<[],si8>, !torch.float, !torch.int -> !torch.vtensor<[],!torch.qint8> | |
| %139 = torch.aten.dequantize.self %138 : !torch.vtensor<[],!torch.qint8> -> !torch.vtensor<[],f32> | |
| %140 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x2x8400xf32>) : !torch.vtensor<[1,2,8400],f32> | |
| %141 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %142 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_0 = torch.constant.int 12 | |
| %143 = torch.aten.item %141 : !torch.vtensor<[],f32> -> !torch.float | |
| %144 = torch.aten.item %142 : !torch.vtensor<[],si8> -> !torch.int | |
| %145 = torch.aten.quantize_per_tensor %140, %143, %144, %int12_0 : !torch.vtensor<[1,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,2,8400],!torch.qint8> | |
| %146 = torch.aten.int_repr %145 : !torch.vtensor<[1,2,8400],!torch.qint8> -> !torch.vtensor<[1,2,8400],si8> | |
| %147 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %148 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %149 = torch.aten.item %147 : !torch.vtensor<[],f32> -> !torch.float | |
| %150 = torch.aten.item %148 : !torch.vtensor<[],si8> -> !torch.int | |
| %151 = torch.aten._make_per_tensor_quantized_tensor %146, %149, %150 : !torch.vtensor<[1,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,2,8400],!torch.qint8> | |
| %152 = torch.aten.dequantize.self %151 : !torch.vtensor<[1,2,8400],!torch.qint8> -> !torch.vtensor<[1,2,8400],f32> | |
| %153 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x8400xf32>) : !torch.vtensor<[1,8400],f32> | |
| %154 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %155 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1 = torch.constant.int 12 | |
| %156 = torch.aten.item %154 : !torch.vtensor<[],f32> -> !torch.float | |
| %157 = torch.aten.item %155 : !torch.vtensor<[],si8> -> !torch.int | |
| %158 = torch.aten.quantize_per_tensor %153, %156, %157, %int12_1 : !torch.vtensor<[1,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,8400],!torch.qint8> | |
| %159 = torch.aten.int_repr %158 : !torch.vtensor<[1,8400],!torch.qint8> -> !torch.vtensor<[1,8400],si8> | |
| %160 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %161 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %162 = torch.aten.item %160 : !torch.vtensor<[],f32> -> !torch.float | |
| %163 = torch.aten.item %161 : !torch.vtensor<[],si8> -> !torch.int | |
| %164 = torch.aten._make_per_tensor_quantized_tensor %159, %162, %163 : !torch.vtensor<[1,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,8400],!torch.qint8> | |
| %165 = torch.aten.dequantize.self %164 : !torch.vtensor<[1,8400],!torch.qint8> -> !torch.vtensor<[1,8400],f32> | |
| %166 = torch.vtensor.literal(dense<2.44140625E-4> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %167 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_2 = torch.constant.int 12 | |
| %168 = torch.aten.item %166 : !torch.vtensor<[],f32> -> !torch.float | |
| %169 = torch.aten.item %167 : !torch.vtensor<[],si8> -> !torch.int | |
| %170 = torch.aten.quantize_per_tensor %arg0, %168, %169, %int12_2 : !torch.vtensor<[1,3,640,640],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,3,640,640],!torch.qint8> | |
| %171 = torch.aten.int_repr %170 : !torch.vtensor<[1,3,640,640],!torch.qint8> -> !torch.vtensor<[1,3,640,640],si8> | |
| %172 = torch.vtensor.literal(dense<2.44140625E-4> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %174 = torch.aten.item %172 : !torch.vtensor<[],f32> -> !torch.float | |
| %175 = torch.aten.item %173 : !torch.vtensor<[],si8> -> !torch.int | |
| %176 = torch.aten._make_per_tensor_quantized_tensor %171, %174, %175 : !torch.vtensor<[1,3,640,640],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,3,640,640],!torch.qint8> | |
| %177 = torch.aten.dequantize.self %176 : !torch.vtensor<[1,3,640,640],!torch.qint8> -> !torch.vtensor<[1,3,640,640],f32> | |
| %178 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_3 = torch.constant.int 12 | |
| %180 = torch.aten.item %178 : !torch.vtensor<[],f32> -> !torch.float | |
| %181 = torch.aten.item %179 : !torch.vtensor<[],si8> -> !torch.int | |
| %182 = torch.aten.quantize_per_tensor %0, %180, %181, %int12_3 : !torch.vtensor<[16,3,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,3,3,3],!torch.qint8> | |
| %183 = torch.aten.int_repr %182 : !torch.vtensor<[16,3,3,3],!torch.qint8> -> !torch.vtensor<[16,3,3,3],si8> | |
| %184 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %185 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %186 = torch.aten.item %184 : !torch.vtensor<[],f32> -> !torch.float | |
| %187 = torch.aten.item %185 : !torch.vtensor<[],si8> -> !torch.int | |
| %188 = torch.aten._make_per_tensor_quantized_tensor %183, %186, %187 : !torch.vtensor<[16,3,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,3,3,3],!torch.qint8> | |
| %189 = torch.aten.dequantize.self %188 : !torch.vtensor<[16,3,3,3],!torch.qint8> -> !torch.vtensor<[16,3,3,3],f32> | |
| %190 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %191 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_4 = torch.constant.int 12 | |
| %192 = torch.aten.item %190 : !torch.vtensor<[],f32> -> !torch.float | |
| %193 = torch.aten.item %191 : !torch.vtensor<[],si8> -> !torch.int | |
| %194 = torch.aten.quantize_per_tensor %1, %192, %193, %int12_4 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %195 = torch.aten.int_repr %194 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
| %196 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %197 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %198 = torch.aten.item %196 : !torch.vtensor<[],f32> -> !torch.float | |
| %199 = torch.aten.item %197 : !torch.vtensor<[],si8> -> !torch.int | |
| %200 = torch.aten._make_per_tensor_quantized_tensor %195, %198, %199 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %201 = torch.aten.dequantize.self %200 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
| %int1 = torch.constant.int 1 | |
| %int1_5 = torch.constant.int 1 | |
| %int1_6 = torch.constant.int 1 | |
| %int1_7 = torch.constant.int 1 | |
| %int2 = torch.constant.int 2 | |
| %int2_8 = torch.constant.int 2 | |
| %int0 = torch.constant.int 0 | |
| %202 = torch.prim.ListConstruct %int1, %int1_5 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %203 = torch.prim.ListConstruct %int1_6, %int1_7 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %204 = torch.prim.ListConstruct %int2, %int2_8 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %205 = torch.prim.ListConstruct %int0, %int0 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false = torch.constant.bool false | |
| %int1_9 = torch.constant.int 1 | |
| %206 = torch.aten.convolution %177, %189, %201, %204, %202, %203, %false, %205, %int1_9 : !torch.vtensor<[1,3,640,640],f32>, !torch.vtensor<[16,3,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,320,320],f32> | |
| %207 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %208 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_10 = torch.constant.int 12 | |
| %209 = torch.aten.item %207 : !torch.vtensor<[],f32> -> !torch.float | |
| %210 = torch.aten.item %208 : !torch.vtensor<[],si8> -> !torch.int | |
| %211 = torch.aten.quantize_per_tensor %206, %209, %210, %int12_10 : !torch.vtensor<[1,16,320,320],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
| %212 = torch.aten.int_repr %211 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],si8> | |
| %213 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %214 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %215 = torch.aten.item %213 : !torch.vtensor<[],f32> -> !torch.float | |
| %216 = torch.aten.item %214 : !torch.vtensor<[],si8> -> !torch.int | |
| %217 = torch.aten._make_per_tensor_quantized_tensor %212, %215, %216 : !torch.vtensor<[1,16,320,320],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
| %218 = torch.aten.dequantize.self %217 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],f32> | |
| %219 = torch.aten.sigmoid %218 : !torch.vtensor<[1,16,320,320],f32> -> !torch.vtensor<[1,16,320,320],f32> | |
| %220 = torch.aten.mul.Tensor %218, %219 : !torch.vtensor<[1,16,320,320],f32>, !torch.vtensor<[1,16,320,320],f32> -> !torch.vtensor<[1,16,320,320],f32> | |
| %221 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_11 = torch.constant.int 12 | |
| %223 = torch.aten.item %221 : !torch.vtensor<[],f32> -> !torch.float | |
| %224 = torch.aten.item %222 : !torch.vtensor<[],si8> -> !torch.int | |
| %225 = torch.aten.quantize_per_tensor %220, %223, %224, %int12_11 : !torch.vtensor<[1,16,320,320],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
| %226 = torch.aten.int_repr %225 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],si8> | |
| %227 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %228 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %229 = torch.aten.item %227 : !torch.vtensor<[],f32> -> !torch.float | |
| %230 = torch.aten.item %228 : !torch.vtensor<[],si8> -> !torch.int | |
| %231 = torch.aten._make_per_tensor_quantized_tensor %226, %229, %230 : !torch.vtensor<[1,16,320,320],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,320,320],!torch.qint8> | |
| %232 = torch.aten.dequantize.self %231 : !torch.vtensor<[1,16,320,320],!torch.qint8> -> !torch.vtensor<[1,16,320,320],f32> | |
| %233 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %234 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_12 = torch.constant.int 12 | |
| %235 = torch.aten.item %233 : !torch.vtensor<[],f32> -> !torch.float | |
| %236 = torch.aten.item %234 : !torch.vtensor<[],si8> -> !torch.int | |
| %237 = torch.aten.quantize_per_tensor %2, %235, %236, %int12_12 : !torch.vtensor<[32,16,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,16,3,3],!torch.qint8> | |
| %238 = torch.aten.int_repr %237 : !torch.vtensor<[32,16,3,3],!torch.qint8> -> !torch.vtensor<[32,16,3,3],si8> | |
| %239 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %240 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %241 = torch.aten.item %239 : !torch.vtensor<[],f32> -> !torch.float | |
| %242 = torch.aten.item %240 : !torch.vtensor<[],si8> -> !torch.int | |
| %243 = torch.aten._make_per_tensor_quantized_tensor %238, %241, %242 : !torch.vtensor<[32,16,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,16,3,3],!torch.qint8> | |
| %244 = torch.aten.dequantize.self %243 : !torch.vtensor<[32,16,3,3],!torch.qint8> -> !torch.vtensor<[32,16,3,3],f32> | |
| %245 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %246 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_13 = torch.constant.int 12 | |
| %247 = torch.aten.item %245 : !torch.vtensor<[],f32> -> !torch.float | |
| %248 = torch.aten.item %246 : !torch.vtensor<[],si8> -> !torch.int | |
| %249 = torch.aten.quantize_per_tensor %3, %247, %248, %int12_13 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %250 = torch.aten.int_repr %249 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %251 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %252 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %253 = torch.aten.item %251 : !torch.vtensor<[],f32> -> !torch.float | |
| %254 = torch.aten.item %252 : !torch.vtensor<[],si8> -> !torch.int | |
| %255 = torch.aten._make_per_tensor_quantized_tensor %250, %253, %254 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %256 = torch.aten.dequantize.self %255 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_14 = torch.constant.int 1 | |
| %int1_15 = torch.constant.int 1 | |
| %int1_16 = torch.constant.int 1 | |
| %int1_17 = torch.constant.int 1 | |
| %int2_18 = torch.constant.int 2 | |
| %int2_19 = torch.constant.int 2 | |
| %int0_20 = torch.constant.int 0 | |
| %257 = torch.prim.ListConstruct %int1_14, %int1_15 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %258 = torch.prim.ListConstruct %int1_16, %int1_17 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %259 = torch.prim.ListConstruct %int2_18, %int2_19 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %260 = torch.prim.ListConstruct %int0_20, %int0_20 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_21 = torch.constant.bool false | |
| %int1_22 = torch.constant.int 1 | |
| %261 = torch.aten.convolution %232, %244, %256, %259, %257, %258, %false_21, %260, %int1_22 : !torch.vtensor<[1,16,320,320],f32>, !torch.vtensor<[32,16,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,160,160],f32> | |
| %262 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %263 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_23 = torch.constant.int 12 | |
| %264 = torch.aten.item %262 : !torch.vtensor<[],f32> -> !torch.float | |
| %265 = torch.aten.item %263 : !torch.vtensor<[],si8> -> !torch.int | |
| %266 = torch.aten.quantize_per_tensor %261, %264, %265, %int12_23 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %267 = torch.aten.int_repr %266 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
| %268 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %269 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %270 = torch.aten.item %268 : !torch.vtensor<[],f32> -> !torch.float | |
| %271 = torch.aten.item %269 : !torch.vtensor<[],si8> -> !torch.int | |
| %272 = torch.aten._make_per_tensor_quantized_tensor %267, %270, %271 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %273 = torch.aten.dequantize.self %272 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
| %274 = torch.aten.sigmoid %273 : !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
| %275 = torch.aten.mul.Tensor %273, %274 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
| %276 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %277 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_24 = torch.constant.int 12 | |
| %278 = torch.aten.item %276 : !torch.vtensor<[],f32> -> !torch.float | |
| %279 = torch.aten.item %277 : !torch.vtensor<[],si8> -> !torch.int | |
| %280 = torch.aten.quantize_per_tensor %275, %278, %279, %int12_24 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %281 = torch.aten.int_repr %280 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
| %282 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %283 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %284 = torch.aten.item %282 : !torch.vtensor<[],f32> -> !torch.float | |
| %285 = torch.aten.item %283 : !torch.vtensor<[],si8> -> !torch.int | |
| %286 = torch.aten._make_per_tensor_quantized_tensor %281, %284, %285 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %287 = torch.aten.dequantize.self %286 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
| %288 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %289 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_25 = torch.constant.int 12 | |
| %290 = torch.aten.item %288 : !torch.vtensor<[],f32> -> !torch.float | |
| %291 = torch.aten.item %289 : !torch.vtensor<[],si8> -> !torch.int | |
| %292 = torch.aten.quantize_per_tensor %4, %290, %291, %int12_25 : !torch.vtensor<[32,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,1,1],!torch.qint8> | |
| %293 = torch.aten.int_repr %292 : !torch.vtensor<[32,32,1,1],!torch.qint8> -> !torch.vtensor<[32,32,1,1],si8> | |
| %294 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %295 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %296 = torch.aten.item %294 : !torch.vtensor<[],f32> -> !torch.float | |
| %297 = torch.aten.item %295 : !torch.vtensor<[],si8> -> !torch.int | |
| %298 = torch.aten._make_per_tensor_quantized_tensor %293, %296, %297 : !torch.vtensor<[32,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,1,1],!torch.qint8> | |
| %299 = torch.aten.dequantize.self %298 : !torch.vtensor<[32,32,1,1],!torch.qint8> -> !torch.vtensor<[32,32,1,1],f32> | |
| %300 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %301 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_26 = torch.constant.int 12 | |
| %302 = torch.aten.item %300 : !torch.vtensor<[],f32> -> !torch.float | |
| %303 = torch.aten.item %301 : !torch.vtensor<[],si8> -> !torch.int | |
| %304 = torch.aten.quantize_per_tensor %5, %302, %303, %int12_26 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %305 = torch.aten.int_repr %304 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %306 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %307 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %308 = torch.aten.item %306 : !torch.vtensor<[],f32> -> !torch.float | |
| %309 = torch.aten.item %307 : !torch.vtensor<[],si8> -> !torch.int | |
| %310 = torch.aten._make_per_tensor_quantized_tensor %305, %308, %309 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %311 = torch.aten.dequantize.self %310 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int0_27 = torch.constant.int 0 | |
| %int0_28 = torch.constant.int 0 | |
| %int1_29 = torch.constant.int 1 | |
| %int1_30 = torch.constant.int 1 | |
| %int1_31 = torch.constant.int 1 | |
| %int1_32 = torch.constant.int 1 | |
| %int0_33 = torch.constant.int 0 | |
| %312 = torch.prim.ListConstruct %int0_27, %int0_28 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %313 = torch.prim.ListConstruct %int1_29, %int1_30 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %314 = torch.prim.ListConstruct %int1_31, %int1_32 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %315 = torch.prim.ListConstruct %int0_33, %int0_33 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_34 = torch.constant.bool false | |
| %int1_35 = torch.constant.int 1 | |
| %316 = torch.aten.convolution %287, %299, %311, %314, %312, %313, %false_34, %315, %int1_35 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[32,32,1,1],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,160,160],f32> | |
| %317 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %318 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_36 = torch.constant.int 12 | |
| %319 = torch.aten.item %317 : !torch.vtensor<[],f32> -> !torch.float | |
| %320 = torch.aten.item %318 : !torch.vtensor<[],si8> -> !torch.int | |
| %321 = torch.aten.quantize_per_tensor %316, %319, %320, %int12_36 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %322 = torch.aten.int_repr %321 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
| %323 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %324 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %325 = torch.aten.item %323 : !torch.vtensor<[],f32> -> !torch.float | |
| %326 = torch.aten.item %324 : !torch.vtensor<[],si8> -> !torch.int | |
| %327 = torch.aten._make_per_tensor_quantized_tensor %322, %325, %326 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %328 = torch.aten.dequantize.self %327 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
| %329 = torch.aten.sigmoid %328 : !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
| %330 = torch.aten.mul.Tensor %328, %329 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
| %331 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %332 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_37 = torch.constant.int 12 | |
| %333 = torch.aten.item %331 : !torch.vtensor<[],f32> -> !torch.float | |
| %334 = torch.aten.item %332 : !torch.vtensor<[],si8> -> !torch.int | |
| %335 = torch.aten.quantize_per_tensor %330, %333, %334, %int12_37 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %336 = torch.aten.int_repr %335 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
| %337 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %338 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %339 = torch.aten.item %337 : !torch.vtensor<[],f32> -> !torch.float | |
| %340 = torch.aten.item %338 : !torch.vtensor<[],si8> -> !torch.int | |
| %341 = torch.aten._make_per_tensor_quantized_tensor %336, %339, %340 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %342 = torch.aten.dequantize.self %341 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
| %343 = torch.vtensor.literal(dense<16> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %344 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %345 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %346 = torch.vtensor.literal(dense<16> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %347 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_38 = torch.constant.int 0 | |
| %int0_39 = torch.constant.int 0 | |
| %348 = torch.prim.NumToTensor.Scalar %int0_39 : !torch.int -> !torch.vtensor<[1],si64> | |
| %349 = torch.aten.index_select %345, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %350 = torch.aten.item %349 : !torch.vtensor<[1],si64> -> !torch.int | |
| %351 = torch.aten.index_select %346, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %352 = torch.aten.item %351 : !torch.vtensor<[1],si64> -> !torch.int | |
| %353 = torch.aten.index_select %344, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %354 = torch.aten.item %353 : !torch.vtensor<[1],si64> -> !torch.int | |
| %355 = torch.aten.index_select %347, %int0_38, %348 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %356 = torch.aten.item %355 : !torch.vtensor<[1],si64> -> !torch.int | |
| %357 = torch.aten.slice.Tensor %342, %354, %350, %352, %356 : !torch.vtensor<[1,32,160,160],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
| %358 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %359 = torch.vtensor.literal(dense<16> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %360 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %361 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_40 = torch.constant.int 0 | |
| %int0_41 = torch.constant.int 0 | |
| %362 = torch.prim.NumToTensor.Scalar %int0_41 : !torch.int -> !torch.vtensor<[1],si64> | |
| %363 = torch.aten.index_select %359, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %364 = torch.aten.item %363 : !torch.vtensor<[1],si64> -> !torch.int | |
| %365 = torch.aten.index_select %360, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %366 = torch.aten.item %365 : !torch.vtensor<[1],si64> -> !torch.int | |
| %367 = torch.aten.index_select %358, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %368 = torch.aten.item %367 : !torch.vtensor<[1],si64> -> !torch.int | |
| %369 = torch.aten.index_select %361, %int0_40, %362 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %370 = torch.aten.item %369 : !torch.vtensor<[1],si64> -> !torch.int | |
| %371 = torch.aten.slice.Tensor %342, %368, %364, %366, %370 : !torch.vtensor<[1,32,160,160],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
| %372 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %373 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_42 = torch.constant.int 12 | |
| %374 = torch.aten.item %372 : !torch.vtensor<[],f32> -> !torch.float | |
| %375 = torch.aten.item %373 : !torch.vtensor<[],si8> -> !torch.int | |
| %376 = torch.aten.quantize_per_tensor %6, %374, %375, %int12_42 : !torch.vtensor<[16,16,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
| %377 = torch.aten.int_repr %376 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],si8> | |
| %378 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %379 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %380 = torch.aten.item %378 : !torch.vtensor<[],f32> -> !torch.float | |
| %381 = torch.aten.item %379 : !torch.vtensor<[],si8> -> !torch.int | |
| %382 = torch.aten._make_per_tensor_quantized_tensor %377, %380, %381 : !torch.vtensor<[16,16,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
| %383 = torch.aten.dequantize.self %382 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],f32> | |
| %384 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %385 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_43 = torch.constant.int 12 | |
| %386 = torch.aten.item %384 : !torch.vtensor<[],f32> -> !torch.float | |
| %387 = torch.aten.item %385 : !torch.vtensor<[],si8> -> !torch.int | |
| %388 = torch.aten.quantize_per_tensor %7, %386, %387, %int12_43 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %389 = torch.aten.int_repr %388 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
| %390 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %391 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %392 = torch.aten.item %390 : !torch.vtensor<[],f32> -> !torch.float | |
| %393 = torch.aten.item %391 : !torch.vtensor<[],si8> -> !torch.int | |
| %394 = torch.aten._make_per_tensor_quantized_tensor %389, %392, %393 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %395 = torch.aten.dequantize.self %394 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
| %int1_44 = torch.constant.int 1 | |
| %int1_45 = torch.constant.int 1 | |
| %int1_46 = torch.constant.int 1 | |
| %int1_47 = torch.constant.int 1 | |
| %int1_48 = torch.constant.int 1 | |
| %int1_49 = torch.constant.int 1 | |
| %int0_50 = torch.constant.int 0 | |
| %396 = torch.prim.ListConstruct %int1_44, %int1_45 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %397 = torch.prim.ListConstruct %int1_46, %int1_47 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %398 = torch.prim.ListConstruct %int1_48, %int1_49 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %399 = torch.prim.ListConstruct %int0_50, %int0_50 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_51 = torch.constant.bool false | |
| %int1_52 = torch.constant.int 1 | |
| %400 = torch.aten.convolution %371, %383, %395, %398, %396, %397, %false_51, %399, %int1_52 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[16,16,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
| %401 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %402 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_53 = torch.constant.int 12 | |
| %403 = torch.aten.item %401 : !torch.vtensor<[],f32> -> !torch.float | |
| %404 = torch.aten.item %402 : !torch.vtensor<[],si8> -> !torch.int | |
| %405 = torch.aten.quantize_per_tensor %400, %403, %404, %int12_53 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %406 = torch.aten.int_repr %405 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
| %407 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %408 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %409 = torch.aten.item %407 : !torch.vtensor<[],f32> -> !torch.float | |
| %410 = torch.aten.item %408 : !torch.vtensor<[],si8> -> !torch.int | |
| %411 = torch.aten._make_per_tensor_quantized_tensor %406, %409, %410 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %412 = torch.aten.dequantize.self %411 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
| %413 = torch.aten.sigmoid %412 : !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
| %414 = torch.aten.mul.Tensor %412, %413 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
| %415 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %416 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_54 = torch.constant.int 12 | |
| %417 = torch.aten.item %415 : !torch.vtensor<[],f32> -> !torch.float | |
| %418 = torch.aten.item %416 : !torch.vtensor<[],si8> -> !torch.int | |
| %419 = torch.aten.quantize_per_tensor %414, %417, %418, %int12_54 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %420 = torch.aten.int_repr %419 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
| %421 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %422 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %423 = torch.aten.item %421 : !torch.vtensor<[],f32> -> !torch.float | |
| %424 = torch.aten.item %422 : !torch.vtensor<[],si8> -> !torch.int | |
| %425 = torch.aten._make_per_tensor_quantized_tensor %420, %423, %424 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %426 = torch.aten.dequantize.self %425 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
| %427 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %428 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_55 = torch.constant.int 12 | |
| %429 = torch.aten.item %427 : !torch.vtensor<[],f32> -> !torch.float | |
| %430 = torch.aten.item %428 : !torch.vtensor<[],si8> -> !torch.int | |
| %431 = torch.aten.quantize_per_tensor %8, %429, %430, %int12_55 : !torch.vtensor<[16,16,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
| %432 = torch.aten.int_repr %431 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],si8> | |
| %433 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %434 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %435 = torch.aten.item %433 : !torch.vtensor<[],f32> -> !torch.float | |
| %436 = torch.aten.item %434 : !torch.vtensor<[],si8> -> !torch.int | |
| %437 = torch.aten._make_per_tensor_quantized_tensor %432, %435, %436 : !torch.vtensor<[16,16,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,16,3,3],!torch.qint8> | |
| %438 = torch.aten.dequantize.self %437 : !torch.vtensor<[16,16,3,3],!torch.qint8> -> !torch.vtensor<[16,16,3,3],f32> | |
| %439 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %440 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_56 = torch.constant.int 12 | |
| %441 = torch.aten.item %439 : !torch.vtensor<[],f32> -> !torch.float | |
| %442 = torch.aten.item %440 : !torch.vtensor<[],si8> -> !torch.int | |
| %443 = torch.aten.quantize_per_tensor %9, %441, %442, %int12_56 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %444 = torch.aten.int_repr %443 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
| %445 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %446 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %447 = torch.aten.item %445 : !torch.vtensor<[],f32> -> !torch.float | |
| %448 = torch.aten.item %446 : !torch.vtensor<[],si8> -> !torch.int | |
| %449 = torch.aten._make_per_tensor_quantized_tensor %444, %447, %448 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %450 = torch.aten.dequantize.self %449 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
| %int1_57 = torch.constant.int 1 | |
| %int1_58 = torch.constant.int 1 | |
| %int1_59 = torch.constant.int 1 | |
| %int1_60 = torch.constant.int 1 | |
| %int1_61 = torch.constant.int 1 | |
| %int1_62 = torch.constant.int 1 | |
| %int0_63 = torch.constant.int 0 | |
| %451 = torch.prim.ListConstruct %int1_57, %int1_58 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %452 = torch.prim.ListConstruct %int1_59, %int1_60 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %453 = torch.prim.ListConstruct %int1_61, %int1_62 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %454 = torch.prim.ListConstruct %int0_63, %int0_63 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_64 = torch.constant.bool false | |
| %int1_65 = torch.constant.int 1 | |
| %455 = torch.aten.convolution %426, %438, %450, %453, %451, %452, %false_64, %454, %int1_65 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[16,16,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
| %456 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %457 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_66 = torch.constant.int 12 | |
| %458 = torch.aten.item %456 : !torch.vtensor<[],f32> -> !torch.float | |
| %459 = torch.aten.item %457 : !torch.vtensor<[],si8> -> !torch.int | |
| %460 = torch.aten.quantize_per_tensor %455, %458, %459, %int12_66 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %461 = torch.aten.int_repr %460 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
| %462 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %463 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %464 = torch.aten.item %462 : !torch.vtensor<[],f32> -> !torch.float | |
| %465 = torch.aten.item %463 : !torch.vtensor<[],si8> -> !torch.int | |
| %466 = torch.aten._make_per_tensor_quantized_tensor %461, %464, %465 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %467 = torch.aten.dequantize.self %466 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
| %468 = torch.aten.sigmoid %467 : !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
| %469 = torch.aten.mul.Tensor %467, %468 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32> -> !torch.vtensor<[1,16,160,160],f32> | |
| %470 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %471 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_67 = torch.constant.int 12 | |
| %472 = torch.aten.item %470 : !torch.vtensor<[],f32> -> !torch.float | |
| %473 = torch.aten.item %471 : !torch.vtensor<[],si8> -> !torch.int | |
| %474 = torch.aten.quantize_per_tensor %469, %472, %473, %int12_67 : !torch.vtensor<[1,16,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %475 = torch.aten.int_repr %474 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],si8> | |
| %476 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %477 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %478 = torch.aten.item %476 : !torch.vtensor<[],f32> -> !torch.float | |
| %479 = torch.aten.item %477 : !torch.vtensor<[],si8> -> !torch.int | |
| %480 = torch.aten._make_per_tensor_quantized_tensor %475, %478, %479 : !torch.vtensor<[1,16,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,160,160],!torch.qint8> | |
| %481 = torch.aten.dequantize.self %480 : !torch.vtensor<[1,16,160,160],!torch.qint8> -> !torch.vtensor<[1,16,160,160],f32> | |
| %int1_68 = torch.constant.int 1 | |
| %482 = torch.aten.add.Tensor %371, %481, %int1_68 : !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32>, !torch.int -> !torch.vtensor<[1,16,160,160],f32> | |
| %483 = torch.prim.ListConstruct %357, %371, %482 : (!torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32>, !torch.vtensor<[1,16,160,160],f32>) -> !torch.list<vtensor> | |
| %int1_69 = torch.constant.int 1 | |
| %484 = torch.aten.cat %483, %int1_69 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,48,160,160],f32> | |
| %485 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %486 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_70 = torch.constant.int 12 | |
| %487 = torch.aten.item %485 : !torch.vtensor<[],f32> -> !torch.float | |
| %488 = torch.aten.item %486 : !torch.vtensor<[],si8> -> !torch.int | |
| %489 = torch.aten.quantize_per_tensor %484, %487, %488, %int12_70 : !torch.vtensor<[1,48,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,48,160,160],!torch.qint8> | |
| %490 = torch.aten.int_repr %489 : !torch.vtensor<[1,48,160,160],!torch.qint8> -> !torch.vtensor<[1,48,160,160],si8> | |
| %491 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %492 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %493 = torch.aten.item %491 : !torch.vtensor<[],f32> -> !torch.float | |
| %494 = torch.aten.item %492 : !torch.vtensor<[],si8> -> !torch.int | |
| %495 = torch.aten._make_per_tensor_quantized_tensor %490, %493, %494 : !torch.vtensor<[1,48,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,48,160,160],!torch.qint8> | |
| %496 = torch.aten.dequantize.self %495 : !torch.vtensor<[1,48,160,160],!torch.qint8> -> !torch.vtensor<[1,48,160,160],f32> | |
| %497 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %498 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_71 = torch.constant.int 12 | |
| %499 = torch.aten.item %497 : !torch.vtensor<[],f32> -> !torch.float | |
| %500 = torch.aten.item %498 : !torch.vtensor<[],si8> -> !torch.int | |
| %501 = torch.aten.quantize_per_tensor %10, %499, %500, %int12_71 : !torch.vtensor<[32,48,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,48,1,1],!torch.qint8> | |
| %502 = torch.aten.int_repr %501 : !torch.vtensor<[32,48,1,1],!torch.qint8> -> !torch.vtensor<[32,48,1,1],si8> | |
| %503 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %504 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %505 = torch.aten.item %503 : !torch.vtensor<[],f32> -> !torch.float | |
| %506 = torch.aten.item %504 : !torch.vtensor<[],si8> -> !torch.int | |
| %507 = torch.aten._make_per_tensor_quantized_tensor %502, %505, %506 : !torch.vtensor<[32,48,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,48,1,1],!torch.qint8> | |
| %508 = torch.aten.dequantize.self %507 : !torch.vtensor<[32,48,1,1],!torch.qint8> -> !torch.vtensor<[32,48,1,1],f32> | |
| %509 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %510 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_72 = torch.constant.int 12 | |
| %511 = torch.aten.item %509 : !torch.vtensor<[],f32> -> !torch.float | |
| %512 = torch.aten.item %510 : !torch.vtensor<[],si8> -> !torch.int | |
| %513 = torch.aten.quantize_per_tensor %11, %511, %512, %int12_72 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %514 = torch.aten.int_repr %513 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %515 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %516 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %517 = torch.aten.item %515 : !torch.vtensor<[],f32> -> !torch.float | |
| %518 = torch.aten.item %516 : !torch.vtensor<[],si8> -> !torch.int | |
| %519 = torch.aten._make_per_tensor_quantized_tensor %514, %517, %518 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %520 = torch.aten.dequantize.self %519 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int0_73 = torch.constant.int 0 | |
| %int0_74 = torch.constant.int 0 | |
| %int1_75 = torch.constant.int 1 | |
| %int1_76 = torch.constant.int 1 | |
| %int1_77 = torch.constant.int 1 | |
| %int1_78 = torch.constant.int 1 | |
| %int0_79 = torch.constant.int 0 | |
| %521 = torch.prim.ListConstruct %int0_73, %int0_74 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %522 = torch.prim.ListConstruct %int1_75, %int1_76 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %523 = torch.prim.ListConstruct %int1_77, %int1_78 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %524 = torch.prim.ListConstruct %int0_79, %int0_79 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_80 = torch.constant.bool false | |
| %int1_81 = torch.constant.int 1 | |
| %525 = torch.aten.convolution %496, %508, %520, %523, %521, %522, %false_80, %524, %int1_81 : !torch.vtensor<[1,48,160,160],f32>, !torch.vtensor<[32,48,1,1],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,160,160],f32> | |
| %526 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %527 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_82 = torch.constant.int 12 | |
| %528 = torch.aten.item %526 : !torch.vtensor<[],f32> -> !torch.float | |
| %529 = torch.aten.item %527 : !torch.vtensor<[],si8> -> !torch.int | |
| %530 = torch.aten.quantize_per_tensor %525, %528, %529, %int12_82 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %531 = torch.aten.int_repr %530 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
| %532 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %533 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %534 = torch.aten.item %532 : !torch.vtensor<[],f32> -> !torch.float | |
| %535 = torch.aten.item %533 : !torch.vtensor<[],si8> -> !torch.int | |
| %536 = torch.aten._make_per_tensor_quantized_tensor %531, %534, %535 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %537 = torch.aten.dequantize.self %536 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
| %538 = torch.aten.sigmoid %537 : !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
| %539 = torch.aten.mul.Tensor %537, %538 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[1,32,160,160],f32> -> !torch.vtensor<[1,32,160,160],f32> | |
| %540 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %541 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_83 = torch.constant.int 12 | |
| %542 = torch.aten.item %540 : !torch.vtensor<[],f32> -> !torch.float | |
| %543 = torch.aten.item %541 : !torch.vtensor<[],si8> -> !torch.int | |
| %544 = torch.aten.quantize_per_tensor %539, %542, %543, %int12_83 : !torch.vtensor<[1,32,160,160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %545 = torch.aten.int_repr %544 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],si8> | |
| %546 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %547 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %548 = torch.aten.item %546 : !torch.vtensor<[],f32> -> !torch.float | |
| %549 = torch.aten.item %547 : !torch.vtensor<[],si8> -> !torch.int | |
| %550 = torch.aten._make_per_tensor_quantized_tensor %545, %548, %549 : !torch.vtensor<[1,32,160,160],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,160,160],!torch.qint8> | |
| %551 = torch.aten.dequantize.self %550 : !torch.vtensor<[1,32,160,160],!torch.qint8> -> !torch.vtensor<[1,32,160,160],f32> | |
| %552 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %553 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_84 = torch.constant.int 12 | |
| %554 = torch.aten.item %552 : !torch.vtensor<[],f32> -> !torch.float | |
| %555 = torch.aten.item %553 : !torch.vtensor<[],si8> -> !torch.int | |
| %556 = torch.aten.quantize_per_tensor %12, %554, %555, %int12_84 : !torch.vtensor<[64,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,32,3,3],!torch.qint8> | |
| %557 = torch.aten.int_repr %556 : !torch.vtensor<[64,32,3,3],!torch.qint8> -> !torch.vtensor<[64,32,3,3],si8> | |
| %558 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %559 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %560 = torch.aten.item %558 : !torch.vtensor<[],f32> -> !torch.float | |
| %561 = torch.aten.item %559 : !torch.vtensor<[],si8> -> !torch.int | |
| %562 = torch.aten._make_per_tensor_quantized_tensor %557, %560, %561 : !torch.vtensor<[64,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,32,3,3],!torch.qint8> | |
| %563 = torch.aten.dequantize.self %562 : !torch.vtensor<[64,32,3,3],!torch.qint8> -> !torch.vtensor<[64,32,3,3],f32> | |
| %564 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %565 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_85 = torch.constant.int 12 | |
| %566 = torch.aten.item %564 : !torch.vtensor<[],f32> -> !torch.float | |
| %567 = torch.aten.item %565 : !torch.vtensor<[],si8> -> !torch.int | |
| %568 = torch.aten.quantize_per_tensor %13, %566, %567, %int12_85 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %569 = torch.aten.int_repr %568 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %570 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %571 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %572 = torch.aten.item %570 : !torch.vtensor<[],f32> -> !torch.float | |
| %573 = torch.aten.item %571 : !torch.vtensor<[],si8> -> !torch.int | |
| %574 = torch.aten._make_per_tensor_quantized_tensor %569, %572, %573 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %575 = torch.aten.dequantize.self %574 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_86 = torch.constant.int 1 | |
| %int1_87 = torch.constant.int 1 | |
| %int1_88 = torch.constant.int 1 | |
| %int1_89 = torch.constant.int 1 | |
| %int2_90 = torch.constant.int 2 | |
| %int2_91 = torch.constant.int 2 | |
| %int0_92 = torch.constant.int 0 | |
| %576 = torch.prim.ListConstruct %int1_86, %int1_87 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %577 = torch.prim.ListConstruct %int1_88, %int1_89 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %578 = torch.prim.ListConstruct %int2_90, %int2_91 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %579 = torch.prim.ListConstruct %int0_92, %int0_92 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_93 = torch.constant.bool false | |
| %int1_94 = torch.constant.int 1 | |
| %580 = torch.aten.convolution %551, %563, %575, %578, %576, %577, %false_93, %579, %int1_94 : !torch.vtensor<[1,32,160,160],f32>, !torch.vtensor<[64,32,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %581 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %582 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_95 = torch.constant.int 12 | |
| %583 = torch.aten.item %581 : !torch.vtensor<[],f32> -> !torch.float | |
| %584 = torch.aten.item %582 : !torch.vtensor<[],si8> -> !torch.int | |
| %585 = torch.aten.quantize_per_tensor %580, %583, %584, %int12_95 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %586 = torch.aten.int_repr %585 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %587 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %588 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %589 = torch.aten.item %587 : !torch.vtensor<[],f32> -> !torch.float | |
| %590 = torch.aten.item %588 : !torch.vtensor<[],si8> -> !torch.int | |
| %591 = torch.aten._make_per_tensor_quantized_tensor %586, %589, %590 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %592 = torch.aten.dequantize.self %591 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %593 = torch.aten.sigmoid %592 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %594 = torch.aten.mul.Tensor %592, %593 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %595 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %596 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_96 = torch.constant.int 12 | |
| %597 = torch.aten.item %595 : !torch.vtensor<[],f32> -> !torch.float | |
| %598 = torch.aten.item %596 : !torch.vtensor<[],si8> -> !torch.int | |
| %599 = torch.aten.quantize_per_tensor %594, %597, %598, %int12_96 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %600 = torch.aten.int_repr %599 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %601 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %602 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %603 = torch.aten.item %601 : !torch.vtensor<[],f32> -> !torch.float | |
| %604 = torch.aten.item %602 : !torch.vtensor<[],si8> -> !torch.int | |
| %605 = torch.aten._make_per_tensor_quantized_tensor %600, %603, %604 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %606 = torch.aten.dequantize.self %605 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %607 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %608 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_97 = torch.constant.int 12 | |
| %609 = torch.aten.item %607 : !torch.vtensor<[],f32> -> !torch.float | |
| %610 = torch.aten.item %608 : !torch.vtensor<[],si8> -> !torch.int | |
| %611 = torch.aten.quantize_per_tensor %14, %609, %610, %int12_97 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %612 = torch.aten.int_repr %611 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
| %613 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %614 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %615 = torch.aten.item %613 : !torch.vtensor<[],f32> -> !torch.float | |
| %616 = torch.aten.item %614 : !torch.vtensor<[],si8> -> !torch.int | |
| %617 = torch.aten._make_per_tensor_quantized_tensor %612, %615, %616 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %618 = torch.aten.dequantize.self %617 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
| %619 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %620 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_98 = torch.constant.int 12 | |
| %621 = torch.aten.item %619 : !torch.vtensor<[],f32> -> !torch.float | |
| %622 = torch.aten.item %620 : !torch.vtensor<[],si8> -> !torch.int | |
| %623 = torch.aten.quantize_per_tensor %15, %621, %622, %int12_98 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %624 = torch.aten.int_repr %623 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %625 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %626 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %627 = torch.aten.item %625 : !torch.vtensor<[],f32> -> !torch.float | |
| %628 = torch.aten.item %626 : !torch.vtensor<[],si8> -> !torch.int | |
| %629 = torch.aten._make_per_tensor_quantized_tensor %624, %627, %628 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %630 = torch.aten.dequantize.self %629 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_99 = torch.constant.int 0 | |
| %int0_100 = torch.constant.int 0 | |
| %int1_101 = torch.constant.int 1 | |
| %int1_102 = torch.constant.int 1 | |
| %int1_103 = torch.constant.int 1 | |
| %int1_104 = torch.constant.int 1 | |
| %int0_105 = torch.constant.int 0 | |
| %631 = torch.prim.ListConstruct %int0_99, %int0_100 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %632 = torch.prim.ListConstruct %int1_101, %int1_102 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %633 = torch.prim.ListConstruct %int1_103, %int1_104 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %634 = torch.prim.ListConstruct %int0_105, %int0_105 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_106 = torch.constant.bool false | |
| %int1_107 = torch.constant.int 1 | |
| %635 = torch.aten.convolution %606, %618, %630, %633, %631, %632, %false_106, %634, %int1_107 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %636 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %637 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_108 = torch.constant.int 12 | |
| %638 = torch.aten.item %636 : !torch.vtensor<[],f32> -> !torch.float | |
| %639 = torch.aten.item %637 : !torch.vtensor<[],si8> -> !torch.int | |
| %640 = torch.aten.quantize_per_tensor %635, %638, %639, %int12_108 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %641 = torch.aten.int_repr %640 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %642 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %643 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %644 = torch.aten.item %642 : !torch.vtensor<[],f32> -> !torch.float | |
| %645 = torch.aten.item %643 : !torch.vtensor<[],si8> -> !torch.int | |
| %646 = torch.aten._make_per_tensor_quantized_tensor %641, %644, %645 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %647 = torch.aten.dequantize.self %646 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %648 = torch.aten.sigmoid %647 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %649 = torch.aten.mul.Tensor %647, %648 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %650 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %651 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_109 = torch.constant.int 12 | |
| %652 = torch.aten.item %650 : !torch.vtensor<[],f32> -> !torch.float | |
| %653 = torch.aten.item %651 : !torch.vtensor<[],si8> -> !torch.int | |
| %654 = torch.aten.quantize_per_tensor %649, %652, %653, %int12_109 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %655 = torch.aten.int_repr %654 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %656 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %657 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %658 = torch.aten.item %656 : !torch.vtensor<[],f32> -> !torch.float | |
| %659 = torch.aten.item %657 : !torch.vtensor<[],si8> -> !torch.int | |
| %660 = torch.aten._make_per_tensor_quantized_tensor %655, %658, %659 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %661 = torch.aten.dequantize.self %660 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %662 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %663 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %664 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %665 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_110 = torch.constant.int 0 | |
| %int0_111 = torch.constant.int 0 | |
| %666 = torch.prim.NumToTensor.Scalar %int0_111 : !torch.int -> !torch.vtensor<[1],si64> | |
| %667 = torch.aten.index_select %663, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %668 = torch.aten.item %667 : !torch.vtensor<[1],si64> -> !torch.int | |
| %669 = torch.aten.index_select %664, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %670 = torch.aten.item %669 : !torch.vtensor<[1],si64> -> !torch.int | |
| %671 = torch.aten.index_select %662, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %672 = torch.aten.item %671 : !torch.vtensor<[1],si64> -> !torch.int | |
| %673 = torch.aten.index_select %665, %int0_110, %666 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %674 = torch.aten.item %673 : !torch.vtensor<[1],si64> -> !torch.int | |
| %675 = torch.aten.slice.Tensor %661, %672, %668, %670, %674 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %676 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %677 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %678 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %679 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_112 = torch.constant.int 0 | |
| %int0_113 = torch.constant.int 0 | |
| %680 = torch.prim.NumToTensor.Scalar %int0_113 : !torch.int -> !torch.vtensor<[1],si64> | |
| %681 = torch.aten.index_select %677, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %682 = torch.aten.item %681 : !torch.vtensor<[1],si64> -> !torch.int | |
| %683 = torch.aten.index_select %678, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %684 = torch.aten.item %683 : !torch.vtensor<[1],si64> -> !torch.int | |
| %685 = torch.aten.index_select %676, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %686 = torch.aten.item %685 : !torch.vtensor<[1],si64> -> !torch.int | |
| %687 = torch.aten.index_select %679, %int0_112, %680 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %688 = torch.aten.item %687 : !torch.vtensor<[1],si64> -> !torch.int | |
| %689 = torch.aten.slice.Tensor %661, %686, %682, %684, %688 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %690 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %691 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_114 = torch.constant.int 12 | |
| %692 = torch.aten.item %690 : !torch.vtensor<[],f32> -> !torch.float | |
| %693 = torch.aten.item %691 : !torch.vtensor<[],si8> -> !torch.int | |
| %694 = torch.aten.quantize_per_tensor %16, %692, %693, %int12_114 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %695 = torch.aten.int_repr %694 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %696 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %697 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %698 = torch.aten.item %696 : !torch.vtensor<[],f32> -> !torch.float | |
| %699 = torch.aten.item %697 : !torch.vtensor<[],si8> -> !torch.int | |
| %700 = torch.aten._make_per_tensor_quantized_tensor %695, %698, %699 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %701 = torch.aten.dequantize.self %700 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %702 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %703 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_115 = torch.constant.int 12 | |
| %704 = torch.aten.item %702 : !torch.vtensor<[],f32> -> !torch.float | |
| %705 = torch.aten.item %703 : !torch.vtensor<[],si8> -> !torch.int | |
| %706 = torch.aten.quantize_per_tensor %17, %704, %705, %int12_115 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %707 = torch.aten.int_repr %706 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %708 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %709 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %710 = torch.aten.item %708 : !torch.vtensor<[],f32> -> !torch.float | |
| %711 = torch.aten.item %709 : !torch.vtensor<[],si8> -> !torch.int | |
| %712 = torch.aten._make_per_tensor_quantized_tensor %707, %710, %711 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %713 = torch.aten.dequantize.self %712 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_116 = torch.constant.int 1 | |
| %int1_117 = torch.constant.int 1 | |
| %int1_118 = torch.constant.int 1 | |
| %int1_119 = torch.constant.int 1 | |
| %int1_120 = torch.constant.int 1 | |
| %int1_121 = torch.constant.int 1 | |
| %int0_122 = torch.constant.int 0 | |
| %714 = torch.prim.ListConstruct %int1_116, %int1_117 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %715 = torch.prim.ListConstruct %int1_118, %int1_119 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %716 = torch.prim.ListConstruct %int1_120, %int1_121 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %717 = torch.prim.ListConstruct %int0_122, %int0_122 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_123 = torch.constant.bool false | |
| %int1_124 = torch.constant.int 1 | |
| %718 = torch.aten.convolution %689, %701, %713, %716, %714, %715, %false_123, %717, %int1_124 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %719 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %720 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_125 = torch.constant.int 12 | |
| %721 = torch.aten.item %719 : !torch.vtensor<[],f32> -> !torch.float | |
| %722 = torch.aten.item %720 : !torch.vtensor<[],si8> -> !torch.int | |
| %723 = torch.aten.quantize_per_tensor %718, %721, %722, %int12_125 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %724 = torch.aten.int_repr %723 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %725 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %726 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %727 = torch.aten.item %725 : !torch.vtensor<[],f32> -> !torch.float | |
| %728 = torch.aten.item %726 : !torch.vtensor<[],si8> -> !torch.int | |
| %729 = torch.aten._make_per_tensor_quantized_tensor %724, %727, %728 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %730 = torch.aten.dequantize.self %729 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %731 = torch.aten.sigmoid %730 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %732 = torch.aten.mul.Tensor %730, %731 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %733 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %734 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_126 = torch.constant.int 12 | |
| %735 = torch.aten.item %733 : !torch.vtensor<[],f32> -> !torch.float | |
| %736 = torch.aten.item %734 : !torch.vtensor<[],si8> -> !torch.int | |
| %737 = torch.aten.quantize_per_tensor %732, %735, %736, %int12_126 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %738 = torch.aten.int_repr %737 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %739 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %740 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %741 = torch.aten.item %739 : !torch.vtensor<[],f32> -> !torch.float | |
| %742 = torch.aten.item %740 : !torch.vtensor<[],si8> -> !torch.int | |
| %743 = torch.aten._make_per_tensor_quantized_tensor %738, %741, %742 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %744 = torch.aten.dequantize.self %743 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %745 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %746 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_127 = torch.constant.int 12 | |
| %747 = torch.aten.item %745 : !torch.vtensor<[],f32> -> !torch.float | |
| %748 = torch.aten.item %746 : !torch.vtensor<[],si8> -> !torch.int | |
| %749 = torch.aten.quantize_per_tensor %18, %747, %748, %int12_127 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %750 = torch.aten.int_repr %749 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %751 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %752 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %753 = torch.aten.item %751 : !torch.vtensor<[],f32> -> !torch.float | |
| %754 = torch.aten.item %752 : !torch.vtensor<[],si8> -> !torch.int | |
| %755 = torch.aten._make_per_tensor_quantized_tensor %750, %753, %754 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %756 = torch.aten.dequantize.self %755 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %757 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %758 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_128 = torch.constant.int 12 | |
| %759 = torch.aten.item %757 : !torch.vtensor<[],f32> -> !torch.float | |
| %760 = torch.aten.item %758 : !torch.vtensor<[],si8> -> !torch.int | |
| %761 = torch.aten.quantize_per_tensor %19, %759, %760, %int12_128 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %762 = torch.aten.int_repr %761 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %763 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %764 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %765 = torch.aten.item %763 : !torch.vtensor<[],f32> -> !torch.float | |
| %766 = torch.aten.item %764 : !torch.vtensor<[],si8> -> !torch.int | |
| %767 = torch.aten._make_per_tensor_quantized_tensor %762, %765, %766 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %768 = torch.aten.dequantize.self %767 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_129 = torch.constant.int 1 | |
| %int1_130 = torch.constant.int 1 | |
| %int1_131 = torch.constant.int 1 | |
| %int1_132 = torch.constant.int 1 | |
| %int1_133 = torch.constant.int 1 | |
| %int1_134 = torch.constant.int 1 | |
| %int0_135 = torch.constant.int 0 | |
| %769 = torch.prim.ListConstruct %int1_129, %int1_130 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %770 = torch.prim.ListConstruct %int1_131, %int1_132 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %771 = torch.prim.ListConstruct %int1_133, %int1_134 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %772 = torch.prim.ListConstruct %int0_135, %int0_135 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_136 = torch.constant.bool false | |
| %int1_137 = torch.constant.int 1 | |
| %773 = torch.aten.convolution %744, %756, %768, %771, %769, %770, %false_136, %772, %int1_137 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %774 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %775 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_138 = torch.constant.int 12 | |
| %776 = torch.aten.item %774 : !torch.vtensor<[],f32> -> !torch.float | |
| %777 = torch.aten.item %775 : !torch.vtensor<[],si8> -> !torch.int | |
| %778 = torch.aten.quantize_per_tensor %773, %776, %777, %int12_138 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %779 = torch.aten.int_repr %778 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %780 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %781 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %782 = torch.aten.item %780 : !torch.vtensor<[],f32> -> !torch.float | |
| %783 = torch.aten.item %781 : !torch.vtensor<[],si8> -> !torch.int | |
| %784 = torch.aten._make_per_tensor_quantized_tensor %779, %782, %783 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %785 = torch.aten.dequantize.self %784 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %786 = torch.aten.sigmoid %785 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %787 = torch.aten.mul.Tensor %785, %786 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %788 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %789 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_139 = torch.constant.int 12 | |
| %790 = torch.aten.item %788 : !torch.vtensor<[],f32> -> !torch.float | |
| %791 = torch.aten.item %789 : !torch.vtensor<[],si8> -> !torch.int | |
| %792 = torch.aten.quantize_per_tensor %787, %790, %791, %int12_139 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %793 = torch.aten.int_repr %792 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %794 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %795 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %796 = torch.aten.item %794 : !torch.vtensor<[],f32> -> !torch.float | |
| %797 = torch.aten.item %795 : !torch.vtensor<[],si8> -> !torch.int | |
| %798 = torch.aten._make_per_tensor_quantized_tensor %793, %796, %797 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %799 = torch.aten.dequantize.self %798 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %int1_140 = torch.constant.int 1 | |
| %800 = torch.aten.add.Tensor %689, %799, %int1_140 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %801 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %802 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_141 = torch.constant.int 12 | |
| %803 = torch.aten.item %801 : !torch.vtensor<[],f32> -> !torch.float | |
| %804 = torch.aten.item %802 : !torch.vtensor<[],si8> -> !torch.int | |
| %805 = torch.aten.quantize_per_tensor %800, %803, %804, %int12_141 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %806 = torch.aten.int_repr %805 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %807 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %808 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %809 = torch.aten.item %807 : !torch.vtensor<[],f32> -> !torch.float | |
| %810 = torch.aten.item %808 : !torch.vtensor<[],si8> -> !torch.int | |
| %811 = torch.aten._make_per_tensor_quantized_tensor %806, %809, %810 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %812 = torch.aten.dequantize.self %811 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %813 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %814 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_142 = torch.constant.int 12 | |
| %815 = torch.aten.item %813 : !torch.vtensor<[],f32> -> !torch.float | |
| %816 = torch.aten.item %814 : !torch.vtensor<[],si8> -> !torch.int | |
| %817 = torch.aten.quantize_per_tensor %20, %815, %816, %int12_142 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %818 = torch.aten.int_repr %817 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %819 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %820 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %821 = torch.aten.item %819 : !torch.vtensor<[],f32> -> !torch.float | |
| %822 = torch.aten.item %820 : !torch.vtensor<[],si8> -> !torch.int | |
| %823 = torch.aten._make_per_tensor_quantized_tensor %818, %821, %822 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %824 = torch.aten.dequantize.self %823 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %825 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %826 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_143 = torch.constant.int 12 | |
| %827 = torch.aten.item %825 : !torch.vtensor<[],f32> -> !torch.float | |
| %828 = torch.aten.item %826 : !torch.vtensor<[],si8> -> !torch.int | |
| %829 = torch.aten.quantize_per_tensor %21, %827, %828, %int12_143 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %830 = torch.aten.int_repr %829 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %831 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %832 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %833 = torch.aten.item %831 : !torch.vtensor<[],f32> -> !torch.float | |
| %834 = torch.aten.item %832 : !torch.vtensor<[],si8> -> !torch.int | |
| %835 = torch.aten._make_per_tensor_quantized_tensor %830, %833, %834 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %836 = torch.aten.dequantize.self %835 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_144 = torch.constant.int 1 | |
| %int1_145 = torch.constant.int 1 | |
| %int1_146 = torch.constant.int 1 | |
| %int1_147 = torch.constant.int 1 | |
| %int1_148 = torch.constant.int 1 | |
| %int1_149 = torch.constant.int 1 | |
| %int0_150 = torch.constant.int 0 | |
| %837 = torch.prim.ListConstruct %int1_144, %int1_145 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %838 = torch.prim.ListConstruct %int1_146, %int1_147 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %839 = torch.prim.ListConstruct %int1_148, %int1_149 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %840 = torch.prim.ListConstruct %int0_150, %int0_150 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_151 = torch.constant.bool false | |
| %int1_152 = torch.constant.int 1 | |
| %841 = torch.aten.convolution %812, %824, %836, %839, %837, %838, %false_151, %840, %int1_152 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %842 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %843 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_153 = torch.constant.int 12 | |
| %844 = torch.aten.item %842 : !torch.vtensor<[],f32> -> !torch.float | |
| %845 = torch.aten.item %843 : !torch.vtensor<[],si8> -> !torch.int | |
| %846 = torch.aten.quantize_per_tensor %841, %844, %845, %int12_153 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %847 = torch.aten.int_repr %846 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %848 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %849 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %850 = torch.aten.item %848 : !torch.vtensor<[],f32> -> !torch.float | |
| %851 = torch.aten.item %849 : !torch.vtensor<[],si8> -> !torch.int | |
| %852 = torch.aten._make_per_tensor_quantized_tensor %847, %850, %851 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %853 = torch.aten.dequantize.self %852 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %854 = torch.aten.sigmoid %853 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %855 = torch.aten.mul.Tensor %853, %854 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %856 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %857 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_154 = torch.constant.int 12 | |
| %858 = torch.aten.item %856 : !torch.vtensor<[],f32> -> !torch.float | |
| %859 = torch.aten.item %857 : !torch.vtensor<[],si8> -> !torch.int | |
| %860 = torch.aten.quantize_per_tensor %855, %858, %859, %int12_154 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %861 = torch.aten.int_repr %860 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %862 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %863 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %864 = torch.aten.item %862 : !torch.vtensor<[],f32> -> !torch.float | |
| %865 = torch.aten.item %863 : !torch.vtensor<[],si8> -> !torch.int | |
| %866 = torch.aten._make_per_tensor_quantized_tensor %861, %864, %865 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %867 = torch.aten.dequantize.self %866 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %868 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %869 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_155 = torch.constant.int 12 | |
| %870 = torch.aten.item %868 : !torch.vtensor<[],f32> -> !torch.float | |
| %871 = torch.aten.item %869 : !torch.vtensor<[],si8> -> !torch.int | |
| %872 = torch.aten.quantize_per_tensor %22, %870, %871, %int12_155 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %873 = torch.aten.int_repr %872 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %874 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %875 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %876 = torch.aten.item %874 : !torch.vtensor<[],f32> -> !torch.float | |
| %877 = torch.aten.item %875 : !torch.vtensor<[],si8> -> !torch.int | |
| %878 = torch.aten._make_per_tensor_quantized_tensor %873, %876, %877 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %879 = torch.aten.dequantize.self %878 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %880 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %881 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_156 = torch.constant.int 12 | |
| %882 = torch.aten.item %880 : !torch.vtensor<[],f32> -> !torch.float | |
| %883 = torch.aten.item %881 : !torch.vtensor<[],si8> -> !torch.int | |
| %884 = torch.aten.quantize_per_tensor %23, %882, %883, %int12_156 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %885 = torch.aten.int_repr %884 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %886 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %887 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %888 = torch.aten.item %886 : !torch.vtensor<[],f32> -> !torch.float | |
| %889 = torch.aten.item %887 : !torch.vtensor<[],si8> -> !torch.int | |
| %890 = torch.aten._make_per_tensor_quantized_tensor %885, %888, %889 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %891 = torch.aten.dequantize.self %890 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_157 = torch.constant.int 1 | |
| %int1_158 = torch.constant.int 1 | |
| %int1_159 = torch.constant.int 1 | |
| %int1_160 = torch.constant.int 1 | |
| %int1_161 = torch.constant.int 1 | |
| %int1_162 = torch.constant.int 1 | |
| %int0_163 = torch.constant.int 0 | |
| %892 = torch.prim.ListConstruct %int1_157, %int1_158 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %893 = torch.prim.ListConstruct %int1_159, %int1_160 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %894 = torch.prim.ListConstruct %int1_161, %int1_162 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %895 = torch.prim.ListConstruct %int0_163, %int0_163 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_164 = torch.constant.bool false | |
| %int1_165 = torch.constant.int 1 | |
| %896 = torch.aten.convolution %867, %879, %891, %894, %892, %893, %false_164, %895, %int1_165 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %897 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %898 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_166 = torch.constant.int 12 | |
| %899 = torch.aten.item %897 : !torch.vtensor<[],f32> -> !torch.float | |
| %900 = torch.aten.item %898 : !torch.vtensor<[],si8> -> !torch.int | |
| %901 = torch.aten.quantize_per_tensor %896, %899, %900, %int12_166 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %902 = torch.aten.int_repr %901 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %903 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %904 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %905 = torch.aten.item %903 : !torch.vtensor<[],f32> -> !torch.float | |
| %906 = torch.aten.item %904 : !torch.vtensor<[],si8> -> !torch.int | |
| %907 = torch.aten._make_per_tensor_quantized_tensor %902, %905, %906 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %908 = torch.aten.dequantize.self %907 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %909 = torch.aten.sigmoid %908 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %910 = torch.aten.mul.Tensor %908, %909 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %911 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %912 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_167 = torch.constant.int 12 | |
| %913 = torch.aten.item %911 : !torch.vtensor<[],f32> -> !torch.float | |
| %914 = torch.aten.item %912 : !torch.vtensor<[],si8> -> !torch.int | |
| %915 = torch.aten.quantize_per_tensor %910, %913, %914, %int12_167 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %916 = torch.aten.int_repr %915 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %917 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %918 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %919 = torch.aten.item %917 : !torch.vtensor<[],f32> -> !torch.float | |
| %920 = torch.aten.item %918 : !torch.vtensor<[],si8> -> !torch.int | |
| %921 = torch.aten._make_per_tensor_quantized_tensor %916, %919, %920 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %922 = torch.aten.dequantize.self %921 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %int1_168 = torch.constant.int 1 | |
| %923 = torch.aten.add.Tensor %812, %922, %int1_168 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %924 = torch.prim.ListConstruct %675, %689, %812, %923 : (!torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>) -> !torch.list<vtensor> | |
| %int1_169 = torch.constant.int 1 | |
| %925 = torch.aten.cat %924, %int1_169 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,128,80,80],f32> | |
| %926 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %927 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_170 = torch.constant.int 12 | |
| %928 = torch.aten.item %926 : !torch.vtensor<[],f32> -> !torch.float | |
| %929 = torch.aten.item %927 : !torch.vtensor<[],si8> -> !torch.int | |
| %930 = torch.aten.quantize_per_tensor %925, %928, %929, %int12_170 : !torch.vtensor<[1,128,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,80,80],!torch.qint8> | |
| %931 = torch.aten.int_repr %930 : !torch.vtensor<[1,128,80,80],!torch.qint8> -> !torch.vtensor<[1,128,80,80],si8> | |
| %932 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %933 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %934 = torch.aten.item %932 : !torch.vtensor<[],f32> -> !torch.float | |
| %935 = torch.aten.item %933 : !torch.vtensor<[],si8> -> !torch.int | |
| %936 = torch.aten._make_per_tensor_quantized_tensor %931, %934, %935 : !torch.vtensor<[1,128,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,80,80],!torch.qint8> | |
| %937 = torch.aten.dequantize.self %936 : !torch.vtensor<[1,128,80,80],!torch.qint8> -> !torch.vtensor<[1,128,80,80],f32> | |
| %938 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %939 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_171 = torch.constant.int 12 | |
| %940 = torch.aten.item %938 : !torch.vtensor<[],f32> -> !torch.float | |
| %941 = torch.aten.item %939 : !torch.vtensor<[],si8> -> !torch.int | |
| %942 = torch.aten.quantize_per_tensor %24, %940, %941, %int12_171 : !torch.vtensor<[64,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,128,1,1],!torch.qint8> | |
| %943 = torch.aten.int_repr %942 : !torch.vtensor<[64,128,1,1],!torch.qint8> -> !torch.vtensor<[64,128,1,1],si8> | |
| %944 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %945 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %946 = torch.aten.item %944 : !torch.vtensor<[],f32> -> !torch.float | |
| %947 = torch.aten.item %945 : !torch.vtensor<[],si8> -> !torch.int | |
| %948 = torch.aten._make_per_tensor_quantized_tensor %943, %946, %947 : !torch.vtensor<[64,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,128,1,1],!torch.qint8> | |
| %949 = torch.aten.dequantize.self %948 : !torch.vtensor<[64,128,1,1],!torch.qint8> -> !torch.vtensor<[64,128,1,1],f32> | |
| %950 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %951 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_172 = torch.constant.int 12 | |
| %952 = torch.aten.item %950 : !torch.vtensor<[],f32> -> !torch.float | |
| %953 = torch.aten.item %951 : !torch.vtensor<[],si8> -> !torch.int | |
| %954 = torch.aten.quantize_per_tensor %25, %952, %953, %int12_172 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %955 = torch.aten.int_repr %954 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %956 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %957 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %958 = torch.aten.item %956 : !torch.vtensor<[],f32> -> !torch.float | |
| %959 = torch.aten.item %957 : !torch.vtensor<[],si8> -> !torch.int | |
| %960 = torch.aten._make_per_tensor_quantized_tensor %955, %958, %959 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %961 = torch.aten.dequantize.self %960 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_173 = torch.constant.int 0 | |
| %int0_174 = torch.constant.int 0 | |
| %int1_175 = torch.constant.int 1 | |
| %int1_176 = torch.constant.int 1 | |
| %int1_177 = torch.constant.int 1 | |
| %int1_178 = torch.constant.int 1 | |
| %int0_179 = torch.constant.int 0 | |
| %962 = torch.prim.ListConstruct %int0_173, %int0_174 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %963 = torch.prim.ListConstruct %int1_175, %int1_176 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %964 = torch.prim.ListConstruct %int1_177, %int1_178 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %965 = torch.prim.ListConstruct %int0_179, %int0_179 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_180 = torch.constant.bool false | |
| %int1_181 = torch.constant.int 1 | |
| %966 = torch.aten.convolution %937, %949, %961, %964, %962, %963, %false_180, %965, %int1_181 : !torch.vtensor<[1,128,80,80],f32>, !torch.vtensor<[64,128,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %967 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %968 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_182 = torch.constant.int 12 | |
| %969 = torch.aten.item %967 : !torch.vtensor<[],f32> -> !torch.float | |
| %970 = torch.aten.item %968 : !torch.vtensor<[],si8> -> !torch.int | |
| %971 = torch.aten.quantize_per_tensor %966, %969, %970, %int12_182 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %972 = torch.aten.int_repr %971 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %973 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %974 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %975 = torch.aten.item %973 : !torch.vtensor<[],f32> -> !torch.float | |
| %976 = torch.aten.item %974 : !torch.vtensor<[],si8> -> !torch.int | |
| %977 = torch.aten._make_per_tensor_quantized_tensor %972, %975, %976 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %978 = torch.aten.dequantize.self %977 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %979 = torch.aten.sigmoid %978 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %980 = torch.aten.mul.Tensor %978, %979 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %981 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %982 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_183 = torch.constant.int 12 | |
| %983 = torch.aten.item %981 : !torch.vtensor<[],f32> -> !torch.float | |
| %984 = torch.aten.item %982 : !torch.vtensor<[],si8> -> !torch.int | |
| %985 = torch.aten.quantize_per_tensor %980, %983, %984, %int12_183 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %986 = torch.aten.int_repr %985 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %987 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %988 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %989 = torch.aten.item %987 : !torch.vtensor<[],f32> -> !torch.float | |
| %990 = torch.aten.item %988 : !torch.vtensor<[],si8> -> !torch.int | |
| %991 = torch.aten._make_per_tensor_quantized_tensor %986, %989, %990 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %992 = torch.aten.dequantize.self %991 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %993 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %994 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_184 = torch.constant.int 12 | |
| %995 = torch.aten.item %993 : !torch.vtensor<[],f32> -> !torch.float | |
| %996 = torch.aten.item %994 : !torch.vtensor<[],si8> -> !torch.int | |
| %997 = torch.aten.quantize_per_tensor %26, %995, %996, %int12_184 : !torch.vtensor<[128,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,64,3,3],!torch.qint8> | |
| %998 = torch.aten.int_repr %997 : !torch.vtensor<[128,64,3,3],!torch.qint8> -> !torch.vtensor<[128,64,3,3],si8> | |
| %999 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1000 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1001 = torch.aten.item %999 : !torch.vtensor<[],f32> -> !torch.float | |
| %1002 = torch.aten.item %1000 : !torch.vtensor<[],si8> -> !torch.int | |
| %1003 = torch.aten._make_per_tensor_quantized_tensor %998, %1001, %1002 : !torch.vtensor<[128,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,64,3,3],!torch.qint8> | |
| %1004 = torch.aten.dequantize.self %1003 : !torch.vtensor<[128,64,3,3],!torch.qint8> -> !torch.vtensor<[128,64,3,3],f32> | |
| %1005 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1006 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_185 = torch.constant.int 12 | |
| %1007 = torch.aten.item %1005 : !torch.vtensor<[],f32> -> !torch.float | |
| %1008 = torch.aten.item %1006 : !torch.vtensor<[],si8> -> !torch.int | |
| %1009 = torch.aten.quantize_per_tensor %27, %1007, %1008, %int12_185 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1010 = torch.aten.int_repr %1009 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1011 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1012 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1013 = torch.aten.item %1011 : !torch.vtensor<[],f32> -> !torch.float | |
| %1014 = torch.aten.item %1012 : !torch.vtensor<[],si8> -> !torch.int | |
| %1015 = torch.aten._make_per_tensor_quantized_tensor %1010, %1013, %1014 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1016 = torch.aten.dequantize.self %1015 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int1_186 = torch.constant.int 1 | |
| %int1_187 = torch.constant.int 1 | |
| %int1_188 = torch.constant.int 1 | |
| %int1_189 = torch.constant.int 1 | |
| %int2_190 = torch.constant.int 2 | |
| %int2_191 = torch.constant.int 2 | |
| %int0_192 = torch.constant.int 0 | |
| %1017 = torch.prim.ListConstruct %int1_186, %int1_187 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1018 = torch.prim.ListConstruct %int1_188, %int1_189 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1019 = torch.prim.ListConstruct %int2_190, %int2_191 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1020 = torch.prim.ListConstruct %int0_192, %int0_192 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_193 = torch.constant.bool false | |
| %int1_194 = torch.constant.int 1 | |
| %1021 = torch.aten.convolution %992, %1004, %1016, %1019, %1017, %1018, %false_193, %1020, %int1_194 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[128,64,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %1022 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1023 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_195 = torch.constant.int 12 | |
| %1024 = torch.aten.item %1022 : !torch.vtensor<[],f32> -> !torch.float | |
| %1025 = torch.aten.item %1023 : !torch.vtensor<[],si8> -> !torch.int | |
| %1026 = torch.aten.quantize_per_tensor %1021, %1024, %1025, %int12_195 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1027 = torch.aten.int_repr %1026 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1028 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1029 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1030 = torch.aten.item %1028 : !torch.vtensor<[],f32> -> !torch.float | |
| %1031 = torch.aten.item %1029 : !torch.vtensor<[],si8> -> !torch.int | |
| %1032 = torch.aten._make_per_tensor_quantized_tensor %1027, %1030, %1031 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1033 = torch.aten.dequantize.self %1032 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1034 = torch.aten.sigmoid %1033 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1035 = torch.aten.mul.Tensor %1033, %1034 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1036 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1037 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_196 = torch.constant.int 12 | |
| %1038 = torch.aten.item %1036 : !torch.vtensor<[],f32> -> !torch.float | |
| %1039 = torch.aten.item %1037 : !torch.vtensor<[],si8> -> !torch.int | |
| %1040 = torch.aten.quantize_per_tensor %1035, %1038, %1039, %int12_196 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1041 = torch.aten.int_repr %1040 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1042 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1043 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1044 = torch.aten.item %1042 : !torch.vtensor<[],f32> -> !torch.float | |
| %1045 = torch.aten.item %1043 : !torch.vtensor<[],si8> -> !torch.int | |
| %1046 = torch.aten._make_per_tensor_quantized_tensor %1041, %1044, %1045 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1047 = torch.aten.dequantize.self %1046 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1048 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1049 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_197 = torch.constant.int 12 | |
| %1050 = torch.aten.item %1048 : !torch.vtensor<[],f32> -> !torch.float | |
| %1051 = torch.aten.item %1049 : !torch.vtensor<[],si8> -> !torch.int | |
| %1052 = torch.aten.quantize_per_tensor %28, %1050, %1051, %int12_197 : !torch.vtensor<[128,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,1,1],!torch.qint8> | |
| %1053 = torch.aten.int_repr %1052 : !torch.vtensor<[128,128,1,1],!torch.qint8> -> !torch.vtensor<[128,128,1,1],si8> | |
| %1054 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1055 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1056 = torch.aten.item %1054 : !torch.vtensor<[],f32> -> !torch.float | |
| %1057 = torch.aten.item %1055 : !torch.vtensor<[],si8> -> !torch.int | |
| %1058 = torch.aten._make_per_tensor_quantized_tensor %1053, %1056, %1057 : !torch.vtensor<[128,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,1,1],!torch.qint8> | |
| %1059 = torch.aten.dequantize.self %1058 : !torch.vtensor<[128,128,1,1],!torch.qint8> -> !torch.vtensor<[128,128,1,1],f32> | |
| %1060 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1061 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_198 = torch.constant.int 12 | |
| %1062 = torch.aten.item %1060 : !torch.vtensor<[],f32> -> !torch.float | |
| %1063 = torch.aten.item %1061 : !torch.vtensor<[],si8> -> !torch.int | |
| %1064 = torch.aten.quantize_per_tensor %29, %1062, %1063, %int12_198 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1065 = torch.aten.int_repr %1064 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1066 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1067 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1068 = torch.aten.item %1066 : !torch.vtensor<[],f32> -> !torch.float | |
| %1069 = torch.aten.item %1067 : !torch.vtensor<[],si8> -> !torch.int | |
| %1070 = torch.aten._make_per_tensor_quantized_tensor %1065, %1068, %1069 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1071 = torch.aten.dequantize.self %1070 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_199 = torch.constant.int 0 | |
| %int0_200 = torch.constant.int 0 | |
| %int1_201 = torch.constant.int 1 | |
| %int1_202 = torch.constant.int 1 | |
| %int1_203 = torch.constant.int 1 | |
| %int1_204 = torch.constant.int 1 | |
| %int0_205 = torch.constant.int 0 | |
| %1072 = torch.prim.ListConstruct %int0_199, %int0_200 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1073 = torch.prim.ListConstruct %int1_201, %int1_202 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1074 = torch.prim.ListConstruct %int1_203, %int1_204 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1075 = torch.prim.ListConstruct %int0_205, %int0_205 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_206 = torch.constant.bool false | |
| %int1_207 = torch.constant.int 1 | |
| %1076 = torch.aten.convolution %1047, %1059, %1071, %1074, %1072, %1073, %false_206, %1075, %int1_207 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[128,128,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %1077 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1078 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_208 = torch.constant.int 12 | |
| %1079 = torch.aten.item %1077 : !torch.vtensor<[],f32> -> !torch.float | |
| %1080 = torch.aten.item %1078 : !torch.vtensor<[],si8> -> !torch.int | |
| %1081 = torch.aten.quantize_per_tensor %1076, %1079, %1080, %int12_208 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1082 = torch.aten.int_repr %1081 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1083 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1084 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1085 = torch.aten.item %1083 : !torch.vtensor<[],f32> -> !torch.float | |
| %1086 = torch.aten.item %1084 : !torch.vtensor<[],si8> -> !torch.int | |
| %1087 = torch.aten._make_per_tensor_quantized_tensor %1082, %1085, %1086 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1088 = torch.aten.dequantize.self %1087 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1089 = torch.aten.sigmoid %1088 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1090 = torch.aten.mul.Tensor %1088, %1089 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1091 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1092 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_209 = torch.constant.int 12 | |
| %1093 = torch.aten.item %1091 : !torch.vtensor<[],f32> -> !torch.float | |
| %1094 = torch.aten.item %1092 : !torch.vtensor<[],si8> -> !torch.int | |
| %1095 = torch.aten.quantize_per_tensor %1090, %1093, %1094, %int12_209 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1096 = torch.aten.int_repr %1095 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1097 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1098 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1099 = torch.aten.item %1097 : !torch.vtensor<[],f32> -> !torch.float | |
| %1100 = torch.aten.item %1098 : !torch.vtensor<[],si8> -> !torch.int | |
| %1101 = torch.aten._make_per_tensor_quantized_tensor %1096, %1099, %1100 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1102 = torch.aten.dequantize.self %1101 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1103 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1104 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1105 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1106 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_210 = torch.constant.int 0 | |
| %int0_211 = torch.constant.int 0 | |
| %1107 = torch.prim.NumToTensor.Scalar %int0_211 : !torch.int -> !torch.vtensor<[1],si64> | |
| %1108 = torch.aten.index_select %1104, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1109 = torch.aten.item %1108 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1110 = torch.aten.index_select %1105, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1111 = torch.aten.item %1110 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1112 = torch.aten.index_select %1103, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1113 = torch.aten.item %1112 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1114 = torch.aten.index_select %1106, %int0_210, %1107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1115 = torch.aten.item %1114 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1116 = torch.aten.slice.Tensor %1102, %1113, %1109, %1111, %1115 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1117 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1118 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1119 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1120 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_212 = torch.constant.int 0 | |
| %int0_213 = torch.constant.int 0 | |
| %1121 = torch.prim.NumToTensor.Scalar %int0_213 : !torch.int -> !torch.vtensor<[1],si64> | |
| %1122 = torch.aten.index_select %1118, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1123 = torch.aten.item %1122 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1124 = torch.aten.index_select %1119, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1125 = torch.aten.item %1124 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1126 = torch.aten.index_select %1117, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1127 = torch.aten.item %1126 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1128 = torch.aten.index_select %1120, %int0_212, %1121 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1129 = torch.aten.item %1128 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1130 = torch.aten.slice.Tensor %1102, %1127, %1123, %1125, %1129 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1131 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1132 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_214 = torch.constant.int 12 | |
| %1133 = torch.aten.item %1131 : !torch.vtensor<[],f32> -> !torch.float | |
| %1134 = torch.aten.item %1132 : !torch.vtensor<[],si8> -> !torch.int | |
| %1135 = torch.aten.quantize_per_tensor %30, %1133, %1134, %int12_214 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1136 = torch.aten.int_repr %1135 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %1137 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1138 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1139 = torch.aten.item %1137 : !torch.vtensor<[],f32> -> !torch.float | |
| %1140 = torch.aten.item %1138 : !torch.vtensor<[],si8> -> !torch.int | |
| %1141 = torch.aten._make_per_tensor_quantized_tensor %1136, %1139, %1140 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1142 = torch.aten.dequantize.self %1141 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %1143 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1144 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_215 = torch.constant.int 12 | |
| %1145 = torch.aten.item %1143 : !torch.vtensor<[],f32> -> !torch.float | |
| %1146 = torch.aten.item %1144 : !torch.vtensor<[],si8> -> !torch.int | |
| %1147 = torch.aten.quantize_per_tensor %31, %1145, %1146, %int12_215 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1148 = torch.aten.int_repr %1147 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %1149 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1150 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1151 = torch.aten.item %1149 : !torch.vtensor<[],f32> -> !torch.float | |
| %1152 = torch.aten.item %1150 : !torch.vtensor<[],si8> -> !torch.int | |
| %1153 = torch.aten._make_per_tensor_quantized_tensor %1148, %1151, %1152 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1154 = torch.aten.dequantize.self %1153 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_216 = torch.constant.int 1 | |
| %int1_217 = torch.constant.int 1 | |
| %int1_218 = torch.constant.int 1 | |
| %int1_219 = torch.constant.int 1 | |
| %int1_220 = torch.constant.int 1 | |
| %int1_221 = torch.constant.int 1 | |
| %int0_222 = torch.constant.int 0 | |
| %1155 = torch.prim.ListConstruct %int1_216, %int1_217 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1156 = torch.prim.ListConstruct %int1_218, %int1_219 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1157 = torch.prim.ListConstruct %int1_220, %int1_221 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1158 = torch.prim.ListConstruct %int0_222, %int0_222 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_223 = torch.constant.bool false | |
| %int1_224 = torch.constant.int 1 | |
| %1159 = torch.aten.convolution %1130, %1142, %1154, %1157, %1155, %1156, %false_223, %1158, %int1_224 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1160 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1161 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_225 = torch.constant.int 12 | |
| %1162 = torch.aten.item %1160 : !torch.vtensor<[],f32> -> !torch.float | |
| %1163 = torch.aten.item %1161 : !torch.vtensor<[],si8> -> !torch.int | |
| %1164 = torch.aten.quantize_per_tensor %1159, %1162, %1163, %int12_225 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1165 = torch.aten.int_repr %1164 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1166 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1167 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1168 = torch.aten.item %1166 : !torch.vtensor<[],f32> -> !torch.float | |
| %1169 = torch.aten.item %1167 : !torch.vtensor<[],si8> -> !torch.int | |
| %1170 = torch.aten._make_per_tensor_quantized_tensor %1165, %1168, %1169 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1171 = torch.aten.dequantize.self %1170 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1172 = torch.aten.sigmoid %1171 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1173 = torch.aten.mul.Tensor %1171, %1172 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1174 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1175 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_226 = torch.constant.int 12 | |
| %1176 = torch.aten.item %1174 : !torch.vtensor<[],f32> -> !torch.float | |
| %1177 = torch.aten.item %1175 : !torch.vtensor<[],si8> -> !torch.int | |
| %1178 = torch.aten.quantize_per_tensor %1173, %1176, %1177, %int12_226 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1179 = torch.aten.int_repr %1178 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1180 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1181 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1182 = torch.aten.item %1180 : !torch.vtensor<[],f32> -> !torch.float | |
| %1183 = torch.aten.item %1181 : !torch.vtensor<[],si8> -> !torch.int | |
| %1184 = torch.aten._make_per_tensor_quantized_tensor %1179, %1182, %1183 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1185 = torch.aten.dequantize.self %1184 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1186 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1187 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_227 = torch.constant.int 12 | |
| %1188 = torch.aten.item %1186 : !torch.vtensor<[],f32> -> !torch.float | |
| %1189 = torch.aten.item %1187 : !torch.vtensor<[],si8> -> !torch.int | |
| %1190 = torch.aten.quantize_per_tensor %32, %1188, %1189, %int12_227 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1191 = torch.aten.int_repr %1190 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %1192 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1193 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1194 = torch.aten.item %1192 : !torch.vtensor<[],f32> -> !torch.float | |
| %1195 = torch.aten.item %1193 : !torch.vtensor<[],si8> -> !torch.int | |
| %1196 = torch.aten._make_per_tensor_quantized_tensor %1191, %1194, %1195 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1197 = torch.aten.dequantize.self %1196 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %1198 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1199 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_228 = torch.constant.int 12 | |
| %1200 = torch.aten.item %1198 : !torch.vtensor<[],f32> -> !torch.float | |
| %1201 = torch.aten.item %1199 : !torch.vtensor<[],si8> -> !torch.int | |
| %1202 = torch.aten.quantize_per_tensor %33, %1200, %1201, %int12_228 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1203 = torch.aten.int_repr %1202 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %1204 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1205 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1206 = torch.aten.item %1204 : !torch.vtensor<[],f32> -> !torch.float | |
| %1207 = torch.aten.item %1205 : !torch.vtensor<[],si8> -> !torch.int | |
| %1208 = torch.aten._make_per_tensor_quantized_tensor %1203, %1206, %1207 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1209 = torch.aten.dequantize.self %1208 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_229 = torch.constant.int 1 | |
| %int1_230 = torch.constant.int 1 | |
| %int1_231 = torch.constant.int 1 | |
| %int1_232 = torch.constant.int 1 | |
| %int1_233 = torch.constant.int 1 | |
| %int1_234 = torch.constant.int 1 | |
| %int0_235 = torch.constant.int 0 | |
| %1210 = torch.prim.ListConstruct %int1_229, %int1_230 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1211 = torch.prim.ListConstruct %int1_231, %int1_232 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1212 = torch.prim.ListConstruct %int1_233, %int1_234 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1213 = torch.prim.ListConstruct %int0_235, %int0_235 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_236 = torch.constant.bool false | |
| %int1_237 = torch.constant.int 1 | |
| %1214 = torch.aten.convolution %1185, %1197, %1209, %1212, %1210, %1211, %false_236, %1213, %int1_237 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1215 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1216 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_238 = torch.constant.int 12 | |
| %1217 = torch.aten.item %1215 : !torch.vtensor<[],f32> -> !torch.float | |
| %1218 = torch.aten.item %1216 : !torch.vtensor<[],si8> -> !torch.int | |
| %1219 = torch.aten.quantize_per_tensor %1214, %1217, %1218, %int12_238 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1220 = torch.aten.int_repr %1219 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1221 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1223 = torch.aten.item %1221 : !torch.vtensor<[],f32> -> !torch.float | |
| %1224 = torch.aten.item %1222 : !torch.vtensor<[],si8> -> !torch.int | |
| %1225 = torch.aten._make_per_tensor_quantized_tensor %1220, %1223, %1224 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1226 = torch.aten.dequantize.self %1225 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1227 = torch.aten.sigmoid %1226 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1228 = torch.aten.mul.Tensor %1226, %1227 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1229 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1230 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_239 = torch.constant.int 12 | |
| %1231 = torch.aten.item %1229 : !torch.vtensor<[],f32> -> !torch.float | |
| %1232 = torch.aten.item %1230 : !torch.vtensor<[],si8> -> !torch.int | |
| %1233 = torch.aten.quantize_per_tensor %1228, %1231, %1232, %int12_239 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1234 = torch.aten.int_repr %1233 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1235 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1236 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1237 = torch.aten.item %1235 : !torch.vtensor<[],f32> -> !torch.float | |
| %1238 = torch.aten.item %1236 : !torch.vtensor<[],si8> -> !torch.int | |
| %1239 = torch.aten._make_per_tensor_quantized_tensor %1234, %1237, %1238 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1240 = torch.aten.dequantize.self %1239 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %int1_240 = torch.constant.int 1 | |
| %1241 = torch.aten.add.Tensor %1130, %1240, %int1_240 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1242 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1243 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_241 = torch.constant.int 12 | |
| %1244 = torch.aten.item %1242 : !torch.vtensor<[],f32> -> !torch.float | |
| %1245 = torch.aten.item %1243 : !torch.vtensor<[],si8> -> !torch.int | |
| %1246 = torch.aten.quantize_per_tensor %1241, %1244, %1245, %int12_241 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1247 = torch.aten.int_repr %1246 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1248 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1249 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1250 = torch.aten.item %1248 : !torch.vtensor<[],f32> -> !torch.float | |
| %1251 = torch.aten.item %1249 : !torch.vtensor<[],si8> -> !torch.int | |
| %1252 = torch.aten._make_per_tensor_quantized_tensor %1247, %1250, %1251 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1253 = torch.aten.dequantize.self %1252 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1254 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1255 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_242 = torch.constant.int 12 | |
| %1256 = torch.aten.item %1254 : !torch.vtensor<[],f32> -> !torch.float | |
| %1257 = torch.aten.item %1255 : !torch.vtensor<[],si8> -> !torch.int | |
| %1258 = torch.aten.quantize_per_tensor %34, %1256, %1257, %int12_242 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1259 = torch.aten.int_repr %1258 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %1260 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1261 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1262 = torch.aten.item %1260 : !torch.vtensor<[],f32> -> !torch.float | |
| %1263 = torch.aten.item %1261 : !torch.vtensor<[],si8> -> !torch.int | |
| %1264 = torch.aten._make_per_tensor_quantized_tensor %1259, %1262, %1263 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1265 = torch.aten.dequantize.self %1264 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %1266 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1267 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_243 = torch.constant.int 12 | |
| %1268 = torch.aten.item %1266 : !torch.vtensor<[],f32> -> !torch.float | |
| %1269 = torch.aten.item %1267 : !torch.vtensor<[],si8> -> !torch.int | |
| %1270 = torch.aten.quantize_per_tensor %35, %1268, %1269, %int12_243 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1271 = torch.aten.int_repr %1270 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %1272 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1273 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1274 = torch.aten.item %1272 : !torch.vtensor<[],f32> -> !torch.float | |
| %1275 = torch.aten.item %1273 : !torch.vtensor<[],si8> -> !torch.int | |
| %1276 = torch.aten._make_per_tensor_quantized_tensor %1271, %1274, %1275 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1277 = torch.aten.dequantize.self %1276 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_244 = torch.constant.int 1 | |
| %int1_245 = torch.constant.int 1 | |
| %int1_246 = torch.constant.int 1 | |
| %int1_247 = torch.constant.int 1 | |
| %int1_248 = torch.constant.int 1 | |
| %int1_249 = torch.constant.int 1 | |
| %int0_250 = torch.constant.int 0 | |
| %1278 = torch.prim.ListConstruct %int1_244, %int1_245 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1279 = torch.prim.ListConstruct %int1_246, %int1_247 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1280 = torch.prim.ListConstruct %int1_248, %int1_249 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1281 = torch.prim.ListConstruct %int0_250, %int0_250 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_251 = torch.constant.bool false | |
| %int1_252 = torch.constant.int 1 | |
| %1282 = torch.aten.convolution %1253, %1265, %1277, %1280, %1278, %1279, %false_251, %1281, %int1_252 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1283 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1284 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_253 = torch.constant.int 12 | |
| %1285 = torch.aten.item %1283 : !torch.vtensor<[],f32> -> !torch.float | |
| %1286 = torch.aten.item %1284 : !torch.vtensor<[],si8> -> !torch.int | |
| %1287 = torch.aten.quantize_per_tensor %1282, %1285, %1286, %int12_253 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1288 = torch.aten.int_repr %1287 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1289 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1290 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1291 = torch.aten.item %1289 : !torch.vtensor<[],f32> -> !torch.float | |
| %1292 = torch.aten.item %1290 : !torch.vtensor<[],si8> -> !torch.int | |
| %1293 = torch.aten._make_per_tensor_quantized_tensor %1288, %1291, %1292 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1294 = torch.aten.dequantize.self %1293 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1295 = torch.aten.sigmoid %1294 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1296 = torch.aten.mul.Tensor %1294, %1295 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1297 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1298 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_254 = torch.constant.int 12 | |
| %1299 = torch.aten.item %1297 : !torch.vtensor<[],f32> -> !torch.float | |
| %1300 = torch.aten.item %1298 : !torch.vtensor<[],si8> -> !torch.int | |
| %1301 = torch.aten.quantize_per_tensor %1296, %1299, %1300, %int12_254 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1302 = torch.aten.int_repr %1301 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1303 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1304 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1305 = torch.aten.item %1303 : !torch.vtensor<[],f32> -> !torch.float | |
| %1306 = torch.aten.item %1304 : !torch.vtensor<[],si8> -> !torch.int | |
| %1307 = torch.aten._make_per_tensor_quantized_tensor %1302, %1305, %1306 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1308 = torch.aten.dequantize.self %1307 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1309 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1310 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_255 = torch.constant.int 12 | |
| %1311 = torch.aten.item %1309 : !torch.vtensor<[],f32> -> !torch.float | |
| %1312 = torch.aten.item %1310 : !torch.vtensor<[],si8> -> !torch.int | |
| %1313 = torch.aten.quantize_per_tensor %36, %1311, %1312, %int12_255 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1314 = torch.aten.int_repr %1313 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %1315 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1316 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1317 = torch.aten.item %1315 : !torch.vtensor<[],f32> -> !torch.float | |
| %1318 = torch.aten.item %1316 : !torch.vtensor<[],si8> -> !torch.int | |
| %1319 = torch.aten._make_per_tensor_quantized_tensor %1314, %1317, %1318 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %1320 = torch.aten.dequantize.self %1319 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %1321 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1322 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_256 = torch.constant.int 12 | |
| %1323 = torch.aten.item %1321 : !torch.vtensor<[],f32> -> !torch.float | |
| %1324 = torch.aten.item %1322 : !torch.vtensor<[],si8> -> !torch.int | |
| %1325 = torch.aten.quantize_per_tensor %37, %1323, %1324, %int12_256 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1326 = torch.aten.int_repr %1325 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %1327 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1328 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1329 = torch.aten.item %1327 : !torch.vtensor<[],f32> -> !torch.float | |
| %1330 = torch.aten.item %1328 : !torch.vtensor<[],si8> -> !torch.int | |
| %1331 = torch.aten._make_per_tensor_quantized_tensor %1326, %1329, %1330 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %1332 = torch.aten.dequantize.self %1331 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_257 = torch.constant.int 1 | |
| %int1_258 = torch.constant.int 1 | |
| %int1_259 = torch.constant.int 1 | |
| %int1_260 = torch.constant.int 1 | |
| %int1_261 = torch.constant.int 1 | |
| %int1_262 = torch.constant.int 1 | |
| %int0_263 = torch.constant.int 0 | |
| %1333 = torch.prim.ListConstruct %int1_257, %int1_258 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1334 = torch.prim.ListConstruct %int1_259, %int1_260 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1335 = torch.prim.ListConstruct %int1_261, %int1_262 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1336 = torch.prim.ListConstruct %int0_263, %int0_263 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_264 = torch.constant.bool false | |
| %int1_265 = torch.constant.int 1 | |
| %1337 = torch.aten.convolution %1308, %1320, %1332, %1335, %1333, %1334, %false_264, %1336, %int1_265 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1338 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1339 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_266 = torch.constant.int 12 | |
| %1340 = torch.aten.item %1338 : !torch.vtensor<[],f32> -> !torch.float | |
| %1341 = torch.aten.item %1339 : !torch.vtensor<[],si8> -> !torch.int | |
| %1342 = torch.aten.quantize_per_tensor %1337, %1340, %1341, %int12_266 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1343 = torch.aten.int_repr %1342 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1344 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1345 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1346 = torch.aten.item %1344 : !torch.vtensor<[],f32> -> !torch.float | |
| %1347 = torch.aten.item %1345 : !torch.vtensor<[],si8> -> !torch.int | |
| %1348 = torch.aten._make_per_tensor_quantized_tensor %1343, %1346, %1347 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1349 = torch.aten.dequantize.self %1348 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1350 = torch.aten.sigmoid %1349 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1351 = torch.aten.mul.Tensor %1349, %1350 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %1352 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1353 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_267 = torch.constant.int 12 | |
| %1354 = torch.aten.item %1352 : !torch.vtensor<[],f32> -> !torch.float | |
| %1355 = torch.aten.item %1353 : !torch.vtensor<[],si8> -> !torch.int | |
| %1356 = torch.aten.quantize_per_tensor %1351, %1354, %1355, %int12_267 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1357 = torch.aten.int_repr %1356 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %1358 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1359 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1360 = torch.aten.item %1358 : !torch.vtensor<[],f32> -> !torch.float | |
| %1361 = torch.aten.item %1359 : !torch.vtensor<[],si8> -> !torch.int | |
| %1362 = torch.aten._make_per_tensor_quantized_tensor %1357, %1360, %1361 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %1363 = torch.aten.dequantize.self %1362 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %int1_268 = torch.constant.int 1 | |
| %1364 = torch.aten.add.Tensor %1253, %1363, %int1_268 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %1365 = torch.prim.ListConstruct %1116, %1130, %1253, %1364 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>) -> !torch.list<vtensor> | |
| %int1_269 = torch.constant.int 1 | |
| %1366 = torch.aten.cat %1365, %int1_269 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,256,40,40],f32> | |
| %1367 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1368 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_270 = torch.constant.int 12 | |
| %1369 = torch.aten.item %1367 : !torch.vtensor<[],f32> -> !torch.float | |
| %1370 = torch.aten.item %1368 : !torch.vtensor<[],si8> -> !torch.int | |
| %1371 = torch.aten.quantize_per_tensor %1366, %1369, %1370, %int12_270 : !torch.vtensor<[1,256,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,40,40],!torch.qint8> | |
| %1372 = torch.aten.int_repr %1371 : !torch.vtensor<[1,256,40,40],!torch.qint8> -> !torch.vtensor<[1,256,40,40],si8> | |
| %1373 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1374 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1375 = torch.aten.item %1373 : !torch.vtensor<[],f32> -> !torch.float | |
| %1376 = torch.aten.item %1374 : !torch.vtensor<[],si8> -> !torch.int | |
| %1377 = torch.aten._make_per_tensor_quantized_tensor %1372, %1375, %1376 : !torch.vtensor<[1,256,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,40,40],!torch.qint8> | |
| %1378 = torch.aten.dequantize.self %1377 : !torch.vtensor<[1,256,40,40],!torch.qint8> -> !torch.vtensor<[1,256,40,40],f32> | |
| %1379 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1380 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_271 = torch.constant.int 12 | |
| %1381 = torch.aten.item %1379 : !torch.vtensor<[],f32> -> !torch.float | |
| %1382 = torch.aten.item %1380 : !torch.vtensor<[],si8> -> !torch.int | |
| %1383 = torch.aten.quantize_per_tensor %38, %1381, %1382, %int12_271 : !torch.vtensor<[128,256,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
| %1384 = torch.aten.int_repr %1383 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],si8> | |
| %1385 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1386 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1387 = torch.aten.item %1385 : !torch.vtensor<[],f32> -> !torch.float | |
| %1388 = torch.aten.item %1386 : !torch.vtensor<[],si8> -> !torch.int | |
| %1389 = torch.aten._make_per_tensor_quantized_tensor %1384, %1387, %1388 : !torch.vtensor<[128,256,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
| %1390 = torch.aten.dequantize.self %1389 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],f32> | |
| %1391 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1392 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_272 = torch.constant.int 12 | |
| %1393 = torch.aten.item %1391 : !torch.vtensor<[],f32> -> !torch.float | |
| %1394 = torch.aten.item %1392 : !torch.vtensor<[],si8> -> !torch.int | |
| %1395 = torch.aten.quantize_per_tensor %39, %1393, %1394, %int12_272 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1396 = torch.aten.int_repr %1395 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1397 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1398 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1399 = torch.aten.item %1397 : !torch.vtensor<[],f32> -> !torch.float | |
| %1400 = torch.aten.item %1398 : !torch.vtensor<[],si8> -> !torch.int | |
| %1401 = torch.aten._make_per_tensor_quantized_tensor %1396, %1399, %1400 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1402 = torch.aten.dequantize.self %1401 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_273 = torch.constant.int 0 | |
| %int0_274 = torch.constant.int 0 | |
| %int1_275 = torch.constant.int 1 | |
| %int1_276 = torch.constant.int 1 | |
| %int1_277 = torch.constant.int 1 | |
| %int1_278 = torch.constant.int 1 | |
| %int0_279 = torch.constant.int 0 | |
| %1403 = torch.prim.ListConstruct %int0_273, %int0_274 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1404 = torch.prim.ListConstruct %int1_275, %int1_276 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1405 = torch.prim.ListConstruct %int1_277, %int1_278 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1406 = torch.prim.ListConstruct %int0_279, %int0_279 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_280 = torch.constant.bool false | |
| %int1_281 = torch.constant.int 1 | |
| %1407 = torch.aten.convolution %1378, %1390, %1402, %1405, %1403, %1404, %false_280, %1406, %int1_281 : !torch.vtensor<[1,256,40,40],f32>, !torch.vtensor<[128,256,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %1408 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1409 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_282 = torch.constant.int 12 | |
| %1410 = torch.aten.item %1408 : !torch.vtensor<[],f32> -> !torch.float | |
| %1411 = torch.aten.item %1409 : !torch.vtensor<[],si8> -> !torch.int | |
| %1412 = torch.aten.quantize_per_tensor %1407, %1410, %1411, %int12_282 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1413 = torch.aten.int_repr %1412 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1414 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1415 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1416 = torch.aten.item %1414 : !torch.vtensor<[],f32> -> !torch.float | |
| %1417 = torch.aten.item %1415 : !torch.vtensor<[],si8> -> !torch.int | |
| %1418 = torch.aten._make_per_tensor_quantized_tensor %1413, %1416, %1417 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1419 = torch.aten.dequantize.self %1418 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1420 = torch.aten.sigmoid %1419 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1421 = torch.aten.mul.Tensor %1419, %1420 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1422 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1423 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_283 = torch.constant.int 12 | |
| %1424 = torch.aten.item %1422 : !torch.vtensor<[],f32> -> !torch.float | |
| %1425 = torch.aten.item %1423 : !torch.vtensor<[],si8> -> !torch.int | |
| %1426 = torch.aten.quantize_per_tensor %1421, %1424, %1425, %int12_283 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1427 = torch.aten.int_repr %1426 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1428 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1429 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1430 = torch.aten.item %1428 : !torch.vtensor<[],f32> -> !torch.float | |
| %1431 = torch.aten.item %1429 : !torch.vtensor<[],si8> -> !torch.int | |
| %1432 = torch.aten._make_per_tensor_quantized_tensor %1427, %1430, %1431 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1433 = torch.aten.dequantize.self %1432 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1434 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1435 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_284 = torch.constant.int 12 | |
| %1436 = torch.aten.item %1434 : !torch.vtensor<[],f32> -> !torch.float | |
| %1437 = torch.aten.item %1435 : !torch.vtensor<[],si8> -> !torch.int | |
| %1438 = torch.aten.quantize_per_tensor %40, %1436, %1437, %int12_284 : !torch.vtensor<[256,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,128,3,3],!torch.qint8> | |
| %1439 = torch.aten.int_repr %1438 : !torch.vtensor<[256,128,3,3],!torch.qint8> -> !torch.vtensor<[256,128,3,3],si8> | |
| %1440 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1441 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1442 = torch.aten.item %1440 : !torch.vtensor<[],f32> -> !torch.float | |
| %1443 = torch.aten.item %1441 : !torch.vtensor<[],si8> -> !torch.int | |
| %1444 = torch.aten._make_per_tensor_quantized_tensor %1439, %1442, %1443 : !torch.vtensor<[256,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,128,3,3],!torch.qint8> | |
| %1445 = torch.aten.dequantize.self %1444 : !torch.vtensor<[256,128,3,3],!torch.qint8> -> !torch.vtensor<[256,128,3,3],f32> | |
| %1446 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1447 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_285 = torch.constant.int 12 | |
| %1448 = torch.aten.item %1446 : !torch.vtensor<[],f32> -> !torch.float | |
| %1449 = torch.aten.item %1447 : !torch.vtensor<[],si8> -> !torch.int | |
| %1450 = torch.aten.quantize_per_tensor %41, %1448, %1449, %int12_285 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1451 = torch.aten.int_repr %1450 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %1452 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1453 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1454 = torch.aten.item %1452 : !torch.vtensor<[],f32> -> !torch.float | |
| %1455 = torch.aten.item %1453 : !torch.vtensor<[],si8> -> !torch.int | |
| %1456 = torch.aten._make_per_tensor_quantized_tensor %1451, %1454, %1455 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1457 = torch.aten.dequantize.self %1456 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %int1_286 = torch.constant.int 1 | |
| %int1_287 = torch.constant.int 1 | |
| %int1_288 = torch.constant.int 1 | |
| %int1_289 = torch.constant.int 1 | |
| %int2_290 = torch.constant.int 2 | |
| %int2_291 = torch.constant.int 2 | |
| %int0_292 = torch.constant.int 0 | |
| %1458 = torch.prim.ListConstruct %int1_286, %int1_287 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1459 = torch.prim.ListConstruct %int1_288, %int1_289 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1460 = torch.prim.ListConstruct %int2_290, %int2_291 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1461 = torch.prim.ListConstruct %int0_292, %int0_292 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_293 = torch.constant.bool false | |
| %int1_294 = torch.constant.int 1 | |
| %1462 = torch.aten.convolution %1433, %1445, %1457, %1460, %1458, %1459, %false_293, %1461, %int1_294 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[256,128,3,3],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
| %1463 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1464 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_295 = torch.constant.int 12 | |
| %1465 = torch.aten.item %1463 : !torch.vtensor<[],f32> -> !torch.float | |
| %1466 = torch.aten.item %1464 : !torch.vtensor<[],si8> -> !torch.int | |
| %1467 = torch.aten.quantize_per_tensor %1462, %1465, %1466, %int12_295 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1468 = torch.aten.int_repr %1467 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1469 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1470 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1471 = torch.aten.item %1469 : !torch.vtensor<[],f32> -> !torch.float | |
| %1472 = torch.aten.item %1470 : !torch.vtensor<[],si8> -> !torch.int | |
| %1473 = torch.aten._make_per_tensor_quantized_tensor %1468, %1471, %1472 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1474 = torch.aten.dequantize.self %1473 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1475 = torch.aten.sigmoid %1474 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1476 = torch.aten.mul.Tensor %1474, %1475 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1477 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1478 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_296 = torch.constant.int 12 | |
| %1479 = torch.aten.item %1477 : !torch.vtensor<[],f32> -> !torch.float | |
| %1480 = torch.aten.item %1478 : !torch.vtensor<[],si8> -> !torch.int | |
| %1481 = torch.aten.quantize_per_tensor %1476, %1479, %1480, %int12_296 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1482 = torch.aten.int_repr %1481 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1483 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1484 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1485 = torch.aten.item %1483 : !torch.vtensor<[],f32> -> !torch.float | |
| %1486 = torch.aten.item %1484 : !torch.vtensor<[],si8> -> !torch.int | |
| %1487 = torch.aten._make_per_tensor_quantized_tensor %1482, %1485, %1486 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1488 = torch.aten.dequantize.self %1487 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1489 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1490 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_297 = torch.constant.int 12 | |
| %1491 = torch.aten.item %1489 : !torch.vtensor<[],f32> -> !torch.float | |
| %1492 = torch.aten.item %1490 : !torch.vtensor<[],si8> -> !torch.int | |
| %1493 = torch.aten.quantize_per_tensor %42, %1491, %1492, %int12_297 : !torch.vtensor<[256,256,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,256,1,1],!torch.qint8> | |
| %1494 = torch.aten.int_repr %1493 : !torch.vtensor<[256,256,1,1],!torch.qint8> -> !torch.vtensor<[256,256,1,1],si8> | |
| %1495 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1496 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1497 = torch.aten.item %1495 : !torch.vtensor<[],f32> -> !torch.float | |
| %1498 = torch.aten.item %1496 : !torch.vtensor<[],si8> -> !torch.int | |
| %1499 = torch.aten._make_per_tensor_quantized_tensor %1494, %1497, %1498 : !torch.vtensor<[256,256,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,256,1,1],!torch.qint8> | |
| %1500 = torch.aten.dequantize.self %1499 : !torch.vtensor<[256,256,1,1],!torch.qint8> -> !torch.vtensor<[256,256,1,1],f32> | |
| %1501 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1502 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_298 = torch.constant.int 12 | |
| %1503 = torch.aten.item %1501 : !torch.vtensor<[],f32> -> !torch.float | |
| %1504 = torch.aten.item %1502 : !torch.vtensor<[],si8> -> !torch.int | |
| %1505 = torch.aten.quantize_per_tensor %43, %1503, %1504, %int12_298 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1506 = torch.aten.int_repr %1505 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %1507 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1508 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1509 = torch.aten.item %1507 : !torch.vtensor<[],f32> -> !torch.float | |
| %1510 = torch.aten.item %1508 : !torch.vtensor<[],si8> -> !torch.int | |
| %1511 = torch.aten._make_per_tensor_quantized_tensor %1506, %1509, %1510 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1512 = torch.aten.dequantize.self %1511 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %int0_299 = torch.constant.int 0 | |
| %int0_300 = torch.constant.int 0 | |
| %int1_301 = torch.constant.int 1 | |
| %int1_302 = torch.constant.int 1 | |
| %int1_303 = torch.constant.int 1 | |
| %int1_304 = torch.constant.int 1 | |
| %int0_305 = torch.constant.int 0 | |
| %1513 = torch.prim.ListConstruct %int0_299, %int0_300 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1514 = torch.prim.ListConstruct %int1_301, %int1_302 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1515 = torch.prim.ListConstruct %int1_303, %int1_304 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1516 = torch.prim.ListConstruct %int0_305, %int0_305 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_306 = torch.constant.bool false | |
| %int1_307 = torch.constant.int 1 | |
| %1517 = torch.aten.convolution %1488, %1500, %1512, %1515, %1513, %1514, %false_306, %1516, %int1_307 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[256,256,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
| %1518 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1519 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_308 = torch.constant.int 12 | |
| %1520 = torch.aten.item %1518 : !torch.vtensor<[],f32> -> !torch.float | |
| %1521 = torch.aten.item %1519 : !torch.vtensor<[],si8> -> !torch.int | |
| %1522 = torch.aten.quantize_per_tensor %1517, %1520, %1521, %int12_308 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1523 = torch.aten.int_repr %1522 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1524 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1525 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1526 = torch.aten.item %1524 : !torch.vtensor<[],f32> -> !torch.float | |
| %1527 = torch.aten.item %1525 : !torch.vtensor<[],si8> -> !torch.int | |
| %1528 = torch.aten._make_per_tensor_quantized_tensor %1523, %1526, %1527 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1529 = torch.aten.dequantize.self %1528 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1530 = torch.aten.sigmoid %1529 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1531 = torch.aten.mul.Tensor %1529, %1530 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1532 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1533 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_309 = torch.constant.int 12 | |
| %1534 = torch.aten.item %1532 : !torch.vtensor<[],f32> -> !torch.float | |
| %1535 = torch.aten.item %1533 : !torch.vtensor<[],si8> -> !torch.int | |
| %1536 = torch.aten.quantize_per_tensor %1531, %1534, %1535, %int12_309 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1537 = torch.aten.int_repr %1536 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1538 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1539 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1540 = torch.aten.item %1538 : !torch.vtensor<[],f32> -> !torch.float | |
| %1541 = torch.aten.item %1539 : !torch.vtensor<[],si8> -> !torch.int | |
| %1542 = torch.aten._make_per_tensor_quantized_tensor %1537, %1540, %1541 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1543 = torch.aten.dequantize.self %1542 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1544 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1545 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1546 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1547 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_310 = torch.constant.int 0 | |
| %int0_311 = torch.constant.int 0 | |
| %1548 = torch.prim.NumToTensor.Scalar %int0_311 : !torch.int -> !torch.vtensor<[1],si64> | |
| %1549 = torch.aten.index_select %1545, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1550 = torch.aten.item %1549 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1551 = torch.aten.index_select %1546, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1552 = torch.aten.item %1551 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1553 = torch.aten.index_select %1544, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1554 = torch.aten.item %1553 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1555 = torch.aten.index_select %1547, %int0_310, %1548 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1556 = torch.aten.item %1555 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1557 = torch.aten.slice.Tensor %1543, %1554, %1550, %1552, %1556 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %1558 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1559 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1560 = torch.vtensor.literal(dense<256> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1561 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_312 = torch.constant.int 0 | |
| %int0_313 = torch.constant.int 0 | |
| %1562 = torch.prim.NumToTensor.Scalar %int0_313 : !torch.int -> !torch.vtensor<[1],si64> | |
| %1563 = torch.aten.index_select %1559, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1564 = torch.aten.item %1563 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1565 = torch.aten.index_select %1560, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1566 = torch.aten.item %1565 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1567 = torch.aten.index_select %1558, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1568 = torch.aten.item %1567 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1569 = torch.aten.index_select %1561, %int0_312, %1562 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1570 = torch.aten.item %1569 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1571 = torch.aten.slice.Tensor %1543, %1568, %1564, %1566, %1570 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %1572 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1573 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_314 = torch.constant.int 12 | |
| %1574 = torch.aten.item %1572 : !torch.vtensor<[],f32> -> !torch.float | |
| %1575 = torch.aten.item %1573 : !torch.vtensor<[],si8> -> !torch.int | |
| %1576 = torch.aten.quantize_per_tensor %44, %1574, %1575, %int12_314 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %1577 = torch.aten.int_repr %1576 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %1578 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1579 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1580 = torch.aten.item %1578 : !torch.vtensor<[],f32> -> !torch.float | |
| %1581 = torch.aten.item %1579 : !torch.vtensor<[],si8> -> !torch.int | |
| %1582 = torch.aten._make_per_tensor_quantized_tensor %1577, %1580, %1581 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %1583 = torch.aten.dequantize.self %1582 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %1584 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1585 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_315 = torch.constant.int 12 | |
| %1586 = torch.aten.item %1584 : !torch.vtensor<[],f32> -> !torch.float | |
| %1587 = torch.aten.item %1585 : !torch.vtensor<[],si8> -> !torch.int | |
| %1588 = torch.aten.quantize_per_tensor %45, %1586, %1587, %int12_315 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1589 = torch.aten.int_repr %1588 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1590 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1591 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1592 = torch.aten.item %1590 : !torch.vtensor<[],f32> -> !torch.float | |
| %1593 = torch.aten.item %1591 : !torch.vtensor<[],si8> -> !torch.int | |
| %1594 = torch.aten._make_per_tensor_quantized_tensor %1589, %1592, %1593 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1595 = torch.aten.dequantize.self %1594 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int1_316 = torch.constant.int 1 | |
| %int1_317 = torch.constant.int 1 | |
| %int1_318 = torch.constant.int 1 | |
| %int1_319 = torch.constant.int 1 | |
| %int1_320 = torch.constant.int 1 | |
| %int1_321 = torch.constant.int 1 | |
| %int0_322 = torch.constant.int 0 | |
| %1596 = torch.prim.ListConstruct %int1_316, %int1_317 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1597 = torch.prim.ListConstruct %int1_318, %int1_319 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1598 = torch.prim.ListConstruct %int1_320, %int1_321 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1599 = torch.prim.ListConstruct %int0_322, %int0_322 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_323 = torch.constant.bool false | |
| %int1_324 = torch.constant.int 1 | |
| %1600 = torch.aten.convolution %1571, %1583, %1595, %1598, %1596, %1597, %false_323, %1599, %int1_324 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %1601 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1602 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_325 = torch.constant.int 12 | |
| %1603 = torch.aten.item %1601 : !torch.vtensor<[],f32> -> !torch.float | |
| %1604 = torch.aten.item %1602 : !torch.vtensor<[],si8> -> !torch.int | |
| %1605 = torch.aten.quantize_per_tensor %1600, %1603, %1604, %int12_325 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1606 = torch.aten.int_repr %1605 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1607 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1608 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1609 = torch.aten.item %1607 : !torch.vtensor<[],f32> -> !torch.float | |
| %1610 = torch.aten.item %1608 : !torch.vtensor<[],si8> -> !torch.int | |
| %1611 = torch.aten._make_per_tensor_quantized_tensor %1606, %1609, %1610 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1612 = torch.aten.dequantize.self %1611 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1613 = torch.aten.sigmoid %1612 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1614 = torch.aten.mul.Tensor %1612, %1613 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1615 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1616 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_326 = torch.constant.int 12 | |
| %1617 = torch.aten.item %1615 : !torch.vtensor<[],f32> -> !torch.float | |
| %1618 = torch.aten.item %1616 : !torch.vtensor<[],si8> -> !torch.int | |
| %1619 = torch.aten.quantize_per_tensor %1614, %1617, %1618, %int12_326 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1620 = torch.aten.int_repr %1619 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1621 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1622 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1623 = torch.aten.item %1621 : !torch.vtensor<[],f32> -> !torch.float | |
| %1624 = torch.aten.item %1622 : !torch.vtensor<[],si8> -> !torch.int | |
| %1625 = torch.aten._make_per_tensor_quantized_tensor %1620, %1623, %1624 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1626 = torch.aten.dequantize.self %1625 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1627 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1628 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_327 = torch.constant.int 12 | |
| %1629 = torch.aten.item %1627 : !torch.vtensor<[],f32> -> !torch.float | |
| %1630 = torch.aten.item %1628 : !torch.vtensor<[],si8> -> !torch.int | |
| %1631 = torch.aten.quantize_per_tensor %46, %1629, %1630, %int12_327 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %1632 = torch.aten.int_repr %1631 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %1633 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1634 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1635 = torch.aten.item %1633 : !torch.vtensor<[],f32> -> !torch.float | |
| %1636 = torch.aten.item %1634 : !torch.vtensor<[],si8> -> !torch.int | |
| %1637 = torch.aten._make_per_tensor_quantized_tensor %1632, %1635, %1636 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %1638 = torch.aten.dequantize.self %1637 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %1639 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1640 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_328 = torch.constant.int 12 | |
| %1641 = torch.aten.item %1639 : !torch.vtensor<[],f32> -> !torch.float | |
| %1642 = torch.aten.item %1640 : !torch.vtensor<[],si8> -> !torch.int | |
| %1643 = torch.aten.quantize_per_tensor %47, %1641, %1642, %int12_328 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1644 = torch.aten.int_repr %1643 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1645 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1646 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1647 = torch.aten.item %1645 : !torch.vtensor<[],f32> -> !torch.float | |
| %1648 = torch.aten.item %1646 : !torch.vtensor<[],si8> -> !torch.int | |
| %1649 = torch.aten._make_per_tensor_quantized_tensor %1644, %1647, %1648 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1650 = torch.aten.dequantize.self %1649 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int1_329 = torch.constant.int 1 | |
| %int1_330 = torch.constant.int 1 | |
| %int1_331 = torch.constant.int 1 | |
| %int1_332 = torch.constant.int 1 | |
| %int1_333 = torch.constant.int 1 | |
| %int1_334 = torch.constant.int 1 | |
| %int0_335 = torch.constant.int 0 | |
| %1651 = torch.prim.ListConstruct %int1_329, %int1_330 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1652 = torch.prim.ListConstruct %int1_331, %int1_332 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1653 = torch.prim.ListConstruct %int1_333, %int1_334 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1654 = torch.prim.ListConstruct %int0_335, %int0_335 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_336 = torch.constant.bool false | |
| %int1_337 = torch.constant.int 1 | |
| %1655 = torch.aten.convolution %1626, %1638, %1650, %1653, %1651, %1652, %false_336, %1654, %int1_337 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %1656 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1657 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_338 = torch.constant.int 12 | |
| %1658 = torch.aten.item %1656 : !torch.vtensor<[],f32> -> !torch.float | |
| %1659 = torch.aten.item %1657 : !torch.vtensor<[],si8> -> !torch.int | |
| %1660 = torch.aten.quantize_per_tensor %1655, %1658, %1659, %int12_338 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1661 = torch.aten.int_repr %1660 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1662 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1663 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1664 = torch.aten.item %1662 : !torch.vtensor<[],f32> -> !torch.float | |
| %1665 = torch.aten.item %1663 : !torch.vtensor<[],si8> -> !torch.int | |
| %1666 = torch.aten._make_per_tensor_quantized_tensor %1661, %1664, %1665 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1667 = torch.aten.dequantize.self %1666 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1668 = torch.aten.sigmoid %1667 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1669 = torch.aten.mul.Tensor %1667, %1668 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1670 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1671 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_339 = torch.constant.int 12 | |
| %1672 = torch.aten.item %1670 : !torch.vtensor<[],f32> -> !torch.float | |
| %1673 = torch.aten.item %1671 : !torch.vtensor<[],si8> -> !torch.int | |
| %1674 = torch.aten.quantize_per_tensor %1669, %1672, %1673, %int12_339 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1675 = torch.aten.int_repr %1674 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1676 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1677 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1678 = torch.aten.item %1676 : !torch.vtensor<[],f32> -> !torch.float | |
| %1679 = torch.aten.item %1677 : !torch.vtensor<[],si8> -> !torch.int | |
| %1680 = torch.aten._make_per_tensor_quantized_tensor %1675, %1678, %1679 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1681 = torch.aten.dequantize.self %1680 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %int1_340 = torch.constant.int 1 | |
| %1682 = torch.aten.add.Tensor %1571, %1681, %int1_340 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %1683 = torch.prim.ListConstruct %1557, %1571, %1682 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>) -> !torch.list<vtensor> | |
| %int1_341 = torch.constant.int 1 | |
| %1684 = torch.aten.cat %1683, %int1_341 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,384,20,20],f32> | |
| %1685 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1686 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_342 = torch.constant.int 12 | |
| %1687 = torch.aten.item %1685 : !torch.vtensor<[],f32> -> !torch.float | |
| %1688 = torch.aten.item %1686 : !torch.vtensor<[],si8> -> !torch.int | |
| %1689 = torch.aten.quantize_per_tensor %1684, %1687, %1688, %int12_342 : !torch.vtensor<[1,384,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
| %1690 = torch.aten.int_repr %1689 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],si8> | |
| %1691 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1692 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1693 = torch.aten.item %1691 : !torch.vtensor<[],f32> -> !torch.float | |
| %1694 = torch.aten.item %1692 : !torch.vtensor<[],si8> -> !torch.int | |
| %1695 = torch.aten._make_per_tensor_quantized_tensor %1690, %1693, %1694 : !torch.vtensor<[1,384,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
| %1696 = torch.aten.dequantize.self %1695 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],f32> | |
| %1697 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1698 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_343 = torch.constant.int 12 | |
| %1699 = torch.aten.item %1697 : !torch.vtensor<[],f32> -> !torch.float | |
| %1700 = torch.aten.item %1698 : !torch.vtensor<[],si8> -> !torch.int | |
| %1701 = torch.aten.quantize_per_tensor %48, %1699, %1700, %int12_343 : !torch.vtensor<[256,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
| %1702 = torch.aten.int_repr %1701 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],si8> | |
| %1703 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1704 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1705 = torch.aten.item %1703 : !torch.vtensor<[],f32> -> !torch.float | |
| %1706 = torch.aten.item %1704 : !torch.vtensor<[],si8> -> !torch.int | |
| %1707 = torch.aten._make_per_tensor_quantized_tensor %1702, %1705, %1706 : !torch.vtensor<[256,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
| %1708 = torch.aten.dequantize.self %1707 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],f32> | |
| %1709 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1710 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_344 = torch.constant.int 12 | |
| %1711 = torch.aten.item %1709 : !torch.vtensor<[],f32> -> !torch.float | |
| %1712 = torch.aten.item %1710 : !torch.vtensor<[],si8> -> !torch.int | |
| %1713 = torch.aten.quantize_per_tensor %49, %1711, %1712, %int12_344 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1714 = torch.aten.int_repr %1713 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %1715 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1716 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1717 = torch.aten.item %1715 : !torch.vtensor<[],f32> -> !torch.float | |
| %1718 = torch.aten.item %1716 : !torch.vtensor<[],si8> -> !torch.int | |
| %1719 = torch.aten._make_per_tensor_quantized_tensor %1714, %1717, %1718 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1720 = torch.aten.dequantize.self %1719 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %int0_345 = torch.constant.int 0 | |
| %int0_346 = torch.constant.int 0 | |
| %int1_347 = torch.constant.int 1 | |
| %int1_348 = torch.constant.int 1 | |
| %int1_349 = torch.constant.int 1 | |
| %int1_350 = torch.constant.int 1 | |
| %int0_351 = torch.constant.int 0 | |
| %1721 = torch.prim.ListConstruct %int0_345, %int0_346 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1722 = torch.prim.ListConstruct %int1_347, %int1_348 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1723 = torch.prim.ListConstruct %int1_349, %int1_350 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1724 = torch.prim.ListConstruct %int0_351, %int0_351 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_352 = torch.constant.bool false | |
| %int1_353 = torch.constant.int 1 | |
| %1725 = torch.aten.convolution %1696, %1708, %1720, %1723, %1721, %1722, %false_352, %1724, %int1_353 : !torch.vtensor<[1,384,20,20],f32>, !torch.vtensor<[256,384,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
| %1726 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1727 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_354 = torch.constant.int 12 | |
| %1728 = torch.aten.item %1726 : !torch.vtensor<[],f32> -> !torch.float | |
| %1729 = torch.aten.item %1727 : !torch.vtensor<[],si8> -> !torch.int | |
| %1730 = torch.aten.quantize_per_tensor %1725, %1728, %1729, %int12_354 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1731 = torch.aten.int_repr %1730 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1732 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1733 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1734 = torch.aten.item %1732 : !torch.vtensor<[],f32> -> !torch.float | |
| %1735 = torch.aten.item %1733 : !torch.vtensor<[],si8> -> !torch.int | |
| %1736 = torch.aten._make_per_tensor_quantized_tensor %1731, %1734, %1735 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1737 = torch.aten.dequantize.self %1736 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1738 = torch.aten.sigmoid %1737 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1739 = torch.aten.mul.Tensor %1737, %1738 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1740 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1741 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_355 = torch.constant.int 12 | |
| %1742 = torch.aten.item %1740 : !torch.vtensor<[],f32> -> !torch.float | |
| %1743 = torch.aten.item %1741 : !torch.vtensor<[],si8> -> !torch.int | |
| %1744 = torch.aten.quantize_per_tensor %1739, %1742, %1743, %int12_355 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1745 = torch.aten.int_repr %1744 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1746 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1747 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1748 = torch.aten.item %1746 : !torch.vtensor<[],f32> -> !torch.float | |
| %1749 = torch.aten.item %1747 : !torch.vtensor<[],si8> -> !torch.int | |
| %1750 = torch.aten._make_per_tensor_quantized_tensor %1745, %1748, %1749 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1751 = torch.aten.dequantize.self %1750 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1752 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1753 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_356 = torch.constant.int 12 | |
| %1754 = torch.aten.item %1752 : !torch.vtensor<[],f32> -> !torch.float | |
| %1755 = torch.aten.item %1753 : !torch.vtensor<[],si8> -> !torch.int | |
| %1756 = torch.aten.quantize_per_tensor %50, %1754, %1755, %int12_356 : !torch.vtensor<[128,256,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
| %1757 = torch.aten.int_repr %1756 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],si8> | |
| %1758 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1759 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1760 = torch.aten.item %1758 : !torch.vtensor<[],f32> -> !torch.float | |
| %1761 = torch.aten.item %1759 : !torch.vtensor<[],si8> -> !torch.int | |
| %1762 = torch.aten._make_per_tensor_quantized_tensor %1757, %1760, %1761 : !torch.vtensor<[128,256,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,256,1,1],!torch.qint8> | |
| %1763 = torch.aten.dequantize.self %1762 : !torch.vtensor<[128,256,1,1],!torch.qint8> -> !torch.vtensor<[128,256,1,1],f32> | |
| %1764 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1765 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_357 = torch.constant.int 12 | |
| %1766 = torch.aten.item %1764 : !torch.vtensor<[],f32> -> !torch.float | |
| %1767 = torch.aten.item %1765 : !torch.vtensor<[],si8> -> !torch.int | |
| %1768 = torch.aten.quantize_per_tensor %51, %1766, %1767, %int12_357 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1769 = torch.aten.int_repr %1768 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1770 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1771 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1772 = torch.aten.item %1770 : !torch.vtensor<[],f32> -> !torch.float | |
| %1773 = torch.aten.item %1771 : !torch.vtensor<[],si8> -> !torch.int | |
| %1774 = torch.aten._make_per_tensor_quantized_tensor %1769, %1772, %1773 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1775 = torch.aten.dequantize.self %1774 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_358 = torch.constant.int 0 | |
| %int0_359 = torch.constant.int 0 | |
| %int1_360 = torch.constant.int 1 | |
| %int1_361 = torch.constant.int 1 | |
| %int1_362 = torch.constant.int 1 | |
| %int1_363 = torch.constant.int 1 | |
| %int0_364 = torch.constant.int 0 | |
| %1776 = torch.prim.ListConstruct %int0_358, %int0_359 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1777 = torch.prim.ListConstruct %int1_360, %int1_361 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1778 = torch.prim.ListConstruct %int1_362, %int1_363 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1779 = torch.prim.ListConstruct %int0_364, %int0_364 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_365 = torch.constant.bool false | |
| %int1_366 = torch.constant.int 1 | |
| %1780 = torch.aten.convolution %1751, %1763, %1775, %1778, %1776, %1777, %false_365, %1779, %int1_366 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[128,256,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %1781 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1782 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_367 = torch.constant.int 12 | |
| %1783 = torch.aten.item %1781 : !torch.vtensor<[],f32> -> !torch.float | |
| %1784 = torch.aten.item %1782 : !torch.vtensor<[],si8> -> !torch.int | |
| %1785 = torch.aten.quantize_per_tensor %1780, %1783, %1784, %int12_367 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1786 = torch.aten.int_repr %1785 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1787 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1788 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1789 = torch.aten.item %1787 : !torch.vtensor<[],f32> -> !torch.float | |
| %1790 = torch.aten.item %1788 : !torch.vtensor<[],si8> -> !torch.int | |
| %1791 = torch.aten._make_per_tensor_quantized_tensor %1786, %1789, %1790 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1792 = torch.aten.dequantize.self %1791 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1793 = torch.aten.sigmoid %1792 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1794 = torch.aten.mul.Tensor %1792, %1793 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %1795 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1796 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_368 = torch.constant.int 12 | |
| %1797 = torch.aten.item %1795 : !torch.vtensor<[],f32> -> !torch.float | |
| %1798 = torch.aten.item %1796 : !torch.vtensor<[],si8> -> !torch.int | |
| %1799 = torch.aten.quantize_per_tensor %1794, %1797, %1798, %int12_368 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1800 = torch.aten.int_repr %1799 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1801 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1802 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1803 = torch.aten.item %1801 : !torch.vtensor<[],f32> -> !torch.float | |
| %1804 = torch.aten.item %1802 : !torch.vtensor<[],si8> -> !torch.int | |
| %1805 = torch.aten._make_per_tensor_quantized_tensor %1800, %1803, %1804 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1806 = torch.aten.dequantize.self %1805 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %int5 = torch.constant.int 5 | |
| %int5_369 = torch.constant.int 5 | |
| %1807 = torch.prim.ListConstruct %int5, %int5_369 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int2_370 = torch.constant.int 2 | |
| %int2_371 = torch.constant.int 2 | |
| %1808 = torch.prim.ListConstruct %int2_370, %int2_371 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int1_372 = torch.constant.int 1 | |
| %int1_373 = torch.constant.int 1 | |
| %1809 = torch.prim.ListConstruct %int1_372, %int1_373 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int1_374 = torch.constant.int 1 | |
| %int1_375 = torch.constant.int 1 | |
| %1810 = torch.prim.ListConstruct %int1_374, %int1_375 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_376 = torch.constant.bool false | |
| %1811 = torch.aten.max_pool2d %1806, %1807, %1809, %1808, %1810, %false_376 : !torch.vtensor<[1,128,20,20],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[1,128,20,20],f32> | |
| %1812 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1813 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_377 = torch.constant.int 12 | |
| %1814 = torch.aten.item %1812 : !torch.vtensor<[],f32> -> !torch.float | |
| %1815 = torch.aten.item %1813 : !torch.vtensor<[],si8> -> !torch.int | |
| %1816 = torch.aten.quantize_per_tensor %1811, %1814, %1815, %int12_377 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1817 = torch.aten.int_repr %1816 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1818 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1819 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1820 = torch.aten.item %1818 : !torch.vtensor<[],f32> -> !torch.float | |
| %1821 = torch.aten.item %1819 : !torch.vtensor<[],si8> -> !torch.int | |
| %1822 = torch.aten._make_per_tensor_quantized_tensor %1817, %1820, %1821 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1823 = torch.aten.dequantize.self %1822 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %int5_378 = torch.constant.int 5 | |
| %int5_379 = torch.constant.int 5 | |
| %1824 = torch.prim.ListConstruct %int5_378, %int5_379 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int2_380 = torch.constant.int 2 | |
| %int2_381 = torch.constant.int 2 | |
| %1825 = torch.prim.ListConstruct %int2_380, %int2_381 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int1_382 = torch.constant.int 1 | |
| %int1_383 = torch.constant.int 1 | |
| %1826 = torch.prim.ListConstruct %int1_382, %int1_383 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int1_384 = torch.constant.int 1 | |
| %int1_385 = torch.constant.int 1 | |
| %1827 = torch.prim.ListConstruct %int1_384, %int1_385 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_386 = torch.constant.bool false | |
| %1828 = torch.aten.max_pool2d %1823, %1824, %1826, %1825, %1827, %false_386 : !torch.vtensor<[1,128,20,20],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[1,128,20,20],f32> | |
| %1829 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1830 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_387 = torch.constant.int 12 | |
| %1831 = torch.aten.item %1829 : !torch.vtensor<[],f32> -> !torch.float | |
| %1832 = torch.aten.item %1830 : !torch.vtensor<[],si8> -> !torch.int | |
| %1833 = torch.aten.quantize_per_tensor %1828, %1831, %1832, %int12_387 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1834 = torch.aten.int_repr %1833 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %1835 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1836 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1837 = torch.aten.item %1835 : !torch.vtensor<[],f32> -> !torch.float | |
| %1838 = torch.aten.item %1836 : !torch.vtensor<[],si8> -> !torch.int | |
| %1839 = torch.aten._make_per_tensor_quantized_tensor %1834, %1837, %1838 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %1840 = torch.aten.dequantize.self %1839 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %int5_388 = torch.constant.int 5 | |
| %int5_389 = torch.constant.int 5 | |
| %1841 = torch.prim.ListConstruct %int5_388, %int5_389 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int2_390 = torch.constant.int 2 | |
| %int2_391 = torch.constant.int 2 | |
| %1842 = torch.prim.ListConstruct %int2_390, %int2_391 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int1_392 = torch.constant.int 1 | |
| %int1_393 = torch.constant.int 1 | |
| %1843 = torch.prim.ListConstruct %int1_392, %int1_393 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %int1_394 = torch.constant.int 1 | |
| %int1_395 = torch.constant.int 1 | |
| %1844 = torch.prim.ListConstruct %int1_394, %int1_395 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_396 = torch.constant.bool false | |
| %1845 = torch.aten.max_pool2d %1840, %1841, %1843, %1842, %1844, %false_396 : !torch.vtensor<[1,128,20,20],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[1,128,20,20],f32> | |
| %1846 = torch.prim.ListConstruct %1806, %1823, %1840, %1845 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>) -> !torch.list<vtensor> | |
| %int1_397 = torch.constant.int 1 | |
| %1847 = torch.aten.cat %1846, %int1_397 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,512,20,20],f32> | |
| %1848 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1849 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_398 = torch.constant.int 12 | |
| %1850 = torch.aten.item %1848 : !torch.vtensor<[],f32> -> !torch.float | |
| %1851 = torch.aten.item %1849 : !torch.vtensor<[],si8> -> !torch.int | |
| %1852 = torch.aten.quantize_per_tensor %1847, %1850, %1851, %int12_398 : !torch.vtensor<[1,512,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,512,20,20],!torch.qint8> | |
| %1853 = torch.aten.int_repr %1852 : !torch.vtensor<[1,512,20,20],!torch.qint8> -> !torch.vtensor<[1,512,20,20],si8> | |
| %1854 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1855 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1856 = torch.aten.item %1854 : !torch.vtensor<[],f32> -> !torch.float | |
| %1857 = torch.aten.item %1855 : !torch.vtensor<[],si8> -> !torch.int | |
| %1858 = torch.aten._make_per_tensor_quantized_tensor %1853, %1856, %1857 : !torch.vtensor<[1,512,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,512,20,20],!torch.qint8> | |
| %1859 = torch.aten.dequantize.self %1858 : !torch.vtensor<[1,512,20,20],!torch.qint8> -> !torch.vtensor<[1,512,20,20],f32> | |
| %1860 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1861 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_399 = torch.constant.int 12 | |
| %1862 = torch.aten.item %1860 : !torch.vtensor<[],f32> -> !torch.float | |
| %1863 = torch.aten.item %1861 : !torch.vtensor<[],si8> -> !torch.int | |
| %1864 = torch.aten.quantize_per_tensor %52, %1862, %1863, %int12_399 : !torch.vtensor<[256,512,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,512,1,1],!torch.qint8> | |
| %1865 = torch.aten.int_repr %1864 : !torch.vtensor<[256,512,1,1],!torch.qint8> -> !torch.vtensor<[256,512,1,1],si8> | |
| %1866 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1867 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1868 = torch.aten.item %1866 : !torch.vtensor<[],f32> -> !torch.float | |
| %1869 = torch.aten.item %1867 : !torch.vtensor<[],si8> -> !torch.int | |
| %1870 = torch.aten._make_per_tensor_quantized_tensor %1865, %1868, %1869 : !torch.vtensor<[256,512,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,512,1,1],!torch.qint8> | |
| %1871 = torch.aten.dequantize.self %1870 : !torch.vtensor<[256,512,1,1],!torch.qint8> -> !torch.vtensor<[256,512,1,1],f32> | |
| %1872 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1873 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_400 = torch.constant.int 12 | |
| %1874 = torch.aten.item %1872 : !torch.vtensor<[],f32> -> !torch.float | |
| %1875 = torch.aten.item %1873 : !torch.vtensor<[],si8> -> !torch.int | |
| %1876 = torch.aten.quantize_per_tensor %53, %1874, %1875, %int12_400 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1877 = torch.aten.int_repr %1876 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %1878 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1879 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1880 = torch.aten.item %1878 : !torch.vtensor<[],f32> -> !torch.float | |
| %1881 = torch.aten.item %1879 : !torch.vtensor<[],si8> -> !torch.int | |
| %1882 = torch.aten._make_per_tensor_quantized_tensor %1877, %1880, %1881 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %1883 = torch.aten.dequantize.self %1882 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %int0_401 = torch.constant.int 0 | |
| %int0_402 = torch.constant.int 0 | |
| %int1_403 = torch.constant.int 1 | |
| %int1_404 = torch.constant.int 1 | |
| %int1_405 = torch.constant.int 1 | |
| %int1_406 = torch.constant.int 1 | |
| %int0_407 = torch.constant.int 0 | |
| %1884 = torch.prim.ListConstruct %int0_401, %int0_402 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1885 = torch.prim.ListConstruct %int1_403, %int1_404 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1886 = torch.prim.ListConstruct %int1_405, %int1_406 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1887 = torch.prim.ListConstruct %int0_407, %int0_407 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_408 = torch.constant.bool false | |
| %int1_409 = torch.constant.int 1 | |
| %1888 = torch.aten.convolution %1859, %1871, %1883, %1886, %1884, %1885, %false_408, %1887, %int1_409 : !torch.vtensor<[1,512,20,20],f32>, !torch.vtensor<[256,512,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
| %1889 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1890 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_410 = torch.constant.int 12 | |
| %1891 = torch.aten.item %1889 : !torch.vtensor<[],f32> -> !torch.float | |
| %1892 = torch.aten.item %1890 : !torch.vtensor<[],si8> -> !torch.int | |
| %1893 = torch.aten.quantize_per_tensor %1888, %1891, %1892, %int12_410 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1894 = torch.aten.int_repr %1893 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1895 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1896 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1897 = torch.aten.item %1895 : !torch.vtensor<[],f32> -> !torch.float | |
| %1898 = torch.aten.item %1896 : !torch.vtensor<[],si8> -> !torch.int | |
| %1899 = torch.aten._make_per_tensor_quantized_tensor %1894, %1897, %1898 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1900 = torch.aten.dequantize.self %1899 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1901 = torch.aten.sigmoid %1900 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1902 = torch.aten.mul.Tensor %1900, %1901 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1903 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1904 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_411 = torch.constant.int 12 | |
| %1905 = torch.aten.item %1903 : !torch.vtensor<[],f32> -> !torch.float | |
| %1906 = torch.aten.item %1904 : !torch.vtensor<[],si8> -> !torch.int | |
| %1907 = torch.aten.quantize_per_tensor %1902, %1905, %1906, %int12_411 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1908 = torch.aten.int_repr %1907 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %1909 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1910 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1911 = torch.aten.item %1909 : !torch.vtensor<[],f32> -> !torch.float | |
| %1912 = torch.aten.item %1910 : !torch.vtensor<[],si8> -> !torch.int | |
| %1913 = torch.aten._make_per_tensor_quantized_tensor %1908, %1911, %1912 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %1914 = torch.aten.dequantize.self %1913 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %1915 = torch.vtensor.literal(dense<1.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
| %1916 = torch.vtensor.literal(dense<2.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
| %1917 = torch.prim.ListConstruct %1915, %1916 : (!torch.vtensor<[2],f32>, !torch.vtensor<[2],f32>) -> !torch.list<vtensor> | |
| %int0_412 = torch.constant.int 0 | |
| %1918 = torch.aten.cat %1917, %int0_412 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],f32> | |
| %1919 = torch.operator "onnx.Resize"(%1914, %none, %1918) {torch.onnx.coordinate_transformation_mode = "asymmetric", torch.onnx.cubic_coeff_a = -7.500000e-01 : f32, torch.onnx.mode = "nearest", torch.onnx.nearest_mode = "floor"} : (!torch.vtensor<[1,256,20,20],f32>, !torch.none, !torch.vtensor<[4],f32>) -> !torch.vtensor<[?,?,?,?],f32> | |
| %1920 = torch.prim.ListConstruct %1919, %1433 : (!torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[1,128,40,40],f32>) -> !torch.list<vtensor> | |
| %int1_413 = torch.constant.int 1 | |
| %1921 = torch.aten.cat %1920, %int1_413 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,?,40,40],f32> | |
| %1922 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1923 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_414 = torch.constant.int 12 | |
| %1924 = torch.aten.item %1922 : !torch.vtensor<[],f32> -> !torch.float | |
| %1925 = torch.aten.item %1923 : !torch.vtensor<[],si8> -> !torch.int | |
| %1926 = torch.aten.quantize_per_tensor %1921, %1924, %1925, %int12_414 : !torch.vtensor<[1,?,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,?,40,40],!torch.qint8> | |
| %1927 = torch.aten.int_repr %1926 : !torch.vtensor<[1,?,40,40],!torch.qint8> -> !torch.vtensor<[1,?,40,40],si8> | |
| %1928 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1929 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1930 = torch.aten.item %1928 : !torch.vtensor<[],f32> -> !torch.float | |
| %1931 = torch.aten.item %1929 : !torch.vtensor<[],si8> -> !torch.int | |
| %1932 = torch.aten._make_per_tensor_quantized_tensor %1927, %1930, %1931 : !torch.vtensor<[1,?,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,?,40,40],!torch.qint8> | |
| %1933 = torch.aten.dequantize.self %1932 : !torch.vtensor<[1,?,40,40],!torch.qint8> -> !torch.vtensor<[1,?,40,40],f32> | |
| %1934 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1935 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_415 = torch.constant.int 12 | |
| %1936 = torch.aten.item %1934 : !torch.vtensor<[],f32> -> !torch.float | |
| %1937 = torch.aten.item %1935 : !torch.vtensor<[],si8> -> !torch.int | |
| %1938 = torch.aten.quantize_per_tensor %54, %1936, %1937, %int12_415 : !torch.vtensor<[128,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,384,1,1],!torch.qint8> | |
| %1939 = torch.aten.int_repr %1938 : !torch.vtensor<[128,384,1,1],!torch.qint8> -> !torch.vtensor<[128,384,1,1],si8> | |
| %1940 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1941 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1942 = torch.aten.item %1940 : !torch.vtensor<[],f32> -> !torch.float | |
| %1943 = torch.aten.item %1941 : !torch.vtensor<[],si8> -> !torch.int | |
| %1944 = torch.aten._make_per_tensor_quantized_tensor %1939, %1942, %1943 : !torch.vtensor<[128,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,384,1,1],!torch.qint8> | |
| %1945 = torch.aten.dequantize.self %1944 : !torch.vtensor<[128,384,1,1],!torch.qint8> -> !torch.vtensor<[128,384,1,1],f32> | |
| %1946 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1947 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_416 = torch.constant.int 12 | |
| %1948 = torch.aten.item %1946 : !torch.vtensor<[],f32> -> !torch.float | |
| %1949 = torch.aten.item %1947 : !torch.vtensor<[],si8> -> !torch.int | |
| %1950 = torch.aten.quantize_per_tensor %55, %1948, %1949, %int12_416 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1951 = torch.aten.int_repr %1950 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %1952 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1953 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1954 = torch.aten.item %1952 : !torch.vtensor<[],f32> -> !torch.float | |
| %1955 = torch.aten.item %1953 : !torch.vtensor<[],si8> -> !torch.int | |
| %1956 = torch.aten._make_per_tensor_quantized_tensor %1951, %1954, %1955 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %1957 = torch.aten.dequantize.self %1956 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_417 = torch.constant.int 0 | |
| %int0_418 = torch.constant.int 0 | |
| %int1_419 = torch.constant.int 1 | |
| %int1_420 = torch.constant.int 1 | |
| %int1_421 = torch.constant.int 1 | |
| %int1_422 = torch.constant.int 1 | |
| %int0_423 = torch.constant.int 0 | |
| %1958 = torch.prim.ListConstruct %int0_417, %int0_418 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1959 = torch.prim.ListConstruct %int1_419, %int1_420 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1960 = torch.prim.ListConstruct %int1_421, %int1_422 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1961 = torch.prim.ListConstruct %int0_423, %int0_423 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_424 = torch.constant.bool false | |
| %int1_425 = torch.constant.int 1 | |
| %1962 = torch.aten.convolution %1933, %1945, %1957, %1960, %1958, %1959, %false_424, %1961, %int1_425 : !torch.vtensor<[1,?,40,40],f32>, !torch.vtensor<[128,384,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %1963 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1964 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_426 = torch.constant.int 12 | |
| %1965 = torch.aten.item %1963 : !torch.vtensor<[],f32> -> !torch.float | |
| %1966 = torch.aten.item %1964 : !torch.vtensor<[],si8> -> !torch.int | |
| %1967 = torch.aten.quantize_per_tensor %1962, %1965, %1966, %int12_426 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1968 = torch.aten.int_repr %1967 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1969 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1970 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1971 = torch.aten.item %1969 : !torch.vtensor<[],f32> -> !torch.float | |
| %1972 = torch.aten.item %1970 : !torch.vtensor<[],si8> -> !torch.int | |
| %1973 = torch.aten._make_per_tensor_quantized_tensor %1968, %1971, %1972 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1974 = torch.aten.dequantize.self %1973 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1975 = torch.aten.sigmoid %1974 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1976 = torch.aten.mul.Tensor %1974, %1975 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1977 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1978 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_427 = torch.constant.int 12 | |
| %1979 = torch.aten.item %1977 : !torch.vtensor<[],f32> -> !torch.float | |
| %1980 = torch.aten.item %1978 : !torch.vtensor<[],si8> -> !torch.int | |
| %1981 = torch.aten.quantize_per_tensor %1976, %1979, %1980, %int12_427 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1982 = torch.aten.int_repr %1981 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %1983 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1984 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1985 = torch.aten.item %1983 : !torch.vtensor<[],f32> -> !torch.float | |
| %1986 = torch.aten.item %1984 : !torch.vtensor<[],si8> -> !torch.int | |
| %1987 = torch.aten._make_per_tensor_quantized_tensor %1982, %1985, %1986 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %1988 = torch.aten.dequantize.self %1987 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %1989 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1990 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1991 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %1992 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_428 = torch.constant.int 0 | |
| %int0_429 = torch.constant.int 0 | |
| %1993 = torch.prim.NumToTensor.Scalar %int0_429 : !torch.int -> !torch.vtensor<[1],si64> | |
| %1994 = torch.aten.index_select %1990, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1995 = torch.aten.item %1994 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1996 = torch.aten.index_select %1991, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1997 = torch.aten.item %1996 : !torch.vtensor<[1],si64> -> !torch.int | |
| %1998 = torch.aten.index_select %1989, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %1999 = torch.aten.item %1998 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2000 = torch.aten.index_select %1992, %int0_428, %1993 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2001 = torch.aten.item %2000 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2002 = torch.aten.slice.Tensor %1988, %1999, %1995, %1997, %2001 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2003 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2004 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2005 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2006 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_430 = torch.constant.int 0 | |
| %int0_431 = torch.constant.int 0 | |
| %2007 = torch.prim.NumToTensor.Scalar %int0_431 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2008 = torch.aten.index_select %2004, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2009 = torch.aten.item %2008 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2010 = torch.aten.index_select %2005, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2011 = torch.aten.item %2010 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2012 = torch.aten.index_select %2003, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2013 = torch.aten.item %2012 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2014 = torch.aten.index_select %2006, %int0_430, %2007 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2015 = torch.aten.item %2014 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2016 = torch.aten.slice.Tensor %1988, %2013, %2009, %2011, %2015 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2017 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2018 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_432 = torch.constant.int 12 | |
| %2019 = torch.aten.item %2017 : !torch.vtensor<[],f32> -> !torch.float | |
| %2020 = torch.aten.item %2018 : !torch.vtensor<[],si8> -> !torch.int | |
| %2021 = torch.aten.quantize_per_tensor %56, %2019, %2020, %int12_432 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2022 = torch.aten.int_repr %2021 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %2023 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2024 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2025 = torch.aten.item %2023 : !torch.vtensor<[],f32> -> !torch.float | |
| %2026 = torch.aten.item %2024 : !torch.vtensor<[],si8> -> !torch.int | |
| %2027 = torch.aten._make_per_tensor_quantized_tensor %2022, %2025, %2026 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2028 = torch.aten.dequantize.self %2027 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %2029 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2030 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_433 = torch.constant.int 12 | |
| %2031 = torch.aten.item %2029 : !torch.vtensor<[],f32> -> !torch.float | |
| %2032 = torch.aten.item %2030 : !torch.vtensor<[],si8> -> !torch.int | |
| %2033 = torch.aten.quantize_per_tensor %57, %2031, %2032, %int12_433 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2034 = torch.aten.int_repr %2033 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2035 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2036 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2037 = torch.aten.item %2035 : !torch.vtensor<[],f32> -> !torch.float | |
| %2038 = torch.aten.item %2036 : !torch.vtensor<[],si8> -> !torch.int | |
| %2039 = torch.aten._make_per_tensor_quantized_tensor %2034, %2037, %2038 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2040 = torch.aten.dequantize.self %2039 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_434 = torch.constant.int 1 | |
| %int1_435 = torch.constant.int 1 | |
| %int1_436 = torch.constant.int 1 | |
| %int1_437 = torch.constant.int 1 | |
| %int1_438 = torch.constant.int 1 | |
| %int1_439 = torch.constant.int 1 | |
| %int0_440 = torch.constant.int 0 | |
| %2041 = torch.prim.ListConstruct %int1_434, %int1_435 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2042 = torch.prim.ListConstruct %int1_436, %int1_437 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2043 = torch.prim.ListConstruct %int1_438, %int1_439 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2044 = torch.prim.ListConstruct %int0_440, %int0_440 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_441 = torch.constant.bool false | |
| %int1_442 = torch.constant.int 1 | |
| %2045 = torch.aten.convolution %2016, %2028, %2040, %2043, %2041, %2042, %false_441, %2044, %int1_442 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2046 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2047 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_443 = torch.constant.int 12 | |
| %2048 = torch.aten.item %2046 : !torch.vtensor<[],f32> -> !torch.float | |
| %2049 = torch.aten.item %2047 : !torch.vtensor<[],si8> -> !torch.int | |
| %2050 = torch.aten.quantize_per_tensor %2045, %2048, %2049, %int12_443 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2051 = torch.aten.int_repr %2050 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2052 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2053 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2054 = torch.aten.item %2052 : !torch.vtensor<[],f32> -> !torch.float | |
| %2055 = torch.aten.item %2053 : !torch.vtensor<[],si8> -> !torch.int | |
| %2056 = torch.aten._make_per_tensor_quantized_tensor %2051, %2054, %2055 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2057 = torch.aten.dequantize.self %2056 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2058 = torch.aten.sigmoid %2057 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2059 = torch.aten.mul.Tensor %2057, %2058 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2060 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2061 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_444 = torch.constant.int 12 | |
| %2062 = torch.aten.item %2060 : !torch.vtensor<[],f32> -> !torch.float | |
| %2063 = torch.aten.item %2061 : !torch.vtensor<[],si8> -> !torch.int | |
| %2064 = torch.aten.quantize_per_tensor %2059, %2062, %2063, %int12_444 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2065 = torch.aten.int_repr %2064 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2066 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2067 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2068 = torch.aten.item %2066 : !torch.vtensor<[],f32> -> !torch.float | |
| %2069 = torch.aten.item %2067 : !torch.vtensor<[],si8> -> !torch.int | |
| %2070 = torch.aten._make_per_tensor_quantized_tensor %2065, %2068, %2069 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2071 = torch.aten.dequantize.self %2070 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2072 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2073 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_445 = torch.constant.int 12 | |
| %2074 = torch.aten.item %2072 : !torch.vtensor<[],f32> -> !torch.float | |
| %2075 = torch.aten.item %2073 : !torch.vtensor<[],si8> -> !torch.int | |
| %2076 = torch.aten.quantize_per_tensor %58, %2074, %2075, %int12_445 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2077 = torch.aten.int_repr %2076 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %2078 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2079 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2080 = torch.aten.item %2078 : !torch.vtensor<[],f32> -> !torch.float | |
| %2081 = torch.aten.item %2079 : !torch.vtensor<[],si8> -> !torch.int | |
| %2082 = torch.aten._make_per_tensor_quantized_tensor %2077, %2080, %2081 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2083 = torch.aten.dequantize.self %2082 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %2084 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2085 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_446 = torch.constant.int 12 | |
| %2086 = torch.aten.item %2084 : !torch.vtensor<[],f32> -> !torch.float | |
| %2087 = torch.aten.item %2085 : !torch.vtensor<[],si8> -> !torch.int | |
| %2088 = torch.aten.quantize_per_tensor %59, %2086, %2087, %int12_446 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2089 = torch.aten.int_repr %2088 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2090 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2091 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2092 = torch.aten.item %2090 : !torch.vtensor<[],f32> -> !torch.float | |
| %2093 = torch.aten.item %2091 : !torch.vtensor<[],si8> -> !torch.int | |
| %2094 = torch.aten._make_per_tensor_quantized_tensor %2089, %2092, %2093 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2095 = torch.aten.dequantize.self %2094 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_447 = torch.constant.int 1 | |
| %int1_448 = torch.constant.int 1 | |
| %int1_449 = torch.constant.int 1 | |
| %int1_450 = torch.constant.int 1 | |
| %int1_451 = torch.constant.int 1 | |
| %int1_452 = torch.constant.int 1 | |
| %int0_453 = torch.constant.int 0 | |
| %2096 = torch.prim.ListConstruct %int1_447, %int1_448 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2097 = torch.prim.ListConstruct %int1_449, %int1_450 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2098 = torch.prim.ListConstruct %int1_451, %int1_452 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2099 = torch.prim.ListConstruct %int0_453, %int0_453 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_454 = torch.constant.bool false | |
| %int1_455 = torch.constant.int 1 | |
| %2100 = torch.aten.convolution %2071, %2083, %2095, %2098, %2096, %2097, %false_454, %2099, %int1_455 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2101 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2102 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_456 = torch.constant.int 12 | |
| %2103 = torch.aten.item %2101 : !torch.vtensor<[],f32> -> !torch.float | |
| %2104 = torch.aten.item %2102 : !torch.vtensor<[],si8> -> !torch.int | |
| %2105 = torch.aten.quantize_per_tensor %2100, %2103, %2104, %int12_456 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2106 = torch.aten.int_repr %2105 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2107 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2108 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2109 = torch.aten.item %2107 : !torch.vtensor<[],f32> -> !torch.float | |
| %2110 = torch.aten.item %2108 : !torch.vtensor<[],si8> -> !torch.int | |
| %2111 = torch.aten._make_per_tensor_quantized_tensor %2106, %2109, %2110 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2112 = torch.aten.dequantize.self %2111 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2113 = torch.aten.sigmoid %2112 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2114 = torch.aten.mul.Tensor %2112, %2113 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2115 = torch.prim.ListConstruct %2002, %2016, %2114 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>) -> !torch.list<vtensor> | |
| %int1_457 = torch.constant.int 1 | |
| %2116 = torch.aten.cat %2115, %int1_457 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,192,40,40],f32> | |
| %2117 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2118 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_458 = torch.constant.int 12 | |
| %2119 = torch.aten.item %2117 : !torch.vtensor<[],f32> -> !torch.float | |
| %2120 = torch.aten.item %2118 : !torch.vtensor<[],si8> -> !torch.int | |
| %2121 = torch.aten.quantize_per_tensor %2116, %2119, %2120, %int12_458 : !torch.vtensor<[1,192,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
| %2122 = torch.aten.int_repr %2121 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],si8> | |
| %2123 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2124 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2125 = torch.aten.item %2123 : !torch.vtensor<[],f32> -> !torch.float | |
| %2126 = torch.aten.item %2124 : !torch.vtensor<[],si8> -> !torch.int | |
| %2127 = torch.aten._make_per_tensor_quantized_tensor %2122, %2125, %2126 : !torch.vtensor<[1,192,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
| %2128 = torch.aten.dequantize.self %2127 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],f32> | |
| %2129 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2130 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_459 = torch.constant.int 12 | |
| %2131 = torch.aten.item %2129 : !torch.vtensor<[],f32> -> !torch.float | |
| %2132 = torch.aten.item %2130 : !torch.vtensor<[],si8> -> !torch.int | |
| %2133 = torch.aten.quantize_per_tensor %60, %2131, %2132, %int12_459 : !torch.vtensor<[128,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
| %2134 = torch.aten.int_repr %2133 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],si8> | |
| %2135 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2136 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2137 = torch.aten.item %2135 : !torch.vtensor<[],f32> -> !torch.float | |
| %2138 = torch.aten.item %2136 : !torch.vtensor<[],si8> -> !torch.int | |
| %2139 = torch.aten._make_per_tensor_quantized_tensor %2134, %2137, %2138 : !torch.vtensor<[128,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
| %2140 = torch.aten.dequantize.self %2139 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],f32> | |
| %2141 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2142 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_460 = torch.constant.int 12 | |
| %2143 = torch.aten.item %2141 : !torch.vtensor<[],f32> -> !torch.float | |
| %2144 = torch.aten.item %2142 : !torch.vtensor<[],si8> -> !torch.int | |
| %2145 = torch.aten.quantize_per_tensor %61, %2143, %2144, %int12_460 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2146 = torch.aten.int_repr %2145 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %2147 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2148 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2149 = torch.aten.item %2147 : !torch.vtensor<[],f32> -> !torch.float | |
| %2150 = torch.aten.item %2148 : !torch.vtensor<[],si8> -> !torch.int | |
| %2151 = torch.aten._make_per_tensor_quantized_tensor %2146, %2149, %2150 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2152 = torch.aten.dequantize.self %2151 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_461 = torch.constant.int 0 | |
| %int0_462 = torch.constant.int 0 | |
| %int1_463 = torch.constant.int 1 | |
| %int1_464 = torch.constant.int 1 | |
| %int1_465 = torch.constant.int 1 | |
| %int1_466 = torch.constant.int 1 | |
| %int0_467 = torch.constant.int 0 | |
| %2153 = torch.prim.ListConstruct %int0_461, %int0_462 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2154 = torch.prim.ListConstruct %int1_463, %int1_464 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2155 = torch.prim.ListConstruct %int1_465, %int1_466 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2156 = torch.prim.ListConstruct %int0_467, %int0_467 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_468 = torch.constant.bool false | |
| %int1_469 = torch.constant.int 1 | |
| %2157 = torch.aten.convolution %2128, %2140, %2152, %2155, %2153, %2154, %false_468, %2156, %int1_469 : !torch.vtensor<[1,192,40,40],f32>, !torch.vtensor<[128,192,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %2158 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2159 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_470 = torch.constant.int 12 | |
| %2160 = torch.aten.item %2158 : !torch.vtensor<[],f32> -> !torch.float | |
| %2161 = torch.aten.item %2159 : !torch.vtensor<[],si8> -> !torch.int | |
| %2162 = torch.aten.quantize_per_tensor %2157, %2160, %2161, %int12_470 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2163 = torch.aten.int_repr %2162 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %2164 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2165 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2166 = torch.aten.item %2164 : !torch.vtensor<[],f32> -> !torch.float | |
| %2167 = torch.aten.item %2165 : !torch.vtensor<[],si8> -> !torch.int | |
| %2168 = torch.aten._make_per_tensor_quantized_tensor %2163, %2166, %2167 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2169 = torch.aten.dequantize.self %2168 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2170 = torch.aten.sigmoid %2169 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2171 = torch.aten.mul.Tensor %2169, %2170 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2172 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_471 = torch.constant.int 12 | |
| %2174 = torch.aten.item %2172 : !torch.vtensor<[],f32> -> !torch.float | |
| %2175 = torch.aten.item %2173 : !torch.vtensor<[],si8> -> !torch.int | |
| %2176 = torch.aten.quantize_per_tensor %2171, %2174, %2175, %int12_471 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2177 = torch.aten.int_repr %2176 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %2178 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2180 = torch.aten.item %2178 : !torch.vtensor<[],f32> -> !torch.float | |
| %2181 = torch.aten.item %2179 : !torch.vtensor<[],si8> -> !torch.int | |
| %2182 = torch.aten._make_per_tensor_quantized_tensor %2177, %2180, %2181 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2183 = torch.aten.dequantize.self %2182 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2184 = torch.vtensor.literal(dense<1.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
| %2185 = torch.vtensor.literal(dense<2.000000e+00> : tensor<2xf32>) : !torch.vtensor<[2],f32> | |
| %2186 = torch.prim.ListConstruct %2184, %2185 : (!torch.vtensor<[2],f32>, !torch.vtensor<[2],f32>) -> !torch.list<vtensor> | |
| %int0_472 = torch.constant.int 0 | |
| %2187 = torch.aten.cat %2186, %int0_472 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],f32> | |
| %2188 = torch.operator "onnx.Resize"(%2183, %none, %2187) {torch.onnx.coordinate_transformation_mode = "asymmetric", torch.onnx.cubic_coeff_a = -7.500000e-01 : f32, torch.onnx.mode = "nearest", torch.onnx.nearest_mode = "floor"} : (!torch.vtensor<[1,128,40,40],f32>, !torch.none, !torch.vtensor<[4],f32>) -> !torch.vtensor<[?,?,?,?],f32> | |
| %2189 = torch.prim.ListConstruct %2188, %992 : (!torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[1,64,80,80],f32>) -> !torch.list<vtensor> | |
| %int1_473 = torch.constant.int 1 | |
| %2190 = torch.aten.cat %2189, %int1_473 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,?,80,80],f32> | |
| %2191 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2192 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_474 = torch.constant.int 12 | |
| %2193 = torch.aten.item %2191 : !torch.vtensor<[],f32> -> !torch.float | |
| %2194 = torch.aten.item %2192 : !torch.vtensor<[],si8> -> !torch.int | |
| %2195 = torch.aten.quantize_per_tensor %2190, %2193, %2194, %int12_474 : !torch.vtensor<[1,?,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,?,80,80],!torch.qint8> | |
| %2196 = torch.aten.int_repr %2195 : !torch.vtensor<[1,?,80,80],!torch.qint8> -> !torch.vtensor<[1,?,80,80],si8> | |
| %2197 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2198 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2199 = torch.aten.item %2197 : !torch.vtensor<[],f32> -> !torch.float | |
| %2200 = torch.aten.item %2198 : !torch.vtensor<[],si8> -> !torch.int | |
| %2201 = torch.aten._make_per_tensor_quantized_tensor %2196, %2199, %2200 : !torch.vtensor<[1,?,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,?,80,80],!torch.qint8> | |
| %2202 = torch.aten.dequantize.self %2201 : !torch.vtensor<[1,?,80,80],!torch.qint8> -> !torch.vtensor<[1,?,80,80],f32> | |
| %2203 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2204 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_475 = torch.constant.int 12 | |
| %2205 = torch.aten.item %2203 : !torch.vtensor<[],f32> -> !torch.float | |
| %2206 = torch.aten.item %2204 : !torch.vtensor<[],si8> -> !torch.int | |
| %2207 = torch.aten.quantize_per_tensor %62, %2205, %2206, %int12_475 : !torch.vtensor<[64,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,192,1,1],!torch.qint8> | |
| %2208 = torch.aten.int_repr %2207 : !torch.vtensor<[64,192,1,1],!torch.qint8> -> !torch.vtensor<[64,192,1,1],si8> | |
| %2209 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2210 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2211 = torch.aten.item %2209 : !torch.vtensor<[],f32> -> !torch.float | |
| %2212 = torch.aten.item %2210 : !torch.vtensor<[],si8> -> !torch.int | |
| %2213 = torch.aten._make_per_tensor_quantized_tensor %2208, %2211, %2212 : !torch.vtensor<[64,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,192,1,1],!torch.qint8> | |
| %2214 = torch.aten.dequantize.self %2213 : !torch.vtensor<[64,192,1,1],!torch.qint8> -> !torch.vtensor<[64,192,1,1],f32> | |
| %2215 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2216 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_476 = torch.constant.int 12 | |
| %2217 = torch.aten.item %2215 : !torch.vtensor<[],f32> -> !torch.float | |
| %2218 = torch.aten.item %2216 : !torch.vtensor<[],si8> -> !torch.int | |
| %2219 = torch.aten.quantize_per_tensor %63, %2217, %2218, %int12_476 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2220 = torch.aten.int_repr %2219 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2221 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2223 = torch.aten.item %2221 : !torch.vtensor<[],f32> -> !torch.float | |
| %2224 = torch.aten.item %2222 : !torch.vtensor<[],si8> -> !torch.int | |
| %2225 = torch.aten._make_per_tensor_quantized_tensor %2220, %2223, %2224 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2226 = torch.aten.dequantize.self %2225 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_477 = torch.constant.int 0 | |
| %int0_478 = torch.constant.int 0 | |
| %int1_479 = torch.constant.int 1 | |
| %int1_480 = torch.constant.int 1 | |
| %int1_481 = torch.constant.int 1 | |
| %int1_482 = torch.constant.int 1 | |
| %int0_483 = torch.constant.int 0 | |
| %2227 = torch.prim.ListConstruct %int0_477, %int0_478 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2228 = torch.prim.ListConstruct %int1_479, %int1_480 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2229 = torch.prim.ListConstruct %int1_481, %int1_482 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2230 = torch.prim.ListConstruct %int0_483, %int0_483 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_484 = torch.constant.bool false | |
| %int1_485 = torch.constant.int 1 | |
| %2231 = torch.aten.convolution %2202, %2214, %2226, %2229, %2227, %2228, %false_484, %2230, %int1_485 : !torch.vtensor<[1,?,80,80],f32>, !torch.vtensor<[64,192,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %2232 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2233 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_486 = torch.constant.int 12 | |
| %2234 = torch.aten.item %2232 : !torch.vtensor<[],f32> -> !torch.float | |
| %2235 = torch.aten.item %2233 : !torch.vtensor<[],si8> -> !torch.int | |
| %2236 = torch.aten.quantize_per_tensor %2231, %2234, %2235, %int12_486 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2237 = torch.aten.int_repr %2236 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %2238 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2239 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2240 = torch.aten.item %2238 : !torch.vtensor<[],f32> -> !torch.float | |
| %2241 = torch.aten.item %2239 : !torch.vtensor<[],si8> -> !torch.int | |
| %2242 = torch.aten._make_per_tensor_quantized_tensor %2237, %2240, %2241 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2243 = torch.aten.dequantize.self %2242 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2244 = torch.aten.sigmoid %2243 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2245 = torch.aten.mul.Tensor %2243, %2244 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2246 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2247 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_487 = torch.constant.int 12 | |
| %2248 = torch.aten.item %2246 : !torch.vtensor<[],f32> -> !torch.float | |
| %2249 = torch.aten.item %2247 : !torch.vtensor<[],si8> -> !torch.int | |
| %2250 = torch.aten.quantize_per_tensor %2245, %2248, %2249, %int12_487 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2251 = torch.aten.int_repr %2250 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %2252 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2253 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2254 = torch.aten.item %2252 : !torch.vtensor<[],f32> -> !torch.float | |
| %2255 = torch.aten.item %2253 : !torch.vtensor<[],si8> -> !torch.int | |
| %2256 = torch.aten._make_per_tensor_quantized_tensor %2251, %2254, %2255 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2257 = torch.aten.dequantize.self %2256 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2258 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2259 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2260 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2261 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_488 = torch.constant.int 0 | |
| %int0_489 = torch.constant.int 0 | |
| %2262 = torch.prim.NumToTensor.Scalar %int0_489 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2263 = torch.aten.index_select %2259, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2264 = torch.aten.item %2263 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2265 = torch.aten.index_select %2260, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2266 = torch.aten.item %2265 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2267 = torch.aten.index_select %2258, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2268 = torch.aten.item %2267 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2269 = torch.aten.index_select %2261, %int0_488, %2262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2270 = torch.aten.item %2269 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2271 = torch.aten.slice.Tensor %2257, %2268, %2264, %2266, %2270 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %2272 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2273 = torch.vtensor.literal(dense<32> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2274 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2275 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_490 = torch.constant.int 0 | |
| %int0_491 = torch.constant.int 0 | |
| %2276 = torch.prim.NumToTensor.Scalar %int0_491 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2277 = torch.aten.index_select %2273, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2278 = torch.aten.item %2277 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2279 = torch.aten.index_select %2274, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2280 = torch.aten.item %2279 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2281 = torch.aten.index_select %2272, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2282 = torch.aten.item %2281 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2283 = torch.aten.index_select %2275, %int0_490, %2276 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2284 = torch.aten.item %2283 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2285 = torch.aten.slice.Tensor %2257, %2282, %2278, %2280, %2284 : !torch.vtensor<[1,64,80,80],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %2286 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2287 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_492 = torch.constant.int 12 | |
| %2288 = torch.aten.item %2286 : !torch.vtensor<[],f32> -> !torch.float | |
| %2289 = torch.aten.item %2287 : !torch.vtensor<[],si8> -> !torch.int | |
| %2290 = torch.aten.quantize_per_tensor %64, %2288, %2289, %int12_492 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %2291 = torch.aten.int_repr %2290 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %2292 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2293 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2294 = torch.aten.item %2292 : !torch.vtensor<[],f32> -> !torch.float | |
| %2295 = torch.aten.item %2293 : !torch.vtensor<[],si8> -> !torch.int | |
| %2296 = torch.aten._make_per_tensor_quantized_tensor %2291, %2294, %2295 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %2297 = torch.aten.dequantize.self %2296 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %2298 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2299 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_493 = torch.constant.int 12 | |
| %2300 = torch.aten.item %2298 : !torch.vtensor<[],f32> -> !torch.float | |
| %2301 = torch.aten.item %2299 : !torch.vtensor<[],si8> -> !torch.int | |
| %2302 = torch.aten.quantize_per_tensor %65, %2300, %2301, %int12_493 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %2303 = torch.aten.int_repr %2302 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %2304 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2305 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2306 = torch.aten.item %2304 : !torch.vtensor<[],f32> -> !torch.float | |
| %2307 = torch.aten.item %2305 : !torch.vtensor<[],si8> -> !torch.int | |
| %2308 = torch.aten._make_per_tensor_quantized_tensor %2303, %2306, %2307 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %2309 = torch.aten.dequantize.self %2308 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_494 = torch.constant.int 1 | |
| %int1_495 = torch.constant.int 1 | |
| %int1_496 = torch.constant.int 1 | |
| %int1_497 = torch.constant.int 1 | |
| %int1_498 = torch.constant.int 1 | |
| %int1_499 = torch.constant.int 1 | |
| %int0_500 = torch.constant.int 0 | |
| %2310 = torch.prim.ListConstruct %int1_494, %int1_495 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2311 = torch.prim.ListConstruct %int1_496, %int1_497 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2312 = torch.prim.ListConstruct %int1_498, %int1_499 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2313 = torch.prim.ListConstruct %int0_500, %int0_500 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_501 = torch.constant.bool false | |
| %int1_502 = torch.constant.int 1 | |
| %2314 = torch.aten.convolution %2285, %2297, %2309, %2312, %2310, %2311, %false_501, %2313, %int1_502 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %2315 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2316 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_503 = torch.constant.int 12 | |
| %2317 = torch.aten.item %2315 : !torch.vtensor<[],f32> -> !torch.float | |
| %2318 = torch.aten.item %2316 : !torch.vtensor<[],si8> -> !torch.int | |
| %2319 = torch.aten.quantize_per_tensor %2314, %2317, %2318, %int12_503 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %2320 = torch.aten.int_repr %2319 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %2321 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2322 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2323 = torch.aten.item %2321 : !torch.vtensor<[],f32> -> !torch.float | |
| %2324 = torch.aten.item %2322 : !torch.vtensor<[],si8> -> !torch.int | |
| %2325 = torch.aten._make_per_tensor_quantized_tensor %2320, %2323, %2324 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %2326 = torch.aten.dequantize.self %2325 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2327 = torch.aten.sigmoid %2326 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2328 = torch.aten.mul.Tensor %2326, %2327 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2329 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2330 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_504 = torch.constant.int 12 | |
| %2331 = torch.aten.item %2329 : !torch.vtensor<[],f32> -> !torch.float | |
| %2332 = torch.aten.item %2330 : !torch.vtensor<[],si8> -> !torch.int | |
| %2333 = torch.aten.quantize_per_tensor %2328, %2331, %2332, %int12_504 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %2334 = torch.aten.int_repr %2333 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %2335 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2336 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2337 = torch.aten.item %2335 : !torch.vtensor<[],f32> -> !torch.float | |
| %2338 = torch.aten.item %2336 : !torch.vtensor<[],si8> -> !torch.int | |
| %2339 = torch.aten._make_per_tensor_quantized_tensor %2334, %2337, %2338 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %2340 = torch.aten.dequantize.self %2339 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2341 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2342 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_505 = torch.constant.int 12 | |
| %2343 = torch.aten.item %2341 : !torch.vtensor<[],f32> -> !torch.float | |
| %2344 = torch.aten.item %2342 : !torch.vtensor<[],si8> -> !torch.int | |
| %2345 = torch.aten.quantize_per_tensor %66, %2343, %2344, %int12_505 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %2346 = torch.aten.int_repr %2345 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %2347 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2348 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2349 = torch.aten.item %2347 : !torch.vtensor<[],f32> -> !torch.float | |
| %2350 = torch.aten.item %2348 : !torch.vtensor<[],si8> -> !torch.int | |
| %2351 = torch.aten._make_per_tensor_quantized_tensor %2346, %2349, %2350 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %2352 = torch.aten.dequantize.self %2351 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %2353 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2354 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_506 = torch.constant.int 12 | |
| %2355 = torch.aten.item %2353 : !torch.vtensor<[],f32> -> !torch.float | |
| %2356 = torch.aten.item %2354 : !torch.vtensor<[],si8> -> !torch.int | |
| %2357 = torch.aten.quantize_per_tensor %67, %2355, %2356, %int12_506 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %2358 = torch.aten.int_repr %2357 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %2359 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2360 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2361 = torch.aten.item %2359 : !torch.vtensor<[],f32> -> !torch.float | |
| %2362 = torch.aten.item %2360 : !torch.vtensor<[],si8> -> !torch.int | |
| %2363 = torch.aten._make_per_tensor_quantized_tensor %2358, %2361, %2362 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %2364 = torch.aten.dequantize.self %2363 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int1_507 = torch.constant.int 1 | |
| %int1_508 = torch.constant.int 1 | |
| %int1_509 = torch.constant.int 1 | |
| %int1_510 = torch.constant.int 1 | |
| %int1_511 = torch.constant.int 1 | |
| %int1_512 = torch.constant.int 1 | |
| %int0_513 = torch.constant.int 0 | |
| %2365 = torch.prim.ListConstruct %int1_507, %int1_508 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2366 = torch.prim.ListConstruct %int1_509, %int1_510 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2367 = torch.prim.ListConstruct %int1_511, %int1_512 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2368 = torch.prim.ListConstruct %int0_513, %int0_513 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_514 = torch.constant.bool false | |
| %int1_515 = torch.constant.int 1 | |
| %2369 = torch.aten.convolution %2340, %2352, %2364, %2367, %2365, %2366, %false_514, %2368, %int1_515 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,80,80],f32> | |
| %2370 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2371 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_516 = torch.constant.int 12 | |
| %2372 = torch.aten.item %2370 : !torch.vtensor<[],f32> -> !torch.float | |
| %2373 = torch.aten.item %2371 : !torch.vtensor<[],si8> -> !torch.int | |
| %2374 = torch.aten.quantize_per_tensor %2369, %2372, %2373, %int12_516 : !torch.vtensor<[1,32,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %2375 = torch.aten.int_repr %2374 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],si8> | |
| %2376 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2377 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2378 = torch.aten.item %2376 : !torch.vtensor<[],f32> -> !torch.float | |
| %2379 = torch.aten.item %2377 : !torch.vtensor<[],si8> -> !torch.int | |
| %2380 = torch.aten._make_per_tensor_quantized_tensor %2375, %2378, %2379 : !torch.vtensor<[1,32,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,80,80],!torch.qint8> | |
| %2381 = torch.aten.dequantize.self %2380 : !torch.vtensor<[1,32,80,80],!torch.qint8> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2382 = torch.aten.sigmoid %2381 : !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2383 = torch.aten.mul.Tensor %2381, %2382 : !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32> -> !torch.vtensor<[1,32,80,80],f32> | |
| %2384 = torch.prim.ListConstruct %2271, %2285, %2383 : (!torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>, !torch.vtensor<[1,32,80,80],f32>) -> !torch.list<vtensor> | |
| %int1_517 = torch.constant.int 1 | |
| %2385 = torch.aten.cat %2384, %int1_517 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,96,80,80],f32> | |
| %2386 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2387 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_518 = torch.constant.int 12 | |
| %2388 = torch.aten.item %2386 : !torch.vtensor<[],f32> -> !torch.float | |
| %2389 = torch.aten.item %2387 : !torch.vtensor<[],si8> -> !torch.int | |
| %2390 = torch.aten.quantize_per_tensor %2385, %2388, %2389, %int12_518 : !torch.vtensor<[1,96,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,96,80,80],!torch.qint8> | |
| %2391 = torch.aten.int_repr %2390 : !torch.vtensor<[1,96,80,80],!torch.qint8> -> !torch.vtensor<[1,96,80,80],si8> | |
| %2392 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2393 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2394 = torch.aten.item %2392 : !torch.vtensor<[],f32> -> !torch.float | |
| %2395 = torch.aten.item %2393 : !torch.vtensor<[],si8> -> !torch.int | |
| %2396 = torch.aten._make_per_tensor_quantized_tensor %2391, %2394, %2395 : !torch.vtensor<[1,96,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,96,80,80],!torch.qint8> | |
| %2397 = torch.aten.dequantize.self %2396 : !torch.vtensor<[1,96,80,80],!torch.qint8> -> !torch.vtensor<[1,96,80,80],f32> | |
| %2398 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2399 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_519 = torch.constant.int 12 | |
| %2400 = torch.aten.item %2398 : !torch.vtensor<[],f32> -> !torch.float | |
| %2401 = torch.aten.item %2399 : !torch.vtensor<[],si8> -> !torch.int | |
| %2402 = torch.aten.quantize_per_tensor %68, %2400, %2401, %int12_519 : !torch.vtensor<[64,96,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,96,1,1],!torch.qint8> | |
| %2403 = torch.aten.int_repr %2402 : !torch.vtensor<[64,96,1,1],!torch.qint8> -> !torch.vtensor<[64,96,1,1],si8> | |
| %2404 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2405 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2406 = torch.aten.item %2404 : !torch.vtensor<[],f32> -> !torch.float | |
| %2407 = torch.aten.item %2405 : !torch.vtensor<[],si8> -> !torch.int | |
| %2408 = torch.aten._make_per_tensor_quantized_tensor %2403, %2406, %2407 : !torch.vtensor<[64,96,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,96,1,1],!torch.qint8> | |
| %2409 = torch.aten.dequantize.self %2408 : !torch.vtensor<[64,96,1,1],!torch.qint8> -> !torch.vtensor<[64,96,1,1],f32> | |
| %2410 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2411 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_520 = torch.constant.int 12 | |
| %2412 = torch.aten.item %2410 : !torch.vtensor<[],f32> -> !torch.float | |
| %2413 = torch.aten.item %2411 : !torch.vtensor<[],si8> -> !torch.int | |
| %2414 = torch.aten.quantize_per_tensor %69, %2412, %2413, %int12_520 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2415 = torch.aten.int_repr %2414 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2416 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2417 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2418 = torch.aten.item %2416 : !torch.vtensor<[],f32> -> !torch.float | |
| %2419 = torch.aten.item %2417 : !torch.vtensor<[],si8> -> !torch.int | |
| %2420 = torch.aten._make_per_tensor_quantized_tensor %2415, %2418, %2419 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2421 = torch.aten.dequantize.self %2420 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_521 = torch.constant.int 0 | |
| %int0_522 = torch.constant.int 0 | |
| %int1_523 = torch.constant.int 1 | |
| %int1_524 = torch.constant.int 1 | |
| %int1_525 = torch.constant.int 1 | |
| %int1_526 = torch.constant.int 1 | |
| %int0_527 = torch.constant.int 0 | |
| %2422 = torch.prim.ListConstruct %int0_521, %int0_522 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2423 = torch.prim.ListConstruct %int1_523, %int1_524 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2424 = torch.prim.ListConstruct %int1_525, %int1_526 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2425 = torch.prim.ListConstruct %int0_527, %int0_527 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_528 = torch.constant.bool false | |
| %int1_529 = torch.constant.int 1 | |
| %2426 = torch.aten.convolution %2397, %2409, %2421, %2424, %2422, %2423, %false_528, %2425, %int1_529 : !torch.vtensor<[1,96,80,80],f32>, !torch.vtensor<[64,96,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %2427 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2428 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_530 = torch.constant.int 12 | |
| %2429 = torch.aten.item %2427 : !torch.vtensor<[],f32> -> !torch.float | |
| %2430 = torch.aten.item %2428 : !torch.vtensor<[],si8> -> !torch.int | |
| %2431 = torch.aten.quantize_per_tensor %2426, %2429, %2430, %int12_530 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2432 = torch.aten.int_repr %2431 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %2433 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2434 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2435 = torch.aten.item %2433 : !torch.vtensor<[],f32> -> !torch.float | |
| %2436 = torch.aten.item %2434 : !torch.vtensor<[],si8> -> !torch.int | |
| %2437 = torch.aten._make_per_tensor_quantized_tensor %2432, %2435, %2436 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2438 = torch.aten.dequantize.self %2437 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2439 = torch.aten.sigmoid %2438 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2440 = torch.aten.mul.Tensor %2438, %2439 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2441 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2442 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_531 = torch.constant.int 12 | |
| %2443 = torch.aten.item %2441 : !torch.vtensor<[],f32> -> !torch.float | |
| %2444 = torch.aten.item %2442 : !torch.vtensor<[],si8> -> !torch.int | |
| %2445 = torch.aten.quantize_per_tensor %2440, %2443, %2444, %int12_531 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2446 = torch.aten.int_repr %2445 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %2447 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2448 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2449 = torch.aten.item %2447 : !torch.vtensor<[],f32> -> !torch.float | |
| %2450 = torch.aten.item %2448 : !torch.vtensor<[],si8> -> !torch.int | |
| %2451 = torch.aten._make_per_tensor_quantized_tensor %2446, %2449, %2450 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %2452 = torch.aten.dequantize.self %2451 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %2453 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2454 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_532 = torch.constant.int 12 | |
| %2455 = torch.aten.item %2453 : !torch.vtensor<[],f32> -> !torch.float | |
| %2456 = torch.aten.item %2454 : !torch.vtensor<[],si8> -> !torch.int | |
| %2457 = torch.aten.quantize_per_tensor %70, %2455, %2456, %int12_532 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2458 = torch.aten.int_repr %2457 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %2459 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2460 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2461 = torch.aten.item %2459 : !torch.vtensor<[],f32> -> !torch.float | |
| %2462 = torch.aten.item %2460 : !torch.vtensor<[],si8> -> !torch.int | |
| %2463 = torch.aten._make_per_tensor_quantized_tensor %2458, %2461, %2462 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2464 = torch.aten.dequantize.self %2463 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %2465 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2466 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_533 = torch.constant.int 12 | |
| %2467 = torch.aten.item %2465 : !torch.vtensor<[],f32> -> !torch.float | |
| %2468 = torch.aten.item %2466 : !torch.vtensor<[],si8> -> !torch.int | |
| %2469 = torch.aten.quantize_per_tensor %71, %2467, %2468, %int12_533 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2470 = torch.aten.int_repr %2469 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2471 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2472 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2473 = torch.aten.item %2471 : !torch.vtensor<[],f32> -> !torch.float | |
| %2474 = torch.aten.item %2472 : !torch.vtensor<[],si8> -> !torch.int | |
| %2475 = torch.aten._make_per_tensor_quantized_tensor %2470, %2473, %2474 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2476 = torch.aten.dequantize.self %2475 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_534 = torch.constant.int 1 | |
| %int1_535 = torch.constant.int 1 | |
| %int1_536 = torch.constant.int 1 | |
| %int1_537 = torch.constant.int 1 | |
| %int2_538 = torch.constant.int 2 | |
| %int2_539 = torch.constant.int 2 | |
| %int0_540 = torch.constant.int 0 | |
| %2477 = torch.prim.ListConstruct %int1_534, %int1_535 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2478 = torch.prim.ListConstruct %int1_536, %int1_537 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2479 = torch.prim.ListConstruct %int2_538, %int2_539 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2480 = torch.prim.ListConstruct %int0_540, %int0_540 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_541 = torch.constant.bool false | |
| %int1_542 = torch.constant.int 1 | |
| %2481 = torch.aten.convolution %2452, %2464, %2476, %2479, %2477, %2478, %false_541, %2480, %int1_542 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2482 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2483 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_543 = torch.constant.int 12 | |
| %2484 = torch.aten.item %2482 : !torch.vtensor<[],f32> -> !torch.float | |
| %2485 = torch.aten.item %2483 : !torch.vtensor<[],si8> -> !torch.int | |
| %2486 = torch.aten.quantize_per_tensor %2481, %2484, %2485, %int12_543 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2487 = torch.aten.int_repr %2486 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2488 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2489 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2490 = torch.aten.item %2488 : !torch.vtensor<[],f32> -> !torch.float | |
| %2491 = torch.aten.item %2489 : !torch.vtensor<[],si8> -> !torch.int | |
| %2492 = torch.aten._make_per_tensor_quantized_tensor %2487, %2490, %2491 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2493 = torch.aten.dequantize.self %2492 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2494 = torch.aten.sigmoid %2493 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2495 = torch.aten.mul.Tensor %2493, %2494 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2496 = torch.prim.ListConstruct %2495, %2183 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,128,40,40],f32>) -> !torch.list<vtensor> | |
| %int1_544 = torch.constant.int 1 | |
| %2497 = torch.aten.cat %2496, %int1_544 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,192,40,40],f32> | |
| %2498 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2499 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_545 = torch.constant.int 12 | |
| %2500 = torch.aten.item %2498 : !torch.vtensor<[],f32> -> !torch.float | |
| %2501 = torch.aten.item %2499 : !torch.vtensor<[],si8> -> !torch.int | |
| %2502 = torch.aten.quantize_per_tensor %2497, %2500, %2501, %int12_545 : !torch.vtensor<[1,192,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
| %2503 = torch.aten.int_repr %2502 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],si8> | |
| %2504 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2505 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2506 = torch.aten.item %2504 : !torch.vtensor<[],f32> -> !torch.float | |
| %2507 = torch.aten.item %2505 : !torch.vtensor<[],si8> -> !torch.int | |
| %2508 = torch.aten._make_per_tensor_quantized_tensor %2503, %2506, %2507 : !torch.vtensor<[1,192,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
| %2509 = torch.aten.dequantize.self %2508 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],f32> | |
| %2510 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2511 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_546 = torch.constant.int 12 | |
| %2512 = torch.aten.item %2510 : !torch.vtensor<[],f32> -> !torch.float | |
| %2513 = torch.aten.item %2511 : !torch.vtensor<[],si8> -> !torch.int | |
| %2514 = torch.aten.quantize_per_tensor %72, %2512, %2513, %int12_546 : !torch.vtensor<[128,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
| %2515 = torch.aten.int_repr %2514 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],si8> | |
| %2516 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2517 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2518 = torch.aten.item %2516 : !torch.vtensor<[],f32> -> !torch.float | |
| %2519 = torch.aten.item %2517 : !torch.vtensor<[],si8> -> !torch.int | |
| %2520 = torch.aten._make_per_tensor_quantized_tensor %2515, %2518, %2519 : !torch.vtensor<[128,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
| %2521 = torch.aten.dequantize.self %2520 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],f32> | |
| %2522 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2523 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_547 = torch.constant.int 12 | |
| %2524 = torch.aten.item %2522 : !torch.vtensor<[],f32> -> !torch.float | |
| %2525 = torch.aten.item %2523 : !torch.vtensor<[],si8> -> !torch.int | |
| %2526 = torch.aten.quantize_per_tensor %73, %2524, %2525, %int12_547 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2527 = torch.aten.int_repr %2526 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %2528 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2529 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2530 = torch.aten.item %2528 : !torch.vtensor<[],f32> -> !torch.float | |
| %2531 = torch.aten.item %2529 : !torch.vtensor<[],si8> -> !torch.int | |
| %2532 = torch.aten._make_per_tensor_quantized_tensor %2527, %2530, %2531 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2533 = torch.aten.dequantize.self %2532 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_548 = torch.constant.int 0 | |
| %int0_549 = torch.constant.int 0 | |
| %int1_550 = torch.constant.int 1 | |
| %int1_551 = torch.constant.int 1 | |
| %int1_552 = torch.constant.int 1 | |
| %int1_553 = torch.constant.int 1 | |
| %int0_554 = torch.constant.int 0 | |
| %2534 = torch.prim.ListConstruct %int0_548, %int0_549 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2535 = torch.prim.ListConstruct %int1_550, %int1_551 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2536 = torch.prim.ListConstruct %int1_552, %int1_553 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2537 = torch.prim.ListConstruct %int0_554, %int0_554 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_555 = torch.constant.bool false | |
| %int1_556 = torch.constant.int 1 | |
| %2538 = torch.aten.convolution %2509, %2521, %2533, %2536, %2534, %2535, %false_555, %2537, %int1_556 : !torch.vtensor<[1,192,40,40],f32>, !torch.vtensor<[128,192,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %2539 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2540 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_557 = torch.constant.int 12 | |
| %2541 = torch.aten.item %2539 : !torch.vtensor<[],f32> -> !torch.float | |
| %2542 = torch.aten.item %2540 : !torch.vtensor<[],si8> -> !torch.int | |
| %2543 = torch.aten.quantize_per_tensor %2538, %2541, %2542, %int12_557 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2544 = torch.aten.int_repr %2543 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %2545 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2546 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2547 = torch.aten.item %2545 : !torch.vtensor<[],f32> -> !torch.float | |
| %2548 = torch.aten.item %2546 : !torch.vtensor<[],si8> -> !torch.int | |
| %2549 = torch.aten._make_per_tensor_quantized_tensor %2544, %2547, %2548 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2550 = torch.aten.dequantize.self %2549 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2551 = torch.aten.sigmoid %2550 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2552 = torch.aten.mul.Tensor %2550, %2551 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2553 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2554 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_558 = torch.constant.int 12 | |
| %2555 = torch.aten.item %2553 : !torch.vtensor<[],f32> -> !torch.float | |
| %2556 = torch.aten.item %2554 : !torch.vtensor<[],si8> -> !torch.int | |
| %2557 = torch.aten.quantize_per_tensor %2552, %2555, %2556, %int12_558 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2558 = torch.aten.int_repr %2557 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %2559 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2560 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2561 = torch.aten.item %2559 : !torch.vtensor<[],f32> -> !torch.float | |
| %2562 = torch.aten.item %2560 : !torch.vtensor<[],si8> -> !torch.int | |
| %2563 = torch.aten._make_per_tensor_quantized_tensor %2558, %2561, %2562 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2564 = torch.aten.dequantize.self %2563 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2565 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2566 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2567 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2568 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_559 = torch.constant.int 0 | |
| %int0_560 = torch.constant.int 0 | |
| %2569 = torch.prim.NumToTensor.Scalar %int0_560 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2570 = torch.aten.index_select %2566, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2571 = torch.aten.item %2570 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2572 = torch.aten.index_select %2567, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2573 = torch.aten.item %2572 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2574 = torch.aten.index_select %2565, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2575 = torch.aten.item %2574 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2576 = torch.aten.index_select %2568, %int0_559, %2569 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2577 = torch.aten.item %2576 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2578 = torch.aten.slice.Tensor %2564, %2575, %2571, %2573, %2577 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2579 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2580 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2581 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2582 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_561 = torch.constant.int 0 | |
| %int0_562 = torch.constant.int 0 | |
| %2583 = torch.prim.NumToTensor.Scalar %int0_562 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2584 = torch.aten.index_select %2580, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2585 = torch.aten.item %2584 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2586 = torch.aten.index_select %2581, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2587 = torch.aten.item %2586 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2588 = torch.aten.index_select %2579, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2589 = torch.aten.item %2588 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2590 = torch.aten.index_select %2582, %int0_561, %2583 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2591 = torch.aten.item %2590 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2592 = torch.aten.slice.Tensor %2564, %2589, %2585, %2587, %2591 : !torch.vtensor<[1,128,40,40],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2593 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2594 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_563 = torch.constant.int 12 | |
| %2595 = torch.aten.item %2593 : !torch.vtensor<[],f32> -> !torch.float | |
| %2596 = torch.aten.item %2594 : !torch.vtensor<[],si8> -> !torch.int | |
| %2597 = torch.aten.quantize_per_tensor %74, %2595, %2596, %int12_563 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2598 = torch.aten.int_repr %2597 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %2599 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2600 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2601 = torch.aten.item %2599 : !torch.vtensor<[],f32> -> !torch.float | |
| %2602 = torch.aten.item %2600 : !torch.vtensor<[],si8> -> !torch.int | |
| %2603 = torch.aten._make_per_tensor_quantized_tensor %2598, %2601, %2602 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2604 = torch.aten.dequantize.self %2603 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %2605 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2606 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_564 = torch.constant.int 12 | |
| %2607 = torch.aten.item %2605 : !torch.vtensor<[],f32> -> !torch.float | |
| %2608 = torch.aten.item %2606 : !torch.vtensor<[],si8> -> !torch.int | |
| %2609 = torch.aten.quantize_per_tensor %75, %2607, %2608, %int12_564 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2610 = torch.aten.int_repr %2609 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2611 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2612 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2613 = torch.aten.item %2611 : !torch.vtensor<[],f32> -> !torch.float | |
| %2614 = torch.aten.item %2612 : !torch.vtensor<[],si8> -> !torch.int | |
| %2615 = torch.aten._make_per_tensor_quantized_tensor %2610, %2613, %2614 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2616 = torch.aten.dequantize.self %2615 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_565 = torch.constant.int 1 | |
| %int1_566 = torch.constant.int 1 | |
| %int1_567 = torch.constant.int 1 | |
| %int1_568 = torch.constant.int 1 | |
| %int1_569 = torch.constant.int 1 | |
| %int1_570 = torch.constant.int 1 | |
| %int0_571 = torch.constant.int 0 | |
| %2617 = torch.prim.ListConstruct %int1_565, %int1_566 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2618 = torch.prim.ListConstruct %int1_567, %int1_568 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2619 = torch.prim.ListConstruct %int1_569, %int1_570 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2620 = torch.prim.ListConstruct %int0_571, %int0_571 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_572 = torch.constant.bool false | |
| %int1_573 = torch.constant.int 1 | |
| %2621 = torch.aten.convolution %2592, %2604, %2616, %2619, %2617, %2618, %false_572, %2620, %int1_573 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2622 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2623 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_574 = torch.constant.int 12 | |
| %2624 = torch.aten.item %2622 : !torch.vtensor<[],f32> -> !torch.float | |
| %2625 = torch.aten.item %2623 : !torch.vtensor<[],si8> -> !torch.int | |
| %2626 = torch.aten.quantize_per_tensor %2621, %2624, %2625, %int12_574 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2627 = torch.aten.int_repr %2626 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2628 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2629 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2630 = torch.aten.item %2628 : !torch.vtensor<[],f32> -> !torch.float | |
| %2631 = torch.aten.item %2629 : !torch.vtensor<[],si8> -> !torch.int | |
| %2632 = torch.aten._make_per_tensor_quantized_tensor %2627, %2630, %2631 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2633 = torch.aten.dequantize.self %2632 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2634 = torch.aten.sigmoid %2633 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2635 = torch.aten.mul.Tensor %2633, %2634 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2636 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2637 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_575 = torch.constant.int 12 | |
| %2638 = torch.aten.item %2636 : !torch.vtensor<[],f32> -> !torch.float | |
| %2639 = torch.aten.item %2637 : !torch.vtensor<[],si8> -> !torch.int | |
| %2640 = torch.aten.quantize_per_tensor %2635, %2638, %2639, %int12_575 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2641 = torch.aten.int_repr %2640 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2642 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2643 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2644 = torch.aten.item %2642 : !torch.vtensor<[],f32> -> !torch.float | |
| %2645 = torch.aten.item %2643 : !torch.vtensor<[],si8> -> !torch.int | |
| %2646 = torch.aten._make_per_tensor_quantized_tensor %2641, %2644, %2645 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2647 = torch.aten.dequantize.self %2646 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2648 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2649 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_576 = torch.constant.int 12 | |
| %2650 = torch.aten.item %2648 : !torch.vtensor<[],f32> -> !torch.float | |
| %2651 = torch.aten.item %2649 : !torch.vtensor<[],si8> -> !torch.int | |
| %2652 = torch.aten.quantize_per_tensor %76, %2650, %2651, %int12_576 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2653 = torch.aten.int_repr %2652 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %2654 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2655 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2656 = torch.aten.item %2654 : !torch.vtensor<[],f32> -> !torch.float | |
| %2657 = torch.aten.item %2655 : !torch.vtensor<[],si8> -> !torch.int | |
| %2658 = torch.aten._make_per_tensor_quantized_tensor %2653, %2656, %2657 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %2659 = torch.aten.dequantize.self %2658 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %2660 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2661 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_577 = torch.constant.int 12 | |
| %2662 = torch.aten.item %2660 : !torch.vtensor<[],f32> -> !torch.float | |
| %2663 = torch.aten.item %2661 : !torch.vtensor<[],si8> -> !torch.int | |
| %2664 = torch.aten.quantize_per_tensor %77, %2662, %2663, %int12_577 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2665 = torch.aten.int_repr %2664 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %2666 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2667 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2668 = torch.aten.item %2666 : !torch.vtensor<[],f32> -> !torch.float | |
| %2669 = torch.aten.item %2667 : !torch.vtensor<[],si8> -> !torch.int | |
| %2670 = torch.aten._make_per_tensor_quantized_tensor %2665, %2668, %2669 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %2671 = torch.aten.dequantize.self %2670 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_578 = torch.constant.int 1 | |
| %int1_579 = torch.constant.int 1 | |
| %int1_580 = torch.constant.int 1 | |
| %int1_581 = torch.constant.int 1 | |
| %int1_582 = torch.constant.int 1 | |
| %int1_583 = torch.constant.int 1 | |
| %int0_584 = torch.constant.int 0 | |
| %2672 = torch.prim.ListConstruct %int1_578, %int1_579 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2673 = torch.prim.ListConstruct %int1_580, %int1_581 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2674 = torch.prim.ListConstruct %int1_582, %int1_583 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2675 = torch.prim.ListConstruct %int0_584, %int0_584 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_585 = torch.constant.bool false | |
| %int1_586 = torch.constant.int 1 | |
| %2676 = torch.aten.convolution %2647, %2659, %2671, %2674, %2672, %2673, %false_585, %2675, %int1_586 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %2677 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2678 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_587 = torch.constant.int 12 | |
| %2679 = torch.aten.item %2677 : !torch.vtensor<[],f32> -> !torch.float | |
| %2680 = torch.aten.item %2678 : !torch.vtensor<[],si8> -> !torch.int | |
| %2681 = torch.aten.quantize_per_tensor %2676, %2679, %2680, %int12_587 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2682 = torch.aten.int_repr %2681 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %2683 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2684 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2685 = torch.aten.item %2683 : !torch.vtensor<[],f32> -> !torch.float | |
| %2686 = torch.aten.item %2684 : !torch.vtensor<[],si8> -> !torch.int | |
| %2687 = torch.aten._make_per_tensor_quantized_tensor %2682, %2685, %2686 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %2688 = torch.aten.dequantize.self %2687 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2689 = torch.aten.sigmoid %2688 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2690 = torch.aten.mul.Tensor %2688, %2689 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %2691 = torch.prim.ListConstruct %2578, %2592, %2690 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32>) -> !torch.list<vtensor> | |
| %int1_588 = torch.constant.int 1 | |
| %2692 = torch.aten.cat %2691, %int1_588 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,192,40,40],f32> | |
| %2693 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2694 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_589 = torch.constant.int 12 | |
| %2695 = torch.aten.item %2693 : !torch.vtensor<[],f32> -> !torch.float | |
| %2696 = torch.aten.item %2694 : !torch.vtensor<[],si8> -> !torch.int | |
| %2697 = torch.aten.quantize_per_tensor %2692, %2695, %2696, %int12_589 : !torch.vtensor<[1,192,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
| %2698 = torch.aten.int_repr %2697 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],si8> | |
| %2699 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2700 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2701 = torch.aten.item %2699 : !torch.vtensor<[],f32> -> !torch.float | |
| %2702 = torch.aten.item %2700 : !torch.vtensor<[],si8> -> !torch.int | |
| %2703 = torch.aten._make_per_tensor_quantized_tensor %2698, %2701, %2702 : !torch.vtensor<[1,192,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,192,40,40],!torch.qint8> | |
| %2704 = torch.aten.dequantize.self %2703 : !torch.vtensor<[1,192,40,40],!torch.qint8> -> !torch.vtensor<[1,192,40,40],f32> | |
| %2705 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2706 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_590 = torch.constant.int 12 | |
| %2707 = torch.aten.item %2705 : !torch.vtensor<[],f32> -> !torch.float | |
| %2708 = torch.aten.item %2706 : !torch.vtensor<[],si8> -> !torch.int | |
| %2709 = torch.aten.quantize_per_tensor %78, %2707, %2708, %int12_590 : !torch.vtensor<[128,192,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
| %2710 = torch.aten.int_repr %2709 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],si8> | |
| %2711 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2712 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2713 = torch.aten.item %2711 : !torch.vtensor<[],f32> -> !torch.float | |
| %2714 = torch.aten.item %2712 : !torch.vtensor<[],si8> -> !torch.int | |
| %2715 = torch.aten._make_per_tensor_quantized_tensor %2710, %2713, %2714 : !torch.vtensor<[128,192,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,192,1,1],!torch.qint8> | |
| %2716 = torch.aten.dequantize.self %2715 : !torch.vtensor<[128,192,1,1],!torch.qint8> -> !torch.vtensor<[128,192,1,1],f32> | |
| %2717 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2718 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_591 = torch.constant.int 12 | |
| %2719 = torch.aten.item %2717 : !torch.vtensor<[],f32> -> !torch.float | |
| %2720 = torch.aten.item %2718 : !torch.vtensor<[],si8> -> !torch.int | |
| %2721 = torch.aten.quantize_per_tensor %79, %2719, %2720, %int12_591 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2722 = torch.aten.int_repr %2721 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %2723 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2724 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2725 = torch.aten.item %2723 : !torch.vtensor<[],f32> -> !torch.float | |
| %2726 = torch.aten.item %2724 : !torch.vtensor<[],si8> -> !torch.int | |
| %2727 = torch.aten._make_per_tensor_quantized_tensor %2722, %2725, %2726 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2728 = torch.aten.dequantize.self %2727 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_592 = torch.constant.int 0 | |
| %int0_593 = torch.constant.int 0 | |
| %int1_594 = torch.constant.int 1 | |
| %int1_595 = torch.constant.int 1 | |
| %int1_596 = torch.constant.int 1 | |
| %int1_597 = torch.constant.int 1 | |
| %int0_598 = torch.constant.int 0 | |
| %2729 = torch.prim.ListConstruct %int0_592, %int0_593 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2730 = torch.prim.ListConstruct %int1_594, %int1_595 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2731 = torch.prim.ListConstruct %int1_596, %int1_597 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2732 = torch.prim.ListConstruct %int0_598, %int0_598 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_599 = torch.constant.bool false | |
| %int1_600 = torch.constant.int 1 | |
| %2733 = torch.aten.convolution %2704, %2716, %2728, %2731, %2729, %2730, %false_599, %2732, %int1_600 : !torch.vtensor<[1,192,40,40],f32>, !torch.vtensor<[128,192,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,40,40],f32> | |
| %2734 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2735 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_601 = torch.constant.int 12 | |
| %2736 = torch.aten.item %2734 : !torch.vtensor<[],f32> -> !torch.float | |
| %2737 = torch.aten.item %2735 : !torch.vtensor<[],si8> -> !torch.int | |
| %2738 = torch.aten.quantize_per_tensor %2733, %2736, %2737, %int12_601 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2739 = torch.aten.int_repr %2738 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %2740 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2741 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2742 = torch.aten.item %2740 : !torch.vtensor<[],f32> -> !torch.float | |
| %2743 = torch.aten.item %2741 : !torch.vtensor<[],si8> -> !torch.int | |
| %2744 = torch.aten._make_per_tensor_quantized_tensor %2739, %2742, %2743 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2745 = torch.aten.dequantize.self %2744 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2746 = torch.aten.sigmoid %2745 : !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2747 = torch.aten.mul.Tensor %2745, %2746 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[1,128,40,40],f32> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2748 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2749 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_602 = torch.constant.int 12 | |
| %2750 = torch.aten.item %2748 : !torch.vtensor<[],f32> -> !torch.float | |
| %2751 = torch.aten.item %2749 : !torch.vtensor<[],si8> -> !torch.int | |
| %2752 = torch.aten.quantize_per_tensor %2747, %2750, %2751, %int12_602 : !torch.vtensor<[1,128,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2753 = torch.aten.int_repr %2752 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],si8> | |
| %2754 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2755 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2756 = torch.aten.item %2754 : !torch.vtensor<[],f32> -> !torch.float | |
| %2757 = torch.aten.item %2755 : !torch.vtensor<[],si8> -> !torch.int | |
| %2758 = torch.aten._make_per_tensor_quantized_tensor %2753, %2756, %2757 : !torch.vtensor<[1,128,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,40,40],!torch.qint8> | |
| %2759 = torch.aten.dequantize.self %2758 : !torch.vtensor<[1,128,40,40],!torch.qint8> -> !torch.vtensor<[1,128,40,40],f32> | |
| %2760 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2761 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_603 = torch.constant.int 12 | |
| %2762 = torch.aten.item %2760 : !torch.vtensor<[],f32> -> !torch.float | |
| %2763 = torch.aten.item %2761 : !torch.vtensor<[],si8> -> !torch.int | |
| %2764 = torch.aten.quantize_per_tensor %80, %2762, %2763, %int12_603 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %2765 = torch.aten.int_repr %2764 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %2766 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2767 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2768 = torch.aten.item %2766 : !torch.vtensor<[],f32> -> !torch.float | |
| %2769 = torch.aten.item %2767 : !torch.vtensor<[],si8> -> !torch.int | |
| %2770 = torch.aten._make_per_tensor_quantized_tensor %2765, %2768, %2769 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %2771 = torch.aten.dequantize.self %2770 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %2772 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2773 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_604 = torch.constant.int 12 | |
| %2774 = torch.aten.item %2772 : !torch.vtensor<[],f32> -> !torch.float | |
| %2775 = torch.aten.item %2773 : !torch.vtensor<[],si8> -> !torch.int | |
| %2776 = torch.aten.quantize_per_tensor %81, %2774, %2775, %int12_604 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2777 = torch.aten.int_repr %2776 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %2778 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2779 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2780 = torch.aten.item %2778 : !torch.vtensor<[],f32> -> !torch.float | |
| %2781 = torch.aten.item %2779 : !torch.vtensor<[],si8> -> !torch.int | |
| %2782 = torch.aten._make_per_tensor_quantized_tensor %2777, %2780, %2781 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2783 = torch.aten.dequantize.self %2782 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int1_605 = torch.constant.int 1 | |
| %int1_606 = torch.constant.int 1 | |
| %int1_607 = torch.constant.int 1 | |
| %int1_608 = torch.constant.int 1 | |
| %int2_609 = torch.constant.int 2 | |
| %int2_610 = torch.constant.int 2 | |
| %int0_611 = torch.constant.int 0 | |
| %2784 = torch.prim.ListConstruct %int1_605, %int1_606 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2785 = torch.prim.ListConstruct %int1_607, %int1_608 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2786 = torch.prim.ListConstruct %int2_609, %int2_610 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2787 = torch.prim.ListConstruct %int0_611, %int0_611 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_612 = torch.constant.bool false | |
| %int1_613 = torch.constant.int 1 | |
| %2788 = torch.aten.convolution %2759, %2771, %2783, %2786, %2784, %2785, %false_612, %2787, %int1_613 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %2789 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2790 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_614 = torch.constant.int 12 | |
| %2791 = torch.aten.item %2789 : !torch.vtensor<[],f32> -> !torch.float | |
| %2792 = torch.aten.item %2790 : !torch.vtensor<[],si8> -> !torch.int | |
| %2793 = torch.aten.quantize_per_tensor %2788, %2791, %2792, %int12_614 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2794 = torch.aten.int_repr %2793 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %2795 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2796 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2797 = torch.aten.item %2795 : !torch.vtensor<[],f32> -> !torch.float | |
| %2798 = torch.aten.item %2796 : !torch.vtensor<[],si8> -> !torch.int | |
| %2799 = torch.aten._make_per_tensor_quantized_tensor %2794, %2797, %2798 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2800 = torch.aten.dequantize.self %2799 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2801 = torch.aten.sigmoid %2800 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2802 = torch.aten.mul.Tensor %2800, %2801 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2803 = torch.prim.ListConstruct %2802, %1914 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,256,20,20],f32>) -> !torch.list<vtensor> | |
| %int1_615 = torch.constant.int 1 | |
| %2804 = torch.aten.cat %2803, %int1_615 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,384,20,20],f32> | |
| %2805 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2806 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_616 = torch.constant.int 12 | |
| %2807 = torch.aten.item %2805 : !torch.vtensor<[],f32> -> !torch.float | |
| %2808 = torch.aten.item %2806 : !torch.vtensor<[],si8> -> !torch.int | |
| %2809 = torch.aten.quantize_per_tensor %2804, %2807, %2808, %int12_616 : !torch.vtensor<[1,384,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
| %2810 = torch.aten.int_repr %2809 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],si8> | |
| %2811 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2812 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2813 = torch.aten.item %2811 : !torch.vtensor<[],f32> -> !torch.float | |
| %2814 = torch.aten.item %2812 : !torch.vtensor<[],si8> -> !torch.int | |
| %2815 = torch.aten._make_per_tensor_quantized_tensor %2810, %2813, %2814 : !torch.vtensor<[1,384,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
| %2816 = torch.aten.dequantize.self %2815 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],f32> | |
| %2817 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2818 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_617 = torch.constant.int 12 | |
| %2819 = torch.aten.item %2817 : !torch.vtensor<[],f32> -> !torch.float | |
| %2820 = torch.aten.item %2818 : !torch.vtensor<[],si8> -> !torch.int | |
| %2821 = torch.aten.quantize_per_tensor %82, %2819, %2820, %int12_617 : !torch.vtensor<[256,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
| %2822 = torch.aten.int_repr %2821 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],si8> | |
| %2823 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2824 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2825 = torch.aten.item %2823 : !torch.vtensor<[],f32> -> !torch.float | |
| %2826 = torch.aten.item %2824 : !torch.vtensor<[],si8> -> !torch.int | |
| %2827 = torch.aten._make_per_tensor_quantized_tensor %2822, %2825, %2826 : !torch.vtensor<[256,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
| %2828 = torch.aten.dequantize.self %2827 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],f32> | |
| %2829 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2830 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_618 = torch.constant.int 12 | |
| %2831 = torch.aten.item %2829 : !torch.vtensor<[],f32> -> !torch.float | |
| %2832 = torch.aten.item %2830 : !torch.vtensor<[],si8> -> !torch.int | |
| %2833 = torch.aten.quantize_per_tensor %83, %2831, %2832, %int12_618 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %2834 = torch.aten.int_repr %2833 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %2835 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2836 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2837 = torch.aten.item %2835 : !torch.vtensor<[],f32> -> !torch.float | |
| %2838 = torch.aten.item %2836 : !torch.vtensor<[],si8> -> !torch.int | |
| %2839 = torch.aten._make_per_tensor_quantized_tensor %2834, %2837, %2838 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %2840 = torch.aten.dequantize.self %2839 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %int0_619 = torch.constant.int 0 | |
| %int0_620 = torch.constant.int 0 | |
| %int1_621 = torch.constant.int 1 | |
| %int1_622 = torch.constant.int 1 | |
| %int1_623 = torch.constant.int 1 | |
| %int1_624 = torch.constant.int 1 | |
| %int0_625 = torch.constant.int 0 | |
| %2841 = torch.prim.ListConstruct %int0_619, %int0_620 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2842 = torch.prim.ListConstruct %int1_621, %int1_622 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2843 = torch.prim.ListConstruct %int1_623, %int1_624 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2844 = torch.prim.ListConstruct %int0_625, %int0_625 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_626 = torch.constant.bool false | |
| %int1_627 = torch.constant.int 1 | |
| %2845 = torch.aten.convolution %2816, %2828, %2840, %2843, %2841, %2842, %false_626, %2844, %int1_627 : !torch.vtensor<[1,384,20,20],f32>, !torch.vtensor<[256,384,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
| %2846 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2847 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_628 = torch.constant.int 12 | |
| %2848 = torch.aten.item %2846 : !torch.vtensor<[],f32> -> !torch.float | |
| %2849 = torch.aten.item %2847 : !torch.vtensor<[],si8> -> !torch.int | |
| %2850 = torch.aten.quantize_per_tensor %2845, %2848, %2849, %int12_628 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %2851 = torch.aten.int_repr %2850 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %2852 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2853 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2854 = torch.aten.item %2852 : !torch.vtensor<[],f32> -> !torch.float | |
| %2855 = torch.aten.item %2853 : !torch.vtensor<[],si8> -> !torch.int | |
| %2856 = torch.aten._make_per_tensor_quantized_tensor %2851, %2854, %2855 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %2857 = torch.aten.dequantize.self %2856 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %2858 = torch.aten.sigmoid %2857 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %2859 = torch.aten.mul.Tensor %2857, %2858 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %2860 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2861 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_629 = torch.constant.int 12 | |
| %2862 = torch.aten.item %2860 : !torch.vtensor<[],f32> -> !torch.float | |
| %2863 = torch.aten.item %2861 : !torch.vtensor<[],si8> -> !torch.int | |
| %2864 = torch.aten.quantize_per_tensor %2859, %2862, %2863, %int12_629 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %2865 = torch.aten.int_repr %2864 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %2866 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2867 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2868 = torch.aten.item %2866 : !torch.vtensor<[],f32> -> !torch.float | |
| %2869 = torch.aten.item %2867 : !torch.vtensor<[],si8> -> !torch.int | |
| %2870 = torch.aten._make_per_tensor_quantized_tensor %2865, %2868, %2869 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %2871 = torch.aten.dequantize.self %2870 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %2872 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2873 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2874 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2875 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_630 = torch.constant.int 0 | |
| %int0_631 = torch.constant.int 0 | |
| %2876 = torch.prim.NumToTensor.Scalar %int0_631 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2877 = torch.aten.index_select %2873, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2878 = torch.aten.item %2877 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2879 = torch.aten.index_select %2874, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2880 = torch.aten.item %2879 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2881 = torch.aten.index_select %2872, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2882 = torch.aten.item %2881 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2883 = torch.aten.index_select %2875, %int0_630, %2876 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2884 = torch.aten.item %2883 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2885 = torch.aten.slice.Tensor %2871, %2882, %2878, %2880, %2884 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %2886 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2887 = torch.vtensor.literal(dense<128> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2888 = torch.vtensor.literal(dense<256> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %2889 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_632 = torch.constant.int 0 | |
| %int0_633 = torch.constant.int 0 | |
| %2890 = torch.prim.NumToTensor.Scalar %int0_633 : !torch.int -> !torch.vtensor<[1],si64> | |
| %2891 = torch.aten.index_select %2887, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2892 = torch.aten.item %2891 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2893 = torch.aten.index_select %2888, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2894 = torch.aten.item %2893 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2895 = torch.aten.index_select %2886, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2896 = torch.aten.item %2895 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2897 = torch.aten.index_select %2889, %int0_632, %2890 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %2898 = torch.aten.item %2897 : !torch.vtensor<[1],si64> -> !torch.int | |
| %2899 = torch.aten.slice.Tensor %2871, %2896, %2892, %2894, %2898 : !torch.vtensor<[1,256,20,20],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %2900 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2901 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_634 = torch.constant.int 12 | |
| %2902 = torch.aten.item %2900 : !torch.vtensor<[],f32> -> !torch.float | |
| %2903 = torch.aten.item %2901 : !torch.vtensor<[],si8> -> !torch.int | |
| %2904 = torch.aten.quantize_per_tensor %84, %2902, %2903, %int12_634 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %2905 = torch.aten.int_repr %2904 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %2906 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2907 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2908 = torch.aten.item %2906 : !torch.vtensor<[],f32> -> !torch.float | |
| %2909 = torch.aten.item %2907 : !torch.vtensor<[],si8> -> !torch.int | |
| %2910 = torch.aten._make_per_tensor_quantized_tensor %2905, %2908, %2909 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %2911 = torch.aten.dequantize.self %2910 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %2912 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2913 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_635 = torch.constant.int 12 | |
| %2914 = torch.aten.item %2912 : !torch.vtensor<[],f32> -> !torch.float | |
| %2915 = torch.aten.item %2913 : !torch.vtensor<[],si8> -> !torch.int | |
| %2916 = torch.aten.quantize_per_tensor %85, %2914, %2915, %int12_635 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2917 = torch.aten.int_repr %2916 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %2918 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2919 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2920 = torch.aten.item %2918 : !torch.vtensor<[],f32> -> !torch.float | |
| %2921 = torch.aten.item %2919 : !torch.vtensor<[],si8> -> !torch.int | |
| %2922 = torch.aten._make_per_tensor_quantized_tensor %2917, %2920, %2921 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2923 = torch.aten.dequantize.self %2922 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int1_636 = torch.constant.int 1 | |
| %int1_637 = torch.constant.int 1 | |
| %int1_638 = torch.constant.int 1 | |
| %int1_639 = torch.constant.int 1 | |
| %int1_640 = torch.constant.int 1 | |
| %int1_641 = torch.constant.int 1 | |
| %int0_642 = torch.constant.int 0 | |
| %2924 = torch.prim.ListConstruct %int1_636, %int1_637 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2925 = torch.prim.ListConstruct %int1_638, %int1_639 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2926 = torch.prim.ListConstruct %int1_640, %int1_641 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2927 = torch.prim.ListConstruct %int0_642, %int0_642 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_643 = torch.constant.bool false | |
| %int1_644 = torch.constant.int 1 | |
| %2928 = torch.aten.convolution %2899, %2911, %2923, %2926, %2924, %2925, %false_643, %2927, %int1_644 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %2929 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2930 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_645 = torch.constant.int 12 | |
| %2931 = torch.aten.item %2929 : !torch.vtensor<[],f32> -> !torch.float | |
| %2932 = torch.aten.item %2930 : !torch.vtensor<[],si8> -> !torch.int | |
| %2933 = torch.aten.quantize_per_tensor %2928, %2931, %2932, %int12_645 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2934 = torch.aten.int_repr %2933 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %2935 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2936 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2937 = torch.aten.item %2935 : !torch.vtensor<[],f32> -> !torch.float | |
| %2938 = torch.aten.item %2936 : !torch.vtensor<[],si8> -> !torch.int | |
| %2939 = torch.aten._make_per_tensor_quantized_tensor %2934, %2937, %2938 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2940 = torch.aten.dequantize.self %2939 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2941 = torch.aten.sigmoid %2940 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2942 = torch.aten.mul.Tensor %2940, %2941 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2943 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2944 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_646 = torch.constant.int 12 | |
| %2945 = torch.aten.item %2943 : !torch.vtensor<[],f32> -> !torch.float | |
| %2946 = torch.aten.item %2944 : !torch.vtensor<[],si8> -> !torch.int | |
| %2947 = torch.aten.quantize_per_tensor %2942, %2945, %2946, %int12_646 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2948 = torch.aten.int_repr %2947 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %2949 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2950 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2951 = torch.aten.item %2949 : !torch.vtensor<[],f32> -> !torch.float | |
| %2952 = torch.aten.item %2950 : !torch.vtensor<[],si8> -> !torch.int | |
| %2953 = torch.aten._make_per_tensor_quantized_tensor %2948, %2951, %2952 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2954 = torch.aten.dequantize.self %2953 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2955 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2956 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_647 = torch.constant.int 12 | |
| %2957 = torch.aten.item %2955 : !torch.vtensor<[],f32> -> !torch.float | |
| %2958 = torch.aten.item %2956 : !torch.vtensor<[],si8> -> !torch.int | |
| %2959 = torch.aten.quantize_per_tensor %86, %2957, %2958, %int12_647 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %2960 = torch.aten.int_repr %2959 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %2961 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2962 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2963 = torch.aten.item %2961 : !torch.vtensor<[],f32> -> !torch.float | |
| %2964 = torch.aten.item %2962 : !torch.vtensor<[],si8> -> !torch.int | |
| %2965 = torch.aten._make_per_tensor_quantized_tensor %2960, %2963, %2964 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %2966 = torch.aten.dequantize.self %2965 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %2967 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2968 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_648 = torch.constant.int 12 | |
| %2969 = torch.aten.item %2967 : !torch.vtensor<[],f32> -> !torch.float | |
| %2970 = torch.aten.item %2968 : !torch.vtensor<[],si8> -> !torch.int | |
| %2971 = torch.aten.quantize_per_tensor %87, %2969, %2970, %int12_648 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2972 = torch.aten.int_repr %2971 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %2973 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2974 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2975 = torch.aten.item %2973 : !torch.vtensor<[],f32> -> !torch.float | |
| %2976 = torch.aten.item %2974 : !torch.vtensor<[],si8> -> !torch.int | |
| %2977 = torch.aten._make_per_tensor_quantized_tensor %2972, %2975, %2976 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %2978 = torch.aten.dequantize.self %2977 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int1_649 = torch.constant.int 1 | |
| %int1_650 = torch.constant.int 1 | |
| %int1_651 = torch.constant.int 1 | |
| %int1_652 = torch.constant.int 1 | |
| %int1_653 = torch.constant.int 1 | |
| %int1_654 = torch.constant.int 1 | |
| %int0_655 = torch.constant.int 0 | |
| %2979 = torch.prim.ListConstruct %int1_649, %int1_650 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2980 = torch.prim.ListConstruct %int1_651, %int1_652 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2981 = torch.prim.ListConstruct %int1_653, %int1_654 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2982 = torch.prim.ListConstruct %int0_655, %int0_655 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_656 = torch.constant.bool false | |
| %int1_657 = torch.constant.int 1 | |
| %2983 = torch.aten.convolution %2954, %2966, %2978, %2981, %2979, %2980, %false_656, %2982, %int1_657 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,20,20],f32> | |
| %2984 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2985 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_658 = torch.constant.int 12 | |
| %2986 = torch.aten.item %2984 : !torch.vtensor<[],f32> -> !torch.float | |
| %2987 = torch.aten.item %2985 : !torch.vtensor<[],si8> -> !torch.int | |
| %2988 = torch.aten.quantize_per_tensor %2983, %2986, %2987, %int12_658 : !torch.vtensor<[1,128,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2989 = torch.aten.int_repr %2988 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],si8> | |
| %2990 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2991 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2992 = torch.aten.item %2990 : !torch.vtensor<[],f32> -> !torch.float | |
| %2993 = torch.aten.item %2991 : !torch.vtensor<[],si8> -> !torch.int | |
| %2994 = torch.aten._make_per_tensor_quantized_tensor %2989, %2992, %2993 : !torch.vtensor<[1,128,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,20,20],!torch.qint8> | |
| %2995 = torch.aten.dequantize.self %2994 : !torch.vtensor<[1,128,20,20],!torch.qint8> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2996 = torch.aten.sigmoid %2995 : !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2997 = torch.aten.mul.Tensor %2995, %2996 : !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32> -> !torch.vtensor<[1,128,20,20],f32> | |
| %2998 = torch.prim.ListConstruct %2885, %2899, %2997 : (!torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>, !torch.vtensor<[1,128,20,20],f32>) -> !torch.list<vtensor> | |
| %int1_659 = torch.constant.int 1 | |
| %2999 = torch.aten.cat %2998, %int1_659 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,384,20,20],f32> | |
| %3000 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3001 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_660 = torch.constant.int 12 | |
| %3002 = torch.aten.item %3000 : !torch.vtensor<[],f32> -> !torch.float | |
| %3003 = torch.aten.item %3001 : !torch.vtensor<[],si8> -> !torch.int | |
| %3004 = torch.aten.quantize_per_tensor %2999, %3002, %3003, %int12_660 : !torch.vtensor<[1,384,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
| %3005 = torch.aten.int_repr %3004 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],si8> | |
| %3006 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3007 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3008 = torch.aten.item %3006 : !torch.vtensor<[],f32> -> !torch.float | |
| %3009 = torch.aten.item %3007 : !torch.vtensor<[],si8> -> !torch.int | |
| %3010 = torch.aten._make_per_tensor_quantized_tensor %3005, %3008, %3009 : !torch.vtensor<[1,384,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,384,20,20],!torch.qint8> | |
| %3011 = torch.aten.dequantize.self %3010 : !torch.vtensor<[1,384,20,20],!torch.qint8> -> !torch.vtensor<[1,384,20,20],f32> | |
| %3012 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3013 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_661 = torch.constant.int 12 | |
| %3014 = torch.aten.item %3012 : !torch.vtensor<[],f32> -> !torch.float | |
| %3015 = torch.aten.item %3013 : !torch.vtensor<[],si8> -> !torch.int | |
| %3016 = torch.aten.quantize_per_tensor %88, %3014, %3015, %int12_661 : !torch.vtensor<[256,384,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
| %3017 = torch.aten.int_repr %3016 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],si8> | |
| %3018 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3019 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3020 = torch.aten.item %3018 : !torch.vtensor<[],f32> -> !torch.float | |
| %3021 = torch.aten.item %3019 : !torch.vtensor<[],si8> -> !torch.int | |
| %3022 = torch.aten._make_per_tensor_quantized_tensor %3017, %3020, %3021 : !torch.vtensor<[256,384,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,384,1,1],!torch.qint8> | |
| %3023 = torch.aten.dequantize.self %3022 : !torch.vtensor<[256,384,1,1],!torch.qint8> -> !torch.vtensor<[256,384,1,1],f32> | |
| %3024 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3025 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_662 = torch.constant.int 12 | |
| %3026 = torch.aten.item %3024 : !torch.vtensor<[],f32> -> !torch.float | |
| %3027 = torch.aten.item %3025 : !torch.vtensor<[],si8> -> !torch.int | |
| %3028 = torch.aten.quantize_per_tensor %89, %3026, %3027, %int12_662 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %3029 = torch.aten.int_repr %3028 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %3030 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3031 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3032 = torch.aten.item %3030 : !torch.vtensor<[],f32> -> !torch.float | |
| %3033 = torch.aten.item %3031 : !torch.vtensor<[],si8> -> !torch.int | |
| %3034 = torch.aten._make_per_tensor_quantized_tensor %3029, %3032, %3033 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %3035 = torch.aten.dequantize.self %3034 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %int0_663 = torch.constant.int 0 | |
| %int0_664 = torch.constant.int 0 | |
| %int1_665 = torch.constant.int 1 | |
| %int1_666 = torch.constant.int 1 | |
| %int1_667 = torch.constant.int 1 | |
| %int1_668 = torch.constant.int 1 | |
| %int0_669 = torch.constant.int 0 | |
| %3036 = torch.prim.ListConstruct %int0_663, %int0_664 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3037 = torch.prim.ListConstruct %int1_665, %int1_666 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3038 = torch.prim.ListConstruct %int1_667, %int1_668 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3039 = torch.prim.ListConstruct %int0_669, %int0_669 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_670 = torch.constant.bool false | |
| %int1_671 = torch.constant.int 1 | |
| %3040 = torch.aten.convolution %3011, %3023, %3035, %3038, %3036, %3037, %false_670, %3039, %int1_671 : !torch.vtensor<[1,384,20,20],f32>, !torch.vtensor<[256,384,1,1],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,256,20,20],f32> | |
| %3041 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3042 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_672 = torch.constant.int 12 | |
| %3043 = torch.aten.item %3041 : !torch.vtensor<[],f32> -> !torch.float | |
| %3044 = torch.aten.item %3042 : !torch.vtensor<[],si8> -> !torch.int | |
| %3045 = torch.aten.quantize_per_tensor %3040, %3043, %3044, %int12_672 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %3046 = torch.aten.int_repr %3045 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %3047 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3048 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3049 = torch.aten.item %3047 : !torch.vtensor<[],f32> -> !torch.float | |
| %3050 = torch.aten.item %3048 : !torch.vtensor<[],si8> -> !torch.int | |
| %3051 = torch.aten._make_per_tensor_quantized_tensor %3046, %3049, %3050 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %3052 = torch.aten.dequantize.self %3051 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %3053 = torch.aten.sigmoid %3052 : !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %3054 = torch.aten.mul.Tensor %3052, %3053 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[1,256,20,20],f32> -> !torch.vtensor<[1,256,20,20],f32> | |
| %3055 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3056 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_673 = torch.constant.int 12 | |
| %3057 = torch.aten.item %3055 : !torch.vtensor<[],f32> -> !torch.float | |
| %3058 = torch.aten.item %3056 : !torch.vtensor<[],si8> -> !torch.int | |
| %3059 = torch.aten.quantize_per_tensor %3054, %3057, %3058, %int12_673 : !torch.vtensor<[1,256,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %3060 = torch.aten.int_repr %3059 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],si8> | |
| %3061 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3062 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3063 = torch.aten.item %3061 : !torch.vtensor<[],f32> -> !torch.float | |
| %3064 = torch.aten.item %3062 : !torch.vtensor<[],si8> -> !torch.int | |
| %3065 = torch.aten._make_per_tensor_quantized_tensor %3060, %3063, %3064 : !torch.vtensor<[1,256,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,256,20,20],!torch.qint8> | |
| %3066 = torch.aten.dequantize.self %3065 : !torch.vtensor<[1,256,20,20],!torch.qint8> -> !torch.vtensor<[1,256,20,20],f32> | |
| %3067 = torch.vtensor.literal(dense<1> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %3068 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3069 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_674 = torch.constant.int 12 | |
| %3070 = torch.aten.item %3068 : !torch.vtensor<[],f32> -> !torch.float | |
| %3071 = torch.aten.item %3069 : !torch.vtensor<[],si8> -> !torch.int | |
| %3072 = torch.aten.quantize_per_tensor %90, %3070, %3071, %int12_674 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3073 = torch.aten.int_repr %3072 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %3074 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3075 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3076 = torch.aten.item %3074 : !torch.vtensor<[],f32> -> !torch.float | |
| %3077 = torch.aten.item %3075 : !torch.vtensor<[],si8> -> !torch.int | |
| %3078 = torch.aten._make_per_tensor_quantized_tensor %3073, %3076, %3077 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3079 = torch.aten.dequantize.self %3078 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %3080 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3081 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_675 = torch.constant.int 12 | |
| %3082 = torch.aten.item %3080 : !torch.vtensor<[],f32> -> !torch.float | |
| %3083 = torch.aten.item %3081 : !torch.vtensor<[],si8> -> !torch.int | |
| %3084 = torch.aten.quantize_per_tensor %91, %3082, %3083, %int12_675 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3085 = torch.aten.int_repr %3084 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3086 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3087 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3088 = torch.aten.item %3086 : !torch.vtensor<[],f32> -> !torch.float | |
| %3089 = torch.aten.item %3087 : !torch.vtensor<[],si8> -> !torch.int | |
| %3090 = torch.aten._make_per_tensor_quantized_tensor %3085, %3088, %3089 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3091 = torch.aten.dequantize.self %3090 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_676 = torch.constant.int 1 | |
| %int1_677 = torch.constant.int 1 | |
| %int1_678 = torch.constant.int 1 | |
| %int1_679 = torch.constant.int 1 | |
| %int1_680 = torch.constant.int 1 | |
| %int1_681 = torch.constant.int 1 | |
| %int0_682 = torch.constant.int 0 | |
| %3092 = torch.prim.ListConstruct %int1_676, %int1_677 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3093 = torch.prim.ListConstruct %int1_678, %int1_679 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3094 = torch.prim.ListConstruct %int1_680, %int1_681 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3095 = torch.prim.ListConstruct %int0_682, %int0_682 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_683 = torch.constant.bool false | |
| %int1_684 = torch.constant.int 1 | |
| %3096 = torch.aten.convolution %2452, %3079, %3091, %3094, %3092, %3093, %false_683, %3095, %int1_684 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %3097 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3098 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_685 = torch.constant.int 12 | |
| %3099 = torch.aten.item %3097 : !torch.vtensor<[],f32> -> !torch.float | |
| %3100 = torch.aten.item %3098 : !torch.vtensor<[],si8> -> !torch.int | |
| %3101 = torch.aten.quantize_per_tensor %3096, %3099, %3100, %int12_685 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3102 = torch.aten.int_repr %3101 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %3103 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3104 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3105 = torch.aten.item %3103 : !torch.vtensor<[],f32> -> !torch.float | |
| %3106 = torch.aten.item %3104 : !torch.vtensor<[],si8> -> !torch.int | |
| %3107 = torch.aten._make_per_tensor_quantized_tensor %3102, %3105, %3106 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3108 = torch.aten.dequantize.self %3107 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3109 = torch.aten.sigmoid %3108 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3110 = torch.aten.mul.Tensor %3108, %3109 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3111 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3112 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_686 = torch.constant.int 12 | |
| %3113 = torch.aten.item %3111 : !torch.vtensor<[],f32> -> !torch.float | |
| %3114 = torch.aten.item %3112 : !torch.vtensor<[],si8> -> !torch.int | |
| %3115 = torch.aten.quantize_per_tensor %3110, %3113, %3114, %int12_686 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3116 = torch.aten.int_repr %3115 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %3117 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3118 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3119 = torch.aten.item %3117 : !torch.vtensor<[],f32> -> !torch.float | |
| %3120 = torch.aten.item %3118 : !torch.vtensor<[],si8> -> !torch.int | |
| %3121 = torch.aten._make_per_tensor_quantized_tensor %3116, %3119, %3120 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3122 = torch.aten.dequantize.self %3121 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3123 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3124 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_687 = torch.constant.int 12 | |
| %3125 = torch.aten.item %3123 : !torch.vtensor<[],f32> -> !torch.float | |
| %3126 = torch.aten.item %3124 : !torch.vtensor<[],si8> -> !torch.int | |
| %3127 = torch.aten.quantize_per_tensor %92, %3125, %3126, %int12_687 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3128 = torch.aten.int_repr %3127 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %3129 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3130 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3131 = torch.aten.item %3129 : !torch.vtensor<[],f32> -> !torch.float | |
| %3132 = torch.aten.item %3130 : !torch.vtensor<[],si8> -> !torch.int | |
| %3133 = torch.aten._make_per_tensor_quantized_tensor %3128, %3131, %3132 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3134 = torch.aten.dequantize.self %3133 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %3135 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3136 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_688 = torch.constant.int 12 | |
| %3137 = torch.aten.item %3135 : !torch.vtensor<[],f32> -> !torch.float | |
| %3138 = torch.aten.item %3136 : !torch.vtensor<[],si8> -> !torch.int | |
| %3139 = torch.aten.quantize_per_tensor %93, %3137, %3138, %int12_688 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3140 = torch.aten.int_repr %3139 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3141 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3142 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3143 = torch.aten.item %3141 : !torch.vtensor<[],f32> -> !torch.float | |
| %3144 = torch.aten.item %3142 : !torch.vtensor<[],si8> -> !torch.int | |
| %3145 = torch.aten._make_per_tensor_quantized_tensor %3140, %3143, %3144 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3146 = torch.aten.dequantize.self %3145 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_689 = torch.constant.int 1 | |
| %int1_690 = torch.constant.int 1 | |
| %int1_691 = torch.constant.int 1 | |
| %int1_692 = torch.constant.int 1 | |
| %int1_693 = torch.constant.int 1 | |
| %int1_694 = torch.constant.int 1 | |
| %int0_695 = torch.constant.int 0 | |
| %3147 = torch.prim.ListConstruct %int1_689, %int1_690 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3148 = torch.prim.ListConstruct %int1_691, %int1_692 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3149 = torch.prim.ListConstruct %int1_693, %int1_694 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3150 = torch.prim.ListConstruct %int0_695, %int0_695 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_696 = torch.constant.bool false | |
| %int1_697 = torch.constant.int 1 | |
| %3151 = torch.aten.convolution %3122, %3134, %3146, %3149, %3147, %3148, %false_696, %3150, %int1_697 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %3152 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3153 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_698 = torch.constant.int 12 | |
| %3154 = torch.aten.item %3152 : !torch.vtensor<[],f32> -> !torch.float | |
| %3155 = torch.aten.item %3153 : !torch.vtensor<[],si8> -> !torch.int | |
| %3156 = torch.aten.quantize_per_tensor %3151, %3154, %3155, %int12_698 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3157 = torch.aten.int_repr %3156 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %3158 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3159 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3160 = torch.aten.item %3158 : !torch.vtensor<[],f32> -> !torch.float | |
| %3161 = torch.aten.item %3159 : !torch.vtensor<[],si8> -> !torch.int | |
| %3162 = torch.aten._make_per_tensor_quantized_tensor %3157, %3160, %3161 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3163 = torch.aten.dequantize.self %3162 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3164 = torch.aten.sigmoid %3163 : !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3165 = torch.aten.mul.Tensor %3163, %3164 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,64,80,80],f32> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3166 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3167 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_699 = torch.constant.int 12 | |
| %3168 = torch.aten.item %3166 : !torch.vtensor<[],f32> -> !torch.float | |
| %3169 = torch.aten.item %3167 : !torch.vtensor<[],si8> -> !torch.int | |
| %3170 = torch.aten.quantize_per_tensor %3165, %3168, %3169, %int12_699 : !torch.vtensor<[1,64,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3171 = torch.aten.int_repr %3170 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],si8> | |
| %3172 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3174 = torch.aten.item %3172 : !torch.vtensor<[],f32> -> !torch.float | |
| %3175 = torch.aten.item %3173 : !torch.vtensor<[],si8> -> !torch.int | |
| %3176 = torch.aten._make_per_tensor_quantized_tensor %3171, %3174, %3175 : !torch.vtensor<[1,64,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,80,80],!torch.qint8> | |
| %3177 = torch.aten.dequantize.self %3176 : !torch.vtensor<[1,64,80,80],!torch.qint8> -> !torch.vtensor<[1,64,80,80],f32> | |
| %3178 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_700 = torch.constant.int 12 | |
| %3180 = torch.aten.item %3178 : !torch.vtensor<[],f32> -> !torch.float | |
| %3181 = torch.aten.item %3179 : !torch.vtensor<[],si8> -> !torch.int | |
| %3182 = torch.aten.quantize_per_tensor %94, %3180, %3181, %int12_700 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %3183 = torch.aten.int_repr %3182 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
| %3184 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3185 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3186 = torch.aten.item %3184 : !torch.vtensor<[],f32> -> !torch.float | |
| %3187 = torch.aten.item %3185 : !torch.vtensor<[],si8> -> !torch.int | |
| %3188 = torch.aten._make_per_tensor_quantized_tensor %3183, %3186, %3187 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %3189 = torch.aten.dequantize.self %3188 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
| %3190 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3191 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_701 = torch.constant.int 12 | |
| %3192 = torch.aten.item %3190 : !torch.vtensor<[],f32> -> !torch.float | |
| %3193 = torch.aten.item %3191 : !torch.vtensor<[],si8> -> !torch.int | |
| %3194 = torch.aten.quantize_per_tensor %95, %3192, %3193, %int12_701 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3195 = torch.aten.int_repr %3194 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3196 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3197 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3198 = torch.aten.item %3196 : !torch.vtensor<[],f32> -> !torch.float | |
| %3199 = torch.aten.item %3197 : !torch.vtensor<[],si8> -> !torch.int | |
| %3200 = torch.aten._make_per_tensor_quantized_tensor %3195, %3198, %3199 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3201 = torch.aten.dequantize.self %3200 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_702 = torch.constant.int 0 | |
| %int0_703 = torch.constant.int 0 | |
| %int1_704 = torch.constant.int 1 | |
| %int1_705 = torch.constant.int 1 | |
| %int1_706 = torch.constant.int 1 | |
| %int1_707 = torch.constant.int 1 | |
| %int0_708 = torch.constant.int 0 | |
| %3202 = torch.prim.ListConstruct %int0_702, %int0_703 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3203 = torch.prim.ListConstruct %int1_704, %int1_705 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3204 = torch.prim.ListConstruct %int1_706, %int1_707 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3205 = torch.prim.ListConstruct %int0_708, %int0_708 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_709 = torch.constant.bool false | |
| %int1_710 = torch.constant.int 1 | |
| %3206 = torch.aten.convolution %3177, %3189, %3201, %3204, %3202, %3203, %false_709, %3205, %int1_710 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,80,80],f32> | |
| %3207 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3208 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_711 = torch.constant.int 12 | |
| %3209 = torch.aten.item %3207 : !torch.vtensor<[],f32> -> !torch.float | |
| %3210 = torch.aten.item %3208 : !torch.vtensor<[],si8> -> !torch.int | |
| %3211 = torch.aten.quantize_per_tensor %96, %3209, %3210, %int12_711 : !torch.vtensor<[80,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,64,3,3],!torch.qint8> | |
| %3212 = torch.aten.int_repr %3211 : !torch.vtensor<[80,64,3,3],!torch.qint8> -> !torch.vtensor<[80,64,3,3],si8> | |
| %3213 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3214 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3215 = torch.aten.item %3213 : !torch.vtensor<[],f32> -> !torch.float | |
| %3216 = torch.aten.item %3214 : !torch.vtensor<[],si8> -> !torch.int | |
| %3217 = torch.aten._make_per_tensor_quantized_tensor %3212, %3215, %3216 : !torch.vtensor<[80,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,64,3,3],!torch.qint8> | |
| %3218 = torch.aten.dequantize.self %3217 : !torch.vtensor<[80,64,3,3],!torch.qint8> -> !torch.vtensor<[80,64,3,3],f32> | |
| %3219 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3220 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_712 = torch.constant.int 12 | |
| %3221 = torch.aten.item %3219 : !torch.vtensor<[],f32> -> !torch.float | |
| %3222 = torch.aten.item %3220 : !torch.vtensor<[],si8> -> !torch.int | |
| %3223 = torch.aten.quantize_per_tensor %97, %3221, %3222, %int12_712 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3224 = torch.aten.int_repr %3223 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3225 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3226 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3227 = torch.aten.item %3225 : !torch.vtensor<[],f32> -> !torch.float | |
| %3228 = torch.aten.item %3226 : !torch.vtensor<[],si8> -> !torch.int | |
| %3229 = torch.aten._make_per_tensor_quantized_tensor %3224, %3227, %3228 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3230 = torch.aten.dequantize.self %3229 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int1_713 = torch.constant.int 1 | |
| %int1_714 = torch.constant.int 1 | |
| %int1_715 = torch.constant.int 1 | |
| %int1_716 = torch.constant.int 1 | |
| %int1_717 = torch.constant.int 1 | |
| %int1_718 = torch.constant.int 1 | |
| %int0_719 = torch.constant.int 0 | |
| %3231 = torch.prim.ListConstruct %int1_713, %int1_714 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3232 = torch.prim.ListConstruct %int1_715, %int1_716 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3233 = torch.prim.ListConstruct %int1_717, %int1_718 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3234 = torch.prim.ListConstruct %int0_719, %int0_719 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_720 = torch.constant.bool false | |
| %int1_721 = torch.constant.int 1 | |
| %3235 = torch.aten.convolution %2452, %3218, %3230, %3233, %3231, %3232, %false_720, %3234, %int1_721 : !torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[80,64,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,80,80],f32> | |
| %3236 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3237 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_722 = torch.constant.int 12 | |
| %3238 = torch.aten.item %3236 : !torch.vtensor<[],f32> -> !torch.float | |
| %3239 = torch.aten.item %3237 : !torch.vtensor<[],si8> -> !torch.int | |
| %3240 = torch.aten.quantize_per_tensor %3235, %3238, %3239, %int12_722 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3241 = torch.aten.int_repr %3240 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
| %3242 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3243 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3244 = torch.aten.item %3242 : !torch.vtensor<[],f32> -> !torch.float | |
| %3245 = torch.aten.item %3243 : !torch.vtensor<[],si8> -> !torch.int | |
| %3246 = torch.aten._make_per_tensor_quantized_tensor %3241, %3244, %3245 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3247 = torch.aten.dequantize.self %3246 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3248 = torch.aten.sigmoid %3247 : !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3249 = torch.aten.mul.Tensor %3247, %3248 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3250 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3251 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_723 = torch.constant.int 12 | |
| %3252 = torch.aten.item %3250 : !torch.vtensor<[],f32> -> !torch.float | |
| %3253 = torch.aten.item %3251 : !torch.vtensor<[],si8> -> !torch.int | |
| %3254 = torch.aten.quantize_per_tensor %3249, %3252, %3253, %int12_723 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3255 = torch.aten.int_repr %3254 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
| %3256 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3257 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3258 = torch.aten.item %3256 : !torch.vtensor<[],f32> -> !torch.float | |
| %3259 = torch.aten.item %3257 : !torch.vtensor<[],si8> -> !torch.int | |
| %3260 = torch.aten._make_per_tensor_quantized_tensor %3255, %3258, %3259 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3261 = torch.aten.dequantize.self %3260 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3262 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3263 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_724 = torch.constant.int 12 | |
| %3264 = torch.aten.item %3262 : !torch.vtensor<[],f32> -> !torch.float | |
| %3265 = torch.aten.item %3263 : !torch.vtensor<[],si8> -> !torch.int | |
| %3266 = torch.aten.quantize_per_tensor %98, %3264, %3265, %int12_724 : !torch.vtensor<[80,80,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
| %3267 = torch.aten.int_repr %3266 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],si8> | |
| %3268 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3269 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3270 = torch.aten.item %3268 : !torch.vtensor<[],f32> -> !torch.float | |
| %3271 = torch.aten.item %3269 : !torch.vtensor<[],si8> -> !torch.int | |
| %3272 = torch.aten._make_per_tensor_quantized_tensor %3267, %3270, %3271 : !torch.vtensor<[80,80,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
| %3273 = torch.aten.dequantize.self %3272 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],f32> | |
| %3274 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3275 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_725 = torch.constant.int 12 | |
| %3276 = torch.aten.item %3274 : !torch.vtensor<[],f32> -> !torch.float | |
| %3277 = torch.aten.item %3275 : !torch.vtensor<[],si8> -> !torch.int | |
| %3278 = torch.aten.quantize_per_tensor %99, %3276, %3277, %int12_725 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3279 = torch.aten.int_repr %3278 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3280 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3281 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3282 = torch.aten.item %3280 : !torch.vtensor<[],f32> -> !torch.float | |
| %3283 = torch.aten.item %3281 : !torch.vtensor<[],si8> -> !torch.int | |
| %3284 = torch.aten._make_per_tensor_quantized_tensor %3279, %3282, %3283 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3285 = torch.aten.dequantize.self %3284 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int1_726 = torch.constant.int 1 | |
| %int1_727 = torch.constant.int 1 | |
| %int1_728 = torch.constant.int 1 | |
| %int1_729 = torch.constant.int 1 | |
| %int1_730 = torch.constant.int 1 | |
| %int1_731 = torch.constant.int 1 | |
| %int0_732 = torch.constant.int 0 | |
| %3286 = torch.prim.ListConstruct %int1_726, %int1_727 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3287 = torch.prim.ListConstruct %int1_728, %int1_729 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3288 = torch.prim.ListConstruct %int1_730, %int1_731 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3289 = torch.prim.ListConstruct %int0_732, %int0_732 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_733 = torch.constant.bool false | |
| %int1_734 = torch.constant.int 1 | |
| %3290 = torch.aten.convolution %3261, %3273, %3285, %3288, %3286, %3287, %false_733, %3289, %int1_734 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[80,80,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,80,80],f32> | |
| %3291 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3292 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_735 = torch.constant.int 12 | |
| %3293 = torch.aten.item %3291 : !torch.vtensor<[],f32> -> !torch.float | |
| %3294 = torch.aten.item %3292 : !torch.vtensor<[],si8> -> !torch.int | |
| %3295 = torch.aten.quantize_per_tensor %3290, %3293, %3294, %int12_735 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3296 = torch.aten.int_repr %3295 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
| %3297 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3298 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3299 = torch.aten.item %3297 : !torch.vtensor<[],f32> -> !torch.float | |
| %3300 = torch.aten.item %3298 : !torch.vtensor<[],si8> -> !torch.int | |
| %3301 = torch.aten._make_per_tensor_quantized_tensor %3296, %3299, %3300 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3302 = torch.aten.dequantize.self %3301 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3303 = torch.aten.sigmoid %3302 : !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3304 = torch.aten.mul.Tensor %3302, %3303 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[1,80,80,80],f32> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3305 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3306 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_736 = torch.constant.int 12 | |
| %3307 = torch.aten.item %3305 : !torch.vtensor<[],f32> -> !torch.float | |
| %3308 = torch.aten.item %3306 : !torch.vtensor<[],si8> -> !torch.int | |
| %3309 = torch.aten.quantize_per_tensor %3304, %3307, %3308, %int12_736 : !torch.vtensor<[1,80,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3310 = torch.aten.int_repr %3309 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],si8> | |
| %3311 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3312 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3313 = torch.aten.item %3311 : !torch.vtensor<[],f32> -> !torch.float | |
| %3314 = torch.aten.item %3312 : !torch.vtensor<[],si8> -> !torch.int | |
| %3315 = torch.aten._make_per_tensor_quantized_tensor %3310, %3313, %3314 : !torch.vtensor<[1,80,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,80,80],!torch.qint8> | |
| %3316 = torch.aten.dequantize.self %3315 : !torch.vtensor<[1,80,80,80],!torch.qint8> -> !torch.vtensor<[1,80,80,80],f32> | |
| %3317 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3318 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_737 = torch.constant.int 12 | |
| %3319 = torch.aten.item %3317 : !torch.vtensor<[],f32> -> !torch.float | |
| %3320 = torch.aten.item %3318 : !torch.vtensor<[],si8> -> !torch.int | |
| %3321 = torch.aten.quantize_per_tensor %100, %3319, %3320, %int12_737 : !torch.vtensor<[80,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
| %3322 = torch.aten.int_repr %3321 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],si8> | |
| %3323 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3324 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3325 = torch.aten.item %3323 : !torch.vtensor<[],f32> -> !torch.float | |
| %3326 = torch.aten.item %3324 : !torch.vtensor<[],si8> -> !torch.int | |
| %3327 = torch.aten._make_per_tensor_quantized_tensor %3322, %3325, %3326 : !torch.vtensor<[80,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
| %3328 = torch.aten.dequantize.self %3327 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],f32> | |
| %3329 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3330 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_738 = torch.constant.int 12 | |
| %3331 = torch.aten.item %3329 : !torch.vtensor<[],f32> -> !torch.float | |
| %3332 = torch.aten.item %3330 : !torch.vtensor<[],si8> -> !torch.int | |
| %3333 = torch.aten.quantize_per_tensor %101, %3331, %3332, %int12_738 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3334 = torch.aten.int_repr %3333 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3335 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3336 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3337 = torch.aten.item %3335 : !torch.vtensor<[],f32> -> !torch.float | |
| %3338 = torch.aten.item %3336 : !torch.vtensor<[],si8> -> !torch.int | |
| %3339 = torch.aten._make_per_tensor_quantized_tensor %3334, %3337, %3338 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3340 = torch.aten.dequantize.self %3339 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_739 = torch.constant.int 0 | |
| %int0_740 = torch.constant.int 0 | |
| %int1_741 = torch.constant.int 1 | |
| %int1_742 = torch.constant.int 1 | |
| %int1_743 = torch.constant.int 1 | |
| %int1_744 = torch.constant.int 1 | |
| %int0_745 = torch.constant.int 0 | |
| %3341 = torch.prim.ListConstruct %int0_739, %int0_740 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3342 = torch.prim.ListConstruct %int1_741, %int1_742 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3343 = torch.prim.ListConstruct %int1_743, %int1_744 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3344 = torch.prim.ListConstruct %int0_745, %int0_745 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_746 = torch.constant.bool false | |
| %int1_747 = torch.constant.int 1 | |
| %3345 = torch.aten.convolution %3316, %3328, %3340, %3343, %3341, %3342, %false_746, %3344, %int1_747 : !torch.vtensor<[1,80,80,80],f32>, !torch.vtensor<[80,80,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,80,80],f32> | |
| %3346 = torch.prim.ListConstruct %3206, %3345 : (!torch.vtensor<[1,64,80,80],f32>, !torch.vtensor<[1,80,80,80],f32>) -> !torch.list<vtensor> | |
| %int1_748 = torch.constant.int 1 | |
| %3347 = torch.aten.cat %3346, %int1_748 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,144,80,80],f32> | |
| %3348 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3349 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_749 = torch.constant.int 12 | |
| %3350 = torch.aten.item %3348 : !torch.vtensor<[],f32> -> !torch.float | |
| %3351 = torch.aten.item %3349 : !torch.vtensor<[],si8> -> !torch.int | |
| %3352 = torch.aten.quantize_per_tensor %3347, %3350, %3351, %int12_749 : !torch.vtensor<[1,144,80,80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,144,80,80],!torch.qint8> | |
| %3353 = torch.aten.int_repr %3352 : !torch.vtensor<[1,144,80,80],!torch.qint8> -> !torch.vtensor<[1,144,80,80],si8> | |
| %3354 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3355 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3356 = torch.aten.item %3354 : !torch.vtensor<[],f32> -> !torch.float | |
| %3357 = torch.aten.item %3355 : !torch.vtensor<[],si8> -> !torch.int | |
| %3358 = torch.aten._make_per_tensor_quantized_tensor %3353, %3356, %3357 : !torch.vtensor<[1,144,80,80],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,144,80,80],!torch.qint8> | |
| %3359 = torch.aten.dequantize.self %3358 : !torch.vtensor<[1,144,80,80],!torch.qint8> -> !torch.vtensor<[1,144,80,80],f32> | |
| %3360 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3361 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_750 = torch.constant.int 12 | |
| %3362 = torch.aten.item %3360 : !torch.vtensor<[],f32> -> !torch.float | |
| %3363 = torch.aten.item %3361 : !torch.vtensor<[],si8> -> !torch.int | |
| %3364 = torch.aten.quantize_per_tensor %102, %3362, %3363, %int12_750 : !torch.vtensor<[64,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,128,3,3],!torch.qint8> | |
| %3365 = torch.aten.int_repr %3364 : !torch.vtensor<[64,128,3,3],!torch.qint8> -> !torch.vtensor<[64,128,3,3],si8> | |
| %3366 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3367 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3368 = torch.aten.item %3366 : !torch.vtensor<[],f32> -> !torch.float | |
| %3369 = torch.aten.item %3367 : !torch.vtensor<[],si8> -> !torch.int | |
| %3370 = torch.aten._make_per_tensor_quantized_tensor %3365, %3368, %3369 : !torch.vtensor<[64,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,128,3,3],!torch.qint8> | |
| %3371 = torch.aten.dequantize.self %3370 : !torch.vtensor<[64,128,3,3],!torch.qint8> -> !torch.vtensor<[64,128,3,3],f32> | |
| %3372 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3373 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_751 = torch.constant.int 12 | |
| %3374 = torch.aten.item %3372 : !torch.vtensor<[],f32> -> !torch.float | |
| %3375 = torch.aten.item %3373 : !torch.vtensor<[],si8> -> !torch.int | |
| %3376 = torch.aten.quantize_per_tensor %103, %3374, %3375, %int12_751 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3377 = torch.aten.int_repr %3376 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3378 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3379 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3380 = torch.aten.item %3378 : !torch.vtensor<[],f32> -> !torch.float | |
| %3381 = torch.aten.item %3379 : !torch.vtensor<[],si8> -> !torch.int | |
| %3382 = torch.aten._make_per_tensor_quantized_tensor %3377, %3380, %3381 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3383 = torch.aten.dequantize.self %3382 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_752 = torch.constant.int 1 | |
| %int1_753 = torch.constant.int 1 | |
| %int1_754 = torch.constant.int 1 | |
| %int1_755 = torch.constant.int 1 | |
| %int1_756 = torch.constant.int 1 | |
| %int1_757 = torch.constant.int 1 | |
| %int0_758 = torch.constant.int 0 | |
| %3384 = torch.prim.ListConstruct %int1_752, %int1_753 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3385 = torch.prim.ListConstruct %int1_754, %int1_755 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3386 = torch.prim.ListConstruct %int1_756, %int1_757 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3387 = torch.prim.ListConstruct %int0_758, %int0_758 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_759 = torch.constant.bool false | |
| %int1_760 = torch.constant.int 1 | |
| %3388 = torch.aten.convolution %2759, %3371, %3383, %3386, %3384, %3385, %false_759, %3387, %int1_760 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[64,128,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %3389 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3390 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_761 = torch.constant.int 12 | |
| %3391 = torch.aten.item %3389 : !torch.vtensor<[],f32> -> !torch.float | |
| %3392 = torch.aten.item %3390 : !torch.vtensor<[],si8> -> !torch.int | |
| %3393 = torch.aten.quantize_per_tensor %3388, %3391, %3392, %int12_761 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3394 = torch.aten.int_repr %3393 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %3395 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3396 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3397 = torch.aten.item %3395 : !torch.vtensor<[],f32> -> !torch.float | |
| %3398 = torch.aten.item %3396 : !torch.vtensor<[],si8> -> !torch.int | |
| %3399 = torch.aten._make_per_tensor_quantized_tensor %3394, %3397, %3398 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3400 = torch.aten.dequantize.self %3399 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3401 = torch.aten.sigmoid %3400 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3402 = torch.aten.mul.Tensor %3400, %3401 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3403 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3404 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_762 = torch.constant.int 12 | |
| %3405 = torch.aten.item %3403 : !torch.vtensor<[],f32> -> !torch.float | |
| %3406 = torch.aten.item %3404 : !torch.vtensor<[],si8> -> !torch.int | |
| %3407 = torch.aten.quantize_per_tensor %3402, %3405, %3406, %int12_762 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3408 = torch.aten.int_repr %3407 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %3409 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3410 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3411 = torch.aten.item %3409 : !torch.vtensor<[],f32> -> !torch.float | |
| %3412 = torch.aten.item %3410 : !torch.vtensor<[],si8> -> !torch.int | |
| %3413 = torch.aten._make_per_tensor_quantized_tensor %3408, %3411, %3412 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3414 = torch.aten.dequantize.self %3413 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3415 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3416 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_763 = torch.constant.int 12 | |
| %3417 = torch.aten.item %3415 : !torch.vtensor<[],f32> -> !torch.float | |
| %3418 = torch.aten.item %3416 : !torch.vtensor<[],si8> -> !torch.int | |
| %3419 = torch.aten.quantize_per_tensor %104, %3417, %3418, %int12_763 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3420 = torch.aten.int_repr %3419 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %3421 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3422 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3423 = torch.aten.item %3421 : !torch.vtensor<[],f32> -> !torch.float | |
| %3424 = torch.aten.item %3422 : !torch.vtensor<[],si8> -> !torch.int | |
| %3425 = torch.aten._make_per_tensor_quantized_tensor %3420, %3423, %3424 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3426 = torch.aten.dequantize.self %3425 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %3427 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3428 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_764 = torch.constant.int 12 | |
| %3429 = torch.aten.item %3427 : !torch.vtensor<[],f32> -> !torch.float | |
| %3430 = torch.aten.item %3428 : !torch.vtensor<[],si8> -> !torch.int | |
| %3431 = torch.aten.quantize_per_tensor %105, %3429, %3430, %int12_764 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3432 = torch.aten.int_repr %3431 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3433 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3434 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3435 = torch.aten.item %3433 : !torch.vtensor<[],f32> -> !torch.float | |
| %3436 = torch.aten.item %3434 : !torch.vtensor<[],si8> -> !torch.int | |
| %3437 = torch.aten._make_per_tensor_quantized_tensor %3432, %3435, %3436 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3438 = torch.aten.dequantize.self %3437 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_765 = torch.constant.int 1 | |
| %int1_766 = torch.constant.int 1 | |
| %int1_767 = torch.constant.int 1 | |
| %int1_768 = torch.constant.int 1 | |
| %int1_769 = torch.constant.int 1 | |
| %int1_770 = torch.constant.int 1 | |
| %int0_771 = torch.constant.int 0 | |
| %3439 = torch.prim.ListConstruct %int1_765, %int1_766 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3440 = torch.prim.ListConstruct %int1_767, %int1_768 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3441 = torch.prim.ListConstruct %int1_769, %int1_770 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3442 = torch.prim.ListConstruct %int0_771, %int0_771 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_772 = torch.constant.bool false | |
| %int1_773 = torch.constant.int 1 | |
| %3443 = torch.aten.convolution %3414, %3426, %3438, %3441, %3439, %3440, %false_772, %3442, %int1_773 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %3444 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3445 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_774 = torch.constant.int 12 | |
| %3446 = torch.aten.item %3444 : !torch.vtensor<[],f32> -> !torch.float | |
| %3447 = torch.aten.item %3445 : !torch.vtensor<[],si8> -> !torch.int | |
| %3448 = torch.aten.quantize_per_tensor %3443, %3446, %3447, %int12_774 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3449 = torch.aten.int_repr %3448 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %3450 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3451 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3452 = torch.aten.item %3450 : !torch.vtensor<[],f32> -> !torch.float | |
| %3453 = torch.aten.item %3451 : !torch.vtensor<[],si8> -> !torch.int | |
| %3454 = torch.aten._make_per_tensor_quantized_tensor %3449, %3452, %3453 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3455 = torch.aten.dequantize.self %3454 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3456 = torch.aten.sigmoid %3455 : !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3457 = torch.aten.mul.Tensor %3455, %3456 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,64,40,40],f32> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3458 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3459 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_775 = torch.constant.int 12 | |
| %3460 = torch.aten.item %3458 : !torch.vtensor<[],f32> -> !torch.float | |
| %3461 = torch.aten.item %3459 : !torch.vtensor<[],si8> -> !torch.int | |
| %3462 = torch.aten.quantize_per_tensor %3457, %3460, %3461, %int12_775 : !torch.vtensor<[1,64,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3463 = torch.aten.int_repr %3462 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],si8> | |
| %3464 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3465 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3466 = torch.aten.item %3464 : !torch.vtensor<[],f32> -> !torch.float | |
| %3467 = torch.aten.item %3465 : !torch.vtensor<[],si8> -> !torch.int | |
| %3468 = torch.aten._make_per_tensor_quantized_tensor %3463, %3466, %3467 : !torch.vtensor<[1,64,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,40,40],!torch.qint8> | |
| %3469 = torch.aten.dequantize.self %3468 : !torch.vtensor<[1,64,40,40],!torch.qint8> -> !torch.vtensor<[1,64,40,40],f32> | |
| %3470 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3471 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_776 = torch.constant.int 12 | |
| %3472 = torch.aten.item %3470 : !torch.vtensor<[],f32> -> !torch.float | |
| %3473 = torch.aten.item %3471 : !torch.vtensor<[],si8> -> !torch.int | |
| %3474 = torch.aten.quantize_per_tensor %106, %3472, %3473, %int12_776 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %3475 = torch.aten.int_repr %3474 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
| %3476 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3477 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3478 = torch.aten.item %3476 : !torch.vtensor<[],f32> -> !torch.float | |
| %3479 = torch.aten.item %3477 : !torch.vtensor<[],si8> -> !torch.int | |
| %3480 = torch.aten._make_per_tensor_quantized_tensor %3475, %3478, %3479 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %3481 = torch.aten.dequantize.self %3480 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
| %3482 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3483 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_777 = torch.constant.int 12 | |
| %3484 = torch.aten.item %3482 : !torch.vtensor<[],f32> -> !torch.float | |
| %3485 = torch.aten.item %3483 : !torch.vtensor<[],si8> -> !torch.int | |
| %3486 = torch.aten.quantize_per_tensor %107, %3484, %3485, %int12_777 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3487 = torch.aten.int_repr %3486 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3488 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3489 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3490 = torch.aten.item %3488 : !torch.vtensor<[],f32> -> !torch.float | |
| %3491 = torch.aten.item %3489 : !torch.vtensor<[],si8> -> !torch.int | |
| %3492 = torch.aten._make_per_tensor_quantized_tensor %3487, %3490, %3491 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3493 = torch.aten.dequantize.self %3492 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_778 = torch.constant.int 0 | |
| %int0_779 = torch.constant.int 0 | |
| %int1_780 = torch.constant.int 1 | |
| %int1_781 = torch.constant.int 1 | |
| %int1_782 = torch.constant.int 1 | |
| %int1_783 = torch.constant.int 1 | |
| %int0_784 = torch.constant.int 0 | |
| %3494 = torch.prim.ListConstruct %int0_778, %int0_779 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3495 = torch.prim.ListConstruct %int1_780, %int1_781 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3496 = torch.prim.ListConstruct %int1_782, %int1_783 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3497 = torch.prim.ListConstruct %int0_784, %int0_784 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_785 = torch.constant.bool false | |
| %int1_786 = torch.constant.int 1 | |
| %3498 = torch.aten.convolution %3469, %3481, %3493, %3496, %3494, %3495, %false_785, %3497, %int1_786 : !torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,40,40],f32> | |
| %3499 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3500 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_787 = torch.constant.int 12 | |
| %3501 = torch.aten.item %3499 : !torch.vtensor<[],f32> -> !torch.float | |
| %3502 = torch.aten.item %3500 : !torch.vtensor<[],si8> -> !torch.int | |
| %3503 = torch.aten.quantize_per_tensor %108, %3501, %3502, %int12_787 : !torch.vtensor<[80,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,128,3,3],!torch.qint8> | |
| %3504 = torch.aten.int_repr %3503 : !torch.vtensor<[80,128,3,3],!torch.qint8> -> !torch.vtensor<[80,128,3,3],si8> | |
| %3505 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3506 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3507 = torch.aten.item %3505 : !torch.vtensor<[],f32> -> !torch.float | |
| %3508 = torch.aten.item %3506 : !torch.vtensor<[],si8> -> !torch.int | |
| %3509 = torch.aten._make_per_tensor_quantized_tensor %3504, %3507, %3508 : !torch.vtensor<[80,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,128,3,3],!torch.qint8> | |
| %3510 = torch.aten.dequantize.self %3509 : !torch.vtensor<[80,128,3,3],!torch.qint8> -> !torch.vtensor<[80,128,3,3],f32> | |
| %3511 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3512 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_788 = torch.constant.int 12 | |
| %3513 = torch.aten.item %3511 : !torch.vtensor<[],f32> -> !torch.float | |
| %3514 = torch.aten.item %3512 : !torch.vtensor<[],si8> -> !torch.int | |
| %3515 = torch.aten.quantize_per_tensor %109, %3513, %3514, %int12_788 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3516 = torch.aten.int_repr %3515 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3517 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3518 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3519 = torch.aten.item %3517 : !torch.vtensor<[],f32> -> !torch.float | |
| %3520 = torch.aten.item %3518 : !torch.vtensor<[],si8> -> !torch.int | |
| %3521 = torch.aten._make_per_tensor_quantized_tensor %3516, %3519, %3520 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3522 = torch.aten.dequantize.self %3521 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int1_789 = torch.constant.int 1 | |
| %int1_790 = torch.constant.int 1 | |
| %int1_791 = torch.constant.int 1 | |
| %int1_792 = torch.constant.int 1 | |
| %int1_793 = torch.constant.int 1 | |
| %int1_794 = torch.constant.int 1 | |
| %int0_795 = torch.constant.int 0 | |
| %3523 = torch.prim.ListConstruct %int1_789, %int1_790 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3524 = torch.prim.ListConstruct %int1_791, %int1_792 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3525 = torch.prim.ListConstruct %int1_793, %int1_794 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3526 = torch.prim.ListConstruct %int0_795, %int0_795 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_796 = torch.constant.bool false | |
| %int1_797 = torch.constant.int 1 | |
| %3527 = torch.aten.convolution %2759, %3510, %3522, %3525, %3523, %3524, %false_796, %3526, %int1_797 : !torch.vtensor<[1,128,40,40],f32>, !torch.vtensor<[80,128,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,40,40],f32> | |
| %3528 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3529 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_798 = torch.constant.int 12 | |
| %3530 = torch.aten.item %3528 : !torch.vtensor<[],f32> -> !torch.float | |
| %3531 = torch.aten.item %3529 : !torch.vtensor<[],si8> -> !torch.int | |
| %3532 = torch.aten.quantize_per_tensor %3527, %3530, %3531, %int12_798 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3533 = torch.aten.int_repr %3532 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
| %3534 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3535 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3536 = torch.aten.item %3534 : !torch.vtensor<[],f32> -> !torch.float | |
| %3537 = torch.aten.item %3535 : !torch.vtensor<[],si8> -> !torch.int | |
| %3538 = torch.aten._make_per_tensor_quantized_tensor %3533, %3536, %3537 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3539 = torch.aten.dequantize.self %3538 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3540 = torch.aten.sigmoid %3539 : !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3541 = torch.aten.mul.Tensor %3539, %3540 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3542 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3543 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_799 = torch.constant.int 12 | |
| %3544 = torch.aten.item %3542 : !torch.vtensor<[],f32> -> !torch.float | |
| %3545 = torch.aten.item %3543 : !torch.vtensor<[],si8> -> !torch.int | |
| %3546 = torch.aten.quantize_per_tensor %3541, %3544, %3545, %int12_799 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3547 = torch.aten.int_repr %3546 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
| %3548 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3549 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3550 = torch.aten.item %3548 : !torch.vtensor<[],f32> -> !torch.float | |
| %3551 = torch.aten.item %3549 : !torch.vtensor<[],si8> -> !torch.int | |
| %3552 = torch.aten._make_per_tensor_quantized_tensor %3547, %3550, %3551 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3553 = torch.aten.dequantize.self %3552 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3554 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3555 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_800 = torch.constant.int 12 | |
| %3556 = torch.aten.item %3554 : !torch.vtensor<[],f32> -> !torch.float | |
| %3557 = torch.aten.item %3555 : !torch.vtensor<[],si8> -> !torch.int | |
| %3558 = torch.aten.quantize_per_tensor %110, %3556, %3557, %int12_800 : !torch.vtensor<[80,80,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
| %3559 = torch.aten.int_repr %3558 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],si8> | |
| %3560 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3561 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3562 = torch.aten.item %3560 : !torch.vtensor<[],f32> -> !torch.float | |
| %3563 = torch.aten.item %3561 : !torch.vtensor<[],si8> -> !torch.int | |
| %3564 = torch.aten._make_per_tensor_quantized_tensor %3559, %3562, %3563 : !torch.vtensor<[80,80,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
| %3565 = torch.aten.dequantize.self %3564 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],f32> | |
| %3566 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3567 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_801 = torch.constant.int 12 | |
| %3568 = torch.aten.item %3566 : !torch.vtensor<[],f32> -> !torch.float | |
| %3569 = torch.aten.item %3567 : !torch.vtensor<[],si8> -> !torch.int | |
| %3570 = torch.aten.quantize_per_tensor %111, %3568, %3569, %int12_801 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3571 = torch.aten.int_repr %3570 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3572 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3573 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3574 = torch.aten.item %3572 : !torch.vtensor<[],f32> -> !torch.float | |
| %3575 = torch.aten.item %3573 : !torch.vtensor<[],si8> -> !torch.int | |
| %3576 = torch.aten._make_per_tensor_quantized_tensor %3571, %3574, %3575 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3577 = torch.aten.dequantize.self %3576 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int1_802 = torch.constant.int 1 | |
| %int1_803 = torch.constant.int 1 | |
| %int1_804 = torch.constant.int 1 | |
| %int1_805 = torch.constant.int 1 | |
| %int1_806 = torch.constant.int 1 | |
| %int1_807 = torch.constant.int 1 | |
| %int0_808 = torch.constant.int 0 | |
| %3578 = torch.prim.ListConstruct %int1_802, %int1_803 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3579 = torch.prim.ListConstruct %int1_804, %int1_805 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3580 = torch.prim.ListConstruct %int1_806, %int1_807 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3581 = torch.prim.ListConstruct %int0_808, %int0_808 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_809 = torch.constant.bool false | |
| %int1_810 = torch.constant.int 1 | |
| %3582 = torch.aten.convolution %3553, %3565, %3577, %3580, %3578, %3579, %false_809, %3581, %int1_810 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[80,80,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,40,40],f32> | |
| %3583 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3584 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_811 = torch.constant.int 12 | |
| %3585 = torch.aten.item %3583 : !torch.vtensor<[],f32> -> !torch.float | |
| %3586 = torch.aten.item %3584 : !torch.vtensor<[],si8> -> !torch.int | |
| %3587 = torch.aten.quantize_per_tensor %3582, %3585, %3586, %int12_811 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3588 = torch.aten.int_repr %3587 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
| %3589 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3590 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3591 = torch.aten.item %3589 : !torch.vtensor<[],f32> -> !torch.float | |
| %3592 = torch.aten.item %3590 : !torch.vtensor<[],si8> -> !torch.int | |
| %3593 = torch.aten._make_per_tensor_quantized_tensor %3588, %3591, %3592 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3594 = torch.aten.dequantize.self %3593 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3595 = torch.aten.sigmoid %3594 : !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3596 = torch.aten.mul.Tensor %3594, %3595 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[1,80,40,40],f32> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3597 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3598 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_812 = torch.constant.int 12 | |
| %3599 = torch.aten.item %3597 : !torch.vtensor<[],f32> -> !torch.float | |
| %3600 = torch.aten.item %3598 : !torch.vtensor<[],si8> -> !torch.int | |
| %3601 = torch.aten.quantize_per_tensor %3596, %3599, %3600, %int12_812 : !torch.vtensor<[1,80,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3602 = torch.aten.int_repr %3601 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],si8> | |
| %3603 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3604 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3605 = torch.aten.item %3603 : !torch.vtensor<[],f32> -> !torch.float | |
| %3606 = torch.aten.item %3604 : !torch.vtensor<[],si8> -> !torch.int | |
| %3607 = torch.aten._make_per_tensor_quantized_tensor %3602, %3605, %3606 : !torch.vtensor<[1,80,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,40,40],!torch.qint8> | |
| %3608 = torch.aten.dequantize.self %3607 : !torch.vtensor<[1,80,40,40],!torch.qint8> -> !torch.vtensor<[1,80,40,40],f32> | |
| %3609 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3610 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_813 = torch.constant.int 12 | |
| %3611 = torch.aten.item %3609 : !torch.vtensor<[],f32> -> !torch.float | |
| %3612 = torch.aten.item %3610 : !torch.vtensor<[],si8> -> !torch.int | |
| %3613 = torch.aten.quantize_per_tensor %112, %3611, %3612, %int12_813 : !torch.vtensor<[80,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
| %3614 = torch.aten.int_repr %3613 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],si8> | |
| %3615 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3616 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3617 = torch.aten.item %3615 : !torch.vtensor<[],f32> -> !torch.float | |
| %3618 = torch.aten.item %3616 : !torch.vtensor<[],si8> -> !torch.int | |
| %3619 = torch.aten._make_per_tensor_quantized_tensor %3614, %3617, %3618 : !torch.vtensor<[80,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
| %3620 = torch.aten.dequantize.self %3619 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],f32> | |
| %3621 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3622 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_814 = torch.constant.int 12 | |
| %3623 = torch.aten.item %3621 : !torch.vtensor<[],f32> -> !torch.float | |
| %3624 = torch.aten.item %3622 : !torch.vtensor<[],si8> -> !torch.int | |
| %3625 = torch.aten.quantize_per_tensor %113, %3623, %3624, %int12_814 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3626 = torch.aten.int_repr %3625 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3627 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3628 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3629 = torch.aten.item %3627 : !torch.vtensor<[],f32> -> !torch.float | |
| %3630 = torch.aten.item %3628 : !torch.vtensor<[],si8> -> !torch.int | |
| %3631 = torch.aten._make_per_tensor_quantized_tensor %3626, %3629, %3630 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3632 = torch.aten.dequantize.self %3631 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_815 = torch.constant.int 0 | |
| %int0_816 = torch.constant.int 0 | |
| %int1_817 = torch.constant.int 1 | |
| %int1_818 = torch.constant.int 1 | |
| %int1_819 = torch.constant.int 1 | |
| %int1_820 = torch.constant.int 1 | |
| %int0_821 = torch.constant.int 0 | |
| %3633 = torch.prim.ListConstruct %int0_815, %int0_816 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3634 = torch.prim.ListConstruct %int1_817, %int1_818 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3635 = torch.prim.ListConstruct %int1_819, %int1_820 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3636 = torch.prim.ListConstruct %int0_821, %int0_821 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_822 = torch.constant.bool false | |
| %int1_823 = torch.constant.int 1 | |
| %3637 = torch.aten.convolution %3608, %3620, %3632, %3635, %3633, %3634, %false_822, %3636, %int1_823 : !torch.vtensor<[1,80,40,40],f32>, !torch.vtensor<[80,80,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,40,40],f32> | |
| %3638 = torch.prim.ListConstruct %3498, %3637 : (!torch.vtensor<[1,64,40,40],f32>, !torch.vtensor<[1,80,40,40],f32>) -> !torch.list<vtensor> | |
| %int1_824 = torch.constant.int 1 | |
| %3639 = torch.aten.cat %3638, %int1_824 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,144,40,40],f32> | |
| %3640 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3641 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_825 = torch.constant.int 12 | |
| %3642 = torch.aten.item %3640 : !torch.vtensor<[],f32> -> !torch.float | |
| %3643 = torch.aten.item %3641 : !torch.vtensor<[],si8> -> !torch.int | |
| %3644 = torch.aten.quantize_per_tensor %3639, %3642, %3643, %int12_825 : !torch.vtensor<[1,144,40,40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,144,40,40],!torch.qint8> | |
| %3645 = torch.aten.int_repr %3644 : !torch.vtensor<[1,144,40,40],!torch.qint8> -> !torch.vtensor<[1,144,40,40],si8> | |
| %3646 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3647 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3648 = torch.aten.item %3646 : !torch.vtensor<[],f32> -> !torch.float | |
| %3649 = torch.aten.item %3647 : !torch.vtensor<[],si8> -> !torch.int | |
| %3650 = torch.aten._make_per_tensor_quantized_tensor %3645, %3648, %3649 : !torch.vtensor<[1,144,40,40],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,144,40,40],!torch.qint8> | |
| %3651 = torch.aten.dequantize.self %3650 : !torch.vtensor<[1,144,40,40],!torch.qint8> -> !torch.vtensor<[1,144,40,40],f32> | |
| %3652 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3653 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_826 = torch.constant.int 12 | |
| %3654 = torch.aten.item %3652 : !torch.vtensor<[],f32> -> !torch.float | |
| %3655 = torch.aten.item %3653 : !torch.vtensor<[],si8> -> !torch.int | |
| %3656 = torch.aten.quantize_per_tensor %114, %3654, %3655, %int12_826 : !torch.vtensor<[64,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,256,3,3],!torch.qint8> | |
| %3657 = torch.aten.int_repr %3656 : !torch.vtensor<[64,256,3,3],!torch.qint8> -> !torch.vtensor<[64,256,3,3],si8> | |
| %3658 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3659 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3660 = torch.aten.item %3658 : !torch.vtensor<[],f32> -> !torch.float | |
| %3661 = torch.aten.item %3659 : !torch.vtensor<[],si8> -> !torch.int | |
| %3662 = torch.aten._make_per_tensor_quantized_tensor %3657, %3660, %3661 : !torch.vtensor<[64,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,256,3,3],!torch.qint8> | |
| %3663 = torch.aten.dequantize.self %3662 : !torch.vtensor<[64,256,3,3],!torch.qint8> -> !torch.vtensor<[64,256,3,3],f32> | |
| %3664 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3665 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_827 = torch.constant.int 12 | |
| %3666 = torch.aten.item %3664 : !torch.vtensor<[],f32> -> !torch.float | |
| %3667 = torch.aten.item %3665 : !torch.vtensor<[],si8> -> !torch.int | |
| %3668 = torch.aten.quantize_per_tensor %115, %3666, %3667, %int12_827 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3669 = torch.aten.int_repr %3668 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3670 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3671 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3672 = torch.aten.item %3670 : !torch.vtensor<[],f32> -> !torch.float | |
| %3673 = torch.aten.item %3671 : !torch.vtensor<[],si8> -> !torch.int | |
| %3674 = torch.aten._make_per_tensor_quantized_tensor %3669, %3672, %3673 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3675 = torch.aten.dequantize.self %3674 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_828 = torch.constant.int 1 | |
| %int1_829 = torch.constant.int 1 | |
| %int1_830 = torch.constant.int 1 | |
| %int1_831 = torch.constant.int 1 | |
| %int1_832 = torch.constant.int 1 | |
| %int1_833 = torch.constant.int 1 | |
| %int0_834 = torch.constant.int 0 | |
| %3676 = torch.prim.ListConstruct %int1_828, %int1_829 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3677 = torch.prim.ListConstruct %int1_830, %int1_831 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3678 = torch.prim.ListConstruct %int1_832, %int1_833 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3679 = torch.prim.ListConstruct %int0_834, %int0_834 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_835 = torch.constant.bool false | |
| %int1_836 = torch.constant.int 1 | |
| %3680 = torch.aten.convolution %3066, %3663, %3675, %3678, %3676, %3677, %false_835, %3679, %int1_836 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[64,256,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,20,20],f32> | |
| %3681 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3682 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_837 = torch.constant.int 12 | |
| %3683 = torch.aten.item %3681 : !torch.vtensor<[],f32> -> !torch.float | |
| %3684 = torch.aten.item %3682 : !torch.vtensor<[],si8> -> !torch.int | |
| %3685 = torch.aten.quantize_per_tensor %3680, %3683, %3684, %int12_837 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3686 = torch.aten.int_repr %3685 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
| %3687 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3688 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3689 = torch.aten.item %3687 : !torch.vtensor<[],f32> -> !torch.float | |
| %3690 = torch.aten.item %3688 : !torch.vtensor<[],si8> -> !torch.int | |
| %3691 = torch.aten._make_per_tensor_quantized_tensor %3686, %3689, %3690 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3692 = torch.aten.dequantize.self %3691 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3693 = torch.aten.sigmoid %3692 : !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3694 = torch.aten.mul.Tensor %3692, %3693 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3695 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3696 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_838 = torch.constant.int 12 | |
| %3697 = torch.aten.item %3695 : !torch.vtensor<[],f32> -> !torch.float | |
| %3698 = torch.aten.item %3696 : !torch.vtensor<[],si8> -> !torch.int | |
| %3699 = torch.aten.quantize_per_tensor %3694, %3697, %3698, %int12_838 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3700 = torch.aten.int_repr %3699 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
| %3701 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3702 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3703 = torch.aten.item %3701 : !torch.vtensor<[],f32> -> !torch.float | |
| %3704 = torch.aten.item %3702 : !torch.vtensor<[],si8> -> !torch.int | |
| %3705 = torch.aten._make_per_tensor_quantized_tensor %3700, %3703, %3704 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3706 = torch.aten.dequantize.self %3705 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3707 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3708 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_839 = torch.constant.int 12 | |
| %3709 = torch.aten.item %3707 : !torch.vtensor<[],f32> -> !torch.float | |
| %3710 = torch.aten.item %3708 : !torch.vtensor<[],si8> -> !torch.int | |
| %3711 = torch.aten.quantize_per_tensor %116, %3709, %3710, %int12_839 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3712 = torch.aten.int_repr %3711 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %3713 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3714 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3715 = torch.aten.item %3713 : !torch.vtensor<[],f32> -> !torch.float | |
| %3716 = torch.aten.item %3714 : !torch.vtensor<[],si8> -> !torch.int | |
| %3717 = torch.aten._make_per_tensor_quantized_tensor %3712, %3715, %3716 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %3718 = torch.aten.dequantize.self %3717 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %3719 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3720 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_840 = torch.constant.int 12 | |
| %3721 = torch.aten.item %3719 : !torch.vtensor<[],f32> -> !torch.float | |
| %3722 = torch.aten.item %3720 : !torch.vtensor<[],si8> -> !torch.int | |
| %3723 = torch.aten.quantize_per_tensor %117, %3721, %3722, %int12_840 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3724 = torch.aten.int_repr %3723 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3725 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3726 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3727 = torch.aten.item %3725 : !torch.vtensor<[],f32> -> !torch.float | |
| %3728 = torch.aten.item %3726 : !torch.vtensor<[],si8> -> !torch.int | |
| %3729 = torch.aten._make_per_tensor_quantized_tensor %3724, %3727, %3728 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3730 = torch.aten.dequantize.self %3729 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_841 = torch.constant.int 1 | |
| %int1_842 = torch.constant.int 1 | |
| %int1_843 = torch.constant.int 1 | |
| %int1_844 = torch.constant.int 1 | |
| %int1_845 = torch.constant.int 1 | |
| %int1_846 = torch.constant.int 1 | |
| %int0_847 = torch.constant.int 0 | |
| %3731 = torch.prim.ListConstruct %int1_841, %int1_842 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3732 = torch.prim.ListConstruct %int1_843, %int1_844 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3733 = torch.prim.ListConstruct %int1_845, %int1_846 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3734 = torch.prim.ListConstruct %int0_847, %int0_847 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_848 = torch.constant.bool false | |
| %int1_849 = torch.constant.int 1 | |
| %3735 = torch.aten.convolution %3706, %3718, %3730, %3733, %3731, %3732, %false_848, %3734, %int1_849 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,20,20],f32> | |
| %3736 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3737 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_850 = torch.constant.int 12 | |
| %3738 = torch.aten.item %3736 : !torch.vtensor<[],f32> -> !torch.float | |
| %3739 = torch.aten.item %3737 : !torch.vtensor<[],si8> -> !torch.int | |
| %3740 = torch.aten.quantize_per_tensor %3735, %3738, %3739, %int12_850 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3741 = torch.aten.int_repr %3740 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
| %3742 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3743 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3744 = torch.aten.item %3742 : !torch.vtensor<[],f32> -> !torch.float | |
| %3745 = torch.aten.item %3743 : !torch.vtensor<[],si8> -> !torch.int | |
| %3746 = torch.aten._make_per_tensor_quantized_tensor %3741, %3744, %3745 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3747 = torch.aten.dequantize.self %3746 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3748 = torch.aten.sigmoid %3747 : !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3749 = torch.aten.mul.Tensor %3747, %3748 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[1,64,20,20],f32> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3750 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3751 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_851 = torch.constant.int 12 | |
| %3752 = torch.aten.item %3750 : !torch.vtensor<[],f32> -> !torch.float | |
| %3753 = torch.aten.item %3751 : !torch.vtensor<[],si8> -> !torch.int | |
| %3754 = torch.aten.quantize_per_tensor %3749, %3752, %3753, %int12_851 : !torch.vtensor<[1,64,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3755 = torch.aten.int_repr %3754 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],si8> | |
| %3756 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3757 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3758 = torch.aten.item %3756 : !torch.vtensor<[],f32> -> !torch.float | |
| %3759 = torch.aten.item %3757 : !torch.vtensor<[],si8> -> !torch.int | |
| %3760 = torch.aten._make_per_tensor_quantized_tensor %3755, %3758, %3759 : !torch.vtensor<[1,64,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,20,20],!torch.qint8> | |
| %3761 = torch.aten.dequantize.self %3760 : !torch.vtensor<[1,64,20,20],!torch.qint8> -> !torch.vtensor<[1,64,20,20],f32> | |
| %3762 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3763 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_852 = torch.constant.int 12 | |
| %3764 = torch.aten.item %3762 : !torch.vtensor<[],f32> -> !torch.float | |
| %3765 = torch.aten.item %3763 : !torch.vtensor<[],si8> -> !torch.int | |
| %3766 = torch.aten.quantize_per_tensor %118, %3764, %3765, %int12_852 : !torch.vtensor<[64,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %3767 = torch.aten.int_repr %3766 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],si8> | |
| %3768 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3769 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3770 = torch.aten.item %3768 : !torch.vtensor<[],f32> -> !torch.float | |
| %3771 = torch.aten.item %3769 : !torch.vtensor<[],si8> -> !torch.int | |
| %3772 = torch.aten._make_per_tensor_quantized_tensor %3767, %3770, %3771 : !torch.vtensor<[64,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,1,1],!torch.qint8> | |
| %3773 = torch.aten.dequantize.self %3772 : !torch.vtensor<[64,64,1,1],!torch.qint8> -> !torch.vtensor<[64,64,1,1],f32> | |
| %3774 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3775 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_853 = torch.constant.int 12 | |
| %3776 = torch.aten.item %3774 : !torch.vtensor<[],f32> -> !torch.float | |
| %3777 = torch.aten.item %3775 : !torch.vtensor<[],si8> -> !torch.int | |
| %3778 = torch.aten.quantize_per_tensor %119, %3776, %3777, %int12_853 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3779 = torch.aten.int_repr %3778 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %3780 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3781 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3782 = torch.aten.item %3780 : !torch.vtensor<[],f32> -> !torch.float | |
| %3783 = torch.aten.item %3781 : !torch.vtensor<[],si8> -> !torch.int | |
| %3784 = torch.aten._make_per_tensor_quantized_tensor %3779, %3782, %3783 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %3785 = torch.aten.dequantize.self %3784 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_854 = torch.constant.int 0 | |
| %int0_855 = torch.constant.int 0 | |
| %int1_856 = torch.constant.int 1 | |
| %int1_857 = torch.constant.int 1 | |
| %int1_858 = torch.constant.int 1 | |
| %int1_859 = torch.constant.int 1 | |
| %int0_860 = torch.constant.int 0 | |
| %3786 = torch.prim.ListConstruct %int0_854, %int0_855 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3787 = torch.prim.ListConstruct %int1_856, %int1_857 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3788 = torch.prim.ListConstruct %int1_858, %int1_859 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3789 = torch.prim.ListConstruct %int0_860, %int0_860 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_861 = torch.constant.bool false | |
| %int1_862 = torch.constant.int 1 | |
| %3790 = torch.aten.convolution %3761, %3773, %3785, %3788, %3786, %3787, %false_861, %3789, %int1_862 : !torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[64,64,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,20,20],f32> | |
| %3791 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3792 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_863 = torch.constant.int 12 | |
| %3793 = torch.aten.item %3791 : !torch.vtensor<[],f32> -> !torch.float | |
| %3794 = torch.aten.item %3792 : !torch.vtensor<[],si8> -> !torch.int | |
| %3795 = torch.aten.quantize_per_tensor %120, %3793, %3794, %int12_863 : !torch.vtensor<[80,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,256,3,3],!torch.qint8> | |
| %3796 = torch.aten.int_repr %3795 : !torch.vtensor<[80,256,3,3],!torch.qint8> -> !torch.vtensor<[80,256,3,3],si8> | |
| %3797 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3798 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3799 = torch.aten.item %3797 : !torch.vtensor<[],f32> -> !torch.float | |
| %3800 = torch.aten.item %3798 : !torch.vtensor<[],si8> -> !torch.int | |
| %3801 = torch.aten._make_per_tensor_quantized_tensor %3796, %3799, %3800 : !torch.vtensor<[80,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,256,3,3],!torch.qint8> | |
| %3802 = torch.aten.dequantize.self %3801 : !torch.vtensor<[80,256,3,3],!torch.qint8> -> !torch.vtensor<[80,256,3,3],f32> | |
| %3803 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3804 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_864 = torch.constant.int 12 | |
| %3805 = torch.aten.item %3803 : !torch.vtensor<[],f32> -> !torch.float | |
| %3806 = torch.aten.item %3804 : !torch.vtensor<[],si8> -> !torch.int | |
| %3807 = torch.aten.quantize_per_tensor %121, %3805, %3806, %int12_864 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3808 = torch.aten.int_repr %3807 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3809 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3810 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3811 = torch.aten.item %3809 : !torch.vtensor<[],f32> -> !torch.float | |
| %3812 = torch.aten.item %3810 : !torch.vtensor<[],si8> -> !torch.int | |
| %3813 = torch.aten._make_per_tensor_quantized_tensor %3808, %3811, %3812 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3814 = torch.aten.dequantize.self %3813 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int1_865 = torch.constant.int 1 | |
| %int1_866 = torch.constant.int 1 | |
| %int1_867 = torch.constant.int 1 | |
| %int1_868 = torch.constant.int 1 | |
| %int1_869 = torch.constant.int 1 | |
| %int1_870 = torch.constant.int 1 | |
| %int0_871 = torch.constant.int 0 | |
| %3815 = torch.prim.ListConstruct %int1_865, %int1_866 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3816 = torch.prim.ListConstruct %int1_867, %int1_868 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3817 = torch.prim.ListConstruct %int1_869, %int1_870 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3818 = torch.prim.ListConstruct %int0_871, %int0_871 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_872 = torch.constant.bool false | |
| %int1_873 = torch.constant.int 1 | |
| %3819 = torch.aten.convolution %3066, %3802, %3814, %3817, %3815, %3816, %false_872, %3818, %int1_873 : !torch.vtensor<[1,256,20,20],f32>, !torch.vtensor<[80,256,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,20,20],f32> | |
| %3820 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3821 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_874 = torch.constant.int 12 | |
| %3822 = torch.aten.item %3820 : !torch.vtensor<[],f32> -> !torch.float | |
| %3823 = torch.aten.item %3821 : !torch.vtensor<[],si8> -> !torch.int | |
| %3824 = torch.aten.quantize_per_tensor %3819, %3822, %3823, %int12_874 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3825 = torch.aten.int_repr %3824 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
| %3826 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3827 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3828 = torch.aten.item %3826 : !torch.vtensor<[],f32> -> !torch.float | |
| %3829 = torch.aten.item %3827 : !torch.vtensor<[],si8> -> !torch.int | |
| %3830 = torch.aten._make_per_tensor_quantized_tensor %3825, %3828, %3829 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3831 = torch.aten.dequantize.self %3830 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3832 = torch.aten.sigmoid %3831 : !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3833 = torch.aten.mul.Tensor %3831, %3832 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3834 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3835 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_875 = torch.constant.int 12 | |
| %3836 = torch.aten.item %3834 : !torch.vtensor<[],f32> -> !torch.float | |
| %3837 = torch.aten.item %3835 : !torch.vtensor<[],si8> -> !torch.int | |
| %3838 = torch.aten.quantize_per_tensor %3833, %3836, %3837, %int12_875 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3839 = torch.aten.int_repr %3838 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
| %3840 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3841 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3842 = torch.aten.item %3840 : !torch.vtensor<[],f32> -> !torch.float | |
| %3843 = torch.aten.item %3841 : !torch.vtensor<[],si8> -> !torch.int | |
| %3844 = torch.aten._make_per_tensor_quantized_tensor %3839, %3842, %3843 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3845 = torch.aten.dequantize.self %3844 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3846 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3847 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_876 = torch.constant.int 12 | |
| %3848 = torch.aten.item %3846 : !torch.vtensor<[],f32> -> !torch.float | |
| %3849 = torch.aten.item %3847 : !torch.vtensor<[],si8> -> !torch.int | |
| %3850 = torch.aten.quantize_per_tensor %122, %3848, %3849, %int12_876 : !torch.vtensor<[80,80,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
| %3851 = torch.aten.int_repr %3850 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],si8> | |
| %3852 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3853 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3854 = torch.aten.item %3852 : !torch.vtensor<[],f32> -> !torch.float | |
| %3855 = torch.aten.item %3853 : !torch.vtensor<[],si8> -> !torch.int | |
| %3856 = torch.aten._make_per_tensor_quantized_tensor %3851, %3854, %3855 : !torch.vtensor<[80,80,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,3,3],!torch.qint8> | |
| %3857 = torch.aten.dequantize.self %3856 : !torch.vtensor<[80,80,3,3],!torch.qint8> -> !torch.vtensor<[80,80,3,3],f32> | |
| %3858 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3859 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_877 = torch.constant.int 12 | |
| %3860 = torch.aten.item %3858 : !torch.vtensor<[],f32> -> !torch.float | |
| %3861 = torch.aten.item %3859 : !torch.vtensor<[],si8> -> !torch.int | |
| %3862 = torch.aten.quantize_per_tensor %123, %3860, %3861, %int12_877 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3863 = torch.aten.int_repr %3862 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3864 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3865 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3866 = torch.aten.item %3864 : !torch.vtensor<[],f32> -> !torch.float | |
| %3867 = torch.aten.item %3865 : !torch.vtensor<[],si8> -> !torch.int | |
| %3868 = torch.aten._make_per_tensor_quantized_tensor %3863, %3866, %3867 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3869 = torch.aten.dequantize.self %3868 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int1_878 = torch.constant.int 1 | |
| %int1_879 = torch.constant.int 1 | |
| %int1_880 = torch.constant.int 1 | |
| %int1_881 = torch.constant.int 1 | |
| %int1_882 = torch.constant.int 1 | |
| %int1_883 = torch.constant.int 1 | |
| %int0_884 = torch.constant.int 0 | |
| %3870 = torch.prim.ListConstruct %int1_878, %int1_879 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3871 = torch.prim.ListConstruct %int1_880, %int1_881 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3872 = torch.prim.ListConstruct %int1_882, %int1_883 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3873 = torch.prim.ListConstruct %int0_884, %int0_884 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_885 = torch.constant.bool false | |
| %int1_886 = torch.constant.int 1 | |
| %3874 = torch.aten.convolution %3845, %3857, %3869, %3872, %3870, %3871, %false_885, %3873, %int1_886 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[80,80,3,3],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,20,20],f32> | |
| %3875 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3876 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_887 = torch.constant.int 12 | |
| %3877 = torch.aten.item %3875 : !torch.vtensor<[],f32> -> !torch.float | |
| %3878 = torch.aten.item %3876 : !torch.vtensor<[],si8> -> !torch.int | |
| %3879 = torch.aten.quantize_per_tensor %3874, %3877, %3878, %int12_887 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3880 = torch.aten.int_repr %3879 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
| %3881 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3882 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3883 = torch.aten.item %3881 : !torch.vtensor<[],f32> -> !torch.float | |
| %3884 = torch.aten.item %3882 : !torch.vtensor<[],si8> -> !torch.int | |
| %3885 = torch.aten._make_per_tensor_quantized_tensor %3880, %3883, %3884 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3886 = torch.aten.dequantize.self %3885 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3887 = torch.aten.sigmoid %3886 : !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3888 = torch.aten.mul.Tensor %3886, %3887 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[1,80,20,20],f32> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3889 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3890 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_888 = torch.constant.int 12 | |
| %3891 = torch.aten.item %3889 : !torch.vtensor<[],f32> -> !torch.float | |
| %3892 = torch.aten.item %3890 : !torch.vtensor<[],si8> -> !torch.int | |
| %3893 = torch.aten.quantize_per_tensor %3888, %3891, %3892, %int12_888 : !torch.vtensor<[1,80,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3894 = torch.aten.int_repr %3893 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],si8> | |
| %3895 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3896 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3897 = torch.aten.item %3895 : !torch.vtensor<[],f32> -> !torch.float | |
| %3898 = torch.aten.item %3896 : !torch.vtensor<[],si8> -> !torch.int | |
| %3899 = torch.aten._make_per_tensor_quantized_tensor %3894, %3897, %3898 : !torch.vtensor<[1,80,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,20,20],!torch.qint8> | |
| %3900 = torch.aten.dequantize.self %3899 : !torch.vtensor<[1,80,20,20],!torch.qint8> -> !torch.vtensor<[1,80,20,20],f32> | |
| %3901 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3902 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_889 = torch.constant.int 12 | |
| %3903 = torch.aten.item %3901 : !torch.vtensor<[],f32> -> !torch.float | |
| %3904 = torch.aten.item %3902 : !torch.vtensor<[],si8> -> !torch.int | |
| %3905 = torch.aten.quantize_per_tensor %124, %3903, %3904, %int12_889 : !torch.vtensor<[80,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
| %3906 = torch.aten.int_repr %3905 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],si8> | |
| %3907 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3908 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3909 = torch.aten.item %3907 : !torch.vtensor<[],f32> -> !torch.float | |
| %3910 = torch.aten.item %3908 : !torch.vtensor<[],si8> -> !torch.int | |
| %3911 = torch.aten._make_per_tensor_quantized_tensor %3906, %3909, %3910 : !torch.vtensor<[80,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,80,1,1],!torch.qint8> | |
| %3912 = torch.aten.dequantize.self %3911 : !torch.vtensor<[80,80,1,1],!torch.qint8> -> !torch.vtensor<[80,80,1,1],f32> | |
| %3913 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3914 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_890 = torch.constant.int 12 | |
| %3915 = torch.aten.item %3913 : !torch.vtensor<[],f32> -> !torch.float | |
| %3916 = torch.aten.item %3914 : !torch.vtensor<[],si8> -> !torch.int | |
| %3917 = torch.aten.quantize_per_tensor %125, %3915, %3916, %int12_890 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3918 = torch.aten.int_repr %3917 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %3919 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3920 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3921 = torch.aten.item %3919 : !torch.vtensor<[],f32> -> !torch.float | |
| %3922 = torch.aten.item %3920 : !torch.vtensor<[],si8> -> !torch.int | |
| %3923 = torch.aten._make_per_tensor_quantized_tensor %3918, %3921, %3922 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %3924 = torch.aten.dequantize.self %3923 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_891 = torch.constant.int 0 | |
| %int0_892 = torch.constant.int 0 | |
| %int1_893 = torch.constant.int 1 | |
| %int1_894 = torch.constant.int 1 | |
| %int1_895 = torch.constant.int 1 | |
| %int1_896 = torch.constant.int 1 | |
| %int0_897 = torch.constant.int 0 | |
| %3925 = torch.prim.ListConstruct %int0_891, %int0_892 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3926 = torch.prim.ListConstruct %int1_893, %int1_894 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3927 = torch.prim.ListConstruct %int1_895, %int1_896 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3928 = torch.prim.ListConstruct %int0_897, %int0_897 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_898 = torch.constant.bool false | |
| %int1_899 = torch.constant.int 1 | |
| %3929 = torch.aten.convolution %3900, %3912, %3924, %3927, %3925, %3926, %false_898, %3928, %int1_899 : !torch.vtensor<[1,80,20,20],f32>, !torch.vtensor<[80,80,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,20,20],f32> | |
| %3930 = torch.prim.ListConstruct %3790, %3929 : (!torch.vtensor<[1,64,20,20],f32>, !torch.vtensor<[1,80,20,20],f32>) -> !torch.list<vtensor> | |
| %int1_900 = torch.constant.int 1 | |
| %3931 = torch.aten.cat %3930, %int1_900 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[1,144,20,20],f32> | |
| %3932 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3933 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_901 = torch.constant.int 12 | |
| %3934 = torch.aten.item %3932 : !torch.vtensor<[],f32> -> !torch.float | |
| %3935 = torch.aten.item %3933 : !torch.vtensor<[],si8> -> !torch.int | |
| %3936 = torch.aten.quantize_per_tensor %3931, %3934, %3935, %int12_901 : !torch.vtensor<[1,144,20,20],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,144,20,20],!torch.qint8> | |
| %3937 = torch.aten.int_repr %3936 : !torch.vtensor<[1,144,20,20],!torch.qint8> -> !torch.vtensor<[1,144,20,20],si8> | |
| %3938 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3939 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3940 = torch.aten.item %3938 : !torch.vtensor<[],f32> -> !torch.float | |
| %3941 = torch.aten.item %3939 : !torch.vtensor<[],si8> -> !torch.int | |
| %3942 = torch.aten._make_per_tensor_quantized_tensor %3937, %3940, %3941 : !torch.vtensor<[1,144,20,20],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,144,20,20],!torch.qint8> | |
| %3943 = torch.aten.dequantize.self %3942 : !torch.vtensor<[1,144,20,20],!torch.qint8> -> !torch.vtensor<[1,144,20,20],f32> | |
| %3944 = torch.vtensor.literal(dense<144> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %3945 = torch.vtensor.literal(dense<-1> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %3946 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_902 = torch.constant.int 0 | |
| %int0_903 = torch.constant.int 0 | |
| %int0_904 = torch.constant.int 0 | |
| %3947 = torch.aten.select.int %3946, %int0_902, %int0_904 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3948 = torch.aten.item %3947 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3949 = torch.aten.lt.int %3948, %int0_902 : !torch.int, !torch.int -> !torch.bool | |
| %3950 = torch.aten.Int.bool %3949 : !torch.bool -> !torch.int | |
| %3951 = torch.aten.mul.int %3950, %int0_903 : !torch.int, !torch.int -> !torch.int | |
| %3952 = torch.aten.add.int %3948, %3951 : !torch.int, !torch.int -> !torch.int | |
| %3953 = torch.prim.ListConstruct %3952 : (!torch.int) -> !torch.list<int> | |
| %false_905 = torch.constant.bool false | |
| %none_906 = torch.constant.none | |
| %3954 = torch.aten.tensor %3953, %none_906, %none_906, %false_905 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values, %indices = torch.aten.sort %3954, %int0_902, %false_905 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_907 = torch.constant.int 0 | |
| %3955 = torch.aten.select.int %values, %int0_902, %int0_907 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3956 = torch.aten.item %3955 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3957 = torch.aten.unsqueeze %3067, %3956 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %3958 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_908 = torch.constant.int 0 | |
| %int0_909 = torch.constant.int 0 | |
| %int0_910 = torch.constant.int 0 | |
| %3959 = torch.aten.select.int %3958, %int0_908, %int0_910 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3960 = torch.aten.item %3959 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3961 = torch.aten.lt.int %3960, %int0_908 : !torch.int, !torch.int -> !torch.bool | |
| %3962 = torch.aten.Int.bool %3961 : !torch.bool -> !torch.int | |
| %3963 = torch.aten.mul.int %3962, %int0_909 : !torch.int, !torch.int -> !torch.int | |
| %3964 = torch.aten.add.int %3960, %3963 : !torch.int, !torch.int -> !torch.int | |
| %3965 = torch.prim.ListConstruct %3964 : (!torch.int) -> !torch.list<int> | |
| %false_911 = torch.constant.bool false | |
| %none_912 = torch.constant.none | |
| %3966 = torch.aten.tensor %3965, %none_912, %none_912, %false_911 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_913, %indices_914 = torch.aten.sort %3966, %int0_908, %false_911 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_915 = torch.constant.int 0 | |
| %3967 = torch.aten.select.int %values_913, %int0_908, %int0_915 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3968 = torch.aten.item %3967 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3969 = torch.aten.unsqueeze %3944, %3968 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %3970 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_916 = torch.constant.int 0 | |
| %int0_917 = torch.constant.int 0 | |
| %int0_918 = torch.constant.int 0 | |
| %3971 = torch.aten.select.int %3970, %int0_916, %int0_918 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3972 = torch.aten.item %3971 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3973 = torch.aten.lt.int %3972, %int0_916 : !torch.int, !torch.int -> !torch.bool | |
| %3974 = torch.aten.Int.bool %3973 : !torch.bool -> !torch.int | |
| %3975 = torch.aten.mul.int %3974, %int0_917 : !torch.int, !torch.int -> !torch.int | |
| %3976 = torch.aten.add.int %3972, %3975 : !torch.int, !torch.int -> !torch.int | |
| %3977 = torch.prim.ListConstruct %3976 : (!torch.int) -> !torch.list<int> | |
| %false_919 = torch.constant.bool false | |
| %none_920 = torch.constant.none | |
| %3978 = torch.aten.tensor %3977, %none_920, %none_920, %false_919 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_921, %indices_922 = torch.aten.sort %3978, %int0_916, %false_919 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_923 = torch.constant.int 0 | |
| %3979 = torch.aten.select.int %values_921, %int0_916, %int0_923 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3980 = torch.aten.item %3979 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3981 = torch.aten.unsqueeze %3945, %3980 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %3982 = torch.prim.ListConstruct %3957, %3969, %3981 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_924 = torch.constant.int 0 | |
| %3983 = torch.aten.cat %3982, %int0_924 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
| %3984 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_925 = torch.constant.int 0 | |
| %int0_926 = torch.constant.int 0 | |
| %int0_927 = torch.constant.int 0 | |
| %3985 = torch.aten.select.int %3984, %int0_925, %int0_927 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3986 = torch.aten.item %3985 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3987 = torch.aten.lt.int %3986, %int0_925 : !torch.int, !torch.int -> !torch.bool | |
| %3988 = torch.aten.Int.bool %3987 : !torch.bool -> !torch.int | |
| %3989 = torch.aten.mul.int %3988, %int0_926 : !torch.int, !torch.int -> !torch.int | |
| %3990 = torch.aten.add.int %3986, %3989 : !torch.int, !torch.int -> !torch.int | |
| %3991 = torch.prim.ListConstruct %3990 : (!torch.int) -> !torch.list<int> | |
| %false_928 = torch.constant.bool false | |
| %none_929 = torch.constant.none | |
| %3992 = torch.aten.tensor %3991, %none_929, %none_929, %false_928 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_930, %indices_931 = torch.aten.sort %3992, %int0_925, %false_928 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_932 = torch.constant.int 0 | |
| %3993 = torch.aten.select.int %values_930, %int0_925, %int0_932 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3994 = torch.aten.item %3993 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3995 = torch.aten.unsqueeze %3067, %3994 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %3996 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_933 = torch.constant.int 0 | |
| %int0_934 = torch.constant.int 0 | |
| %int0_935 = torch.constant.int 0 | |
| %3997 = torch.aten.select.int %3996, %int0_933, %int0_935 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %3998 = torch.aten.item %3997 : !torch.vtensor<[1],si64> -> !torch.int | |
| %3999 = torch.aten.lt.int %3998, %int0_933 : !torch.int, !torch.int -> !torch.bool | |
| %4000 = torch.aten.Int.bool %3999 : !torch.bool -> !torch.int | |
| %4001 = torch.aten.mul.int %4000, %int0_934 : !torch.int, !torch.int -> !torch.int | |
| %4002 = torch.aten.add.int %3998, %4001 : !torch.int, !torch.int -> !torch.int | |
| %4003 = torch.prim.ListConstruct %4002 : (!torch.int) -> !torch.list<int> | |
| %false_936 = torch.constant.bool false | |
| %none_937 = torch.constant.none | |
| %4004 = torch.aten.tensor %4003, %none_937, %none_937, %false_936 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_938, %indices_939 = torch.aten.sort %4004, %int0_933, %false_936 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_940 = torch.constant.int 0 | |
| %4005 = torch.aten.select.int %values_938, %int0_933, %int0_940 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4006 = torch.aten.item %4005 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4007 = torch.aten.unsqueeze %3944, %4006 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4008 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_941 = torch.constant.int 0 | |
| %int0_942 = torch.constant.int 0 | |
| %int0_943 = torch.constant.int 0 | |
| %4009 = torch.aten.select.int %4008, %int0_941, %int0_943 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4010 = torch.aten.item %4009 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4011 = torch.aten.lt.int %4010, %int0_941 : !torch.int, !torch.int -> !torch.bool | |
| %4012 = torch.aten.Int.bool %4011 : !torch.bool -> !torch.int | |
| %4013 = torch.aten.mul.int %4012, %int0_942 : !torch.int, !torch.int -> !torch.int | |
| %4014 = torch.aten.add.int %4010, %4013 : !torch.int, !torch.int -> !torch.int | |
| %4015 = torch.prim.ListConstruct %4014 : (!torch.int) -> !torch.list<int> | |
| %false_944 = torch.constant.bool false | |
| %none_945 = torch.constant.none | |
| %4016 = torch.aten.tensor %4015, %none_945, %none_945, %false_944 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_946, %indices_947 = torch.aten.sort %4016, %int0_941, %false_944 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_948 = torch.constant.int 0 | |
| %4017 = torch.aten.select.int %values_946, %int0_941, %int0_948 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4018 = torch.aten.item %4017 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4019 = torch.aten.unsqueeze %3945, %4018 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4020 = torch.prim.ListConstruct %3995, %4007, %4019 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_949 = torch.constant.int 0 | |
| %4021 = torch.aten.cat %4020, %int0_949 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
| %4022 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_950 = torch.constant.int 0 | |
| %int0_951 = torch.constant.int 0 | |
| %int0_952 = torch.constant.int 0 | |
| %4023 = torch.aten.select.int %4022, %int0_950, %int0_952 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4024 = torch.aten.item %4023 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4025 = torch.aten.lt.int %4024, %int0_950 : !torch.int, !torch.int -> !torch.bool | |
| %4026 = torch.aten.Int.bool %4025 : !torch.bool -> !torch.int | |
| %4027 = torch.aten.mul.int %4026, %int0_951 : !torch.int, !torch.int -> !torch.int | |
| %4028 = torch.aten.add.int %4024, %4027 : !torch.int, !torch.int -> !torch.int | |
| %4029 = torch.prim.ListConstruct %4028 : (!torch.int) -> !torch.list<int> | |
| %false_953 = torch.constant.bool false | |
| %none_954 = torch.constant.none | |
| %4030 = torch.aten.tensor %4029, %none_954, %none_954, %false_953 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_955, %indices_956 = torch.aten.sort %4030, %int0_950, %false_953 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_957 = torch.constant.int 0 | |
| %4031 = torch.aten.select.int %values_955, %int0_950, %int0_957 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4032 = torch.aten.item %4031 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4033 = torch.aten.unsqueeze %3067, %4032 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4034 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_958 = torch.constant.int 0 | |
| %int0_959 = torch.constant.int 0 | |
| %int0_960 = torch.constant.int 0 | |
| %4035 = torch.aten.select.int %4034, %int0_958, %int0_960 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4036 = torch.aten.item %4035 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4037 = torch.aten.lt.int %4036, %int0_958 : !torch.int, !torch.int -> !torch.bool | |
| %4038 = torch.aten.Int.bool %4037 : !torch.bool -> !torch.int | |
| %4039 = torch.aten.mul.int %4038, %int0_959 : !torch.int, !torch.int -> !torch.int | |
| %4040 = torch.aten.add.int %4036, %4039 : !torch.int, !torch.int -> !torch.int | |
| %4041 = torch.prim.ListConstruct %4040 : (!torch.int) -> !torch.list<int> | |
| %false_961 = torch.constant.bool false | |
| %none_962 = torch.constant.none | |
| %4042 = torch.aten.tensor %4041, %none_962, %none_962, %false_961 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_963, %indices_964 = torch.aten.sort %4042, %int0_958, %false_961 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_965 = torch.constant.int 0 | |
| %4043 = torch.aten.select.int %values_963, %int0_958, %int0_965 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4044 = torch.aten.item %4043 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4045 = torch.aten.unsqueeze %3944, %4044 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4046 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_966 = torch.constant.int 0 | |
| %int0_967 = torch.constant.int 0 | |
| %int0_968 = torch.constant.int 0 | |
| %4047 = torch.aten.select.int %4046, %int0_966, %int0_968 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4048 = torch.aten.item %4047 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4049 = torch.aten.lt.int %4048, %int0_966 : !torch.int, !torch.int -> !torch.bool | |
| %4050 = torch.aten.Int.bool %4049 : !torch.bool -> !torch.int | |
| %4051 = torch.aten.mul.int %4050, %int0_967 : !torch.int, !torch.int -> !torch.int | |
| %4052 = torch.aten.add.int %4048, %4051 : !torch.int, !torch.int -> !torch.int | |
| %4053 = torch.prim.ListConstruct %4052 : (!torch.int) -> !torch.list<int> | |
| %false_969 = torch.constant.bool false | |
| %none_970 = torch.constant.none | |
| %4054 = torch.aten.tensor %4053, %none_970, %none_970, %false_969 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_971, %indices_972 = torch.aten.sort %4054, %int0_966, %false_969 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_973 = torch.constant.int 0 | |
| %4055 = torch.aten.select.int %values_971, %int0_966, %int0_973 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4056 = torch.aten.item %4055 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4057 = torch.aten.unsqueeze %3945, %4056 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4058 = torch.prim.ListConstruct %4033, %4045, %4057 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_974 = torch.constant.int 0 | |
| %4059 = torch.aten.cat %4058, %int0_974 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
| %int0_975 = torch.constant.int 0 | |
| %int0_976 = torch.constant.int 0 | |
| %4060 = torch.aten.select.int %3983, %int0_975, %int0_976 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4061 = torch.aten.item %4060 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4062 = torch.aten.eq.int %4061, %int0_975 : !torch.int, !torch.int -> !torch.bool | |
| %4063 = torch.aten.Int.bool %4062 : !torch.bool -> !torch.int | |
| %int0_977 = torch.constant.int 0 | |
| %4064 = torch.aten.size.int %3359, %int0_977 : !torch.vtensor<[1,144,80,80],f32>, !torch.int -> !torch.int | |
| %4065 = torch.prim.NumToTensor.Scalar %4063 : !torch.int -> !torch.vtensor<[],i1> | |
| %4066 = torch.prim.NumToTensor.Scalar %4064 : !torch.int -> !torch.vtensor<[],si64> | |
| %4067 = torch.prim.NumToTensor.Scalar %4061 : !torch.int -> !torch.vtensor<[],si64> | |
| %4068 = torch.aten.where.self %4065, %4066, %4067 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4069 = torch.aten.item %4068 : !torch.vtensor<[],si64> -> !torch.int | |
| %int1_978 = torch.constant.int 1 | |
| %4070 = torch.aten.select.int %3983, %int0_975, %int1_978 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4071 = torch.aten.item %4070 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4072 = torch.aten.eq.int %4071, %int0_975 : !torch.int, !torch.int -> !torch.bool | |
| %4073 = torch.aten.Int.bool %4072 : !torch.bool -> !torch.int | |
| %int1_979 = torch.constant.int 1 | |
| %4074 = torch.aten.size.int %3359, %int1_979 : !torch.vtensor<[1,144,80,80],f32>, !torch.int -> !torch.int | |
| %4075 = torch.prim.NumToTensor.Scalar %4073 : !torch.int -> !torch.vtensor<[],i1> | |
| %4076 = torch.prim.NumToTensor.Scalar %4074 : !torch.int -> !torch.vtensor<[],si64> | |
| %4077 = torch.prim.NumToTensor.Scalar %4071 : !torch.int -> !torch.vtensor<[],si64> | |
| %4078 = torch.aten.where.self %4075, %4076, %4077 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4079 = torch.aten.item %4078 : !torch.vtensor<[],si64> -> !torch.int | |
| %int2_980 = torch.constant.int 2 | |
| %4080 = torch.aten.select.int %3983, %int0_975, %int2_980 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4081 = torch.aten.item %4080 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4082 = torch.aten.eq.int %4081, %int0_975 : !torch.int, !torch.int -> !torch.bool | |
| %4083 = torch.aten.Int.bool %4082 : !torch.bool -> !torch.int | |
| %int2_981 = torch.constant.int 2 | |
| %4084 = torch.aten.size.int %3359, %int2_981 : !torch.vtensor<[1,144,80,80],f32>, !torch.int -> !torch.int | |
| %4085 = torch.prim.NumToTensor.Scalar %4083 : !torch.int -> !torch.vtensor<[],i1> | |
| %4086 = torch.prim.NumToTensor.Scalar %4084 : !torch.int -> !torch.vtensor<[],si64> | |
| %4087 = torch.prim.NumToTensor.Scalar %4081 : !torch.int -> !torch.vtensor<[],si64> | |
| %4088 = torch.aten.where.self %4085, %4086, %4087 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4089 = torch.aten.item %4088 : !torch.vtensor<[],si64> -> !torch.int | |
| %4090 = torch.prim.ListConstruct %4069, %4079, %4089 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %4091 = torch.aten.reshape %3359, %4090 : !torch.vtensor<[1,144,80,80],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
| %int0_982 = torch.constant.int 0 | |
| %int0_983 = torch.constant.int 0 | |
| %4092 = torch.aten.select.int %4021, %int0_982, %int0_983 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4093 = torch.aten.item %4092 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4094 = torch.aten.eq.int %4093, %int0_982 : !torch.int, !torch.int -> !torch.bool | |
| %4095 = torch.aten.Int.bool %4094 : !torch.bool -> !torch.int | |
| %int0_984 = torch.constant.int 0 | |
| %4096 = torch.aten.size.int %3651, %int0_984 : !torch.vtensor<[1,144,40,40],f32>, !torch.int -> !torch.int | |
| %4097 = torch.prim.NumToTensor.Scalar %4095 : !torch.int -> !torch.vtensor<[],i1> | |
| %4098 = torch.prim.NumToTensor.Scalar %4096 : !torch.int -> !torch.vtensor<[],si64> | |
| %4099 = torch.prim.NumToTensor.Scalar %4093 : !torch.int -> !torch.vtensor<[],si64> | |
| %4100 = torch.aten.where.self %4097, %4098, %4099 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4101 = torch.aten.item %4100 : !torch.vtensor<[],si64> -> !torch.int | |
| %int1_985 = torch.constant.int 1 | |
| %4102 = torch.aten.select.int %4021, %int0_982, %int1_985 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4103 = torch.aten.item %4102 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4104 = torch.aten.eq.int %4103, %int0_982 : !torch.int, !torch.int -> !torch.bool | |
| %4105 = torch.aten.Int.bool %4104 : !torch.bool -> !torch.int | |
| %int1_986 = torch.constant.int 1 | |
| %4106 = torch.aten.size.int %3651, %int1_986 : !torch.vtensor<[1,144,40,40],f32>, !torch.int -> !torch.int | |
| %4107 = torch.prim.NumToTensor.Scalar %4105 : !torch.int -> !torch.vtensor<[],i1> | |
| %4108 = torch.prim.NumToTensor.Scalar %4106 : !torch.int -> !torch.vtensor<[],si64> | |
| %4109 = torch.prim.NumToTensor.Scalar %4103 : !torch.int -> !torch.vtensor<[],si64> | |
| %4110 = torch.aten.where.self %4107, %4108, %4109 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4111 = torch.aten.item %4110 : !torch.vtensor<[],si64> -> !torch.int | |
| %int2_987 = torch.constant.int 2 | |
| %4112 = torch.aten.select.int %4021, %int0_982, %int2_987 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4113 = torch.aten.item %4112 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4114 = torch.aten.eq.int %4113, %int0_982 : !torch.int, !torch.int -> !torch.bool | |
| %4115 = torch.aten.Int.bool %4114 : !torch.bool -> !torch.int | |
| %int2_988 = torch.constant.int 2 | |
| %4116 = torch.aten.size.int %3651, %int2_988 : !torch.vtensor<[1,144,40,40],f32>, !torch.int -> !torch.int | |
| %4117 = torch.prim.NumToTensor.Scalar %4115 : !torch.int -> !torch.vtensor<[],i1> | |
| %4118 = torch.prim.NumToTensor.Scalar %4116 : !torch.int -> !torch.vtensor<[],si64> | |
| %4119 = torch.prim.NumToTensor.Scalar %4113 : !torch.int -> !torch.vtensor<[],si64> | |
| %4120 = torch.aten.where.self %4117, %4118, %4119 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4121 = torch.aten.item %4120 : !torch.vtensor<[],si64> -> !torch.int | |
| %4122 = torch.prim.ListConstruct %4101, %4111, %4121 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %4123 = torch.aten.reshape %3651, %4122 : !torch.vtensor<[1,144,40,40],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
| %int0_989 = torch.constant.int 0 | |
| %int0_990 = torch.constant.int 0 | |
| %4124 = torch.aten.select.int %4059, %int0_989, %int0_990 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4125 = torch.aten.item %4124 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4126 = torch.aten.eq.int %4125, %int0_989 : !torch.int, !torch.int -> !torch.bool | |
| %4127 = torch.aten.Int.bool %4126 : !torch.bool -> !torch.int | |
| %int0_991 = torch.constant.int 0 | |
| %4128 = torch.aten.size.int %3943, %int0_991 : !torch.vtensor<[1,144,20,20],f32>, !torch.int -> !torch.int | |
| %4129 = torch.prim.NumToTensor.Scalar %4127 : !torch.int -> !torch.vtensor<[],i1> | |
| %4130 = torch.prim.NumToTensor.Scalar %4128 : !torch.int -> !torch.vtensor<[],si64> | |
| %4131 = torch.prim.NumToTensor.Scalar %4125 : !torch.int -> !torch.vtensor<[],si64> | |
| %4132 = torch.aten.where.self %4129, %4130, %4131 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4133 = torch.aten.item %4132 : !torch.vtensor<[],si64> -> !torch.int | |
| %int1_992 = torch.constant.int 1 | |
| %4134 = torch.aten.select.int %4059, %int0_989, %int1_992 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4135 = torch.aten.item %4134 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4136 = torch.aten.eq.int %4135, %int0_989 : !torch.int, !torch.int -> !torch.bool | |
| %4137 = torch.aten.Int.bool %4136 : !torch.bool -> !torch.int | |
| %int1_993 = torch.constant.int 1 | |
| %4138 = torch.aten.size.int %3943, %int1_993 : !torch.vtensor<[1,144,20,20],f32>, !torch.int -> !torch.int | |
| %4139 = torch.prim.NumToTensor.Scalar %4137 : !torch.int -> !torch.vtensor<[],i1> | |
| %4140 = torch.prim.NumToTensor.Scalar %4138 : !torch.int -> !torch.vtensor<[],si64> | |
| %4141 = torch.prim.NumToTensor.Scalar %4135 : !torch.int -> !torch.vtensor<[],si64> | |
| %4142 = torch.aten.where.self %4139, %4140, %4141 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4143 = torch.aten.item %4142 : !torch.vtensor<[],si64> -> !torch.int | |
| %int2_994 = torch.constant.int 2 | |
| %4144 = torch.aten.select.int %4059, %int0_989, %int2_994 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4145 = torch.aten.item %4144 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4146 = torch.aten.eq.int %4145, %int0_989 : !torch.int, !torch.int -> !torch.bool | |
| %4147 = torch.aten.Int.bool %4146 : !torch.bool -> !torch.int | |
| %int2_995 = torch.constant.int 2 | |
| %4148 = torch.aten.size.int %3943, %int2_995 : !torch.vtensor<[1,144,20,20],f32>, !torch.int -> !torch.int | |
| %4149 = torch.prim.NumToTensor.Scalar %4147 : !torch.int -> !torch.vtensor<[],i1> | |
| %4150 = torch.prim.NumToTensor.Scalar %4148 : !torch.int -> !torch.vtensor<[],si64> | |
| %4151 = torch.prim.NumToTensor.Scalar %4145 : !torch.int -> !torch.vtensor<[],si64> | |
| %4152 = torch.aten.where.self %4149, %4150, %4151 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4153 = torch.aten.item %4152 : !torch.vtensor<[],si64> -> !torch.int | |
| %4154 = torch.prim.ListConstruct %4133, %4143, %4153 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %4155 = torch.aten.reshape %3943, %4154 : !torch.vtensor<[1,144,20,20],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
| %4156 = torch.prim.ListConstruct %4091, %4123, %4155 : (!torch.vtensor<[?,?,?],f32>, !torch.vtensor<[?,?,?],f32>, !torch.vtensor<[?,?,?],f32>) -> !torch.list<vtensor> | |
| %int2_996 = torch.constant.int 2 | |
| %4157 = torch.aten.cat %4156, %int2_996 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
| %4158 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4159 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_997 = torch.constant.int 12 | |
| %4160 = torch.aten.item %4158 : !torch.vtensor<[],f32> -> !torch.float | |
| %4161 = torch.aten.item %4159 : !torch.vtensor<[],si8> -> !torch.int | |
| %4162 = torch.aten.quantize_per_tensor %4157, %4160, %4161, %int12_997 : !torch.vtensor<[?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],!torch.qint8> | |
| %4163 = torch.aten.int_repr %4162 : !torch.vtensor<[?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?],si8> | |
| %4164 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4165 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4166 = torch.aten.item %4164 : !torch.vtensor<[],f32> -> !torch.float | |
| %4167 = torch.aten.item %4165 : !torch.vtensor<[],si8> -> !torch.int | |
| %4168 = torch.aten._make_per_tensor_quantized_tensor %4163, %4166, %4167 : !torch.vtensor<[?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?],!torch.qint8> | |
| %4169 = torch.aten.dequantize.self %4168 : !torch.vtensor<[?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?],f32> | |
| %4170 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4171 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4172 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4173 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_998 = torch.constant.int 0 | |
| %int0_999 = torch.constant.int 0 | |
| %4174 = torch.prim.NumToTensor.Scalar %int0_999 : !torch.int -> !torch.vtensor<[1],si64> | |
| %4175 = torch.aten.index_select %4171, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4176 = torch.aten.item %4175 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4177 = torch.aten.index_select %4172, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4178 = torch.aten.item %4177 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4179 = torch.aten.index_select %4170, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4180 = torch.aten.item %4179 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4181 = torch.aten.index_select %4173, %int0_998, %4174 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4182 = torch.aten.item %4181 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4183 = torch.aten.slice.Tensor %4169, %4180, %4176, %4178, %4182 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
| %4184 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4185 = torch.vtensor.literal(dense<64> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4186 = torch.vtensor.literal(dense<144> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4187 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1000 = torch.constant.int 0 | |
| %int0_1001 = torch.constant.int 0 | |
| %4188 = torch.prim.NumToTensor.Scalar %int0_1001 : !torch.int -> !torch.vtensor<[1],si64> | |
| %4189 = torch.aten.index_select %4185, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4190 = torch.aten.item %4189 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4191 = torch.aten.index_select %4186, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4192 = torch.aten.item %4191 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4193 = torch.aten.index_select %4184, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4194 = torch.aten.item %4193 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4195 = torch.aten.index_select %4187, %int0_1000, %4188 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4196 = torch.aten.item %4195 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4197 = torch.aten.slice.Tensor %4169, %4194, %4190, %4192, %4196 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
| %4198 = torch.vtensor.literal(dense<1> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4199 = torch.vtensor.literal(dense<8400> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4200 = torch.vtensor.literal(dense<4> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4201 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1002 = torch.constant.int 0 | |
| %int0_1003 = torch.constant.int 0 | |
| %int0_1004 = torch.constant.int 0 | |
| %4202 = torch.aten.select.int %4201, %int0_1002, %int0_1004 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4203 = torch.aten.item %4202 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4204 = torch.aten.lt.int %4203, %int0_1002 : !torch.int, !torch.int -> !torch.bool | |
| %4205 = torch.aten.Int.bool %4204 : !torch.bool -> !torch.int | |
| %4206 = torch.aten.mul.int %4205, %int0_1003 : !torch.int, !torch.int -> !torch.int | |
| %4207 = torch.aten.add.int %4203, %4206 : !torch.int, !torch.int -> !torch.int | |
| %4208 = torch.prim.ListConstruct %4207 : (!torch.int) -> !torch.list<int> | |
| %false_1005 = torch.constant.bool false | |
| %none_1006 = torch.constant.none | |
| %4209 = torch.aten.tensor %4208, %none_1006, %none_1006, %false_1005 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1007, %indices_1008 = torch.aten.sort %4209, %int0_1002, %false_1005 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1009 = torch.constant.int 0 | |
| %4210 = torch.aten.select.int %values_1007, %int0_1002, %int0_1009 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4211 = torch.aten.item %4210 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4212 = torch.aten.unsqueeze %4198, %4211 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4213 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1010 = torch.constant.int 0 | |
| %int0_1011 = torch.constant.int 0 | |
| %int0_1012 = torch.constant.int 0 | |
| %4214 = torch.aten.select.int %4213, %int0_1010, %int0_1012 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4215 = torch.aten.item %4214 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4216 = torch.aten.lt.int %4215, %int0_1010 : !torch.int, !torch.int -> !torch.bool | |
| %4217 = torch.aten.Int.bool %4216 : !torch.bool -> !torch.int | |
| %4218 = torch.aten.mul.int %4217, %int0_1011 : !torch.int, !torch.int -> !torch.int | |
| %4219 = torch.aten.add.int %4215, %4218 : !torch.int, !torch.int -> !torch.int | |
| %4220 = torch.prim.ListConstruct %4219 : (!torch.int) -> !torch.list<int> | |
| %false_1013 = torch.constant.bool false | |
| %none_1014 = torch.constant.none | |
| %4221 = torch.aten.tensor %4220, %none_1014, %none_1014, %false_1013 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1015, %indices_1016 = torch.aten.sort %4221, %int0_1010, %false_1013 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1017 = torch.constant.int 0 | |
| %4222 = torch.aten.select.int %values_1015, %int0_1010, %int0_1017 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4223 = torch.aten.item %4222 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4224 = torch.aten.unsqueeze %4200, %4223 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4225 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1018 = torch.constant.int 0 | |
| %int0_1019 = torch.constant.int 0 | |
| %int0_1020 = torch.constant.int 0 | |
| %4226 = torch.aten.select.int %4225, %int0_1018, %int0_1020 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4227 = torch.aten.item %4226 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4228 = torch.aten.lt.int %4227, %int0_1018 : !torch.int, !torch.int -> !torch.bool | |
| %4229 = torch.aten.Int.bool %4228 : !torch.bool -> !torch.int | |
| %4230 = torch.aten.mul.int %4229, %int0_1019 : !torch.int, !torch.int -> !torch.int | |
| %4231 = torch.aten.add.int %4227, %4230 : !torch.int, !torch.int -> !torch.int | |
| %4232 = torch.prim.ListConstruct %4231 : (!torch.int) -> !torch.list<int> | |
| %false_1021 = torch.constant.bool false | |
| %none_1022 = torch.constant.none | |
| %4233 = torch.aten.tensor %4232, %none_1022, %none_1022, %false_1021 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1023, %indices_1024 = torch.aten.sort %4233, %int0_1018, %false_1021 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1025 = torch.constant.int 0 | |
| %4234 = torch.aten.select.int %values_1023, %int0_1018, %int0_1025 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4235 = torch.aten.item %4234 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4236 = torch.aten.unsqueeze %343, %4235 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4237 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1026 = torch.constant.int 0 | |
| %int0_1027 = torch.constant.int 0 | |
| %int0_1028 = torch.constant.int 0 | |
| %4238 = torch.aten.select.int %4237, %int0_1026, %int0_1028 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4239 = torch.aten.item %4238 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4240 = torch.aten.lt.int %4239, %int0_1026 : !torch.int, !torch.int -> !torch.bool | |
| %4241 = torch.aten.Int.bool %4240 : !torch.bool -> !torch.int | |
| %4242 = torch.aten.mul.int %4241, %int0_1027 : !torch.int, !torch.int -> !torch.int | |
| %4243 = torch.aten.add.int %4239, %4242 : !torch.int, !torch.int -> !torch.int | |
| %4244 = torch.prim.ListConstruct %4243 : (!torch.int) -> !torch.list<int> | |
| %false_1029 = torch.constant.bool false | |
| %none_1030 = torch.constant.none | |
| %4245 = torch.aten.tensor %4244, %none_1030, %none_1030, %false_1029 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1031, %indices_1032 = torch.aten.sort %4245, %int0_1026, %false_1029 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1033 = torch.constant.int 0 | |
| %4246 = torch.aten.select.int %values_1031, %int0_1026, %int0_1033 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4247 = torch.aten.item %4246 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4248 = torch.aten.unsqueeze %4199, %4247 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4249 = torch.prim.ListConstruct %4212, %4224, %4236, %4248 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_1034 = torch.constant.int 0 | |
| %4250 = torch.aten.cat %4249, %int0_1034 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],si64> | |
| %int0_1035 = torch.constant.int 0 | |
| %int0_1036 = torch.constant.int 0 | |
| %4251 = torch.aten.select.int %4250, %int0_1035, %int0_1036 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4252 = torch.aten.item %4251 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4253 = torch.aten.eq.int %4252, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
| %4254 = torch.aten.Int.bool %4253 : !torch.bool -> !torch.int | |
| %int0_1037 = torch.constant.int 0 | |
| %4255 = torch.aten.size.int %4183, %int0_1037 : !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.int | |
| %4256 = torch.prim.NumToTensor.Scalar %4254 : !torch.int -> !torch.vtensor<[],i1> | |
| %4257 = torch.prim.NumToTensor.Scalar %4255 : !torch.int -> !torch.vtensor<[],si64> | |
| %4258 = torch.prim.NumToTensor.Scalar %4252 : !torch.int -> !torch.vtensor<[],si64> | |
| %4259 = torch.aten.where.self %4256, %4257, %4258 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4260 = torch.aten.item %4259 : !torch.vtensor<[],si64> -> !torch.int | |
| %int1_1038 = torch.constant.int 1 | |
| %4261 = torch.aten.select.int %4250, %int0_1035, %int1_1038 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4262 = torch.aten.item %4261 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4263 = torch.aten.eq.int %4262, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
| %4264 = torch.aten.Int.bool %4263 : !torch.bool -> !torch.int | |
| %int1_1039 = torch.constant.int 1 | |
| %4265 = torch.aten.size.int %4183, %int1_1039 : !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.int | |
| %4266 = torch.prim.NumToTensor.Scalar %4264 : !torch.int -> !torch.vtensor<[],i1> | |
| %4267 = torch.prim.NumToTensor.Scalar %4265 : !torch.int -> !torch.vtensor<[],si64> | |
| %4268 = torch.prim.NumToTensor.Scalar %4262 : !torch.int -> !torch.vtensor<[],si64> | |
| %4269 = torch.aten.where.self %4266, %4267, %4268 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4270 = torch.aten.item %4269 : !torch.vtensor<[],si64> -> !torch.int | |
| %int2_1040 = torch.constant.int 2 | |
| %4271 = torch.aten.select.int %4250, %int0_1035, %int2_1040 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4272 = torch.aten.item %4271 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4273 = torch.aten.eq.int %4272, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
| %4274 = torch.aten.Int.bool %4273 : !torch.bool -> !torch.int | |
| %int2_1041 = torch.constant.int 2 | |
| %4275 = torch.aten.size.int %4183, %int2_1041 : !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.int | |
| %4276 = torch.prim.NumToTensor.Scalar %4274 : !torch.int -> !torch.vtensor<[],i1> | |
| %4277 = torch.prim.NumToTensor.Scalar %4275 : !torch.int -> !torch.vtensor<[],si64> | |
| %4278 = torch.prim.NumToTensor.Scalar %4272 : !torch.int -> !torch.vtensor<[],si64> | |
| %4279 = torch.aten.where.self %4276, %4277, %4278 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4280 = torch.aten.item %4279 : !torch.vtensor<[],si64> -> !torch.int | |
| %int3 = torch.constant.int 3 | |
| %4281 = torch.aten.select.int %4250, %int0_1035, %int3 : !torch.vtensor<[4],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4282 = torch.aten.item %4281 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4283 = torch.aten.eq.int %4282, %int0_1035 : !torch.int, !torch.int -> !torch.bool | |
| %4284 = torch.aten.Int.bool %4283 : !torch.bool -> !torch.int | |
| %4285 = torch.prim.ListConstruct %4260, %4270, %4280, %4282 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %4286 = torch.aten.reshape %4183, %4285 : !torch.vtensor<[?,?,?],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?,?],f32> | |
| %int1_1042 = torch.constant.int 1 | |
| %int2_1043 = torch.constant.int 2 | |
| %4287 = torch.aten.transpose.int %4286, %int1_1042, %int2_1043 : !torch.vtensor<[?,?,?,?],f32>, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],f32> | |
| %4288 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4289 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1044 = torch.constant.int 12 | |
| %4290 = torch.aten.item %4288 : !torch.vtensor<[],f32> -> !torch.float | |
| %4291 = torch.aten.item %4289 : !torch.vtensor<[],si8> -> !torch.int | |
| %4292 = torch.aten.quantize_per_tensor %4287, %4290, %4291, %int12_1044 : !torch.vtensor<[?,?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4293 = torch.aten.int_repr %4292 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],si8> | |
| %4294 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4295 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4296 = torch.aten.item %4294 : !torch.vtensor<[],f32> -> !torch.float | |
| %4297 = torch.aten.item %4295 : !torch.vtensor<[],si8> -> !torch.int | |
| %4298 = torch.aten._make_per_tensor_quantized_tensor %4293, %4296, %4297 : !torch.vtensor<[?,?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4299 = torch.aten.dequantize.self %4298 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],f32> | |
| %int1_1045 = torch.constant.int 1 | |
| %none_1046 = torch.constant.none | |
| %4300 = torch.aten.softmax.int %4299, %int1_1045, %none_1046 : !torch.vtensor<[?,?,?,?],f32>, !torch.int, !torch.none -> !torch.vtensor<[?,?,?,?],f32> | |
| %4301 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4302 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1047 = torch.constant.int 12 | |
| %4303 = torch.aten.item %4301 : !torch.vtensor<[],f32> -> !torch.float | |
| %4304 = torch.aten.item %4302 : !torch.vtensor<[],si8> -> !torch.int | |
| %4305 = torch.aten.quantize_per_tensor %4300, %4303, %4304, %int12_1047 : !torch.vtensor<[?,?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4306 = torch.aten.int_repr %4305 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],si8> | |
| %4307 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4308 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4309 = torch.aten.item %4307 : !torch.vtensor<[],f32> -> !torch.float | |
| %4310 = torch.aten.item %4308 : !torch.vtensor<[],si8> -> !torch.int | |
| %4311 = torch.aten._make_per_tensor_quantized_tensor %4306, %4309, %4310 : !torch.vtensor<[?,?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4312 = torch.aten.dequantize.self %4311 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],f32> | |
| %4313 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4314 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1048 = torch.constant.int 12 | |
| %4315 = torch.aten.item %4313 : !torch.vtensor<[],f32> -> !torch.float | |
| %4316 = torch.aten.item %4314 : !torch.vtensor<[],si8> -> !torch.int | |
| %4317 = torch.aten.quantize_per_tensor %126, %4315, %4316, %int12_1048 : !torch.vtensor<[1,16,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,1,1],!torch.qint8> | |
| %4318 = torch.aten.int_repr %4317 : !torch.vtensor<[1,16,1,1],!torch.qint8> -> !torch.vtensor<[1,16,1,1],si8> | |
| %4319 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4320 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4321 = torch.aten.item %4319 : !torch.vtensor<[],f32> -> !torch.float | |
| %4322 = torch.aten.item %4320 : !torch.vtensor<[],si8> -> !torch.int | |
| %4323 = torch.aten._make_per_tensor_quantized_tensor %4318, %4321, %4322 : !torch.vtensor<[1,16,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,1,1],!torch.qint8> | |
| %4324 = torch.aten.dequantize.self %4323 : !torch.vtensor<[1,16,1,1],!torch.qint8> -> !torch.vtensor<[1,16,1,1],f32> | |
| %int0_1049 = torch.constant.int 0 | |
| %int0_1050 = torch.constant.int 0 | |
| %int1_1051 = torch.constant.int 1 | |
| %int1_1052 = torch.constant.int 1 | |
| %int1_1053 = torch.constant.int 1 | |
| %int1_1054 = torch.constant.int 1 | |
| %int0_1055 = torch.constant.int 0 | |
| %4325 = torch.prim.ListConstruct %int0_1049, %int0_1050 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4326 = torch.prim.ListConstruct %int1_1051, %int1_1052 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4327 = torch.prim.ListConstruct %int1_1053, %int1_1054 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4328 = torch.prim.ListConstruct %int0_1055, %int0_1055 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1056 = torch.constant.bool false | |
| %none_1057 = torch.constant.none | |
| %int1_1058 = torch.constant.int 1 | |
| %4329 = torch.aten.convolution %4312, %4324, %none_1057, %4327, %4325, %4326, %false_1056, %4328, %int1_1058 : !torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[1,16,1,1],f32>, !torch.none, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[?,1,?,?],f32> | |
| %4330 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4331 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1059 = torch.constant.int 12 | |
| %4332 = torch.aten.item %4330 : !torch.vtensor<[],f32> -> !torch.float | |
| %4333 = torch.aten.item %4331 : !torch.vtensor<[],si8> -> !torch.int | |
| %4334 = torch.aten.quantize_per_tensor %4329, %4332, %4333, %int12_1059 : !torch.vtensor<[?,1,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,1,?,?],!torch.qint8> | |
| %4335 = torch.aten.int_repr %4334 : !torch.vtensor<[?,1,?,?],!torch.qint8> -> !torch.vtensor<[?,1,?,?],si8> | |
| %4336 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4337 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4338 = torch.aten.item %4336 : !torch.vtensor<[],f32> -> !torch.float | |
| %4339 = torch.aten.item %4337 : !torch.vtensor<[],si8> -> !torch.int | |
| %4340 = torch.aten._make_per_tensor_quantized_tensor %4335, %4338, %4339 : !torch.vtensor<[?,1,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,1,?,?],!torch.qint8> | |
| %4341 = torch.aten.dequantize.self %4340 : !torch.vtensor<[?,1,?,?],!torch.qint8> -> !torch.vtensor<[?,1,?,?],f32> | |
| %4342 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1060 = torch.constant.int 0 | |
| %int0_1061 = torch.constant.int 0 | |
| %int0_1062 = torch.constant.int 0 | |
| %4343 = torch.aten.select.int %4342, %int0_1060, %int0_1062 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4344 = torch.aten.item %4343 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4345 = torch.aten.lt.int %4344, %int0_1060 : !torch.int, !torch.int -> !torch.bool | |
| %4346 = torch.aten.Int.bool %4345 : !torch.bool -> !torch.int | |
| %4347 = torch.aten.mul.int %4346, %int0_1061 : !torch.int, !torch.int -> !torch.int | |
| %4348 = torch.aten.add.int %4344, %4347 : !torch.int, !torch.int -> !torch.int | |
| %4349 = torch.prim.ListConstruct %4348 : (!torch.int) -> !torch.list<int> | |
| %false_1063 = torch.constant.bool false | |
| %none_1064 = torch.constant.none | |
| %4350 = torch.aten.tensor %4349, %none_1064, %none_1064, %false_1063 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1065, %indices_1066 = torch.aten.sort %4350, %int0_1060, %false_1063 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1067 = torch.constant.int 0 | |
| %4351 = torch.aten.select.int %values_1065, %int0_1060, %int0_1067 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4352 = torch.aten.item %4351 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4353 = torch.aten.unsqueeze %4198, %4352 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4354 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1068 = torch.constant.int 0 | |
| %int0_1069 = torch.constant.int 0 | |
| %int0_1070 = torch.constant.int 0 | |
| %4355 = torch.aten.select.int %4354, %int0_1068, %int0_1070 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4356 = torch.aten.item %4355 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4357 = torch.aten.lt.int %4356, %int0_1068 : !torch.int, !torch.int -> !torch.bool | |
| %4358 = torch.aten.Int.bool %4357 : !torch.bool -> !torch.int | |
| %4359 = torch.aten.mul.int %4358, %int0_1069 : !torch.int, !torch.int -> !torch.int | |
| %4360 = torch.aten.add.int %4356, %4359 : !torch.int, !torch.int -> !torch.int | |
| %4361 = torch.prim.ListConstruct %4360 : (!torch.int) -> !torch.list<int> | |
| %false_1071 = torch.constant.bool false | |
| %none_1072 = torch.constant.none | |
| %4362 = torch.aten.tensor %4361, %none_1072, %none_1072, %false_1071 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1073, %indices_1074 = torch.aten.sort %4362, %int0_1068, %false_1071 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1075 = torch.constant.int 0 | |
| %4363 = torch.aten.select.int %values_1073, %int0_1068, %int0_1075 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4364 = torch.aten.item %4363 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4365 = torch.aten.unsqueeze %4200, %4364 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4366 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1076 = torch.constant.int 0 | |
| %int0_1077 = torch.constant.int 0 | |
| %int0_1078 = torch.constant.int 0 | |
| %4367 = torch.aten.select.int %4366, %int0_1076, %int0_1078 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4368 = torch.aten.item %4367 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4369 = torch.aten.lt.int %4368, %int0_1076 : !torch.int, !torch.int -> !torch.bool | |
| %4370 = torch.aten.Int.bool %4369 : !torch.bool -> !torch.int | |
| %4371 = torch.aten.mul.int %4370, %int0_1077 : !torch.int, !torch.int -> !torch.int | |
| %4372 = torch.aten.add.int %4368, %4371 : !torch.int, !torch.int -> !torch.int | |
| %4373 = torch.prim.ListConstruct %4372 : (!torch.int) -> !torch.list<int> | |
| %false_1079 = torch.constant.bool false | |
| %none_1080 = torch.constant.none | |
| %4374 = torch.aten.tensor %4373, %none_1080, %none_1080, %false_1079 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1081, %indices_1082 = torch.aten.sort %4374, %int0_1076, %false_1079 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1083 = torch.constant.int 0 | |
| %4375 = torch.aten.select.int %values_1081, %int0_1076, %int0_1083 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4376 = torch.aten.item %4375 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4377 = torch.aten.unsqueeze %4199, %4376 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4378 = torch.prim.ListConstruct %4353, %4365, %4377 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_1084 = torch.constant.int 0 | |
| %4379 = torch.aten.cat %4378, %int0_1084 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[3],si64> | |
| %int0_1085 = torch.constant.int 0 | |
| %int0_1086 = torch.constant.int 0 | |
| %4380 = torch.aten.select.int %4379, %int0_1085, %int0_1086 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4381 = torch.aten.item %4380 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4382 = torch.aten.eq.int %4381, %int0_1085 : !torch.int, !torch.int -> !torch.bool | |
| %4383 = torch.aten.Int.bool %4382 : !torch.bool -> !torch.int | |
| %int0_1087 = torch.constant.int 0 | |
| %4384 = torch.aten.size.int %4341, %int0_1087 : !torch.vtensor<[?,1,?,?],f32>, !torch.int -> !torch.int | |
| %4385 = torch.prim.NumToTensor.Scalar %4383 : !torch.int -> !torch.vtensor<[],i1> | |
| %4386 = torch.prim.NumToTensor.Scalar %4384 : !torch.int -> !torch.vtensor<[],si64> | |
| %4387 = torch.prim.NumToTensor.Scalar %4381 : !torch.int -> !torch.vtensor<[],si64> | |
| %4388 = torch.aten.where.self %4385, %4386, %4387 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4389 = torch.aten.item %4388 : !torch.vtensor<[],si64> -> !torch.int | |
| %int1_1088 = torch.constant.int 1 | |
| %4390 = torch.aten.select.int %4379, %int0_1085, %int1_1088 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4391 = torch.aten.item %4390 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4392 = torch.aten.eq.int %4391, %int0_1085 : !torch.int, !torch.int -> !torch.bool | |
| %4393 = torch.aten.Int.bool %4392 : !torch.bool -> !torch.int | |
| %int1_1089 = torch.constant.int 1 | |
| %4394 = torch.aten.size.int %4341, %int1_1089 : !torch.vtensor<[?,1,?,?],f32>, !torch.int -> !torch.int | |
| %4395 = torch.prim.NumToTensor.Scalar %4393 : !torch.int -> !torch.vtensor<[],i1> | |
| %4396 = torch.prim.NumToTensor.Scalar %4394 : !torch.int -> !torch.vtensor<[],si64> | |
| %4397 = torch.prim.NumToTensor.Scalar %4391 : !torch.int -> !torch.vtensor<[],si64> | |
| %4398 = torch.aten.where.self %4395, %4396, %4397 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4399 = torch.aten.item %4398 : !torch.vtensor<[],si64> -> !torch.int | |
| %int2_1090 = torch.constant.int 2 | |
| %4400 = torch.aten.select.int %4379, %int0_1085, %int2_1090 : !torch.vtensor<[3],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4401 = torch.aten.item %4400 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4402 = torch.aten.eq.int %4401, %int0_1085 : !torch.int, !torch.int -> !torch.bool | |
| %4403 = torch.aten.Int.bool %4402 : !torch.bool -> !torch.int | |
| %int2_1091 = torch.constant.int 2 | |
| %4404 = torch.aten.size.int %4341, %int2_1091 : !torch.vtensor<[?,1,?,?],f32>, !torch.int -> !torch.int | |
| %4405 = torch.prim.NumToTensor.Scalar %4403 : !torch.int -> !torch.vtensor<[],i1> | |
| %4406 = torch.prim.NumToTensor.Scalar %4404 : !torch.int -> !torch.vtensor<[],si64> | |
| %4407 = torch.prim.NumToTensor.Scalar %4401 : !torch.int -> !torch.vtensor<[],si64> | |
| %4408 = torch.aten.where.self %4405, %4406, %4407 : !torch.vtensor<[],i1>, !torch.vtensor<[],si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[],si64> | |
| %4409 = torch.aten.item %4408 : !torch.vtensor<[],si64> -> !torch.int | |
| %4410 = torch.prim.ListConstruct %4389, %4399, %4409 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %4411 = torch.aten.reshape %4341, %4410 : !torch.vtensor<[?,1,?,?],f32>, !torch.list<int> -> !torch.vtensor<[?,?,?],f32> | |
| %4412 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4413 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4414 = torch.vtensor.literal(dense<2> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4415 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1092 = torch.constant.int 0 | |
| %int0_1093 = torch.constant.int 0 | |
| %4416 = torch.prim.NumToTensor.Scalar %int0_1093 : !torch.int -> !torch.vtensor<[1],si64> | |
| %4417 = torch.aten.index_select %4413, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4418 = torch.aten.item %4417 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4419 = torch.aten.index_select %4414, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4420 = torch.aten.item %4419 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4421 = torch.aten.index_select %4412, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4422 = torch.aten.item %4421 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4423 = torch.aten.index_select %4415, %int0_1092, %4416 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4424 = torch.aten.item %4423 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4425 = torch.aten.slice.Tensor %4411, %4422, %4418, %4420, %4424 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
| %4426 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4427 = torch.vtensor.literal(dense<2> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4428 = torch.vtensor.literal(dense<4> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4429 = torch.vtensor.literal(dense<1> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1094 = torch.constant.int 0 | |
| %int0_1095 = torch.constant.int 0 | |
| %4430 = torch.prim.NumToTensor.Scalar %int0_1095 : !torch.int -> !torch.vtensor<[1],si64> | |
| %4431 = torch.aten.index_select %4427, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4432 = torch.aten.item %4431 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4433 = torch.aten.index_select %4428, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4434 = torch.aten.item %4433 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4435 = torch.aten.index_select %4426, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4436 = torch.aten.item %4435 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4437 = torch.aten.index_select %4429, %int0_1094, %4430 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4438 = torch.aten.item %4437 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4439 = torch.aten.slice.Tensor %4411, %4436, %4432, %4434, %4438 : !torch.vtensor<[?,?,?],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[?,?,?],f32> | |
| %4440 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4441 = torch.aten.mul.Tensor %4425, %4440 : !torch.vtensor<[?,?,?],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[?,?,?],f32> | |
| %int1_1096 = torch.constant.int 1 | |
| %4442 = torch.aten.sub.Tensor %152, %4441, %int1_1096 : !torch.vtensor<[1,2,8400],f32>, !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
| %4443 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4444 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1097 = torch.constant.int 12 | |
| %4445 = torch.aten.item %4443 : !torch.vtensor<[],f32> -> !torch.float | |
| %4446 = torch.aten.item %4444 : !torch.vtensor<[],si8> -> !torch.int | |
| %4447 = torch.aten.quantize_per_tensor %4442, %4445, %4446, %int12_1097 : !torch.vtensor<[?,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
| %4448 = torch.aten.int_repr %4447 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],si8> | |
| %4449 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4450 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4451 = torch.aten.item %4449 : !torch.vtensor<[],f32> -> !torch.float | |
| %4452 = torch.aten.item %4450 : !torch.vtensor<[],si8> -> !torch.int | |
| %4453 = torch.aten._make_per_tensor_quantized_tensor %4448, %4451, %4452 : !torch.vtensor<[?,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
| %4454 = torch.aten.dequantize.self %4453 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],f32> | |
| %int1_1098 = torch.constant.int 1 | |
| %4455 = torch.aten.add.Tensor %152, %4439, %int1_1098 : !torch.vtensor<[1,2,8400],f32>, !torch.vtensor<[?,?,?],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
| %4456 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4457 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1099 = torch.constant.int 12 | |
| %4458 = torch.aten.item %4456 : !torch.vtensor<[],f32> -> !torch.float | |
| %4459 = torch.aten.item %4457 : !torch.vtensor<[],si8> -> !torch.int | |
| %4460 = torch.aten.quantize_per_tensor %4455, %4458, %4459, %int12_1099 : !torch.vtensor<[?,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
| %4461 = torch.aten.int_repr %4460 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],si8> | |
| %4462 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4463 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4464 = torch.aten.item %4462 : !torch.vtensor<[],f32> -> !torch.float | |
| %4465 = torch.aten.item %4463 : !torch.vtensor<[],si8> -> !torch.int | |
| %4466 = torch.aten._make_per_tensor_quantized_tensor %4461, %4464, %4465 : !torch.vtensor<[?,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
| %4467 = torch.aten.dequantize.self %4466 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],f32> | |
| %int1_1100 = torch.constant.int 1 | |
| %4468 = torch.aten.add.Tensor %4454, %4467, %int1_1100 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[?,2,8400],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
| %4469 = torch.vtensor.literal(dense<2.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4470 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1101 = torch.constant.int 12 | |
| %4471 = torch.aten.item %4469 : !torch.vtensor<[],f32> -> !torch.float | |
| %4472 = torch.aten.item %4470 : !torch.vtensor<[],si8> -> !torch.int | |
| %4473 = torch.aten.quantize_per_tensor %4468, %4471, %4472, %int12_1101 : !torch.vtensor<[?,2,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
| %4474 = torch.aten.int_repr %4473 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],si8> | |
| %4475 = torch.vtensor.literal(dense<2.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4476 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4477 = torch.aten.item %4475 : !torch.vtensor<[],f32> -> !torch.float | |
| %4478 = torch.aten.item %4476 : !torch.vtensor<[],si8> -> !torch.int | |
| %4479 = torch.aten._make_per_tensor_quantized_tensor %4474, %4477, %4478 : !torch.vtensor<[?,2,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,2,8400],!torch.qint8> | |
| %4480 = torch.aten.dequantize.self %4479 : !torch.vtensor<[?,2,8400],!torch.qint8> -> !torch.vtensor<[?,2,8400],f32> | |
| %4481 = torch.aten.div.Tensor %4480, %139 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[?,2,8400],f32> | |
| %4482 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4483 = torch.aten.mul.Tensor %4454, %4482 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[?,2,8400],f32> | |
| %int1_1102 = torch.constant.int 1 | |
| %4484 = torch.aten.sub.Tensor %4467, %4483, %int1_1102 : !torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[?,2,8400],f32>, !torch.int -> !torch.vtensor<[?,2,8400],f32> | |
| %4485 = torch.prim.ListConstruct %4481, %4484 : (!torch.vtensor<[?,2,8400],f32>, !torch.vtensor<[?,2,8400],f32>) -> !torch.list<vtensor> | |
| %int1_1103 = torch.constant.int 1 | |
| %4486 = torch.aten.cat %4485, %int1_1103 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[?,4,8400],f32> | |
| %4487 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4488 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1104 = torch.constant.int 12 | |
| %4489 = torch.aten.item %4487 : !torch.vtensor<[],f32> -> !torch.float | |
| %4490 = torch.aten.item %4488 : !torch.vtensor<[],si8> -> !torch.int | |
| %4491 = torch.aten.quantize_per_tensor %4486, %4489, %4490, %int12_1104 : !torch.vtensor<[?,4,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,4,8400],!torch.qint8> | |
| %4492 = torch.aten.int_repr %4491 : !torch.vtensor<[?,4,8400],!torch.qint8> -> !torch.vtensor<[?,4,8400],si8> | |
| %4493 = torch.vtensor.literal(dense<1.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4494 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4495 = torch.aten.item %4493 : !torch.vtensor<[],f32> -> !torch.float | |
| %4496 = torch.aten.item %4494 : !torch.vtensor<[],si8> -> !torch.int | |
| %4497 = torch.aten._make_per_tensor_quantized_tensor %4492, %4495, %4496 : !torch.vtensor<[?,4,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,4,8400],!torch.qint8> | |
| %4498 = torch.aten.dequantize.self %4497 : !torch.vtensor<[?,4,8400],!torch.qint8> -> !torch.vtensor<[?,4,8400],f32> | |
| %4499 = torch.aten.mul.Tensor %4498, %165 : !torch.vtensor<[?,4,8400],f32>, !torch.vtensor<[1,8400],f32> -> !torch.vtensor<[?,4,8400],f32> | |
| %4500 = torch.aten.sigmoid %4197 : !torch.vtensor<[?,?,?],f32> -> !torch.vtensor<[?,?,?],f32> | |
| %4501 = torch.prim.ListConstruct %4499, %4500 : (!torch.vtensor<[?,4,8400],f32>, !torch.vtensor<[?,?,?],f32>) -> !torch.list<vtensor> | |
| %int1_1105 = torch.constant.int 1 | |
| %4502 = torch.aten.cat %4501, %int1_1105 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[?,?,8400],f32> | |
| %4503 = torch.vtensor.literal(dense<8.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4504 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1106 = torch.constant.int 12 | |
| %4505 = torch.aten.item %4503 : !torch.vtensor<[],f32> -> !torch.float | |
| %4506 = torch.aten.item %4504 : !torch.vtensor<[],si8> -> !torch.int | |
| %4507 = torch.aten.quantize_per_tensor %4502, %4505, %4506, %int12_1106 : !torch.vtensor<[?,?,8400],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,8400],!torch.qint8> | |
| %4508 = torch.aten.int_repr %4507 : !torch.vtensor<[?,?,8400],!torch.qint8> -> !torch.vtensor<[?,?,8400],si8> | |
| %4509 = torch.vtensor.literal(dense<8.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4510 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4511 = torch.aten.item %4509 : !torch.vtensor<[],f32> -> !torch.float | |
| %4512 = torch.aten.item %4510 : !torch.vtensor<[],si8> -> !torch.int | |
| %4513 = torch.aten._make_per_tensor_quantized_tensor %4508, %4511, %4512 : !torch.vtensor<[?,?,8400],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,84,8400],!torch.qint8> | |
| %4514 = torch.aten.dequantize.self %4513 : !torch.vtensor<[1,84,8400],!torch.qint8> -> !torch.vtensor<[1,84,8400],f32> | |
| return %4514, %3359, %3651, %3943 : !torch.vtensor<[1,84,8400],f32>, !torch.vtensor<[1,144,80,80],f32>, !torch.vtensor<[1,144,40,40],f32>, !torch.vtensor<[1,144,20,20],f32> | |
| } | |
| } | |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment