Created
April 19, 2024 21:58
-
-
Save AmosLewis/cb9d7a443e66997a57e2297cf571293a to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| module { | |
| func.func @torch_jit(%arg0: !torch.vtensor<[32,3,256,256],f32>) -> !torch.vtensor<[32,1,256,256],f32> attributes {torch.onnx_meta.ir_version = 8 : si64, torch.onnx_meta.opset_version = 17 : si64, torch.onnx_meta.producer_name = "pytorch", torch.onnx_meta.producer_version = "1.13.1"} { | |
| %float1.250000e-01 = torch.constant.float 1.250000e-01 | |
| %float4.882810e-04 = torch.constant.float 4.8828125E-4 | |
| %float3.125000e-02 = torch.constant.float 3.125000e-02 | |
| %float1.953130e-03 = torch.constant.float 0.001953125 | |
| %float3.906250e-03 = torch.constant.float 3.906250e-03 | |
| %float9.765620e-04 = torch.constant.float 9.765625E-4 | |
| %float7.812500e-03 = torch.constant.float 7.812500e-03 | |
| %true = torch.constant.bool true | |
| %int2 = torch.constant.int 2 | |
| %false = torch.constant.bool false | |
| %int0 = torch.constant.int 0 | |
| %int1 = torch.constant.int 1 | |
| %0 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x3x3x3xf32>) : !torch.vtensor<[32,3,3,3],f32> | |
| %1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %4 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x32x3x3xf32>) : !torch.vtensor<[64,32,3,3],f32> | |
| %5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %6 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %7 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %8 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x64x3x3xf32>) : !torch.vtensor<[128,64,3,3],f32> | |
| %9 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %10 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %11 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %12 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x128x3x3xf32>) : !torch.vtensor<[256,128,3,3],f32> | |
| %13 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %14 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x256x3x3xf32>) : !torch.vtensor<[256,256,3,3],f32> | |
| %15 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %16 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512x256x3x3xf32>) : !torch.vtensor<[512,256,3,3],f32> | |
| %17 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.vtensor<[512],f32> | |
| %18 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512x512x3x3xf32>) : !torch.vtensor<[512,512,3,3],f32> | |
| %19 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.vtensor<[512],f32> | |
| %20 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512x256x2x2xf32>) : !torch.vtensor<[512,256,2,2],f32> | |
| %21 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %22 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x512x3x3xf32>) : !torch.vtensor<[256,512,3,3],f32> | |
| %23 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %24 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x256x3x3xf32>) : !torch.vtensor<[256,256,3,3],f32> | |
| %25 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256xf32>) : !torch.vtensor<[256],f32> | |
| %26 = torch.vtensor.literal(dense_resource<__elided__> : tensor<256x128x2x2xf32>) : !torch.vtensor<[256,128,2,2],f32> | |
| %27 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %28 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x256x3x3xf32>) : !torch.vtensor<[128,256,3,3],f32> | |
| %29 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %30 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x128x3x3xf32>) : !torch.vtensor<[128,128,3,3],f32> | |
| %31 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %32 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x64x2x2xf32>) : !torch.vtensor<[128,64,2,2],f32> | |
| %33 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %34 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x128x3x3xf32>) : !torch.vtensor<[64,128,3,3],f32> | |
| %35 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %36 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x64x3x3xf32>) : !torch.vtensor<[64,64,3,3],f32> | |
| %37 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %38 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x32x2x2xf32>) : !torch.vtensor<[64,32,2,2],f32> | |
| %39 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %40 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x64x3x3xf32>) : !torch.vtensor<[32,64,3,3],f32> | |
| %41 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %42 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x32x3x3xf32>) : !torch.vtensor<[32,32,3,3],f32> | |
| %43 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %44 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x32x1x1xf32>) : !torch.vtensor<[1,32,1,1],f32> | |
| %45 = torch.vtensor.literal(dense<-0.3984375> : tensor<1xf32>) : !torch.vtensor<[1],f32> | |
| %int12 = torch.constant.int 12 | |
| %float1.562500e-02 = torch.constant.float 1.562500e-02 | |
| %46 = torch.aten.quantize_per_tensor %arg0, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,3,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,3,256,256],!torch.qint8> | |
| %47 = torch.aten.int_repr %46 : !torch.vtensor<[32,3,256,256],!torch.qint8> -> !torch.vtensor<[32,3,256,256],si8> | |
| %48 = torch.aten._make_per_tensor_quantized_tensor %47, %float1.562500e-02, %int0 : !torch.vtensor<[32,3,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,3,256,256],!torch.qint8> | |
| %49 = torch.aten.dequantize.self %48 : !torch.vtensor<[32,3,256,256],!torch.qint8> -> !torch.vtensor<[32,3,256,256],f32> | |
| %50 = torch.aten.quantize_per_tensor %0, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[32,3,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,3,3,3],!torch.qint8> | |
| %51 = torch.aten.int_repr %50 : !torch.vtensor<[32,3,3,3],!torch.qint8> -> !torch.vtensor<[32,3,3,3],si8> | |
| %52 = torch.aten._make_per_tensor_quantized_tensor %51, %float7.812500e-03, %int0 : !torch.vtensor<[32,3,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,3,3,3],!torch.qint8> | |
| %53 = torch.aten.dequantize.self %52 : !torch.vtensor<[32,3,3,3],!torch.qint8> -> !torch.vtensor<[32,3,3,3],f32> | |
| %54 = torch.aten.quantize_per_tensor %1, %float9.765620e-04, %int0, %int12 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %55 = torch.aten.int_repr %54 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %56 = torch.aten._make_per_tensor_quantized_tensor %55, %float9.765620e-04, %int0 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %57 = torch.aten.dequantize.self %56 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %58 = torch.prim.ListConstruct %int1, %int1 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %59 = torch.prim.ListConstruct %int0, %int0 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %60 = torch.aten.convolution %49, %53, %57, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,3,256,256],f32>, !torch.vtensor<[32,3,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,32,256,256],f32> | |
| %61 = torch.aten.relu %60 : !torch.vtensor<[32,32,256,256],f32> -> !torch.vtensor<[32,32,256,256],f32> | |
| %62 = torch.aten.quantize_per_tensor %61, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[32,32,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %63 = torch.aten.int_repr %62 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],si8> | |
| %64 = torch.aten._make_per_tensor_quantized_tensor %63, %float7.812500e-03, %int0 : !torch.vtensor<[32,32,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %65 = torch.aten.dequantize.self %64 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],f32> | |
| %66 = torch.aten.quantize_per_tensor %2, %float3.906250e-03, %int0, %int12 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %67 = torch.aten.int_repr %66 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %68 = torch.aten._make_per_tensor_quantized_tensor %67, %float3.906250e-03, %int0 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %69 = torch.aten.dequantize.self %68 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %70 = torch.aten.quantize_per_tensor %3, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %71 = torch.aten.int_repr %70 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %72 = torch.aten._make_per_tensor_quantized_tensor %71, %float7.812500e-03, %int0 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %73 = torch.aten.dequantize.self %72 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %74 = torch.aten.convolution %65, %69, %73, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,32,256,256],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,32,256,256],f32> | |
| %75 = torch.aten.relu %74 : !torch.vtensor<[32,32,256,256],f32> -> !torch.vtensor<[32,32,256,256],f32> | |
| %76 = torch.aten.quantize_per_tensor %75, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,32,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %77 = torch.aten.int_repr %76 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],si8> | |
| %78 = torch.aten._make_per_tensor_quantized_tensor %77, %float1.562500e-02, %int0 : !torch.vtensor<[32,32,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %79 = torch.aten.dequantize.self %78 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],f32> | |
| %80 = torch.prim.ListConstruct %int2, %int2 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %81 = torch.aten.max_pool2d %79, %80, %80, %59, %58, %false : !torch.vtensor<[32,32,256,256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[32,32,128,128],f32> | |
| %82 = torch.aten.quantize_per_tensor %81, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,32,128,128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,128,128],!torch.qint8> | |
| %83 = torch.aten.int_repr %82 : !torch.vtensor<[32,32,128,128],!torch.qint8> -> !torch.vtensor<[32,32,128,128],si8> | |
| %84 = torch.aten._make_per_tensor_quantized_tensor %83, %float1.562500e-02, %int0 : !torch.vtensor<[32,32,128,128],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,128,128],!torch.qint8> | |
| %85 = torch.aten.dequantize.self %84 : !torch.vtensor<[32,32,128,128],!torch.qint8> -> !torch.vtensor<[32,32,128,128],f32> | |
| %86 = torch.aten.quantize_per_tensor %4, %float3.906250e-03, %int0, %int12 : !torch.vtensor<[64,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,32,3,3],!torch.qint8> | |
| %87 = torch.aten.int_repr %86 : !torch.vtensor<[64,32,3,3],!torch.qint8> -> !torch.vtensor<[64,32,3,3],si8> | |
| %88 = torch.aten._make_per_tensor_quantized_tensor %87, %float3.906250e-03, %int0 : !torch.vtensor<[64,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,32,3,3],!torch.qint8> | |
| %89 = torch.aten.dequantize.self %88 : !torch.vtensor<[64,32,3,3],!torch.qint8> -> !torch.vtensor<[64,32,3,3],f32> | |
| %90 = torch.aten.quantize_per_tensor %5, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %91 = torch.aten.int_repr %90 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %92 = torch.aten._make_per_tensor_quantized_tensor %91, %float7.812500e-03, %int0 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %93 = torch.aten.dequantize.self %92 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %94 = torch.aten.convolution %85, %89, %93, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,32,128,128],f32>, !torch.vtensor<[64,32,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,64,128,128],f32> | |
| %95 = torch.aten.relu %94 : !torch.vtensor<[32,64,128,128],f32> -> !torch.vtensor<[32,64,128,128],f32> | |
| %96 = torch.aten.quantize_per_tensor %95, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,64,128,128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %97 = torch.aten.int_repr %96 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],si8> | |
| %98 = torch.aten._make_per_tensor_quantized_tensor %97, %float1.562500e-02, %int0 : !torch.vtensor<[32,64,128,128],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %99 = torch.aten.dequantize.self %98 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],f32> | |
| %100 = torch.aten.quantize_per_tensor %6, %float3.906250e-03, %int0, %int12 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %101 = torch.aten.int_repr %100 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %102 = torch.aten._make_per_tensor_quantized_tensor %101, %float3.906250e-03, %int0 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %103 = torch.aten.dequantize.self %102 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %104 = torch.aten.quantize_per_tensor %7, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %105 = torch.aten.int_repr %104 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %106 = torch.aten._make_per_tensor_quantized_tensor %105, %float7.812500e-03, %int0 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %107 = torch.aten.dequantize.self %106 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %108 = torch.aten.convolution %99, %103, %107, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,64,128,128],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,64,128,128],f32> | |
| %109 = torch.aten.relu %108 : !torch.vtensor<[32,64,128,128],f32> -> !torch.vtensor<[32,64,128,128],f32> | |
| %110 = torch.aten.quantize_per_tensor %109, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,64,128,128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %111 = torch.aten.int_repr %110 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],si8> | |
| %112 = torch.aten._make_per_tensor_quantized_tensor %111, %float1.562500e-02, %int0 : !torch.vtensor<[32,64,128,128],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %113 = torch.aten.dequantize.self %112 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],f32> | |
| %114 = torch.aten.max_pool2d %113, %80, %80, %59, %58, %false : !torch.vtensor<[32,64,128,128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[32,64,64,64],f32> | |
| %115 = torch.aten.quantize_per_tensor %114, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,64,64,64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,64,64],!torch.qint8> | |
| %116 = torch.aten.int_repr %115 : !torch.vtensor<[32,64,64,64],!torch.qint8> -> !torch.vtensor<[32,64,64,64],si8> | |
| %117 = torch.aten._make_per_tensor_quantized_tensor %116, %float1.562500e-02, %int0 : !torch.vtensor<[32,64,64,64],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,64,64],!torch.qint8> | |
| %118 = torch.aten.dequantize.self %117 : !torch.vtensor<[32,64,64,64],!torch.qint8> -> !torch.vtensor<[32,64,64,64],f32> | |
| %119 = torch.aten.quantize_per_tensor %8, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[128,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,64,3,3],!torch.qint8> | |
| %120 = torch.aten.int_repr %119 : !torch.vtensor<[128,64,3,3],!torch.qint8> -> !torch.vtensor<[128,64,3,3],si8> | |
| %121 = torch.aten._make_per_tensor_quantized_tensor %120, %float1.953130e-03, %int0 : !torch.vtensor<[128,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,64,3,3],!torch.qint8> | |
| %122 = torch.aten.dequantize.self %121 : !torch.vtensor<[128,64,3,3],!torch.qint8> -> !torch.vtensor<[128,64,3,3],f32> | |
| %123 = torch.aten.quantize_per_tensor %9, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %124 = torch.aten.int_repr %123 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %125 = torch.aten._make_per_tensor_quantized_tensor %124, %float7.812500e-03, %int0 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %126 = torch.aten.dequantize.self %125 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %127 = torch.aten.convolution %118, %122, %126, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,64,64,64],f32>, !torch.vtensor<[128,64,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,128,64,64],f32> | |
| %128 = torch.aten.relu %127 : !torch.vtensor<[32,128,64,64],f32> -> !torch.vtensor<[32,128,64,64],f32> | |
| %129 = torch.aten.quantize_per_tensor %128, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,128,64,64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %130 = torch.aten.int_repr %129 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],si8> | |
| %131 = torch.aten._make_per_tensor_quantized_tensor %130, %float1.562500e-02, %int0 : !torch.vtensor<[32,128,64,64],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %132 = torch.aten.dequantize.self %131 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],f32> | |
| %133 = torch.aten.quantize_per_tensor %10, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %134 = torch.aten.int_repr %133 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %135 = torch.aten._make_per_tensor_quantized_tensor %134, %float1.953130e-03, %int0 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %136 = torch.aten.dequantize.self %135 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %137 = torch.aten.quantize_per_tensor %11, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %138 = torch.aten.int_repr %137 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %139 = torch.aten._make_per_tensor_quantized_tensor %138, %float1.562500e-02, %int0 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %140 = torch.aten.dequantize.self %139 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %141 = torch.aten.convolution %132, %136, %140, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,128,64,64],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,128,64,64],f32> | |
| %142 = torch.aten.relu %141 : !torch.vtensor<[32,128,64,64],f32> -> !torch.vtensor<[32,128,64,64],f32> | |
| %143 = torch.aten.quantize_per_tensor %142, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,128,64,64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %144 = torch.aten.int_repr %143 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],si8> | |
| %145 = torch.aten._make_per_tensor_quantized_tensor %144, %float3.125000e-02, %int0 : !torch.vtensor<[32,128,64,64],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %146 = torch.aten.dequantize.self %145 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],f32> | |
| %147 = torch.aten.max_pool2d %146, %80, %80, %59, %58, %false : !torch.vtensor<[32,128,64,64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[32,128,32,32],f32> | |
| %148 = torch.aten.quantize_per_tensor %147, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,128,32,32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,128,32,32],!torch.qint8> | |
| %149 = torch.aten.int_repr %148 : !torch.vtensor<[32,128,32,32],!torch.qint8> -> !torch.vtensor<[32,128,32,32],si8> | |
| %150 = torch.aten._make_per_tensor_quantized_tensor %149, %float3.125000e-02, %int0 : !torch.vtensor<[32,128,32,32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,128,32,32],!torch.qint8> | |
| %151 = torch.aten.dequantize.self %150 : !torch.vtensor<[32,128,32,32],!torch.qint8> -> !torch.vtensor<[32,128,32,32],f32> | |
| %152 = torch.aten.quantize_per_tensor %12, %float9.765620e-04, %int0, %int12 : !torch.vtensor<[256,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,128,3,3],!torch.qint8> | |
| %153 = torch.aten.int_repr %152 : !torch.vtensor<[256,128,3,3],!torch.qint8> -> !torch.vtensor<[256,128,3,3],si8> | |
| %154 = torch.aten._make_per_tensor_quantized_tensor %153, %float9.765620e-04, %int0 : !torch.vtensor<[256,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,128,3,3],!torch.qint8> | |
| %155 = torch.aten.dequantize.self %154 : !torch.vtensor<[256,128,3,3],!torch.qint8> -> !torch.vtensor<[256,128,3,3],f32> | |
| %156 = torch.aten.quantize_per_tensor %13, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %157 = torch.aten.int_repr %156 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %158 = torch.aten._make_per_tensor_quantized_tensor %157, %float7.812500e-03, %int0 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %159 = torch.aten.dequantize.self %158 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %160 = torch.aten.convolution %151, %155, %159, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,128,32,32],f32>, !torch.vtensor<[256,128,3,3],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,256,32,32],f32> | |
| %161 = torch.aten.relu %160 : !torch.vtensor<[32,256,32,32],f32> -> !torch.vtensor<[32,256,32,32],f32> | |
| %162 = torch.aten.quantize_per_tensor %161, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,256,32,32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %163 = torch.aten.int_repr %162 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],si8> | |
| %164 = torch.aten._make_per_tensor_quantized_tensor %163, %float3.125000e-02, %int0 : !torch.vtensor<[32,256,32,32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %165 = torch.aten.dequantize.self %164 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],f32> | |
| %166 = torch.aten.quantize_per_tensor %14, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[256,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,256,3,3],!torch.qint8> | |
| %167 = torch.aten.int_repr %166 : !torch.vtensor<[256,256,3,3],!torch.qint8> -> !torch.vtensor<[256,256,3,3],si8> | |
| %168 = torch.aten._make_per_tensor_quantized_tensor %167, %float1.953130e-03, %int0 : !torch.vtensor<[256,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,256,3,3],!torch.qint8> | |
| %169 = torch.aten.dequantize.self %168 : !torch.vtensor<[256,256,3,3],!torch.qint8> -> !torch.vtensor<[256,256,3,3],f32> | |
| %170 = torch.aten.quantize_per_tensor %15, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %171 = torch.aten.int_repr %170 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %172 = torch.aten._make_per_tensor_quantized_tensor %171, %float1.562500e-02, %int0 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %173 = torch.aten.dequantize.self %172 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %174 = torch.aten.convolution %165, %169, %173, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,256,32,32],f32>, !torch.vtensor<[256,256,3,3],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,256,32,32],f32> | |
| %175 = torch.aten.relu %174 : !torch.vtensor<[32,256,32,32],f32> -> !torch.vtensor<[32,256,32,32],f32> | |
| %176 = torch.aten.quantize_per_tensor %175, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,256,32,32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %177 = torch.aten.int_repr %176 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],si8> | |
| %178 = torch.aten._make_per_tensor_quantized_tensor %177, %float1.562500e-02, %int0 : !torch.vtensor<[32,256,32,32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %179 = torch.aten.dequantize.self %178 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],f32> | |
| %180 = torch.aten.max_pool2d %179, %80, %80, %59, %58, %false : !torch.vtensor<[32,256,32,32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool -> !torch.vtensor<[32,256,16,16],f32> | |
| %181 = torch.aten.quantize_per_tensor %180, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,256,16,16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,256,16,16],!torch.qint8> | |
| %182 = torch.aten.int_repr %181 : !torch.vtensor<[32,256,16,16],!torch.qint8> -> !torch.vtensor<[32,256,16,16],si8> | |
| %183 = torch.aten._make_per_tensor_quantized_tensor %182, %float1.562500e-02, %int0 : !torch.vtensor<[32,256,16,16],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,256,16,16],!torch.qint8> | |
| %184 = torch.aten.dequantize.self %183 : !torch.vtensor<[32,256,16,16],!torch.qint8> -> !torch.vtensor<[32,256,16,16],f32> | |
| %185 = torch.aten.quantize_per_tensor %16, %float4.882810e-04, %int0, %int12 : !torch.vtensor<[512,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[512,256,3,3],!torch.qint8> | |
| %186 = torch.aten.int_repr %185 : !torch.vtensor<[512,256,3,3],!torch.qint8> -> !torch.vtensor<[512,256,3,3],si8> | |
| %187 = torch.aten._make_per_tensor_quantized_tensor %186, %float4.882810e-04, %int0 : !torch.vtensor<[512,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[512,256,3,3],!torch.qint8> | |
| %188 = torch.aten.dequantize.self %187 : !torch.vtensor<[512,256,3,3],!torch.qint8> -> !torch.vtensor<[512,256,3,3],f32> | |
| %189 = torch.aten.quantize_per_tensor %17, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[512],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[512],!torch.qint8> | |
| %190 = torch.aten.int_repr %189 : !torch.vtensor<[512],!torch.qint8> -> !torch.vtensor<[512],si8> | |
| %191 = torch.aten._make_per_tensor_quantized_tensor %190, %float7.812500e-03, %int0 : !torch.vtensor<[512],si8>, !torch.float, !torch.int -> !torch.vtensor<[512],!torch.qint8> | |
| %192 = torch.aten.dequantize.self %191 : !torch.vtensor<[512],!torch.qint8> -> !torch.vtensor<[512],f32> | |
| %193 = torch.aten.convolution %184, %188, %192, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,256,16,16],f32>, !torch.vtensor<[512,256,3,3],f32>, !torch.vtensor<[512],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,512,16,16],f32> | |
| %194 = torch.aten.relu %193 : !torch.vtensor<[32,512,16,16],f32> -> !torch.vtensor<[32,512,16,16],f32> | |
| %195 = torch.aten.quantize_per_tensor %194, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,512,16,16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,512,16,16],!torch.qint8> | |
| %196 = torch.aten.int_repr %195 : !torch.vtensor<[32,512,16,16],!torch.qint8> -> !torch.vtensor<[32,512,16,16],si8> | |
| %197 = torch.aten._make_per_tensor_quantized_tensor %196, %float3.125000e-02, %int0 : !torch.vtensor<[32,512,16,16],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,512,16,16],!torch.qint8> | |
| %198 = torch.aten.dequantize.self %197 : !torch.vtensor<[32,512,16,16],!torch.qint8> -> !torch.vtensor<[32,512,16,16],f32> | |
| %199 = torch.aten.quantize_per_tensor %18, %float9.765620e-04, %int0, %int12 : !torch.vtensor<[512,512,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[512,512,3,3],!torch.qint8> | |
| %200 = torch.aten.int_repr %199 : !torch.vtensor<[512,512,3,3],!torch.qint8> -> !torch.vtensor<[512,512,3,3],si8> | |
| %201 = torch.aten._make_per_tensor_quantized_tensor %200, %float9.765620e-04, %int0 : !torch.vtensor<[512,512,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[512,512,3,3],!torch.qint8> | |
| %202 = torch.aten.dequantize.self %201 : !torch.vtensor<[512,512,3,3],!torch.qint8> -> !torch.vtensor<[512,512,3,3],f32> | |
| %203 = torch.aten.quantize_per_tensor %19, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[512],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[512],!torch.qint8> | |
| %204 = torch.aten.int_repr %203 : !torch.vtensor<[512],!torch.qint8> -> !torch.vtensor<[512],si8> | |
| %205 = torch.aten._make_per_tensor_quantized_tensor %204, %float7.812500e-03, %int0 : !torch.vtensor<[512],si8>, !torch.float, !torch.int -> !torch.vtensor<[512],!torch.qint8> | |
| %206 = torch.aten.dequantize.self %205 : !torch.vtensor<[512],!torch.qint8> -> !torch.vtensor<[512],f32> | |
| %207 = torch.aten.convolution %198, %202, %206, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,512,16,16],f32>, !torch.vtensor<[512,512,3,3],f32>, !torch.vtensor<[512],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,512,16,16],f32> | |
| %208 = torch.aten.relu %207 : !torch.vtensor<[32,512,16,16],f32> -> !torch.vtensor<[32,512,16,16],f32> | |
| %209 = torch.aten.quantize_per_tensor %208, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,512,16,16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,512,16,16],!torch.qint8> | |
| %210 = torch.aten.int_repr %209 : !torch.vtensor<[32,512,16,16],!torch.qint8> -> !torch.vtensor<[32,512,16,16],si8> | |
| %211 = torch.aten._make_per_tensor_quantized_tensor %210, %float1.562500e-02, %int0 : !torch.vtensor<[32,512,16,16],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,512,16,16],!torch.qint8> | |
| %212 = torch.aten.dequantize.self %211 : !torch.vtensor<[32,512,16,16],!torch.qint8> -> !torch.vtensor<[32,512,16,16],f32> | |
| %213 = torch.aten.quantize_per_tensor %20, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[512,256,2,2],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[512,256,2,2],!torch.qint8> | |
| %214 = torch.aten.int_repr %213 : !torch.vtensor<[512,256,2,2],!torch.qint8> -> !torch.vtensor<[512,256,2,2],si8> | |
| %215 = torch.aten._make_per_tensor_quantized_tensor %214, %float1.953130e-03, %int0 : !torch.vtensor<[512,256,2,2],si8>, !torch.float, !torch.int -> !torch.vtensor<[512,256,2,2],!torch.qint8> | |
| %216 = torch.aten.dequantize.self %215 : !torch.vtensor<[512,256,2,2],!torch.qint8> -> !torch.vtensor<[512,256,2,2],f32> | |
| %217 = torch.aten.quantize_per_tensor %21, %float4.882810e-04, %int0, %int12 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %218 = torch.aten.int_repr %217 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %219 = torch.aten._make_per_tensor_quantized_tensor %218, %float4.882810e-04, %int0 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %220 = torch.aten.dequantize.self %219 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %221 = torch.aten.convolution %212, %216, %220, %80, %59, %58, %true, %59, %int1 : !torch.vtensor<[32,512,16,16],f32>, !torch.vtensor<[512,256,2,2],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,256,32,32],f32> | |
| %222 = torch.prim.ListConstruct %221, %179 : (!torch.vtensor<[32,256,32,32],f32>, !torch.vtensor<[32,256,32,32],f32>) -> !torch.list<vtensor> | |
| %223 = torch.aten.cat %222, %int1 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[32,512,32,32],f32> | |
| %224 = torch.aten.quantize_per_tensor %223, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,512,32,32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,512,32,32],!torch.qint8> | |
| %225 = torch.aten.int_repr %224 : !torch.vtensor<[32,512,32,32],!torch.qint8> -> !torch.vtensor<[32,512,32,32],si8> | |
| %226 = torch.aten._make_per_tensor_quantized_tensor %225, %float1.562500e-02, %int0 : !torch.vtensor<[32,512,32,32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,512,32,32],!torch.qint8> | |
| %227 = torch.aten.dequantize.self %226 : !torch.vtensor<[32,512,32,32],!torch.qint8> -> !torch.vtensor<[32,512,32,32],f32> | |
| %228 = torch.aten.quantize_per_tensor %22, %float4.882810e-04, %int0, %int12 : !torch.vtensor<[256,512,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,512,3,3],!torch.qint8> | |
| %229 = torch.aten.int_repr %228 : !torch.vtensor<[256,512,3,3],!torch.qint8> -> !torch.vtensor<[256,512,3,3],si8> | |
| %230 = torch.aten._make_per_tensor_quantized_tensor %229, %float4.882810e-04, %int0 : !torch.vtensor<[256,512,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,512,3,3],!torch.qint8> | |
| %231 = torch.aten.dequantize.self %230 : !torch.vtensor<[256,512,3,3],!torch.qint8> -> !torch.vtensor<[256,512,3,3],f32> | |
| %232 = torch.aten.quantize_per_tensor %23, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %233 = torch.aten.int_repr %232 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %234 = torch.aten._make_per_tensor_quantized_tensor %233, %float1.562500e-02, %int0 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %235 = torch.aten.dequantize.self %234 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %236 = torch.aten.convolution %227, %231, %235, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,512,32,32],f32>, !torch.vtensor<[256,512,3,3],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,256,32,32],f32> | |
| %237 = torch.aten.relu %236 : !torch.vtensor<[32,256,32,32],f32> -> !torch.vtensor<[32,256,32,32],f32> | |
| %238 = torch.aten.quantize_per_tensor %237, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,256,32,32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %239 = torch.aten.int_repr %238 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],si8> | |
| %240 = torch.aten._make_per_tensor_quantized_tensor %239, %float1.562500e-02, %int0 : !torch.vtensor<[32,256,32,32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %241 = torch.aten.dequantize.self %240 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],f32> | |
| %242 = torch.aten.quantize_per_tensor %24, %float9.765620e-04, %int0, %int12 : !torch.vtensor<[256,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,256,3,3],!torch.qint8> | |
| %243 = torch.aten.int_repr %242 : !torch.vtensor<[256,256,3,3],!torch.qint8> -> !torch.vtensor<[256,256,3,3],si8> | |
| %244 = torch.aten._make_per_tensor_quantized_tensor %243, %float9.765620e-04, %int0 : !torch.vtensor<[256,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,256,3,3],!torch.qint8> | |
| %245 = torch.aten.dequantize.self %244 : !torch.vtensor<[256,256,3,3],!torch.qint8> -> !torch.vtensor<[256,256,3,3],f32> | |
| %246 = torch.aten.quantize_per_tensor %25, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %247 = torch.aten.int_repr %246 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],si8> | |
| %248 = torch.aten._make_per_tensor_quantized_tensor %247, %float7.812500e-03, %int0 : !torch.vtensor<[256],si8>, !torch.float, !torch.int -> !torch.vtensor<[256],!torch.qint8> | |
| %249 = torch.aten.dequantize.self %248 : !torch.vtensor<[256],!torch.qint8> -> !torch.vtensor<[256],f32> | |
| %250 = torch.aten.convolution %241, %245, %249, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,256,32,32],f32>, !torch.vtensor<[256,256,3,3],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,256,32,32],f32> | |
| %251 = torch.aten.relu %250 : !torch.vtensor<[32,256,32,32],f32> -> !torch.vtensor<[32,256,32,32],f32> | |
| %252 = torch.aten.quantize_per_tensor %251, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,256,32,32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %253 = torch.aten.int_repr %252 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],si8> | |
| %254 = torch.aten._make_per_tensor_quantized_tensor %253, %float1.562500e-02, %int0 : !torch.vtensor<[32,256,32,32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,256,32,32],!torch.qint8> | |
| %255 = torch.aten.dequantize.self %254 : !torch.vtensor<[32,256,32,32],!torch.qint8> -> !torch.vtensor<[32,256,32,32],f32> | |
| %256 = torch.aten.quantize_per_tensor %26, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[256,128,2,2],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[256,128,2,2],!torch.qint8> | |
| %257 = torch.aten.int_repr %256 : !torch.vtensor<[256,128,2,2],!torch.qint8> -> !torch.vtensor<[256,128,2,2],si8> | |
| %258 = torch.aten._make_per_tensor_quantized_tensor %257, %float1.953130e-03, %int0 : !torch.vtensor<[256,128,2,2],si8>, !torch.float, !torch.int -> !torch.vtensor<[256,128,2,2],!torch.qint8> | |
| %259 = torch.aten.dequantize.self %258 : !torch.vtensor<[256,128,2,2],!torch.qint8> -> !torch.vtensor<[256,128,2,2],f32> | |
| %260 = torch.aten.quantize_per_tensor %27, %float9.765620e-04, %int0, %int12 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %261 = torch.aten.int_repr %260 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %262 = torch.aten._make_per_tensor_quantized_tensor %261, %float9.765620e-04, %int0 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %263 = torch.aten.dequantize.self %262 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %264 = torch.aten.convolution %255, %259, %263, %80, %59, %58, %true, %59, %int1 : !torch.vtensor<[32,256,32,32],f32>, !torch.vtensor<[256,128,2,2],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,128,64,64],f32> | |
| %265 = torch.prim.ListConstruct %264, %146 : (!torch.vtensor<[32,128,64,64],f32>, !torch.vtensor<[32,128,64,64],f32>) -> !torch.list<vtensor> | |
| %266 = torch.aten.cat %265, %int1 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[32,256,64,64],f32> | |
| %267 = torch.aten.quantize_per_tensor %266, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,256,64,64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,256,64,64],!torch.qint8> | |
| %268 = torch.aten.int_repr %267 : !torch.vtensor<[32,256,64,64],!torch.qint8> -> !torch.vtensor<[32,256,64,64],si8> | |
| %269 = torch.aten._make_per_tensor_quantized_tensor %268, %float3.125000e-02, %int0 : !torch.vtensor<[32,256,64,64],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,256,64,64],!torch.qint8> | |
| %270 = torch.aten.dequantize.self %269 : !torch.vtensor<[32,256,64,64],!torch.qint8> -> !torch.vtensor<[32,256,64,64],f32> | |
| %271 = torch.aten.quantize_per_tensor %28, %float9.765620e-04, %int0, %int12 : !torch.vtensor<[128,256,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,256,3,3],!torch.qint8> | |
| %272 = torch.aten.int_repr %271 : !torch.vtensor<[128,256,3,3],!torch.qint8> -> !torch.vtensor<[128,256,3,3],si8> | |
| %273 = torch.aten._make_per_tensor_quantized_tensor %272, %float9.765620e-04, %int0 : !torch.vtensor<[128,256,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,256,3,3],!torch.qint8> | |
| %274 = torch.aten.dequantize.self %273 : !torch.vtensor<[128,256,3,3],!torch.qint8> -> !torch.vtensor<[128,256,3,3],f32> | |
| %275 = torch.aten.quantize_per_tensor %29, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %276 = torch.aten.int_repr %275 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %277 = torch.aten._make_per_tensor_quantized_tensor %276, %float7.812500e-03, %int0 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %278 = torch.aten.dequantize.self %277 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %279 = torch.aten.convolution %270, %274, %278, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,256,64,64],f32>, !torch.vtensor<[128,256,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,128,64,64],f32> | |
| %280 = torch.aten.relu %279 : !torch.vtensor<[32,128,64,64],f32> -> !torch.vtensor<[32,128,64,64],f32> | |
| %281 = torch.aten.quantize_per_tensor %280, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,128,64,64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %282 = torch.aten.int_repr %281 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],si8> | |
| %283 = torch.aten._make_per_tensor_quantized_tensor %282, %float1.562500e-02, %int0 : !torch.vtensor<[32,128,64,64],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %284 = torch.aten.dequantize.self %283 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],f32> | |
| %285 = torch.aten.quantize_per_tensor %30, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[128,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %286 = torch.aten.int_repr %285 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],si8> | |
| %287 = torch.aten._make_per_tensor_quantized_tensor %286, %float1.953130e-03, %int0 : !torch.vtensor<[128,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,128,3,3],!torch.qint8> | |
| %288 = torch.aten.dequantize.self %287 : !torch.vtensor<[128,128,3,3],!torch.qint8> -> !torch.vtensor<[128,128,3,3],f32> | |
| %289 = torch.aten.quantize_per_tensor %31, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %290 = torch.aten.int_repr %289 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %291 = torch.aten._make_per_tensor_quantized_tensor %290, %float7.812500e-03, %int0 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %292 = torch.aten.dequantize.self %291 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %293 = torch.aten.convolution %284, %288, %292, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,128,64,64],f32>, !torch.vtensor<[128,128,3,3],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,128,64,64],f32> | |
| %294 = torch.aten.relu %293 : !torch.vtensor<[32,128,64,64],f32> -> !torch.vtensor<[32,128,64,64],f32> | |
| %295 = torch.aten.quantize_per_tensor %294, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,128,64,64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %296 = torch.aten.int_repr %295 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],si8> | |
| %297 = torch.aten._make_per_tensor_quantized_tensor %296, %float1.562500e-02, %int0 : !torch.vtensor<[32,128,64,64],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,128,64,64],!torch.qint8> | |
| %298 = torch.aten.dequantize.self %297 : !torch.vtensor<[32,128,64,64],!torch.qint8> -> !torch.vtensor<[32,128,64,64],f32> | |
| %299 = torch.aten.quantize_per_tensor %32, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[128,64,2,2],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,64,2,2],!torch.qint8> | |
| %300 = torch.aten.int_repr %299 : !torch.vtensor<[128,64,2,2],!torch.qint8> -> !torch.vtensor<[128,64,2,2],si8> | |
| %301 = torch.aten._make_per_tensor_quantized_tensor %300, %float1.953130e-03, %int0 : !torch.vtensor<[128,64,2,2],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,64,2,2],!torch.qint8> | |
| %302 = torch.aten.dequantize.self %301 : !torch.vtensor<[128,64,2,2],!torch.qint8> -> !torch.vtensor<[128,64,2,2],f32> | |
| %303 = torch.aten.quantize_per_tensor %33, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %304 = torch.aten.int_repr %303 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %305 = torch.aten._make_per_tensor_quantized_tensor %304, %float1.953130e-03, %int0 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %306 = torch.aten.dequantize.self %305 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %307 = torch.aten.convolution %298, %302, %306, %80, %59, %58, %true, %59, %int1 : !torch.vtensor<[32,128,64,64],f32>, !torch.vtensor<[128,64,2,2],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,64,128,128],f32> | |
| %308 = torch.prim.ListConstruct %307, %113 : (!torch.vtensor<[32,64,128,128],f32>, !torch.vtensor<[32,64,128,128],f32>) -> !torch.list<vtensor> | |
| %309 = torch.aten.cat %308, %int1 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[32,128,128,128],f32> | |
| %310 = torch.aten.quantize_per_tensor %309, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,128,128,128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,128,128,128],!torch.qint8> | |
| %311 = torch.aten.int_repr %310 : !torch.vtensor<[32,128,128,128],!torch.qint8> -> !torch.vtensor<[32,128,128,128],si8> | |
| %312 = torch.aten._make_per_tensor_quantized_tensor %311, %float1.562500e-02, %int0 : !torch.vtensor<[32,128,128,128],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,128,128,128],!torch.qint8> | |
| %313 = torch.aten.dequantize.self %312 : !torch.vtensor<[32,128,128,128],!torch.qint8> -> !torch.vtensor<[32,128,128,128],f32> | |
| %314 = torch.aten.quantize_per_tensor %34, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[64,128,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,128,3,3],!torch.qint8> | |
| %315 = torch.aten.int_repr %314 : !torch.vtensor<[64,128,3,3],!torch.qint8> -> !torch.vtensor<[64,128,3,3],si8> | |
| %316 = torch.aten._make_per_tensor_quantized_tensor %315, %float1.953130e-03, %int0 : !torch.vtensor<[64,128,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,128,3,3],!torch.qint8> | |
| %317 = torch.aten.dequantize.self %316 : !torch.vtensor<[64,128,3,3],!torch.qint8> -> !torch.vtensor<[64,128,3,3],f32> | |
| %318 = torch.aten.quantize_per_tensor %35, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %319 = torch.aten.int_repr %318 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %320 = torch.aten._make_per_tensor_quantized_tensor %319, %float7.812500e-03, %int0 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %321 = torch.aten.dequantize.self %320 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %322 = torch.aten.convolution %313, %317, %321, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,128,128,128],f32>, !torch.vtensor<[64,128,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,64,128,128],f32> | |
| %323 = torch.aten.relu %322 : !torch.vtensor<[32,64,128,128],f32> -> !torch.vtensor<[32,64,128,128],f32> | |
| %324 = torch.aten.quantize_per_tensor %323, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,64,128,128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %325 = torch.aten.int_repr %324 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],si8> | |
| %326 = torch.aten._make_per_tensor_quantized_tensor %325, %float1.562500e-02, %int0 : !torch.vtensor<[32,64,128,128],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %327 = torch.aten.dequantize.self %326 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],f32> | |
| %328 = torch.aten.quantize_per_tensor %36, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[64,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %329 = torch.aten.int_repr %328 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],si8> | |
| %330 = torch.aten._make_per_tensor_quantized_tensor %329, %float1.953130e-03, %int0 : !torch.vtensor<[64,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,64,3,3],!torch.qint8> | |
| %331 = torch.aten.dequantize.self %330 : !torch.vtensor<[64,64,3,3],!torch.qint8> -> !torch.vtensor<[64,64,3,3],f32> | |
| %332 = torch.aten.quantize_per_tensor %37, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %333 = torch.aten.int_repr %332 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %334 = torch.aten._make_per_tensor_quantized_tensor %333, %float7.812500e-03, %int0 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %335 = torch.aten.dequantize.self %334 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %336 = torch.aten.convolution %327, %331, %335, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,64,128,128],f32>, !torch.vtensor<[64,64,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,64,128,128],f32> | |
| %337 = torch.aten.relu %336 : !torch.vtensor<[32,64,128,128],f32> -> !torch.vtensor<[32,64,128,128],f32> | |
| %338 = torch.aten.quantize_per_tensor %337, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,64,128,128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %339 = torch.aten.int_repr %338 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],si8> | |
| %340 = torch.aten._make_per_tensor_quantized_tensor %339, %float1.562500e-02, %int0 : !torch.vtensor<[32,64,128,128],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,128,128],!torch.qint8> | |
| %341 = torch.aten.dequantize.self %340 : !torch.vtensor<[32,64,128,128],!torch.qint8> -> !torch.vtensor<[32,64,128,128],f32> | |
| %342 = torch.aten.quantize_per_tensor %38, %float3.906250e-03, %int0, %int12 : !torch.vtensor<[64,32,2,2],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,32,2,2],!torch.qint8> | |
| %343 = torch.aten.int_repr %342 : !torch.vtensor<[64,32,2,2],!torch.qint8> -> !torch.vtensor<[64,32,2,2],si8> | |
| %344 = torch.aten._make_per_tensor_quantized_tensor %343, %float3.906250e-03, %int0 : !torch.vtensor<[64,32,2,2],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,32,2,2],!torch.qint8> | |
| %345 = torch.aten.dequantize.self %344 : !torch.vtensor<[64,32,2,2],!torch.qint8> -> !torch.vtensor<[64,32,2,2],f32> | |
| %346 = torch.aten.quantize_per_tensor %39, %float1.953130e-03, %int0, %int12 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %347 = torch.aten.int_repr %346 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %348 = torch.aten._make_per_tensor_quantized_tensor %347, %float1.953130e-03, %int0 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %349 = torch.aten.dequantize.self %348 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %350 = torch.aten.convolution %341, %345, %349, %80, %59, %58, %true, %59, %int1 : !torch.vtensor<[32,64,128,128],f32>, !torch.vtensor<[64,32,2,2],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,32,256,256],f32> | |
| %351 = torch.prim.ListConstruct %350, %79 : (!torch.vtensor<[32,32,256,256],f32>, !torch.vtensor<[32,32,256,256],f32>) -> !torch.list<vtensor> | |
| %352 = torch.aten.cat %351, %int1 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[32,64,256,256],f32> | |
| %353 = torch.aten.quantize_per_tensor %352, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32,64,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,256,256],!torch.qint8> | |
| %354 = torch.aten.int_repr %353 : !torch.vtensor<[32,64,256,256],!torch.qint8> -> !torch.vtensor<[32,64,256,256],si8> | |
| %355 = torch.aten._make_per_tensor_quantized_tensor %354, %float1.562500e-02, %int0 : !torch.vtensor<[32,64,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,256,256],!torch.qint8> | |
| %356 = torch.aten.dequantize.self %355 : !torch.vtensor<[32,64,256,256],!torch.qint8> -> !torch.vtensor<[32,64,256,256],f32> | |
| %357 = torch.aten.quantize_per_tensor %40, %float3.906250e-03, %int0, %int12 : !torch.vtensor<[32,64,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,64,3,3],!torch.qint8> | |
| %358 = torch.aten.int_repr %357 : !torch.vtensor<[32,64,3,3],!torch.qint8> -> !torch.vtensor<[32,64,3,3],si8> | |
| %359 = torch.aten._make_per_tensor_quantized_tensor %358, %float3.906250e-03, %int0 : !torch.vtensor<[32,64,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,64,3,3],!torch.qint8> | |
| %360 = torch.aten.dequantize.self %359 : !torch.vtensor<[32,64,3,3],!torch.qint8> -> !torch.vtensor<[32,64,3,3],f32> | |
| %361 = torch.aten.quantize_per_tensor %41, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %362 = torch.aten.int_repr %361 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %363 = torch.aten._make_per_tensor_quantized_tensor %362, %float1.562500e-02, %int0 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %364 = torch.aten.dequantize.self %363 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %365 = torch.aten.convolution %356, %360, %364, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,64,256,256],f32>, !torch.vtensor<[32,64,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,32,256,256],f32> | |
| %366 = torch.aten.relu %365 : !torch.vtensor<[32,32,256,256],f32> -> !torch.vtensor<[32,32,256,256],f32> | |
| %367 = torch.aten.quantize_per_tensor %366, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,32,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %368 = torch.aten.int_repr %367 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],si8> | |
| %369 = torch.aten._make_per_tensor_quantized_tensor %368, %float3.125000e-02, %int0 : !torch.vtensor<[32,32,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %370 = torch.aten.dequantize.self %369 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],f32> | |
| %371 = torch.aten.quantize_per_tensor %42, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[32,32,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %372 = torch.aten.int_repr %371 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],si8> | |
| %373 = torch.aten._make_per_tensor_quantized_tensor %372, %float7.812500e-03, %int0 : !torch.vtensor<[32,32,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,3,3],!torch.qint8> | |
| %374 = torch.aten.dequantize.self %373 : !torch.vtensor<[32,32,3,3],!torch.qint8> -> !torch.vtensor<[32,32,3,3],f32> | |
| %375 = torch.aten.quantize_per_tensor %43, %float1.562500e-02, %int0, %int12 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %376 = torch.aten.int_repr %375 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %377 = torch.aten._make_per_tensor_quantized_tensor %376, %float1.562500e-02, %int0 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %378 = torch.aten.dequantize.self %377 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %379 = torch.aten.convolution %370, %374, %378, %58, %58, %58, %false, %59, %int1 : !torch.vtensor<[32,32,256,256],f32>, !torch.vtensor<[32,32,3,3],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,32,256,256],f32> | |
| %380 = torch.aten.relu %379 : !torch.vtensor<[32,32,256,256],f32> -> !torch.vtensor<[32,32,256,256],f32> | |
| %381 = torch.aten.quantize_per_tensor %380, %float3.125000e-02, %int0, %int12 : !torch.vtensor<[32,32,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %382 = torch.aten.int_repr %381 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],si8> | |
| %383 = torch.aten._make_per_tensor_quantized_tensor %382, %float3.125000e-02, %int0 : !torch.vtensor<[32,32,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,32,256,256],!torch.qint8> | |
| %384 = torch.aten.dequantize.self %383 : !torch.vtensor<[32,32,256,256],!torch.qint8> -> !torch.vtensor<[32,32,256,256],f32> | |
| %385 = torch.aten.quantize_per_tensor %44, %float7.812500e-03, %int0, %int12 : !torch.vtensor<[1,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,1,1],!torch.qint8> | |
| %386 = torch.aten.int_repr %385 : !torch.vtensor<[1,32,1,1],!torch.qint8> -> !torch.vtensor<[1,32,1,1],si8> | |
| %387 = torch.aten._make_per_tensor_quantized_tensor %386, %float7.812500e-03, %int0 : !torch.vtensor<[1,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,1,1],!torch.qint8> | |
| %388 = torch.aten.dequantize.self %387 : !torch.vtensor<[1,32,1,1],!torch.qint8> -> !torch.vtensor<[1,32,1,1],f32> | |
| %389 = torch.aten.quantize_per_tensor %45, %float3.906250e-03, %int0, %int12 : !torch.vtensor<[1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1],!torch.qint8> | |
| %390 = torch.aten.int_repr %389 : !torch.vtensor<[1],!torch.qint8> -> !torch.vtensor<[1],si8> | |
| %391 = torch.aten._make_per_tensor_quantized_tensor %390, %float3.906250e-03, %int0 : !torch.vtensor<[1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1],!torch.qint8> | |
| %392 = torch.aten.dequantize.self %391 : !torch.vtensor<[1],!torch.qint8> -> !torch.vtensor<[1],f32> | |
| %393 = torch.aten.convolution %384, %388, %392, %58, %59, %58, %false, %59, %int1 : !torch.vtensor<[32,32,256,256],f32>, !torch.vtensor<[1,32,1,1],f32>, !torch.vtensor<[1],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,1,256,256],f32> | |
| %394 = torch.aten.quantize_per_tensor %393, %float1.250000e-01, %int0, %int12 : !torch.vtensor<[32,1,256,256],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,1,256,256],!torch.qint8> | |
| %395 = torch.aten.int_repr %394 : !torch.vtensor<[32,1,256,256],!torch.qint8> -> !torch.vtensor<[32,1,256,256],si8> | |
| %396 = torch.aten._make_per_tensor_quantized_tensor %395, %float1.250000e-01, %int0 : !torch.vtensor<[32,1,256,256],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,1,256,256],!torch.qint8> | |
| %397 = torch.aten.dequantize.self %396 : !torch.vtensor<[32,1,256,256],!torch.qint8> -> !torch.vtensor<[32,1,256,256],f32> | |
| %398 = torch.aten.sigmoid %397 : !torch.vtensor<[32,1,256,256],f32> -> !torch.vtensor<[32,1,256,256],f32> | |
| return %398 : !torch.vtensor<[32,1,256,256],f32> | |
| } | |
| } | |
Author
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
%221 = torch.aten.convolution %212, %216, %220, %80, %59, %58, %true, %59, %int1 : !torch.vtensor<[32,512,16,16],f32>, !torch.vtensor<[512,256,2,2],f32>, !torch.vtensor<[256],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[32,256,32,32],f32>