Created
April 22, 2024 18:26
-
-
Save AmosLewis/eb8849f517b442b92d241ca3bca7baf2 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| module { | |
| func.func @torch_jit(%arg0: !torch.vtensor<[1,3,224,224],f32>) -> !torch.vtensor<[1,21,224,224],f32> attributes {torch.onnx_meta.ir_version = 8 : si64, torch.onnx_meta.opset_version = 17 : si64, torch.onnx_meta.producer_name = "pytorch", torch.onnx_meta.producer_version = "1.13.1"} { | |
| %0 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x3x3x3xf32>) : !torch.vtensor<[16,3,3,3],f32> | |
| %1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
| %2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x1x3x3xf32>) : !torch.vtensor<[16,1,3,3],f32> | |
| %3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
| %4 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16x16x1x1xf32>) : !torch.vtensor<[16,16,1,1],f32> | |
| %5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<16xf32>) : !torch.vtensor<[16],f32> | |
| %6 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x16x1x1xf32>) : !torch.vtensor<[64,16,1,1],f32> | |
| %7 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %8 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64x1x3x3xf32>) : !torch.vtensor<[64,1,3,3],f32> | |
| %9 = torch.vtensor.literal(dense_resource<__elided__> : tensor<64xf32>) : !torch.vtensor<[64],f32> | |
| %10 = torch.vtensor.literal(dense_resource<__elided__> : tensor<24x64x1x1xf32>) : !torch.vtensor<[24,64,1,1],f32> | |
| %11 = torch.vtensor.literal(dense_resource<__elided__> : tensor<24xf32>) : !torch.vtensor<[24],f32> | |
| %12 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72x24x1x1xf32>) : !torch.vtensor<[72,24,1,1],f32> | |
| %13 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72xf32>) : !torch.vtensor<[72],f32> | |
| %14 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72x1x3x3xf32>) : !torch.vtensor<[72,1,3,3],f32> | |
| %15 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72xf32>) : !torch.vtensor<[72],f32> | |
| %16 = torch.vtensor.literal(dense_resource<__elided__> : tensor<24x72x1x1xf32>) : !torch.vtensor<[24,72,1,1],f32> | |
| %17 = torch.vtensor.literal(dense_resource<__elided__> : tensor<24xf32>) : !torch.vtensor<[24],f32> | |
| %18 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72x24x1x1xf32>) : !torch.vtensor<[72,24,1,1],f32> | |
| %19 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72xf32>) : !torch.vtensor<[72],f32> | |
| %20 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72x1x5x5xf32>) : !torch.vtensor<[72,1,5,5],f32> | |
| %21 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72xf32>) : !torch.vtensor<[72],f32> | |
| %22 = torch.vtensor.literal(dense_resource<__elided__> : tensor<24x72x1x1xf32>) : !torch.vtensor<[24,72,1,1],f32> | |
| %23 = torch.vtensor.literal(dense_resource<__elided__> : tensor<24xf32>) : !torch.vtensor<[24],f32> | |
| %24 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72x24x1x1xf32>) : !torch.vtensor<[72,24,1,1],f32> | |
| %25 = torch.vtensor.literal(dense_resource<__elided__> : tensor<72xf32>) : !torch.vtensor<[72],f32> | |
| %26 = torch.vtensor.literal(dense_resource<__elided__> : tensor<40x72x1x1xf32>) : !torch.vtensor<[40,72,1,1],f32> | |
| %27 = torch.vtensor.literal(dense_resource<__elided__> : tensor<40xf32>) : !torch.vtensor<[40],f32> | |
| %28 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x40x1x1xf32>) : !torch.vtensor<[120,40,1,1],f32> | |
| %29 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %30 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x1x5x5xf32>) : !torch.vtensor<[120,1,5,5],f32> | |
| %31 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %32 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x120x1x1xf32>) : !torch.vtensor<[32,120,1,1],f32> | |
| %33 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %34 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x32x1x1xf32>) : !torch.vtensor<[120,32,1,1],f32> | |
| %35 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %36 = torch.vtensor.literal(dense_resource<__elided__> : tensor<40x120x1x1xf32>) : !torch.vtensor<[40,120,1,1],f32> | |
| %37 = torch.vtensor.literal(dense_resource<__elided__> : tensor<40xf32>) : !torch.vtensor<[40],f32> | |
| %38 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x40x1x1xf32>) : !torch.vtensor<[120,40,1,1],f32> | |
| %39 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %40 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x1x5x5xf32>) : !torch.vtensor<[120,1,5,5],f32> | |
| %41 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %42 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x120x1x1xf32>) : !torch.vtensor<[32,120,1,1],f32> | |
| %43 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32xf32>) : !torch.vtensor<[32],f32> | |
| %44 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x32x1x1xf32>) : !torch.vtensor<[120,32,1,1],f32> | |
| %45 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %46 = torch.vtensor.literal(dense_resource<__elided__> : tensor<40x120x1x1xf32>) : !torch.vtensor<[40,120,1,1],f32> | |
| %47 = torch.vtensor.literal(dense_resource<__elided__> : tensor<40xf32>) : !torch.vtensor<[40],f32> | |
| %48 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240x40x1x1xf32>) : !torch.vtensor<[240,40,1,1],f32> | |
| %49 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240xf32>) : !torch.vtensor<[240],f32> | |
| %50 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240x1x3x3xf32>) : !torch.vtensor<[240,1,3,3],f32> | |
| %51 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240xf32>) : !torch.vtensor<[240],f32> | |
| %52 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x240x1x1xf32>) : !torch.vtensor<[80,240,1,1],f32> | |
| %53 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %54 = torch.vtensor.literal(dense_resource<__elided__> : tensor<200x80x1x1xf32>) : !torch.vtensor<[200,80,1,1],f32> | |
| %55 = torch.vtensor.literal(dense_resource<__elided__> : tensor<200xf32>) : !torch.vtensor<[200],f32> | |
| %56 = torch.vtensor.literal(dense_resource<__elided__> : tensor<200x1x3x3xf32>) : !torch.vtensor<[200,1,3,3],f32> | |
| %57 = torch.vtensor.literal(dense_resource<__elided__> : tensor<200xf32>) : !torch.vtensor<[200],f32> | |
| %58 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x200x1x1xf32>) : !torch.vtensor<[80,200,1,1],f32> | |
| %59 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %60 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184x80x1x1xf32>) : !torch.vtensor<[184,80,1,1],f32> | |
| %61 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184xf32>) : !torch.vtensor<[184],f32> | |
| %62 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184x1x3x3xf32>) : !torch.vtensor<[184,1,3,3],f32> | |
| %63 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184xf32>) : !torch.vtensor<[184],f32> | |
| %64 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x184x1x1xf32>) : !torch.vtensor<[80,184,1,1],f32> | |
| %65 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %66 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184x80x1x1xf32>) : !torch.vtensor<[184,80,1,1],f32> | |
| %67 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184xf32>) : !torch.vtensor<[184],f32> | |
| %68 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184x1x3x3xf32>) : !torch.vtensor<[184,1,3,3],f32> | |
| %69 = torch.vtensor.literal(dense_resource<__elided__> : tensor<184xf32>) : !torch.vtensor<[184],f32> | |
| %70 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80x184x1x1xf32>) : !torch.vtensor<[80,184,1,1],f32> | |
| %71 = torch.vtensor.literal(dense_resource<__elided__> : tensor<80xf32>) : !torch.vtensor<[80],f32> | |
| %72 = torch.vtensor.literal(dense_resource<__elided__> : tensor<480x80x1x1xf32>) : !torch.vtensor<[480,80,1,1],f32> | |
| %73 = torch.vtensor.literal(dense_resource<__elided__> : tensor<480xf32>) : !torch.vtensor<[480],f32> | |
| %74 = torch.vtensor.literal(dense_resource<__elided__> : tensor<480x1x3x3xf32>) : !torch.vtensor<[480,1,3,3],f32> | |
| %75 = torch.vtensor.literal(dense_resource<__elided__> : tensor<480xf32>) : !torch.vtensor<[480],f32> | |
| %76 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120x480x1x1xf32>) : !torch.vtensor<[120,480,1,1],f32> | |
| %77 = torch.vtensor.literal(dense_resource<__elided__> : tensor<120xf32>) : !torch.vtensor<[120],f32> | |
| %78 = torch.vtensor.literal(dense_resource<__elided__> : tensor<480x120x1x1xf32>) : !torch.vtensor<[480,120,1,1],f32> | |
| %79 = torch.vtensor.literal(dense_resource<__elided__> : tensor<480xf32>) : !torch.vtensor<[480],f32> | |
| %80 = torch.vtensor.literal(dense_resource<__elided__> : tensor<112x480x1x1xf32>) : !torch.vtensor<[112,480,1,1],f32> | |
| %81 = torch.vtensor.literal(dense_resource<__elided__> : tensor<112xf32>) : !torch.vtensor<[112],f32> | |
| %82 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672x112x1x1xf32>) : !torch.vtensor<[672,112,1,1],f32> | |
| %83 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672xf32>) : !torch.vtensor<[672],f32> | |
| %84 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672x1x3x3xf32>) : !torch.vtensor<[672,1,3,3],f32> | |
| %85 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672xf32>) : !torch.vtensor<[672],f32> | |
| %86 = torch.vtensor.literal(dense_resource<__elided__> : tensor<168x672x1x1xf32>) : !torch.vtensor<[168,672,1,1],f32> | |
| %87 = torch.vtensor.literal(dense_resource<__elided__> : tensor<168xf32>) : !torch.vtensor<[168],f32> | |
| %88 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672x168x1x1xf32>) : !torch.vtensor<[672,168,1,1],f32> | |
| %89 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672xf32>) : !torch.vtensor<[672],f32> | |
| %90 = torch.vtensor.literal(dense_resource<__elided__> : tensor<112x672x1x1xf32>) : !torch.vtensor<[112,672,1,1],f32> | |
| %91 = torch.vtensor.literal(dense_resource<__elided__> : tensor<112xf32>) : !torch.vtensor<[112],f32> | |
| %92 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672x112x1x1xf32>) : !torch.vtensor<[672,112,1,1],f32> | |
| %93 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672xf32>) : !torch.vtensor<[672],f32> | |
| %94 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672x1x5x5xf32>) : !torch.vtensor<[672,1,5,5],f32> | |
| %95 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672xf32>) : !torch.vtensor<[672],f32> | |
| %96 = torch.vtensor.literal(dense_resource<__elided__> : tensor<168x672x1x1xf32>) : !torch.vtensor<[168,672,1,1],f32> | |
| %97 = torch.vtensor.literal(dense_resource<__elided__> : tensor<168xf32>) : !torch.vtensor<[168],f32> | |
| %98 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672x168x1x1xf32>) : !torch.vtensor<[672,168,1,1],f32> | |
| %99 = torch.vtensor.literal(dense_resource<__elided__> : tensor<672xf32>) : !torch.vtensor<[672],f32> | |
| %100 = torch.vtensor.literal(dense_resource<__elided__> : tensor<160x672x1x1xf32>) : !torch.vtensor<[160,672,1,1],f32> | |
| %101 = torch.vtensor.literal(dense_resource<__elided__> : tensor<160xf32>) : !torch.vtensor<[160],f32> | |
| %102 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x160x1x1xf32>) : !torch.vtensor<[960,160,1,1],f32> | |
| %103 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %104 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x1x5x5xf32>) : !torch.vtensor<[960,1,5,5],f32> | |
| %105 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %106 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240x960x1x1xf32>) : !torch.vtensor<[240,960,1,1],f32> | |
| %107 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240xf32>) : !torch.vtensor<[240],f32> | |
| %108 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x240x1x1xf32>) : !torch.vtensor<[960,240,1,1],f32> | |
| %109 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %110 = torch.vtensor.literal(dense_resource<__elided__> : tensor<160x960x1x1xf32>) : !torch.vtensor<[160,960,1,1],f32> | |
| %111 = torch.vtensor.literal(dense_resource<__elided__> : tensor<160xf32>) : !torch.vtensor<[160],f32> | |
| %112 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x160x1x1xf32>) : !torch.vtensor<[960,160,1,1],f32> | |
| %113 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %114 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x1x5x5xf32>) : !torch.vtensor<[960,1,5,5],f32> | |
| %115 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %116 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240x960x1x1xf32>) : !torch.vtensor<[240,960,1,1],f32> | |
| %117 = torch.vtensor.literal(dense_resource<__elided__> : tensor<240xf32>) : !torch.vtensor<[240],f32> | |
| %118 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x240x1x1xf32>) : !torch.vtensor<[960,240,1,1],f32> | |
| %119 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %120 = torch.vtensor.literal(dense_resource<__elided__> : tensor<160x960x1x1xf32>) : !torch.vtensor<[160,960,1,1],f32> | |
| %121 = torch.vtensor.literal(dense_resource<__elided__> : tensor<160xf32>) : !torch.vtensor<[160],f32> | |
| %122 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960x160x1x1xf32>) : !torch.vtensor<[960,160,1,1],f32> | |
| %123 = torch.vtensor.literal(dense_resource<__elided__> : tensor<960xf32>) : !torch.vtensor<[960],f32> | |
| %124 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x960x1x1xf32>) : !torch.vtensor<[128,960,1,1],f32> | |
| %125 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128xf32>) : !torch.vtensor<[128],f32> | |
| %126 = torch.vtensor.literal(dense_resource<__elided__> : tensor<128x960x1x1xf32>) : !torch.vtensor<[128,960,1,1],f32> | |
| %127 = torch.vtensor.literal(dense_resource<__elided__> : tensor<21x40x1x1xf32>) : !torch.vtensor<[21,40,1,1],f32> | |
| %128 = torch.vtensor.literal(dense_resource<__elided__> : tensor<21xf32>) : !torch.vtensor<[21],f32> | |
| %129 = torch.vtensor.literal(dense_resource<__elided__> : tensor<21x128x1x1xf32>) : !torch.vtensor<[21,128,1,1],f32> | |
| %130 = torch.vtensor.literal(dense_resource<__elided__> : tensor<21xf32>) : !torch.vtensor<[21],f32> | |
| %none = torch.constant.none | |
| %131 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %132 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12 = torch.constant.int 12 | |
| %133 = torch.aten.item %131 : !torch.vtensor<[],f32> -> !torch.float | |
| %134 = torch.aten.item %132 : !torch.vtensor<[],si8> -> !torch.int | |
| %135 = torch.aten.quantize_per_tensor %arg0, %133, %134, %int12 : !torch.vtensor<[1,3,224,224],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,3,224,224],!torch.qint8> | |
| %136 = torch.aten.int_repr %135 : !torch.vtensor<[1,3,224,224],!torch.qint8> -> !torch.vtensor<[1,3,224,224],si8> | |
| %137 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %138 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %139 = torch.aten.item %137 : !torch.vtensor<[],f32> -> !torch.float | |
| %140 = torch.aten.item %138 : !torch.vtensor<[],si8> -> !torch.int | |
| %141 = torch.aten._make_per_tensor_quantized_tensor %136, %139, %140 : !torch.vtensor<[1,3,224,224],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,3,224,224],!torch.qint8> | |
| %142 = torch.aten.dequantize.self %141 : !torch.vtensor<[1,3,224,224],!torch.qint8> -> !torch.vtensor<[1,3,224,224],f32> | |
| %143 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %144 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_0 = torch.constant.int 12 | |
| %145 = torch.aten.item %143 : !torch.vtensor<[],f32> -> !torch.float | |
| %146 = torch.aten.item %144 : !torch.vtensor<[],si8> -> !torch.int | |
| %147 = torch.aten.quantize_per_tensor %0, %145, %146, %int12_0 : !torch.vtensor<[16,3,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,3,3,3],!torch.qint8> | |
| %148 = torch.aten.int_repr %147 : !torch.vtensor<[16,3,3,3],!torch.qint8> -> !torch.vtensor<[16,3,3,3],si8> | |
| %149 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %150 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %151 = torch.aten.item %149 : !torch.vtensor<[],f32> -> !torch.float | |
| %152 = torch.aten.item %150 : !torch.vtensor<[],si8> -> !torch.int | |
| %153 = torch.aten._make_per_tensor_quantized_tensor %148, %151, %152 : !torch.vtensor<[16,3,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,3,3,3],!torch.qint8> | |
| %154 = torch.aten.dequantize.self %153 : !torch.vtensor<[16,3,3,3],!torch.qint8> -> !torch.vtensor<[16,3,3,3],f32> | |
| %155 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %156 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1 = torch.constant.int 12 | |
| %157 = torch.aten.item %155 : !torch.vtensor<[],f32> -> !torch.float | |
| %158 = torch.aten.item %156 : !torch.vtensor<[],si8> -> !torch.int | |
| %159 = torch.aten.quantize_per_tensor %1, %157, %158, %int12_1 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %160 = torch.aten.int_repr %159 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
| %161 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %162 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %163 = torch.aten.item %161 : !torch.vtensor<[],f32> -> !torch.float | |
| %164 = torch.aten.item %162 : !torch.vtensor<[],si8> -> !torch.int | |
| %165 = torch.aten._make_per_tensor_quantized_tensor %160, %163, %164 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %166 = torch.aten.dequantize.self %165 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
| %int1 = torch.constant.int 1 | |
| %int1_2 = torch.constant.int 1 | |
| %int1_3 = torch.constant.int 1 | |
| %int1_4 = torch.constant.int 1 | |
| %int2 = torch.constant.int 2 | |
| %int2_5 = torch.constant.int 2 | |
| %int0 = torch.constant.int 0 | |
| %167 = torch.prim.ListConstruct %int1, %int1_2 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %168 = torch.prim.ListConstruct %int1_3, %int1_4 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %169 = torch.prim.ListConstruct %int2, %int2_5 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %170 = torch.prim.ListConstruct %int0, %int0 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false = torch.constant.bool false | |
| %int1_6 = torch.constant.int 1 | |
| %171 = torch.aten.convolution %142, %154, %166, %169, %167, %168, %false, %170, %int1_6 : !torch.vtensor<[1,3,224,224],f32>, !torch.vtensor<[16,3,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,112,112],f32> | |
| %172 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %173 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_7 = torch.constant.int 12 | |
| %174 = torch.aten.item %172 : !torch.vtensor<[],f32> -> !torch.float | |
| %175 = torch.aten.item %173 : !torch.vtensor<[],si8> -> !torch.int | |
| %176 = torch.aten.quantize_per_tensor %171, %174, %175, %int12_7 : !torch.vtensor<[1,16,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %177 = torch.aten.int_repr %176 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],si8> | |
| %178 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %179 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %180 = torch.aten.item %178 : !torch.vtensor<[],f32> -> !torch.float | |
| %181 = torch.aten.item %179 : !torch.vtensor<[],si8> -> !torch.int | |
| %182 = torch.aten._make_per_tensor_quantized_tensor %177, %180, %181 : !torch.vtensor<[1,16,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %183 = torch.aten.dequantize.self %182 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],f32> | |
| %184 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_8 = torch.constant.int 1 | |
| %185 = torch.aten.add.Tensor %183, %184, %int1_8 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,16,112,112],f32> | |
| %186 = torch.aten.relu %185 : !torch.vtensor<[1,16,112,112],f32> -> !torch.vtensor<[1,16,112,112],f32> | |
| %187 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %188 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6 = torch.constant.int 6 | |
| %none_9 = torch.constant.none | |
| %false_10 = torch.constant.bool false | |
| %189 = torch.aten.to.dtype %187, %int6, %false_10, %false_10, %none_9 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_11 = torch.constant.int 6 | |
| %none_12 = torch.constant.none | |
| %false_13 = torch.constant.bool false | |
| %190 = torch.aten.to.dtype %188, %int6_11, %false_13, %false_13, %none_12 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %191 = torch.aten.clamp.Tensor %186, %189, %190 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,16,112,112],f32> | |
| %192 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %193 = torch.aten.mul.Tensor %191, %192 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,16,112,112],f32> | |
| %194 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %195 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_14 = torch.constant.int 12 | |
| %196 = torch.aten.item %194 : !torch.vtensor<[],f32> -> !torch.float | |
| %197 = torch.aten.item %195 : !torch.vtensor<[],si8> -> !torch.int | |
| %198 = torch.aten.quantize_per_tensor %193, %196, %197, %int12_14 : !torch.vtensor<[1,16,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %199 = torch.aten.int_repr %198 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],si8> | |
| %200 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %201 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %202 = torch.aten.item %200 : !torch.vtensor<[],f32> -> !torch.float | |
| %203 = torch.aten.item %201 : !torch.vtensor<[],si8> -> !torch.int | |
| %204 = torch.aten._make_per_tensor_quantized_tensor %199, %202, %203 : !torch.vtensor<[1,16,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %205 = torch.aten.dequantize.self %204 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],f32> | |
| %206 = torch.aten.mul.Tensor %183, %205 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[1,16,112,112],f32> -> !torch.vtensor<[1,16,112,112],f32> | |
| %207 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %208 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_15 = torch.constant.int 12 | |
| %209 = torch.aten.item %207 : !torch.vtensor<[],f32> -> !torch.float | |
| %210 = torch.aten.item %208 : !torch.vtensor<[],si8> -> !torch.int | |
| %211 = torch.aten.quantize_per_tensor %206, %209, %210, %int12_15 : !torch.vtensor<[1,16,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %212 = torch.aten.int_repr %211 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],si8> | |
| %213 = torch.vtensor.literal(dense<5.000000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %214 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %215 = torch.aten.item %213 : !torch.vtensor<[],f32> -> !torch.float | |
| %216 = torch.aten.item %214 : !torch.vtensor<[],si8> -> !torch.int | |
| %217 = torch.aten._make_per_tensor_quantized_tensor %212, %215, %216 : !torch.vtensor<[1,16,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %218 = torch.aten.dequantize.self %217 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],f32> | |
| %219 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %220 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_16 = torch.constant.int 12 | |
| %221 = torch.aten.item %219 : !torch.vtensor<[],f32> -> !torch.float | |
| %222 = torch.aten.item %220 : !torch.vtensor<[],si8> -> !torch.int | |
| %223 = torch.aten.quantize_per_tensor %2, %221, %222, %int12_16 : !torch.vtensor<[16,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,1,3,3],!torch.qint8> | |
| %224 = torch.aten.int_repr %223 : !torch.vtensor<[16,1,3,3],!torch.qint8> -> !torch.vtensor<[16,1,3,3],si8> | |
| %225 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %226 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %227 = torch.aten.item %225 : !torch.vtensor<[],f32> -> !torch.float | |
| %228 = torch.aten.item %226 : !torch.vtensor<[],si8> -> !torch.int | |
| %229 = torch.aten._make_per_tensor_quantized_tensor %224, %227, %228 : !torch.vtensor<[16,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,1,3,3],!torch.qint8> | |
| %230 = torch.aten.dequantize.self %229 : !torch.vtensor<[16,1,3,3],!torch.qint8> -> !torch.vtensor<[16,1,3,3],f32> | |
| %231 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %232 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_17 = torch.constant.int 12 | |
| %233 = torch.aten.item %231 : !torch.vtensor<[],f32> -> !torch.float | |
| %234 = torch.aten.item %232 : !torch.vtensor<[],si8> -> !torch.int | |
| %235 = torch.aten.quantize_per_tensor %3, %233, %234, %int12_17 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %236 = torch.aten.int_repr %235 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
| %237 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %238 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %239 = torch.aten.item %237 : !torch.vtensor<[],f32> -> !torch.float | |
| %240 = torch.aten.item %238 : !torch.vtensor<[],si8> -> !torch.int | |
| %241 = torch.aten._make_per_tensor_quantized_tensor %236, %239, %240 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %242 = torch.aten.dequantize.self %241 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
| %int1_18 = torch.constant.int 1 | |
| %int1_19 = torch.constant.int 1 | |
| %int1_20 = torch.constant.int 1 | |
| %int1_21 = torch.constant.int 1 | |
| %int1_22 = torch.constant.int 1 | |
| %int1_23 = torch.constant.int 1 | |
| %int0_24 = torch.constant.int 0 | |
| %243 = torch.prim.ListConstruct %int1_18, %int1_19 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %244 = torch.prim.ListConstruct %int1_20, %int1_21 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %245 = torch.prim.ListConstruct %int1_22, %int1_23 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %246 = torch.prim.ListConstruct %int0_24, %int0_24 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_25 = torch.constant.bool false | |
| %int16 = torch.constant.int 16 | |
| %247 = torch.aten.convolution %218, %230, %242, %245, %243, %244, %false_25, %246, %int16 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[16,1,3,3],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,112,112],f32> | |
| %248 = torch.aten.relu %247 : !torch.vtensor<[1,16,112,112],f32> -> !torch.vtensor<[1,16,112,112],f32> | |
| %249 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %250 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_26 = torch.constant.int 12 | |
| %251 = torch.aten.item %249 : !torch.vtensor<[],f32> -> !torch.float | |
| %252 = torch.aten.item %250 : !torch.vtensor<[],si8> -> !torch.int | |
| %253 = torch.aten.quantize_per_tensor %248, %251, %252, %int12_26 : !torch.vtensor<[1,16,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %254 = torch.aten.int_repr %253 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],si8> | |
| %255 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %256 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %257 = torch.aten.item %255 : !torch.vtensor<[],f32> -> !torch.float | |
| %258 = torch.aten.item %256 : !torch.vtensor<[],si8> -> !torch.int | |
| %259 = torch.aten._make_per_tensor_quantized_tensor %254, %257, %258 : !torch.vtensor<[1,16,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %260 = torch.aten.dequantize.self %259 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],f32> | |
| %261 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %262 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_27 = torch.constant.int 12 | |
| %263 = torch.aten.item %261 : !torch.vtensor<[],f32> -> !torch.float | |
| %264 = torch.aten.item %262 : !torch.vtensor<[],si8> -> !torch.int | |
| %265 = torch.aten.quantize_per_tensor %4, %263, %264, %int12_27 : !torch.vtensor<[16,16,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16,16,1,1],!torch.qint8> | |
| %266 = torch.aten.int_repr %265 : !torch.vtensor<[16,16,1,1],!torch.qint8> -> !torch.vtensor<[16,16,1,1],si8> | |
| %267 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %268 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %269 = torch.aten.item %267 : !torch.vtensor<[],f32> -> !torch.float | |
| %270 = torch.aten.item %268 : !torch.vtensor<[],si8> -> !torch.int | |
| %271 = torch.aten._make_per_tensor_quantized_tensor %266, %269, %270 : !torch.vtensor<[16,16,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[16,16,1,1],!torch.qint8> | |
| %272 = torch.aten.dequantize.self %271 : !torch.vtensor<[16,16,1,1],!torch.qint8> -> !torch.vtensor<[16,16,1,1],f32> | |
| %273 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %274 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_28 = torch.constant.int 12 | |
| %275 = torch.aten.item %273 : !torch.vtensor<[],f32> -> !torch.float | |
| %276 = torch.aten.item %274 : !torch.vtensor<[],si8> -> !torch.int | |
| %277 = torch.aten.quantize_per_tensor %5, %275, %276, %int12_28 : !torch.vtensor<[16],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %278 = torch.aten.int_repr %277 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],si8> | |
| %279 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %280 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %281 = torch.aten.item %279 : !torch.vtensor<[],f32> -> !torch.float | |
| %282 = torch.aten.item %280 : !torch.vtensor<[],si8> -> !torch.int | |
| %283 = torch.aten._make_per_tensor_quantized_tensor %278, %281, %282 : !torch.vtensor<[16],si8>, !torch.float, !torch.int -> !torch.vtensor<[16],!torch.qint8> | |
| %284 = torch.aten.dequantize.self %283 : !torch.vtensor<[16],!torch.qint8> -> !torch.vtensor<[16],f32> | |
| %int0_29 = torch.constant.int 0 | |
| %int0_30 = torch.constant.int 0 | |
| %int1_31 = torch.constant.int 1 | |
| %int1_32 = torch.constant.int 1 | |
| %int1_33 = torch.constant.int 1 | |
| %int1_34 = torch.constant.int 1 | |
| %int0_35 = torch.constant.int 0 | |
| %285 = torch.prim.ListConstruct %int0_29, %int0_30 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %286 = torch.prim.ListConstruct %int1_31, %int1_32 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %287 = torch.prim.ListConstruct %int1_33, %int1_34 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %288 = torch.prim.ListConstruct %int0_35, %int0_35 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_36 = torch.constant.bool false | |
| %int1_37 = torch.constant.int 1 | |
| %289 = torch.aten.convolution %260, %272, %284, %287, %285, %286, %false_36, %288, %int1_37 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[16,16,1,1],f32>, !torch.vtensor<[16],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,16,112,112],f32> | |
| %290 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %291 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_38 = torch.constant.int 12 | |
| %292 = torch.aten.item %290 : !torch.vtensor<[],f32> -> !torch.float | |
| %293 = torch.aten.item %291 : !torch.vtensor<[],si8> -> !torch.int | |
| %294 = torch.aten.quantize_per_tensor %289, %292, %293, %int12_38 : !torch.vtensor<[1,16,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %295 = torch.aten.int_repr %294 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],si8> | |
| %296 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %297 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %298 = torch.aten.item %296 : !torch.vtensor<[],f32> -> !torch.float | |
| %299 = torch.aten.item %297 : !torch.vtensor<[],si8> -> !torch.int | |
| %300 = torch.aten._make_per_tensor_quantized_tensor %295, %298, %299 : !torch.vtensor<[1,16,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %301 = torch.aten.dequantize.self %300 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],f32> | |
| %int1_39 = torch.constant.int 1 | |
| %302 = torch.aten.add.Tensor %301, %218, %int1_39 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[1,16,112,112],f32>, !torch.int -> !torch.vtensor<[1,16,112,112],f32> | |
| %303 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %304 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_40 = torch.constant.int 12 | |
| %305 = torch.aten.item %303 : !torch.vtensor<[],f32> -> !torch.float | |
| %306 = torch.aten.item %304 : !torch.vtensor<[],si8> -> !torch.int | |
| %307 = torch.aten.quantize_per_tensor %302, %305, %306, %int12_40 : !torch.vtensor<[1,16,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %308 = torch.aten.int_repr %307 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],si8> | |
| %309 = torch.vtensor.literal(dense<2.500000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %310 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %311 = torch.aten.item %309 : !torch.vtensor<[],f32> -> !torch.float | |
| %312 = torch.aten.item %310 : !torch.vtensor<[],si8> -> !torch.int | |
| %313 = torch.aten._make_per_tensor_quantized_tensor %308, %311, %312 : !torch.vtensor<[1,16,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,16,112,112],!torch.qint8> | |
| %314 = torch.aten.dequantize.self %313 : !torch.vtensor<[1,16,112,112],!torch.qint8> -> !torch.vtensor<[1,16,112,112],f32> | |
| %315 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %316 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_41 = torch.constant.int 12 | |
| %317 = torch.aten.item %315 : !torch.vtensor<[],f32> -> !torch.float | |
| %318 = torch.aten.item %316 : !torch.vtensor<[],si8> -> !torch.int | |
| %319 = torch.aten.quantize_per_tensor %6, %317, %318, %int12_41 : !torch.vtensor<[64,16,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,16,1,1],!torch.qint8> | |
| %320 = torch.aten.int_repr %319 : !torch.vtensor<[64,16,1,1],!torch.qint8> -> !torch.vtensor<[64,16,1,1],si8> | |
| %321 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %322 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %323 = torch.aten.item %321 : !torch.vtensor<[],f32> -> !torch.float | |
| %324 = torch.aten.item %322 : !torch.vtensor<[],si8> -> !torch.int | |
| %325 = torch.aten._make_per_tensor_quantized_tensor %320, %323, %324 : !torch.vtensor<[64,16,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,16,1,1],!torch.qint8> | |
| %326 = torch.aten.dequantize.self %325 : !torch.vtensor<[64,16,1,1],!torch.qint8> -> !torch.vtensor<[64,16,1,1],f32> | |
| %327 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %328 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_42 = torch.constant.int 12 | |
| %329 = torch.aten.item %327 : !torch.vtensor<[],f32> -> !torch.float | |
| %330 = torch.aten.item %328 : !torch.vtensor<[],si8> -> !torch.int | |
| %331 = torch.aten.quantize_per_tensor %7, %329, %330, %int12_42 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %332 = torch.aten.int_repr %331 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %333 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %334 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %335 = torch.aten.item %333 : !torch.vtensor<[],f32> -> !torch.float | |
| %336 = torch.aten.item %334 : !torch.vtensor<[],si8> -> !torch.int | |
| %337 = torch.aten._make_per_tensor_quantized_tensor %332, %335, %336 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %338 = torch.aten.dequantize.self %337 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int0_43 = torch.constant.int 0 | |
| %int0_44 = torch.constant.int 0 | |
| %int1_45 = torch.constant.int 1 | |
| %int1_46 = torch.constant.int 1 | |
| %int1_47 = torch.constant.int 1 | |
| %int1_48 = torch.constant.int 1 | |
| %int0_49 = torch.constant.int 0 | |
| %339 = torch.prim.ListConstruct %int0_43, %int0_44 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %340 = torch.prim.ListConstruct %int1_45, %int1_46 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %341 = torch.prim.ListConstruct %int1_47, %int1_48 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %342 = torch.prim.ListConstruct %int0_49, %int0_49 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_50 = torch.constant.bool false | |
| %int1_51 = torch.constant.int 1 | |
| %343 = torch.aten.convolution %314, %326, %338, %341, %339, %340, %false_50, %342, %int1_51 : !torch.vtensor<[1,16,112,112],f32>, !torch.vtensor<[64,16,1,1],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,112,112],f32> | |
| %344 = torch.aten.relu %343 : !torch.vtensor<[1,64,112,112],f32> -> !torch.vtensor<[1,64,112,112],f32> | |
| %345 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %346 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_52 = torch.constant.int 12 | |
| %347 = torch.aten.item %345 : !torch.vtensor<[],f32> -> !torch.float | |
| %348 = torch.aten.item %346 : !torch.vtensor<[],si8> -> !torch.int | |
| %349 = torch.aten.quantize_per_tensor %344, %347, %348, %int12_52 : !torch.vtensor<[1,64,112,112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,112,112],!torch.qint8> | |
| %350 = torch.aten.int_repr %349 : !torch.vtensor<[1,64,112,112],!torch.qint8> -> !torch.vtensor<[1,64,112,112],si8> | |
| %351 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %352 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %353 = torch.aten.item %351 : !torch.vtensor<[],f32> -> !torch.float | |
| %354 = torch.aten.item %352 : !torch.vtensor<[],si8> -> !torch.int | |
| %355 = torch.aten._make_per_tensor_quantized_tensor %350, %353, %354 : !torch.vtensor<[1,64,112,112],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,112,112],!torch.qint8> | |
| %356 = torch.aten.dequantize.self %355 : !torch.vtensor<[1,64,112,112],!torch.qint8> -> !torch.vtensor<[1,64,112,112],f32> | |
| %357 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %358 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_53 = torch.constant.int 12 | |
| %359 = torch.aten.item %357 : !torch.vtensor<[],f32> -> !torch.float | |
| %360 = torch.aten.item %358 : !torch.vtensor<[],si8> -> !torch.int | |
| %361 = torch.aten.quantize_per_tensor %8, %359, %360, %int12_53 : !torch.vtensor<[64,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64,1,3,3],!torch.qint8> | |
| %362 = torch.aten.int_repr %361 : !torch.vtensor<[64,1,3,3],!torch.qint8> -> !torch.vtensor<[64,1,3,3],si8> | |
| %363 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %364 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %365 = torch.aten.item %363 : !torch.vtensor<[],f32> -> !torch.float | |
| %366 = torch.aten.item %364 : !torch.vtensor<[],si8> -> !torch.int | |
| %367 = torch.aten._make_per_tensor_quantized_tensor %362, %365, %366 : !torch.vtensor<[64,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[64,1,3,3],!torch.qint8> | |
| %368 = torch.aten.dequantize.self %367 : !torch.vtensor<[64,1,3,3],!torch.qint8> -> !torch.vtensor<[64,1,3,3],f32> | |
| %369 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %370 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_54 = torch.constant.int 12 | |
| %371 = torch.aten.item %369 : !torch.vtensor<[],f32> -> !torch.float | |
| %372 = torch.aten.item %370 : !torch.vtensor<[],si8> -> !torch.int | |
| %373 = torch.aten.quantize_per_tensor %9, %371, %372, %int12_54 : !torch.vtensor<[64],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %374 = torch.aten.int_repr %373 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],si8> | |
| %375 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %376 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %377 = torch.aten.item %375 : !torch.vtensor<[],f32> -> !torch.float | |
| %378 = torch.aten.item %376 : !torch.vtensor<[],si8> -> !torch.int | |
| %379 = torch.aten._make_per_tensor_quantized_tensor %374, %377, %378 : !torch.vtensor<[64],si8>, !torch.float, !torch.int -> !torch.vtensor<[64],!torch.qint8> | |
| %380 = torch.aten.dequantize.self %379 : !torch.vtensor<[64],!torch.qint8> -> !torch.vtensor<[64],f32> | |
| %int1_55 = torch.constant.int 1 | |
| %int1_56 = torch.constant.int 1 | |
| %int1_57 = torch.constant.int 1 | |
| %int1_58 = torch.constant.int 1 | |
| %int2_59 = torch.constant.int 2 | |
| %int2_60 = torch.constant.int 2 | |
| %int0_61 = torch.constant.int 0 | |
| %381 = torch.prim.ListConstruct %int1_55, %int1_56 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %382 = torch.prim.ListConstruct %int1_57, %int1_58 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %383 = torch.prim.ListConstruct %int2_59, %int2_60 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %384 = torch.prim.ListConstruct %int0_61, %int0_61 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_62 = torch.constant.bool false | |
| %int64 = torch.constant.int 64 | |
| %385 = torch.aten.convolution %356, %368, %380, %383, %381, %382, %false_62, %384, %int64 : !torch.vtensor<[1,64,112,112],f32>, !torch.vtensor<[64,1,3,3],f32>, !torch.vtensor<[64],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,64,56,56],f32> | |
| %386 = torch.aten.relu %385 : !torch.vtensor<[1,64,56,56],f32> -> !torch.vtensor<[1,64,56,56],f32> | |
| %387 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %388 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_63 = torch.constant.int 12 | |
| %389 = torch.aten.item %387 : !torch.vtensor<[],f32> -> !torch.float | |
| %390 = torch.aten.item %388 : !torch.vtensor<[],si8> -> !torch.int | |
| %391 = torch.aten.quantize_per_tensor %386, %389, %390, %int12_63 : !torch.vtensor<[1,64,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,64,56,56],!torch.qint8> | |
| %392 = torch.aten.int_repr %391 : !torch.vtensor<[1,64,56,56],!torch.qint8> -> !torch.vtensor<[1,64,56,56],si8> | |
| %393 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %394 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %395 = torch.aten.item %393 : !torch.vtensor<[],f32> -> !torch.float | |
| %396 = torch.aten.item %394 : !torch.vtensor<[],si8> -> !torch.int | |
| %397 = torch.aten._make_per_tensor_quantized_tensor %392, %395, %396 : !torch.vtensor<[1,64,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,64,56,56],!torch.qint8> | |
| %398 = torch.aten.dequantize.self %397 : !torch.vtensor<[1,64,56,56],!torch.qint8> -> !torch.vtensor<[1,64,56,56],f32> | |
| %399 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %400 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_64 = torch.constant.int 12 | |
| %401 = torch.aten.item %399 : !torch.vtensor<[],f32> -> !torch.float | |
| %402 = torch.aten.item %400 : !torch.vtensor<[],si8> -> !torch.int | |
| %403 = torch.aten.quantize_per_tensor %10, %401, %402, %int12_64 : !torch.vtensor<[24,64,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[24,64,1,1],!torch.qint8> | |
| %404 = torch.aten.int_repr %403 : !torch.vtensor<[24,64,1,1],!torch.qint8> -> !torch.vtensor<[24,64,1,1],si8> | |
| %405 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %406 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %407 = torch.aten.item %405 : !torch.vtensor<[],f32> -> !torch.float | |
| %408 = torch.aten.item %406 : !torch.vtensor<[],si8> -> !torch.int | |
| %409 = torch.aten._make_per_tensor_quantized_tensor %404, %407, %408 : !torch.vtensor<[24,64,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[24,64,1,1],!torch.qint8> | |
| %410 = torch.aten.dequantize.self %409 : !torch.vtensor<[24,64,1,1],!torch.qint8> -> !torch.vtensor<[24,64,1,1],f32> | |
| %411 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %412 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_65 = torch.constant.int 12 | |
| %413 = torch.aten.item %411 : !torch.vtensor<[],f32> -> !torch.float | |
| %414 = torch.aten.item %412 : !torch.vtensor<[],si8> -> !torch.int | |
| %415 = torch.aten.quantize_per_tensor %11, %413, %414, %int12_65 : !torch.vtensor<[24],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[24],!torch.qint8> | |
| %416 = torch.aten.int_repr %415 : !torch.vtensor<[24],!torch.qint8> -> !torch.vtensor<[24],si8> | |
| %417 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %418 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %419 = torch.aten.item %417 : !torch.vtensor<[],f32> -> !torch.float | |
| %420 = torch.aten.item %418 : !torch.vtensor<[],si8> -> !torch.int | |
| %421 = torch.aten._make_per_tensor_quantized_tensor %416, %419, %420 : !torch.vtensor<[24],si8>, !torch.float, !torch.int -> !torch.vtensor<[24],!torch.qint8> | |
| %422 = torch.aten.dequantize.self %421 : !torch.vtensor<[24],!torch.qint8> -> !torch.vtensor<[24],f32> | |
| %int0_66 = torch.constant.int 0 | |
| %int0_67 = torch.constant.int 0 | |
| %int1_68 = torch.constant.int 1 | |
| %int1_69 = torch.constant.int 1 | |
| %int1_70 = torch.constant.int 1 | |
| %int1_71 = torch.constant.int 1 | |
| %int0_72 = torch.constant.int 0 | |
| %423 = torch.prim.ListConstruct %int0_66, %int0_67 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %424 = torch.prim.ListConstruct %int1_68, %int1_69 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %425 = torch.prim.ListConstruct %int1_70, %int1_71 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %426 = torch.prim.ListConstruct %int0_72, %int0_72 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_73 = torch.constant.bool false | |
| %int1_74 = torch.constant.int 1 | |
| %427 = torch.aten.convolution %398, %410, %422, %425, %423, %424, %false_73, %426, %int1_74 : !torch.vtensor<[1,64,56,56],f32>, !torch.vtensor<[24,64,1,1],f32>, !torch.vtensor<[24],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,24,56,56],f32> | |
| %428 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %429 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_75 = torch.constant.int 12 | |
| %430 = torch.aten.item %428 : !torch.vtensor<[],f32> -> !torch.float | |
| %431 = torch.aten.item %429 : !torch.vtensor<[],si8> -> !torch.int | |
| %432 = torch.aten.quantize_per_tensor %427, %430, %431, %int12_75 : !torch.vtensor<[1,24,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,24,56,56],!torch.qint8> | |
| %433 = torch.aten.int_repr %432 : !torch.vtensor<[1,24,56,56],!torch.qint8> -> !torch.vtensor<[1,24,56,56],si8> | |
| %434 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %435 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %436 = torch.aten.item %434 : !torch.vtensor<[],f32> -> !torch.float | |
| %437 = torch.aten.item %435 : !torch.vtensor<[],si8> -> !torch.int | |
| %438 = torch.aten._make_per_tensor_quantized_tensor %433, %436, %437 : !torch.vtensor<[1,24,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,24,56,56],!torch.qint8> | |
| %439 = torch.aten.dequantize.self %438 : !torch.vtensor<[1,24,56,56],!torch.qint8> -> !torch.vtensor<[1,24,56,56],f32> | |
| %440 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %441 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_76 = torch.constant.int 12 | |
| %442 = torch.aten.item %440 : !torch.vtensor<[],f32> -> !torch.float | |
| %443 = torch.aten.item %441 : !torch.vtensor<[],si8> -> !torch.int | |
| %444 = torch.aten.quantize_per_tensor %12, %442, %443, %int12_76 : !torch.vtensor<[72,24,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72,24,1,1],!torch.qint8> | |
| %445 = torch.aten.int_repr %444 : !torch.vtensor<[72,24,1,1],!torch.qint8> -> !torch.vtensor<[72,24,1,1],si8> | |
| %446 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %447 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %448 = torch.aten.item %446 : !torch.vtensor<[],f32> -> !torch.float | |
| %449 = torch.aten.item %447 : !torch.vtensor<[],si8> -> !torch.int | |
| %450 = torch.aten._make_per_tensor_quantized_tensor %445, %448, %449 : !torch.vtensor<[72,24,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[72,24,1,1],!torch.qint8> | |
| %451 = torch.aten.dequantize.self %450 : !torch.vtensor<[72,24,1,1],!torch.qint8> -> !torch.vtensor<[72,24,1,1],f32> | |
| %452 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %453 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_77 = torch.constant.int 12 | |
| %454 = torch.aten.item %452 : !torch.vtensor<[],f32> -> !torch.float | |
| %455 = torch.aten.item %453 : !torch.vtensor<[],si8> -> !torch.int | |
| %456 = torch.aten.quantize_per_tensor %13, %454, %455, %int12_77 : !torch.vtensor<[72],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %457 = torch.aten.int_repr %456 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],si8> | |
| %458 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %459 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %460 = torch.aten.item %458 : !torch.vtensor<[],f32> -> !torch.float | |
| %461 = torch.aten.item %459 : !torch.vtensor<[],si8> -> !torch.int | |
| %462 = torch.aten._make_per_tensor_quantized_tensor %457, %460, %461 : !torch.vtensor<[72],si8>, !torch.float, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %463 = torch.aten.dequantize.self %462 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],f32> | |
| %int0_78 = torch.constant.int 0 | |
| %int0_79 = torch.constant.int 0 | |
| %int1_80 = torch.constant.int 1 | |
| %int1_81 = torch.constant.int 1 | |
| %int1_82 = torch.constant.int 1 | |
| %int1_83 = torch.constant.int 1 | |
| %int0_84 = torch.constant.int 0 | |
| %464 = torch.prim.ListConstruct %int0_78, %int0_79 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %465 = torch.prim.ListConstruct %int1_80, %int1_81 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %466 = torch.prim.ListConstruct %int1_82, %int1_83 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %467 = torch.prim.ListConstruct %int0_84, %int0_84 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_85 = torch.constant.bool false | |
| %int1_86 = torch.constant.int 1 | |
| %468 = torch.aten.convolution %439, %451, %463, %466, %464, %465, %false_85, %467, %int1_86 : !torch.vtensor<[1,24,56,56],f32>, !torch.vtensor<[72,24,1,1],f32>, !torch.vtensor<[72],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,72,56,56],f32> | |
| %469 = torch.aten.relu %468 : !torch.vtensor<[1,72,56,56],f32> -> !torch.vtensor<[1,72,56,56],f32> | |
| %470 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %471 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_87 = torch.constant.int 12 | |
| %472 = torch.aten.item %470 : !torch.vtensor<[],f32> -> !torch.float | |
| %473 = torch.aten.item %471 : !torch.vtensor<[],si8> -> !torch.int | |
| %474 = torch.aten.quantize_per_tensor %469, %472, %473, %int12_87 : !torch.vtensor<[1,72,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,56,56],!torch.qint8> | |
| %475 = torch.aten.int_repr %474 : !torch.vtensor<[1,72,56,56],!torch.qint8> -> !torch.vtensor<[1,72,56,56],si8> | |
| %476 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %477 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %478 = torch.aten.item %476 : !torch.vtensor<[],f32> -> !torch.float | |
| %479 = torch.aten.item %477 : !torch.vtensor<[],si8> -> !torch.int | |
| %480 = torch.aten._make_per_tensor_quantized_tensor %475, %478, %479 : !torch.vtensor<[1,72,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,56,56],!torch.qint8> | |
| %481 = torch.aten.dequantize.self %480 : !torch.vtensor<[1,72,56,56],!torch.qint8> -> !torch.vtensor<[1,72,56,56],f32> | |
| %482 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %483 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_88 = torch.constant.int 12 | |
| %484 = torch.aten.item %482 : !torch.vtensor<[],f32> -> !torch.float | |
| %485 = torch.aten.item %483 : !torch.vtensor<[],si8> -> !torch.int | |
| %486 = torch.aten.quantize_per_tensor %14, %484, %485, %int12_88 : !torch.vtensor<[72,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72,1,3,3],!torch.qint8> | |
| %487 = torch.aten.int_repr %486 : !torch.vtensor<[72,1,3,3],!torch.qint8> -> !torch.vtensor<[72,1,3,3],si8> | |
| %488 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %489 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %490 = torch.aten.item %488 : !torch.vtensor<[],f32> -> !torch.float | |
| %491 = torch.aten.item %489 : !torch.vtensor<[],si8> -> !torch.int | |
| %492 = torch.aten._make_per_tensor_quantized_tensor %487, %490, %491 : !torch.vtensor<[72,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[72,1,3,3],!torch.qint8> | |
| %493 = torch.aten.dequantize.self %492 : !torch.vtensor<[72,1,3,3],!torch.qint8> -> !torch.vtensor<[72,1,3,3],f32> | |
| %494 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %495 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_89 = torch.constant.int 12 | |
| %496 = torch.aten.item %494 : !torch.vtensor<[],f32> -> !torch.float | |
| %497 = torch.aten.item %495 : !torch.vtensor<[],si8> -> !torch.int | |
| %498 = torch.aten.quantize_per_tensor %15, %496, %497, %int12_89 : !torch.vtensor<[72],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %499 = torch.aten.int_repr %498 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],si8> | |
| %500 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %501 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %502 = torch.aten.item %500 : !torch.vtensor<[],f32> -> !torch.float | |
| %503 = torch.aten.item %501 : !torch.vtensor<[],si8> -> !torch.int | |
| %504 = torch.aten._make_per_tensor_quantized_tensor %499, %502, %503 : !torch.vtensor<[72],si8>, !torch.float, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %505 = torch.aten.dequantize.self %504 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],f32> | |
| %int1_90 = torch.constant.int 1 | |
| %int1_91 = torch.constant.int 1 | |
| %int1_92 = torch.constant.int 1 | |
| %int1_93 = torch.constant.int 1 | |
| %int1_94 = torch.constant.int 1 | |
| %int1_95 = torch.constant.int 1 | |
| %int0_96 = torch.constant.int 0 | |
| %506 = torch.prim.ListConstruct %int1_90, %int1_91 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %507 = torch.prim.ListConstruct %int1_92, %int1_93 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %508 = torch.prim.ListConstruct %int1_94, %int1_95 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %509 = torch.prim.ListConstruct %int0_96, %int0_96 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_97 = torch.constant.bool false | |
| %int72 = torch.constant.int 72 | |
| %510 = torch.aten.convolution %481, %493, %505, %508, %506, %507, %false_97, %509, %int72 : !torch.vtensor<[1,72,56,56],f32>, !torch.vtensor<[72,1,3,3],f32>, !torch.vtensor<[72],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,72,56,56],f32> | |
| %511 = torch.aten.relu %510 : !torch.vtensor<[1,72,56,56],f32> -> !torch.vtensor<[1,72,56,56],f32> | |
| %512 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %513 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_98 = torch.constant.int 12 | |
| %514 = torch.aten.item %512 : !torch.vtensor<[],f32> -> !torch.float | |
| %515 = torch.aten.item %513 : !torch.vtensor<[],si8> -> !torch.int | |
| %516 = torch.aten.quantize_per_tensor %511, %514, %515, %int12_98 : !torch.vtensor<[1,72,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,56,56],!torch.qint8> | |
| %517 = torch.aten.int_repr %516 : !torch.vtensor<[1,72,56,56],!torch.qint8> -> !torch.vtensor<[1,72,56,56],si8> | |
| %518 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %519 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %520 = torch.aten.item %518 : !torch.vtensor<[],f32> -> !torch.float | |
| %521 = torch.aten.item %519 : !torch.vtensor<[],si8> -> !torch.int | |
| %522 = torch.aten._make_per_tensor_quantized_tensor %517, %520, %521 : !torch.vtensor<[1,72,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,56,56],!torch.qint8> | |
| %523 = torch.aten.dequantize.self %522 : !torch.vtensor<[1,72,56,56],!torch.qint8> -> !torch.vtensor<[1,72,56,56],f32> | |
| %524 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %525 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_99 = torch.constant.int 12 | |
| %526 = torch.aten.item %524 : !torch.vtensor<[],f32> -> !torch.float | |
| %527 = torch.aten.item %525 : !torch.vtensor<[],si8> -> !torch.int | |
| %528 = torch.aten.quantize_per_tensor %16, %526, %527, %int12_99 : !torch.vtensor<[24,72,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[24,72,1,1],!torch.qint8> | |
| %529 = torch.aten.int_repr %528 : !torch.vtensor<[24,72,1,1],!torch.qint8> -> !torch.vtensor<[24,72,1,1],si8> | |
| %530 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %531 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %532 = torch.aten.item %530 : !torch.vtensor<[],f32> -> !torch.float | |
| %533 = torch.aten.item %531 : !torch.vtensor<[],si8> -> !torch.int | |
| %534 = torch.aten._make_per_tensor_quantized_tensor %529, %532, %533 : !torch.vtensor<[24,72,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[24,72,1,1],!torch.qint8> | |
| %535 = torch.aten.dequantize.self %534 : !torch.vtensor<[24,72,1,1],!torch.qint8> -> !torch.vtensor<[24,72,1,1],f32> | |
| %536 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %537 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_100 = torch.constant.int 12 | |
| %538 = torch.aten.item %536 : !torch.vtensor<[],f32> -> !torch.float | |
| %539 = torch.aten.item %537 : !torch.vtensor<[],si8> -> !torch.int | |
| %540 = torch.aten.quantize_per_tensor %17, %538, %539, %int12_100 : !torch.vtensor<[24],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[24],!torch.qint8> | |
| %541 = torch.aten.int_repr %540 : !torch.vtensor<[24],!torch.qint8> -> !torch.vtensor<[24],si8> | |
| %542 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %543 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %544 = torch.aten.item %542 : !torch.vtensor<[],f32> -> !torch.float | |
| %545 = torch.aten.item %543 : !torch.vtensor<[],si8> -> !torch.int | |
| %546 = torch.aten._make_per_tensor_quantized_tensor %541, %544, %545 : !torch.vtensor<[24],si8>, !torch.float, !torch.int -> !torch.vtensor<[24],!torch.qint8> | |
| %547 = torch.aten.dequantize.self %546 : !torch.vtensor<[24],!torch.qint8> -> !torch.vtensor<[24],f32> | |
| %int0_101 = torch.constant.int 0 | |
| %int0_102 = torch.constant.int 0 | |
| %int1_103 = torch.constant.int 1 | |
| %int1_104 = torch.constant.int 1 | |
| %int1_105 = torch.constant.int 1 | |
| %int1_106 = torch.constant.int 1 | |
| %int0_107 = torch.constant.int 0 | |
| %548 = torch.prim.ListConstruct %int0_101, %int0_102 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %549 = torch.prim.ListConstruct %int1_103, %int1_104 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %550 = torch.prim.ListConstruct %int1_105, %int1_106 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %551 = torch.prim.ListConstruct %int0_107, %int0_107 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_108 = torch.constant.bool false | |
| %int1_109 = torch.constant.int 1 | |
| %552 = torch.aten.convolution %523, %535, %547, %550, %548, %549, %false_108, %551, %int1_109 : !torch.vtensor<[1,72,56,56],f32>, !torch.vtensor<[24,72,1,1],f32>, !torch.vtensor<[24],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,24,56,56],f32> | |
| %553 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %554 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_110 = torch.constant.int 12 | |
| %555 = torch.aten.item %553 : !torch.vtensor<[],f32> -> !torch.float | |
| %556 = torch.aten.item %554 : !torch.vtensor<[],si8> -> !torch.int | |
| %557 = torch.aten.quantize_per_tensor %552, %555, %556, %int12_110 : !torch.vtensor<[1,24,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,24,56,56],!torch.qint8> | |
| %558 = torch.aten.int_repr %557 : !torch.vtensor<[1,24,56,56],!torch.qint8> -> !torch.vtensor<[1,24,56,56],si8> | |
| %559 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %560 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %561 = torch.aten.item %559 : !torch.vtensor<[],f32> -> !torch.float | |
| %562 = torch.aten.item %560 : !torch.vtensor<[],si8> -> !torch.int | |
| %563 = torch.aten._make_per_tensor_quantized_tensor %558, %561, %562 : !torch.vtensor<[1,24,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,24,56,56],!torch.qint8> | |
| %564 = torch.aten.dequantize.self %563 : !torch.vtensor<[1,24,56,56],!torch.qint8> -> !torch.vtensor<[1,24,56,56],f32> | |
| %int1_111 = torch.constant.int 1 | |
| %565 = torch.aten.add.Tensor %564, %439, %int1_111 : !torch.vtensor<[1,24,56,56],f32>, !torch.vtensor<[1,24,56,56],f32>, !torch.int -> !torch.vtensor<[1,24,56,56],f32> | |
| %566 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %567 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_112 = torch.constant.int 12 | |
| %568 = torch.aten.item %566 : !torch.vtensor<[],f32> -> !torch.float | |
| %569 = torch.aten.item %567 : !torch.vtensor<[],si8> -> !torch.int | |
| %570 = torch.aten.quantize_per_tensor %565, %568, %569, %int12_112 : !torch.vtensor<[1,24,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,24,56,56],!torch.qint8> | |
| %571 = torch.aten.int_repr %570 : !torch.vtensor<[1,24,56,56],!torch.qint8> -> !torch.vtensor<[1,24,56,56],si8> | |
| %572 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %573 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %574 = torch.aten.item %572 : !torch.vtensor<[],f32> -> !torch.float | |
| %575 = torch.aten.item %573 : !torch.vtensor<[],si8> -> !torch.int | |
| %576 = torch.aten._make_per_tensor_quantized_tensor %571, %574, %575 : !torch.vtensor<[1,24,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,24,56,56],!torch.qint8> | |
| %577 = torch.aten.dequantize.self %576 : !torch.vtensor<[1,24,56,56],!torch.qint8> -> !torch.vtensor<[1,24,56,56],f32> | |
| %578 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %579 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_113 = torch.constant.int 12 | |
| %580 = torch.aten.item %578 : !torch.vtensor<[],f32> -> !torch.float | |
| %581 = torch.aten.item %579 : !torch.vtensor<[],si8> -> !torch.int | |
| %582 = torch.aten.quantize_per_tensor %18, %580, %581, %int12_113 : !torch.vtensor<[72,24,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72,24,1,1],!torch.qint8> | |
| %583 = torch.aten.int_repr %582 : !torch.vtensor<[72,24,1,1],!torch.qint8> -> !torch.vtensor<[72,24,1,1],si8> | |
| %584 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %585 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %586 = torch.aten.item %584 : !torch.vtensor<[],f32> -> !torch.float | |
| %587 = torch.aten.item %585 : !torch.vtensor<[],si8> -> !torch.int | |
| %588 = torch.aten._make_per_tensor_quantized_tensor %583, %586, %587 : !torch.vtensor<[72,24,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[72,24,1,1],!torch.qint8> | |
| %589 = torch.aten.dequantize.self %588 : !torch.vtensor<[72,24,1,1],!torch.qint8> -> !torch.vtensor<[72,24,1,1],f32> | |
| %590 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %591 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_114 = torch.constant.int 12 | |
| %592 = torch.aten.item %590 : !torch.vtensor<[],f32> -> !torch.float | |
| %593 = torch.aten.item %591 : !torch.vtensor<[],si8> -> !torch.int | |
| %594 = torch.aten.quantize_per_tensor %19, %592, %593, %int12_114 : !torch.vtensor<[72],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %595 = torch.aten.int_repr %594 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],si8> | |
| %596 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %597 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %598 = torch.aten.item %596 : !torch.vtensor<[],f32> -> !torch.float | |
| %599 = torch.aten.item %597 : !torch.vtensor<[],si8> -> !torch.int | |
| %600 = torch.aten._make_per_tensor_quantized_tensor %595, %598, %599 : !torch.vtensor<[72],si8>, !torch.float, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %601 = torch.aten.dequantize.self %600 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],f32> | |
| %int0_115 = torch.constant.int 0 | |
| %int0_116 = torch.constant.int 0 | |
| %int1_117 = torch.constant.int 1 | |
| %int1_118 = torch.constant.int 1 | |
| %int1_119 = torch.constant.int 1 | |
| %int1_120 = torch.constant.int 1 | |
| %int0_121 = torch.constant.int 0 | |
| %602 = torch.prim.ListConstruct %int0_115, %int0_116 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %603 = torch.prim.ListConstruct %int1_117, %int1_118 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %604 = torch.prim.ListConstruct %int1_119, %int1_120 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %605 = torch.prim.ListConstruct %int0_121, %int0_121 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_122 = torch.constant.bool false | |
| %int1_123 = torch.constant.int 1 | |
| %606 = torch.aten.convolution %577, %589, %601, %604, %602, %603, %false_122, %605, %int1_123 : !torch.vtensor<[1,24,56,56],f32>, !torch.vtensor<[72,24,1,1],f32>, !torch.vtensor<[72],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,72,56,56],f32> | |
| %607 = torch.aten.relu %606 : !torch.vtensor<[1,72,56,56],f32> -> !torch.vtensor<[1,72,56,56],f32> | |
| %608 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %609 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_124 = torch.constant.int 12 | |
| %610 = torch.aten.item %608 : !torch.vtensor<[],f32> -> !torch.float | |
| %611 = torch.aten.item %609 : !torch.vtensor<[],si8> -> !torch.int | |
| %612 = torch.aten.quantize_per_tensor %607, %610, %611, %int12_124 : !torch.vtensor<[1,72,56,56],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,56,56],!torch.qint8> | |
| %613 = torch.aten.int_repr %612 : !torch.vtensor<[1,72,56,56],!torch.qint8> -> !torch.vtensor<[1,72,56,56],si8> | |
| %614 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %615 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %616 = torch.aten.item %614 : !torch.vtensor<[],f32> -> !torch.float | |
| %617 = torch.aten.item %615 : !torch.vtensor<[],si8> -> !torch.int | |
| %618 = torch.aten._make_per_tensor_quantized_tensor %613, %616, %617 : !torch.vtensor<[1,72,56,56],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,56,56],!torch.qint8> | |
| %619 = torch.aten.dequantize.self %618 : !torch.vtensor<[1,72,56,56],!torch.qint8> -> !torch.vtensor<[1,72,56,56],f32> | |
| %620 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %621 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_125 = torch.constant.int 12 | |
| %622 = torch.aten.item %620 : !torch.vtensor<[],f32> -> !torch.float | |
| %623 = torch.aten.item %621 : !torch.vtensor<[],si8> -> !torch.int | |
| %624 = torch.aten.quantize_per_tensor %20, %622, %623, %int12_125 : !torch.vtensor<[72,1,5,5],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72,1,5,5],!torch.qint8> | |
| %625 = torch.aten.int_repr %624 : !torch.vtensor<[72,1,5,5],!torch.qint8> -> !torch.vtensor<[72,1,5,5],si8> | |
| %626 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %627 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %628 = torch.aten.item %626 : !torch.vtensor<[],f32> -> !torch.float | |
| %629 = torch.aten.item %627 : !torch.vtensor<[],si8> -> !torch.int | |
| %630 = torch.aten._make_per_tensor_quantized_tensor %625, %628, %629 : !torch.vtensor<[72,1,5,5],si8>, !torch.float, !torch.int -> !torch.vtensor<[72,1,5,5],!torch.qint8> | |
| %631 = torch.aten.dequantize.self %630 : !torch.vtensor<[72,1,5,5],!torch.qint8> -> !torch.vtensor<[72,1,5,5],f32> | |
| %632 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %633 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_126 = torch.constant.int 12 | |
| %634 = torch.aten.item %632 : !torch.vtensor<[],f32> -> !torch.float | |
| %635 = torch.aten.item %633 : !torch.vtensor<[],si8> -> !torch.int | |
| %636 = torch.aten.quantize_per_tensor %21, %634, %635, %int12_126 : !torch.vtensor<[72],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %637 = torch.aten.int_repr %636 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],si8> | |
| %638 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %639 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %640 = torch.aten.item %638 : !torch.vtensor<[],f32> -> !torch.float | |
| %641 = torch.aten.item %639 : !torch.vtensor<[],si8> -> !torch.int | |
| %642 = torch.aten._make_per_tensor_quantized_tensor %637, %640, %641 : !torch.vtensor<[72],si8>, !torch.float, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %643 = torch.aten.dequantize.self %642 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],f32> | |
| %int2_127 = torch.constant.int 2 | |
| %int2_128 = torch.constant.int 2 | |
| %int1_129 = torch.constant.int 1 | |
| %int1_130 = torch.constant.int 1 | |
| %int2_131 = torch.constant.int 2 | |
| %int2_132 = torch.constant.int 2 | |
| %int0_133 = torch.constant.int 0 | |
| %644 = torch.prim.ListConstruct %int2_127, %int2_128 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %645 = torch.prim.ListConstruct %int1_129, %int1_130 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %646 = torch.prim.ListConstruct %int2_131, %int2_132 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %647 = torch.prim.ListConstruct %int0_133, %int0_133 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_134 = torch.constant.bool false | |
| %int72_135 = torch.constant.int 72 | |
| %648 = torch.aten.convolution %619, %631, %643, %646, %644, %645, %false_134, %647, %int72_135 : !torch.vtensor<[1,72,56,56],f32>, !torch.vtensor<[72,1,5,5],f32>, !torch.vtensor<[72],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,72,28,28],f32> | |
| %649 = torch.aten.relu %648 : !torch.vtensor<[1,72,28,28],f32> -> !torch.vtensor<[1,72,28,28],f32> | |
| %650 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %651 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_136 = torch.constant.int 12 | |
| %652 = torch.aten.item %650 : !torch.vtensor<[],f32> -> !torch.float | |
| %653 = torch.aten.item %651 : !torch.vtensor<[],si8> -> !torch.int | |
| %654 = torch.aten.quantize_per_tensor %649, %652, %653, %int12_136 : !torch.vtensor<[1,72,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,28,28],!torch.qint8> | |
| %655 = torch.aten.int_repr %654 : !torch.vtensor<[1,72,28,28],!torch.qint8> -> !torch.vtensor<[1,72,28,28],si8> | |
| %656 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %657 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %658 = torch.aten.item %656 : !torch.vtensor<[],f32> -> !torch.float | |
| %659 = torch.aten.item %657 : !torch.vtensor<[],si8> -> !torch.int | |
| %660 = torch.aten._make_per_tensor_quantized_tensor %655, %658, %659 : !torch.vtensor<[1,72,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,28,28],!torch.qint8> | |
| %661 = torch.aten.dequantize.self %660 : !torch.vtensor<[1,72,28,28],!torch.qint8> -> !torch.vtensor<[1,72,28,28],f32> | |
| %int0_137 = torch.constant.int 0 | |
| %int1_138 = torch.constant.int 1 | |
| %int28 = torch.constant.int 28 | |
| %int28_139 = torch.constant.int 28 | |
| %662 = torch.prim.ListConstruct %int28, %int28_139 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %663 = torch.prim.ListConstruct %int0_137, %int0_137 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %664 = torch.prim.ListConstruct %int1_138, %int1_138 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_140 = torch.constant.bool false | |
| %none_141 = torch.constant.none | |
| %665 = torch.aten.avg_pool2d %661, %662, %664, %663, %false_140, %false_140, %none_141 : !torch.vtensor<[1,72,28,28],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,72,1,1],f32> | |
| %666 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %667 = torch.aten.mul.Tensor %665, %666 : !torch.vtensor<[1,72,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,72,1,1],f32> | |
| %668 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %669 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_142 = torch.constant.int 12 | |
| %670 = torch.aten.item %668 : !torch.vtensor<[],f32> -> !torch.float | |
| %671 = torch.aten.item %669 : !torch.vtensor<[],si8> -> !torch.int | |
| %672 = torch.aten.quantize_per_tensor %667, %670, %671, %int12_142 : !torch.vtensor<[1,72,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,1,1],!torch.qint8> | |
| %673 = torch.aten.int_repr %672 : !torch.vtensor<[1,72,1,1],!torch.qint8> -> !torch.vtensor<[1,72,1,1],si8> | |
| %674 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %675 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %676 = torch.aten.item %674 : !torch.vtensor<[],f32> -> !torch.float | |
| %677 = torch.aten.item %675 : !torch.vtensor<[],si8> -> !torch.int | |
| %678 = torch.aten._make_per_tensor_quantized_tensor %673, %676, %677 : !torch.vtensor<[1,72,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,1,1],!torch.qint8> | |
| %679 = torch.aten.dequantize.self %678 : !torch.vtensor<[1,72,1,1],!torch.qint8> -> !torch.vtensor<[1,72,1,1],f32> | |
| %680 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %681 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_143 = torch.constant.int 12 | |
| %682 = torch.aten.item %680 : !torch.vtensor<[],f32> -> !torch.float | |
| %683 = torch.aten.item %681 : !torch.vtensor<[],si8> -> !torch.int | |
| %684 = torch.aten.quantize_per_tensor %22, %682, %683, %int12_143 : !torch.vtensor<[24,72,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[24,72,1,1],!torch.qint8> | |
| %685 = torch.aten.int_repr %684 : !torch.vtensor<[24,72,1,1],!torch.qint8> -> !torch.vtensor<[24,72,1,1],si8> | |
| %686 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %687 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %688 = torch.aten.item %686 : !torch.vtensor<[],f32> -> !torch.float | |
| %689 = torch.aten.item %687 : !torch.vtensor<[],si8> -> !torch.int | |
| %690 = torch.aten._make_per_tensor_quantized_tensor %685, %688, %689 : !torch.vtensor<[24,72,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[24,72,1,1],!torch.qint8> | |
| %691 = torch.aten.dequantize.self %690 : !torch.vtensor<[24,72,1,1],!torch.qint8> -> !torch.vtensor<[24,72,1,1],f32> | |
| %692 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %693 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_144 = torch.constant.int 12 | |
| %694 = torch.aten.item %692 : !torch.vtensor<[],f32> -> !torch.float | |
| %695 = torch.aten.item %693 : !torch.vtensor<[],si8> -> !torch.int | |
| %696 = torch.aten.quantize_per_tensor %23, %694, %695, %int12_144 : !torch.vtensor<[24],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[24],!torch.qint8> | |
| %697 = torch.aten.int_repr %696 : !torch.vtensor<[24],!torch.qint8> -> !torch.vtensor<[24],si8> | |
| %698 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %699 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %700 = torch.aten.item %698 : !torch.vtensor<[],f32> -> !torch.float | |
| %701 = torch.aten.item %699 : !torch.vtensor<[],si8> -> !torch.int | |
| %702 = torch.aten._make_per_tensor_quantized_tensor %697, %700, %701 : !torch.vtensor<[24],si8>, !torch.float, !torch.int -> !torch.vtensor<[24],!torch.qint8> | |
| %703 = torch.aten.dequantize.self %702 : !torch.vtensor<[24],!torch.qint8> -> !torch.vtensor<[24],f32> | |
| %int0_145 = torch.constant.int 0 | |
| %int0_146 = torch.constant.int 0 | |
| %int1_147 = torch.constant.int 1 | |
| %int1_148 = torch.constant.int 1 | |
| %int1_149 = torch.constant.int 1 | |
| %int1_150 = torch.constant.int 1 | |
| %int0_151 = torch.constant.int 0 | |
| %704 = torch.prim.ListConstruct %int0_145, %int0_146 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %705 = torch.prim.ListConstruct %int1_147, %int1_148 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %706 = torch.prim.ListConstruct %int1_149, %int1_150 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %707 = torch.prim.ListConstruct %int0_151, %int0_151 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_152 = torch.constant.bool false | |
| %int1_153 = torch.constant.int 1 | |
| %708 = torch.aten.convolution %679, %691, %703, %706, %704, %705, %false_152, %707, %int1_153 : !torch.vtensor<[1,72,1,1],f32>, !torch.vtensor<[24,72,1,1],f32>, !torch.vtensor<[24],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,24,1,1],f32> | |
| %709 = torch.aten.relu %708 : !torch.vtensor<[1,24,1,1],f32> -> !torch.vtensor<[1,24,1,1],f32> | |
| %710 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %711 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_154 = torch.constant.int 12 | |
| %712 = torch.aten.item %710 : !torch.vtensor<[],f32> -> !torch.float | |
| %713 = torch.aten.item %711 : !torch.vtensor<[],si8> -> !torch.int | |
| %714 = torch.aten.quantize_per_tensor %709, %712, %713, %int12_154 : !torch.vtensor<[1,24,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,24,1,1],!torch.qint8> | |
| %715 = torch.aten.int_repr %714 : !torch.vtensor<[1,24,1,1],!torch.qint8> -> !torch.vtensor<[1,24,1,1],si8> | |
| %716 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %717 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %718 = torch.aten.item %716 : !torch.vtensor<[],f32> -> !torch.float | |
| %719 = torch.aten.item %717 : !torch.vtensor<[],si8> -> !torch.int | |
| %720 = torch.aten._make_per_tensor_quantized_tensor %715, %718, %719 : !torch.vtensor<[1,24,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,24,1,1],!torch.qint8> | |
| %721 = torch.aten.dequantize.self %720 : !torch.vtensor<[1,24,1,1],!torch.qint8> -> !torch.vtensor<[1,24,1,1],f32> | |
| %722 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %723 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_155 = torch.constant.int 12 | |
| %724 = torch.aten.item %722 : !torch.vtensor<[],f32> -> !torch.float | |
| %725 = torch.aten.item %723 : !torch.vtensor<[],si8> -> !torch.int | |
| %726 = torch.aten.quantize_per_tensor %24, %724, %725, %int12_155 : !torch.vtensor<[72,24,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72,24,1,1],!torch.qint8> | |
| %727 = torch.aten.int_repr %726 : !torch.vtensor<[72,24,1,1],!torch.qint8> -> !torch.vtensor<[72,24,1,1],si8> | |
| %728 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %729 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %730 = torch.aten.item %728 : !torch.vtensor<[],f32> -> !torch.float | |
| %731 = torch.aten.item %729 : !torch.vtensor<[],si8> -> !torch.int | |
| %732 = torch.aten._make_per_tensor_quantized_tensor %727, %730, %731 : !torch.vtensor<[72,24,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[72,24,1,1],!torch.qint8> | |
| %733 = torch.aten.dequantize.self %732 : !torch.vtensor<[72,24,1,1],!torch.qint8> -> !torch.vtensor<[72,24,1,1],f32> | |
| %734 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %735 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_156 = torch.constant.int 12 | |
| %736 = torch.aten.item %734 : !torch.vtensor<[],f32> -> !torch.float | |
| %737 = torch.aten.item %735 : !torch.vtensor<[],si8> -> !torch.int | |
| %738 = torch.aten.quantize_per_tensor %25, %736, %737, %int12_156 : !torch.vtensor<[72],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %739 = torch.aten.int_repr %738 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],si8> | |
| %740 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %741 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %742 = torch.aten.item %740 : !torch.vtensor<[],f32> -> !torch.float | |
| %743 = torch.aten.item %741 : !torch.vtensor<[],si8> -> !torch.int | |
| %744 = torch.aten._make_per_tensor_quantized_tensor %739, %742, %743 : !torch.vtensor<[72],si8>, !torch.float, !torch.int -> !torch.vtensor<[72],!torch.qint8> | |
| %745 = torch.aten.dequantize.self %744 : !torch.vtensor<[72],!torch.qint8> -> !torch.vtensor<[72],f32> | |
| %int0_157 = torch.constant.int 0 | |
| %int0_158 = torch.constant.int 0 | |
| %int1_159 = torch.constant.int 1 | |
| %int1_160 = torch.constant.int 1 | |
| %int1_161 = torch.constant.int 1 | |
| %int1_162 = torch.constant.int 1 | |
| %int0_163 = torch.constant.int 0 | |
| %746 = torch.prim.ListConstruct %int0_157, %int0_158 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %747 = torch.prim.ListConstruct %int1_159, %int1_160 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %748 = torch.prim.ListConstruct %int1_161, %int1_162 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %749 = torch.prim.ListConstruct %int0_163, %int0_163 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_164 = torch.constant.bool false | |
| %int1_165 = torch.constant.int 1 | |
| %750 = torch.aten.convolution %721, %733, %745, %748, %746, %747, %false_164, %749, %int1_165 : !torch.vtensor<[1,24,1,1],f32>, !torch.vtensor<[72,24,1,1],f32>, !torch.vtensor<[72],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,72,1,1],f32> | |
| %751 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %752 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_166 = torch.constant.int 12 | |
| %753 = torch.aten.item %751 : !torch.vtensor<[],f32> -> !torch.float | |
| %754 = torch.aten.item %752 : !torch.vtensor<[],si8> -> !torch.int | |
| %755 = torch.aten.quantize_per_tensor %750, %753, %754, %int12_166 : !torch.vtensor<[1,72,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,1,1],!torch.qint8> | |
| %756 = torch.aten.int_repr %755 : !torch.vtensor<[1,72,1,1],!torch.qint8> -> !torch.vtensor<[1,72,1,1],si8> | |
| %757 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %758 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %759 = torch.aten.item %757 : !torch.vtensor<[],f32> -> !torch.float | |
| %760 = torch.aten.item %758 : !torch.vtensor<[],si8> -> !torch.int | |
| %761 = torch.aten._make_per_tensor_quantized_tensor %756, %759, %760 : !torch.vtensor<[1,72,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,1,1],!torch.qint8> | |
| %762 = torch.aten.dequantize.self %761 : !torch.vtensor<[1,72,1,1],!torch.qint8> -> !torch.vtensor<[1,72,1,1],f32> | |
| %763 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_167 = torch.constant.int 1 | |
| %764 = torch.aten.add.Tensor %762, %763, %int1_167 : !torch.vtensor<[1,72,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,72,1,1],f32> | |
| %765 = torch.aten.relu %764 : !torch.vtensor<[1,72,1,1],f32> -> !torch.vtensor<[1,72,1,1],f32> | |
| %766 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %767 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_168 = torch.constant.int 6 | |
| %none_169 = torch.constant.none | |
| %false_170 = torch.constant.bool false | |
| %768 = torch.aten.to.dtype %766, %int6_168, %false_170, %false_170, %none_169 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_171 = torch.constant.int 6 | |
| %none_172 = torch.constant.none | |
| %false_173 = torch.constant.bool false | |
| %769 = torch.aten.to.dtype %767, %int6_171, %false_173, %false_173, %none_172 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %770 = torch.aten.clamp.Tensor %765, %768, %769 : !torch.vtensor<[1,72,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,72,1,1],f32> | |
| %771 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %772 = torch.aten.mul.Tensor %770, %771 : !torch.vtensor<[1,72,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,72,1,1],f32> | |
| %773 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %774 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_174 = torch.constant.int 12 | |
| %775 = torch.aten.item %773 : !torch.vtensor<[],f32> -> !torch.float | |
| %776 = torch.aten.item %774 : !torch.vtensor<[],si8> -> !torch.int | |
| %777 = torch.aten.quantize_per_tensor %772, %775, %776, %int12_174 : !torch.vtensor<[1,72,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,1,1],!torch.qint8> | |
| %778 = torch.aten.int_repr %777 : !torch.vtensor<[1,72,1,1],!torch.qint8> -> !torch.vtensor<[1,72,1,1],si8> | |
| %779 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %780 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %781 = torch.aten.item %779 : !torch.vtensor<[],f32> -> !torch.float | |
| %782 = torch.aten.item %780 : !torch.vtensor<[],si8> -> !torch.int | |
| %783 = torch.aten._make_per_tensor_quantized_tensor %778, %781, %782 : !torch.vtensor<[1,72,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,1,1],!torch.qint8> | |
| %784 = torch.aten.dequantize.self %783 : !torch.vtensor<[1,72,1,1],!torch.qint8> -> !torch.vtensor<[1,72,1,1],f32> | |
| %785 = torch.aten.mul.Tensor %784, %661 : !torch.vtensor<[1,72,1,1],f32>, !torch.vtensor<[1,72,28,28],f32> -> !torch.vtensor<[1,72,28,28],f32> | |
| %786 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %787 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_175 = torch.constant.int 12 | |
| %788 = torch.aten.item %786 : !torch.vtensor<[],f32> -> !torch.float | |
| %789 = torch.aten.item %787 : !torch.vtensor<[],si8> -> !torch.int | |
| %790 = torch.aten.quantize_per_tensor %785, %788, %789, %int12_175 : !torch.vtensor<[1,72,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,72,28,28],!torch.qint8> | |
| %791 = torch.aten.int_repr %790 : !torch.vtensor<[1,72,28,28],!torch.qint8> -> !torch.vtensor<[1,72,28,28],si8> | |
| %792 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %793 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %794 = torch.aten.item %792 : !torch.vtensor<[],f32> -> !torch.float | |
| %795 = torch.aten.item %793 : !torch.vtensor<[],si8> -> !torch.int | |
| %796 = torch.aten._make_per_tensor_quantized_tensor %791, %794, %795 : !torch.vtensor<[1,72,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,72,28,28],!torch.qint8> | |
| %797 = torch.aten.dequantize.self %796 : !torch.vtensor<[1,72,28,28],!torch.qint8> -> !torch.vtensor<[1,72,28,28],f32> | |
| %798 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %799 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_176 = torch.constant.int 12 | |
| %800 = torch.aten.item %798 : !torch.vtensor<[],f32> -> !torch.float | |
| %801 = torch.aten.item %799 : !torch.vtensor<[],si8> -> !torch.int | |
| %802 = torch.aten.quantize_per_tensor %26, %800, %801, %int12_176 : !torch.vtensor<[40,72,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[40,72,1,1],!torch.qint8> | |
| %803 = torch.aten.int_repr %802 : !torch.vtensor<[40,72,1,1],!torch.qint8> -> !torch.vtensor<[40,72,1,1],si8> | |
| %804 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %805 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %806 = torch.aten.item %804 : !torch.vtensor<[],f32> -> !torch.float | |
| %807 = torch.aten.item %805 : !torch.vtensor<[],si8> -> !torch.int | |
| %808 = torch.aten._make_per_tensor_quantized_tensor %803, %806, %807 : !torch.vtensor<[40,72,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[40,72,1,1],!torch.qint8> | |
| %809 = torch.aten.dequantize.self %808 : !torch.vtensor<[40,72,1,1],!torch.qint8> -> !torch.vtensor<[40,72,1,1],f32> | |
| %810 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %811 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_177 = torch.constant.int 12 | |
| %812 = torch.aten.item %810 : !torch.vtensor<[],f32> -> !torch.float | |
| %813 = torch.aten.item %811 : !torch.vtensor<[],si8> -> !torch.int | |
| %814 = torch.aten.quantize_per_tensor %27, %812, %813, %int12_177 : !torch.vtensor<[40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[40],!torch.qint8> | |
| %815 = torch.aten.int_repr %814 : !torch.vtensor<[40],!torch.qint8> -> !torch.vtensor<[40],si8> | |
| %816 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %817 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %818 = torch.aten.item %816 : !torch.vtensor<[],f32> -> !torch.float | |
| %819 = torch.aten.item %817 : !torch.vtensor<[],si8> -> !torch.int | |
| %820 = torch.aten._make_per_tensor_quantized_tensor %815, %818, %819 : !torch.vtensor<[40],si8>, !torch.float, !torch.int -> !torch.vtensor<[40],!torch.qint8> | |
| %821 = torch.aten.dequantize.self %820 : !torch.vtensor<[40],!torch.qint8> -> !torch.vtensor<[40],f32> | |
| %int0_178 = torch.constant.int 0 | |
| %int0_179 = torch.constant.int 0 | |
| %int1_180 = torch.constant.int 1 | |
| %int1_181 = torch.constant.int 1 | |
| %int1_182 = torch.constant.int 1 | |
| %int1_183 = torch.constant.int 1 | |
| %int0_184 = torch.constant.int 0 | |
| %822 = torch.prim.ListConstruct %int0_178, %int0_179 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %823 = torch.prim.ListConstruct %int1_180, %int1_181 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %824 = torch.prim.ListConstruct %int1_182, %int1_183 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %825 = torch.prim.ListConstruct %int0_184, %int0_184 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_185 = torch.constant.bool false | |
| %int1_186 = torch.constant.int 1 | |
| %826 = torch.aten.convolution %797, %809, %821, %824, %822, %823, %false_185, %825, %int1_186 : !torch.vtensor<[1,72,28,28],f32>, !torch.vtensor<[40,72,1,1],f32>, !torch.vtensor<[40],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,40,28,28],f32> | |
| %827 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %828 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_187 = torch.constant.int 12 | |
| %829 = torch.aten.item %827 : !torch.vtensor<[],f32> -> !torch.float | |
| %830 = torch.aten.item %828 : !torch.vtensor<[],si8> -> !torch.int | |
| %831 = torch.aten.quantize_per_tensor %826, %829, %830, %int12_187 : !torch.vtensor<[1,40,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %832 = torch.aten.int_repr %831 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],si8> | |
| %833 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %834 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %835 = torch.aten.item %833 : !torch.vtensor<[],f32> -> !torch.float | |
| %836 = torch.aten.item %834 : !torch.vtensor<[],si8> -> !torch.int | |
| %837 = torch.aten._make_per_tensor_quantized_tensor %832, %835, %836 : !torch.vtensor<[1,40,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %838 = torch.aten.dequantize.self %837 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],f32> | |
| %839 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %840 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_188 = torch.constant.int 12 | |
| %841 = torch.aten.item %839 : !torch.vtensor<[],f32> -> !torch.float | |
| %842 = torch.aten.item %840 : !torch.vtensor<[],si8> -> !torch.int | |
| %843 = torch.aten.quantize_per_tensor %28, %841, %842, %int12_188 : !torch.vtensor<[120,40,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,40,1,1],!torch.qint8> | |
| %844 = torch.aten.int_repr %843 : !torch.vtensor<[120,40,1,1],!torch.qint8> -> !torch.vtensor<[120,40,1,1],si8> | |
| %845 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %846 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %847 = torch.aten.item %845 : !torch.vtensor<[],f32> -> !torch.float | |
| %848 = torch.aten.item %846 : !torch.vtensor<[],si8> -> !torch.int | |
| %849 = torch.aten._make_per_tensor_quantized_tensor %844, %847, %848 : !torch.vtensor<[120,40,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,40,1,1],!torch.qint8> | |
| %850 = torch.aten.dequantize.self %849 : !torch.vtensor<[120,40,1,1],!torch.qint8> -> !torch.vtensor<[120,40,1,1],f32> | |
| %851 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %852 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_189 = torch.constant.int 12 | |
| %853 = torch.aten.item %851 : !torch.vtensor<[],f32> -> !torch.float | |
| %854 = torch.aten.item %852 : !torch.vtensor<[],si8> -> !torch.int | |
| %855 = torch.aten.quantize_per_tensor %29, %853, %854, %int12_189 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %856 = torch.aten.int_repr %855 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %857 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %858 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %859 = torch.aten.item %857 : !torch.vtensor<[],f32> -> !torch.float | |
| %860 = torch.aten.item %858 : !torch.vtensor<[],si8> -> !torch.int | |
| %861 = torch.aten._make_per_tensor_quantized_tensor %856, %859, %860 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %862 = torch.aten.dequantize.self %861 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int0_190 = torch.constant.int 0 | |
| %int0_191 = torch.constant.int 0 | |
| %int1_192 = torch.constant.int 1 | |
| %int1_193 = torch.constant.int 1 | |
| %int1_194 = torch.constant.int 1 | |
| %int1_195 = torch.constant.int 1 | |
| %int0_196 = torch.constant.int 0 | |
| %863 = torch.prim.ListConstruct %int0_190, %int0_191 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %864 = torch.prim.ListConstruct %int1_192, %int1_193 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %865 = torch.prim.ListConstruct %int1_194, %int1_195 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %866 = torch.prim.ListConstruct %int0_196, %int0_196 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_197 = torch.constant.bool false | |
| %int1_198 = torch.constant.int 1 | |
| %867 = torch.aten.convolution %838, %850, %862, %865, %863, %864, %false_197, %866, %int1_198 : !torch.vtensor<[1,40,28,28],f32>, !torch.vtensor<[120,40,1,1],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,28,28],f32> | |
| %868 = torch.aten.relu %867 : !torch.vtensor<[1,120,28,28],f32> -> !torch.vtensor<[1,120,28,28],f32> | |
| %869 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %870 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_199 = torch.constant.int 12 | |
| %871 = torch.aten.item %869 : !torch.vtensor<[],f32> -> !torch.float | |
| %872 = torch.aten.item %870 : !torch.vtensor<[],si8> -> !torch.int | |
| %873 = torch.aten.quantize_per_tensor %868, %871, %872, %int12_199 : !torch.vtensor<[1,120,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %874 = torch.aten.int_repr %873 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],si8> | |
| %875 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %876 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %877 = torch.aten.item %875 : !torch.vtensor<[],f32> -> !torch.float | |
| %878 = torch.aten.item %876 : !torch.vtensor<[],si8> -> !torch.int | |
| %879 = torch.aten._make_per_tensor_quantized_tensor %874, %877, %878 : !torch.vtensor<[1,120,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %880 = torch.aten.dequantize.self %879 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],f32> | |
| %881 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %882 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_200 = torch.constant.int 12 | |
| %883 = torch.aten.item %881 : !torch.vtensor<[],f32> -> !torch.float | |
| %884 = torch.aten.item %882 : !torch.vtensor<[],si8> -> !torch.int | |
| %885 = torch.aten.quantize_per_tensor %30, %883, %884, %int12_200 : !torch.vtensor<[120,1,5,5],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,1,5,5],!torch.qint8> | |
| %886 = torch.aten.int_repr %885 : !torch.vtensor<[120,1,5,5],!torch.qint8> -> !torch.vtensor<[120,1,5,5],si8> | |
| %887 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %888 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %889 = torch.aten.item %887 : !torch.vtensor<[],f32> -> !torch.float | |
| %890 = torch.aten.item %888 : !torch.vtensor<[],si8> -> !torch.int | |
| %891 = torch.aten._make_per_tensor_quantized_tensor %886, %889, %890 : !torch.vtensor<[120,1,5,5],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,1,5,5],!torch.qint8> | |
| %892 = torch.aten.dequantize.self %891 : !torch.vtensor<[120,1,5,5],!torch.qint8> -> !torch.vtensor<[120,1,5,5],f32> | |
| %893 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %894 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_201 = torch.constant.int 12 | |
| %895 = torch.aten.item %893 : !torch.vtensor<[],f32> -> !torch.float | |
| %896 = torch.aten.item %894 : !torch.vtensor<[],si8> -> !torch.int | |
| %897 = torch.aten.quantize_per_tensor %31, %895, %896, %int12_201 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %898 = torch.aten.int_repr %897 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %899 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %900 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %901 = torch.aten.item %899 : !torch.vtensor<[],f32> -> !torch.float | |
| %902 = torch.aten.item %900 : !torch.vtensor<[],si8> -> !torch.int | |
| %903 = torch.aten._make_per_tensor_quantized_tensor %898, %901, %902 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %904 = torch.aten.dequantize.self %903 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int2_202 = torch.constant.int 2 | |
| %int2_203 = torch.constant.int 2 | |
| %int1_204 = torch.constant.int 1 | |
| %int1_205 = torch.constant.int 1 | |
| %int1_206 = torch.constant.int 1 | |
| %int1_207 = torch.constant.int 1 | |
| %int0_208 = torch.constant.int 0 | |
| %905 = torch.prim.ListConstruct %int2_202, %int2_203 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %906 = torch.prim.ListConstruct %int1_204, %int1_205 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %907 = torch.prim.ListConstruct %int1_206, %int1_207 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %908 = torch.prim.ListConstruct %int0_208, %int0_208 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_209 = torch.constant.bool false | |
| %int120 = torch.constant.int 120 | |
| %909 = torch.aten.convolution %880, %892, %904, %907, %905, %906, %false_209, %908, %int120 : !torch.vtensor<[1,120,28,28],f32>, !torch.vtensor<[120,1,5,5],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,28,28],f32> | |
| %910 = torch.aten.relu %909 : !torch.vtensor<[1,120,28,28],f32> -> !torch.vtensor<[1,120,28,28],f32> | |
| %911 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %912 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_210 = torch.constant.int 12 | |
| %913 = torch.aten.item %911 : !torch.vtensor<[],f32> -> !torch.float | |
| %914 = torch.aten.item %912 : !torch.vtensor<[],si8> -> !torch.int | |
| %915 = torch.aten.quantize_per_tensor %910, %913, %914, %int12_210 : !torch.vtensor<[1,120,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %916 = torch.aten.int_repr %915 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],si8> | |
| %917 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %918 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %919 = torch.aten.item %917 : !torch.vtensor<[],f32> -> !torch.float | |
| %920 = torch.aten.item %918 : !torch.vtensor<[],si8> -> !torch.int | |
| %921 = torch.aten._make_per_tensor_quantized_tensor %916, %919, %920 : !torch.vtensor<[1,120,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %922 = torch.aten.dequantize.self %921 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],f32> | |
| %int0_211 = torch.constant.int 0 | |
| %int1_212 = torch.constant.int 1 | |
| %int28_213 = torch.constant.int 28 | |
| %int28_214 = torch.constant.int 28 | |
| %923 = torch.prim.ListConstruct %int28_213, %int28_214 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %924 = torch.prim.ListConstruct %int0_211, %int0_211 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %925 = torch.prim.ListConstruct %int1_212, %int1_212 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_215 = torch.constant.bool false | |
| %none_216 = torch.constant.none | |
| %926 = torch.aten.avg_pool2d %922, %923, %925, %924, %false_215, %false_215, %none_216 : !torch.vtensor<[1,120,28,28],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,120,1,1],f32> | |
| %927 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %928 = torch.aten.mul.Tensor %926, %927 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %929 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %930 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_217 = torch.constant.int 12 | |
| %931 = torch.aten.item %929 : !torch.vtensor<[],f32> -> !torch.float | |
| %932 = torch.aten.item %930 : !torch.vtensor<[],si8> -> !torch.int | |
| %933 = torch.aten.quantize_per_tensor %928, %931, %932, %int12_217 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %934 = torch.aten.int_repr %933 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %935 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %936 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %937 = torch.aten.item %935 : !torch.vtensor<[],f32> -> !torch.float | |
| %938 = torch.aten.item %936 : !torch.vtensor<[],si8> -> !torch.int | |
| %939 = torch.aten._make_per_tensor_quantized_tensor %934, %937, %938 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %940 = torch.aten.dequantize.self %939 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %941 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %942 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_218 = torch.constant.int 12 | |
| %943 = torch.aten.item %941 : !torch.vtensor<[],f32> -> !torch.float | |
| %944 = torch.aten.item %942 : !torch.vtensor<[],si8> -> !torch.int | |
| %945 = torch.aten.quantize_per_tensor %32, %943, %944, %int12_218 : !torch.vtensor<[32,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,120,1,1],!torch.qint8> | |
| %946 = torch.aten.int_repr %945 : !torch.vtensor<[32,120,1,1],!torch.qint8> -> !torch.vtensor<[32,120,1,1],si8> | |
| %947 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %948 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %949 = torch.aten.item %947 : !torch.vtensor<[],f32> -> !torch.float | |
| %950 = torch.aten.item %948 : !torch.vtensor<[],si8> -> !torch.int | |
| %951 = torch.aten._make_per_tensor_quantized_tensor %946, %949, %950 : !torch.vtensor<[32,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,120,1,1],!torch.qint8> | |
| %952 = torch.aten.dequantize.self %951 : !torch.vtensor<[32,120,1,1],!torch.qint8> -> !torch.vtensor<[32,120,1,1],f32> | |
| %953 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %954 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_219 = torch.constant.int 12 | |
| %955 = torch.aten.item %953 : !torch.vtensor<[],f32> -> !torch.float | |
| %956 = torch.aten.item %954 : !torch.vtensor<[],si8> -> !torch.int | |
| %957 = torch.aten.quantize_per_tensor %33, %955, %956, %int12_219 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %958 = torch.aten.int_repr %957 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %959 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %960 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %961 = torch.aten.item %959 : !torch.vtensor<[],f32> -> !torch.float | |
| %962 = torch.aten.item %960 : !torch.vtensor<[],si8> -> !torch.int | |
| %963 = torch.aten._make_per_tensor_quantized_tensor %958, %961, %962 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %964 = torch.aten.dequantize.self %963 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int0_220 = torch.constant.int 0 | |
| %int0_221 = torch.constant.int 0 | |
| %int1_222 = torch.constant.int 1 | |
| %int1_223 = torch.constant.int 1 | |
| %int1_224 = torch.constant.int 1 | |
| %int1_225 = torch.constant.int 1 | |
| %int0_226 = torch.constant.int 0 | |
| %965 = torch.prim.ListConstruct %int0_220, %int0_221 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %966 = torch.prim.ListConstruct %int1_222, %int1_223 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %967 = torch.prim.ListConstruct %int1_224, %int1_225 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %968 = torch.prim.ListConstruct %int0_226, %int0_226 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_227 = torch.constant.bool false | |
| %int1_228 = torch.constant.int 1 | |
| %969 = torch.aten.convolution %940, %952, %964, %967, %965, %966, %false_227, %968, %int1_228 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[32,120,1,1],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,1,1],f32> | |
| %970 = torch.aten.relu %969 : !torch.vtensor<[1,32,1,1],f32> -> !torch.vtensor<[1,32,1,1],f32> | |
| %971 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %972 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_229 = torch.constant.int 12 | |
| %973 = torch.aten.item %971 : !torch.vtensor<[],f32> -> !torch.float | |
| %974 = torch.aten.item %972 : !torch.vtensor<[],si8> -> !torch.int | |
| %975 = torch.aten.quantize_per_tensor %970, %973, %974, %int12_229 : !torch.vtensor<[1,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,1,1],!torch.qint8> | |
| %976 = torch.aten.int_repr %975 : !torch.vtensor<[1,32,1,1],!torch.qint8> -> !torch.vtensor<[1,32,1,1],si8> | |
| %977 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %978 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %979 = torch.aten.item %977 : !torch.vtensor<[],f32> -> !torch.float | |
| %980 = torch.aten.item %978 : !torch.vtensor<[],si8> -> !torch.int | |
| %981 = torch.aten._make_per_tensor_quantized_tensor %976, %979, %980 : !torch.vtensor<[1,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,1,1],!torch.qint8> | |
| %982 = torch.aten.dequantize.self %981 : !torch.vtensor<[1,32,1,1],!torch.qint8> -> !torch.vtensor<[1,32,1,1],f32> | |
| %983 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %984 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_230 = torch.constant.int 12 | |
| %985 = torch.aten.item %983 : !torch.vtensor<[],f32> -> !torch.float | |
| %986 = torch.aten.item %984 : !torch.vtensor<[],si8> -> !torch.int | |
| %987 = torch.aten.quantize_per_tensor %34, %985, %986, %int12_230 : !torch.vtensor<[120,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,32,1,1],!torch.qint8> | |
| %988 = torch.aten.int_repr %987 : !torch.vtensor<[120,32,1,1],!torch.qint8> -> !torch.vtensor<[120,32,1,1],si8> | |
| %989 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %990 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %991 = torch.aten.item %989 : !torch.vtensor<[],f32> -> !torch.float | |
| %992 = torch.aten.item %990 : !torch.vtensor<[],si8> -> !torch.int | |
| %993 = torch.aten._make_per_tensor_quantized_tensor %988, %991, %992 : !torch.vtensor<[120,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,32,1,1],!torch.qint8> | |
| %994 = torch.aten.dequantize.self %993 : !torch.vtensor<[120,32,1,1],!torch.qint8> -> !torch.vtensor<[120,32,1,1],f32> | |
| %995 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %996 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_231 = torch.constant.int 12 | |
| %997 = torch.aten.item %995 : !torch.vtensor<[],f32> -> !torch.float | |
| %998 = torch.aten.item %996 : !torch.vtensor<[],si8> -> !torch.int | |
| %999 = torch.aten.quantize_per_tensor %35, %997, %998, %int12_231 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1000 = torch.aten.int_repr %999 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %1001 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1002 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1003 = torch.aten.item %1001 : !torch.vtensor<[],f32> -> !torch.float | |
| %1004 = torch.aten.item %1002 : !torch.vtensor<[],si8> -> !torch.int | |
| %1005 = torch.aten._make_per_tensor_quantized_tensor %1000, %1003, %1004 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1006 = torch.aten.dequantize.self %1005 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int0_232 = torch.constant.int 0 | |
| %int0_233 = torch.constant.int 0 | |
| %int1_234 = torch.constant.int 1 | |
| %int1_235 = torch.constant.int 1 | |
| %int1_236 = torch.constant.int 1 | |
| %int1_237 = torch.constant.int 1 | |
| %int0_238 = torch.constant.int 0 | |
| %1007 = torch.prim.ListConstruct %int0_232, %int0_233 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1008 = torch.prim.ListConstruct %int1_234, %int1_235 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1009 = torch.prim.ListConstruct %int1_236, %int1_237 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1010 = torch.prim.ListConstruct %int0_238, %int0_238 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_239 = torch.constant.bool false | |
| %int1_240 = torch.constant.int 1 | |
| %1011 = torch.aten.convolution %982, %994, %1006, %1009, %1007, %1008, %false_239, %1010, %int1_240 : !torch.vtensor<[1,32,1,1],f32>, !torch.vtensor<[120,32,1,1],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,1,1],f32> | |
| %1012 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1013 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_241 = torch.constant.int 12 | |
| %1014 = torch.aten.item %1012 : !torch.vtensor<[],f32> -> !torch.float | |
| %1015 = torch.aten.item %1013 : !torch.vtensor<[],si8> -> !torch.int | |
| %1016 = torch.aten.quantize_per_tensor %1011, %1014, %1015, %int12_241 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1017 = torch.aten.int_repr %1016 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %1018 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1019 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1020 = torch.aten.item %1018 : !torch.vtensor<[],f32> -> !torch.float | |
| %1021 = torch.aten.item %1019 : !torch.vtensor<[],si8> -> !torch.int | |
| %1022 = torch.aten._make_per_tensor_quantized_tensor %1017, %1020, %1021 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1023 = torch.aten.dequantize.self %1022 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1024 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_242 = torch.constant.int 1 | |
| %1025 = torch.aten.add.Tensor %1023, %1024, %int1_242 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,120,1,1],f32> | |
| %1026 = torch.aten.relu %1025 : !torch.vtensor<[1,120,1,1],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1027 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1028 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_243 = torch.constant.int 6 | |
| %none_244 = torch.constant.none | |
| %false_245 = torch.constant.bool false | |
| %1029 = torch.aten.to.dtype %1027, %int6_243, %false_245, %false_245, %none_244 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_246 = torch.constant.int 6 | |
| %none_247 = torch.constant.none | |
| %false_248 = torch.constant.bool false | |
| %1030 = torch.aten.to.dtype %1028, %int6_246, %false_248, %false_248, %none_247 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1031 = torch.aten.clamp.Tensor %1026, %1029, %1030 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1032 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1033 = torch.aten.mul.Tensor %1031, %1032 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1034 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1035 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_249 = torch.constant.int 12 | |
| %1036 = torch.aten.item %1034 : !torch.vtensor<[],f32> -> !torch.float | |
| %1037 = torch.aten.item %1035 : !torch.vtensor<[],si8> -> !torch.int | |
| %1038 = torch.aten.quantize_per_tensor %1033, %1036, %1037, %int12_249 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1039 = torch.aten.int_repr %1038 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %1040 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1041 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1042 = torch.aten.item %1040 : !torch.vtensor<[],f32> -> !torch.float | |
| %1043 = torch.aten.item %1041 : !torch.vtensor<[],si8> -> !torch.int | |
| %1044 = torch.aten._make_per_tensor_quantized_tensor %1039, %1042, %1043 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1045 = torch.aten.dequantize.self %1044 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1046 = torch.aten.mul.Tensor %1045, %922 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[1,120,28,28],f32> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1047 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1048 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_250 = torch.constant.int 12 | |
| %1049 = torch.aten.item %1047 : !torch.vtensor<[],f32> -> !torch.float | |
| %1050 = torch.aten.item %1048 : !torch.vtensor<[],si8> -> !torch.int | |
| %1051 = torch.aten.quantize_per_tensor %1046, %1049, %1050, %int12_250 : !torch.vtensor<[1,120,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1052 = torch.aten.int_repr %1051 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],si8> | |
| %1053 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1054 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1055 = torch.aten.item %1053 : !torch.vtensor<[],f32> -> !torch.float | |
| %1056 = torch.aten.item %1054 : !torch.vtensor<[],si8> -> !torch.int | |
| %1057 = torch.aten._make_per_tensor_quantized_tensor %1052, %1055, %1056 : !torch.vtensor<[1,120,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1058 = torch.aten.dequantize.self %1057 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1059 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1060 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_251 = torch.constant.int 12 | |
| %1061 = torch.aten.item %1059 : !torch.vtensor<[],f32> -> !torch.float | |
| %1062 = torch.aten.item %1060 : !torch.vtensor<[],si8> -> !torch.int | |
| %1063 = torch.aten.quantize_per_tensor %36, %1061, %1062, %int12_251 : !torch.vtensor<[40,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[40,120,1,1],!torch.qint8> | |
| %1064 = torch.aten.int_repr %1063 : !torch.vtensor<[40,120,1,1],!torch.qint8> -> !torch.vtensor<[40,120,1,1],si8> | |
| %1065 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1066 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1067 = torch.aten.item %1065 : !torch.vtensor<[],f32> -> !torch.float | |
| %1068 = torch.aten.item %1066 : !torch.vtensor<[],si8> -> !torch.int | |
| %1069 = torch.aten._make_per_tensor_quantized_tensor %1064, %1067, %1068 : !torch.vtensor<[40,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[40,120,1,1],!torch.qint8> | |
| %1070 = torch.aten.dequantize.self %1069 : !torch.vtensor<[40,120,1,1],!torch.qint8> -> !torch.vtensor<[40,120,1,1],f32> | |
| %1071 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1072 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_252 = torch.constant.int 12 | |
| %1073 = torch.aten.item %1071 : !torch.vtensor<[],f32> -> !torch.float | |
| %1074 = torch.aten.item %1072 : !torch.vtensor<[],si8> -> !torch.int | |
| %1075 = torch.aten.quantize_per_tensor %37, %1073, %1074, %int12_252 : !torch.vtensor<[40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[40],!torch.qint8> | |
| %1076 = torch.aten.int_repr %1075 : !torch.vtensor<[40],!torch.qint8> -> !torch.vtensor<[40],si8> | |
| %1077 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1078 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1079 = torch.aten.item %1077 : !torch.vtensor<[],f32> -> !torch.float | |
| %1080 = torch.aten.item %1078 : !torch.vtensor<[],si8> -> !torch.int | |
| %1081 = torch.aten._make_per_tensor_quantized_tensor %1076, %1079, %1080 : !torch.vtensor<[40],si8>, !torch.float, !torch.int -> !torch.vtensor<[40],!torch.qint8> | |
| %1082 = torch.aten.dequantize.self %1081 : !torch.vtensor<[40],!torch.qint8> -> !torch.vtensor<[40],f32> | |
| %int0_253 = torch.constant.int 0 | |
| %int0_254 = torch.constant.int 0 | |
| %int1_255 = torch.constant.int 1 | |
| %int1_256 = torch.constant.int 1 | |
| %int1_257 = torch.constant.int 1 | |
| %int1_258 = torch.constant.int 1 | |
| %int0_259 = torch.constant.int 0 | |
| %1083 = torch.prim.ListConstruct %int0_253, %int0_254 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1084 = torch.prim.ListConstruct %int1_255, %int1_256 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1085 = torch.prim.ListConstruct %int1_257, %int1_258 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1086 = torch.prim.ListConstruct %int0_259, %int0_259 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_260 = torch.constant.bool false | |
| %int1_261 = torch.constant.int 1 | |
| %1087 = torch.aten.convolution %1058, %1070, %1082, %1085, %1083, %1084, %false_260, %1086, %int1_261 : !torch.vtensor<[1,120,28,28],f32>, !torch.vtensor<[40,120,1,1],f32>, !torch.vtensor<[40],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,40,28,28],f32> | |
| %1088 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1089 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_262 = torch.constant.int 12 | |
| %1090 = torch.aten.item %1088 : !torch.vtensor<[],f32> -> !torch.float | |
| %1091 = torch.aten.item %1089 : !torch.vtensor<[],si8> -> !torch.int | |
| %1092 = torch.aten.quantize_per_tensor %1087, %1090, %1091, %int12_262 : !torch.vtensor<[1,40,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1093 = torch.aten.int_repr %1092 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],si8> | |
| %1094 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1095 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1096 = torch.aten.item %1094 : !torch.vtensor<[],f32> -> !torch.float | |
| %1097 = torch.aten.item %1095 : !torch.vtensor<[],si8> -> !torch.int | |
| %1098 = torch.aten._make_per_tensor_quantized_tensor %1093, %1096, %1097 : !torch.vtensor<[1,40,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1099 = torch.aten.dequantize.self %1098 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],f32> | |
| %int1_263 = torch.constant.int 1 | |
| %1100 = torch.aten.add.Tensor %1099, %838, %int1_263 : !torch.vtensor<[1,40,28,28],f32>, !torch.vtensor<[1,40,28,28],f32>, !torch.int -> !torch.vtensor<[1,40,28,28],f32> | |
| %1101 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1102 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_264 = torch.constant.int 12 | |
| %1103 = torch.aten.item %1101 : !torch.vtensor<[],f32> -> !torch.float | |
| %1104 = torch.aten.item %1102 : !torch.vtensor<[],si8> -> !torch.int | |
| %1105 = torch.aten.quantize_per_tensor %1100, %1103, %1104, %int12_264 : !torch.vtensor<[1,40,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1106 = torch.aten.int_repr %1105 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],si8> | |
| %1107 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1108 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1109 = torch.aten.item %1107 : !torch.vtensor<[],f32> -> !torch.float | |
| %1110 = torch.aten.item %1108 : !torch.vtensor<[],si8> -> !torch.int | |
| %1111 = torch.aten._make_per_tensor_quantized_tensor %1106, %1109, %1110 : !torch.vtensor<[1,40,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1112 = torch.aten.dequantize.self %1111 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],f32> | |
| %1113 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1114 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_265 = torch.constant.int 12 | |
| %1115 = torch.aten.item %1113 : !torch.vtensor<[],f32> -> !torch.float | |
| %1116 = torch.aten.item %1114 : !torch.vtensor<[],si8> -> !torch.int | |
| %1117 = torch.aten.quantize_per_tensor %38, %1115, %1116, %int12_265 : !torch.vtensor<[120,40,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,40,1,1],!torch.qint8> | |
| %1118 = torch.aten.int_repr %1117 : !torch.vtensor<[120,40,1,1],!torch.qint8> -> !torch.vtensor<[120,40,1,1],si8> | |
| %1119 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1120 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1121 = torch.aten.item %1119 : !torch.vtensor<[],f32> -> !torch.float | |
| %1122 = torch.aten.item %1120 : !torch.vtensor<[],si8> -> !torch.int | |
| %1123 = torch.aten._make_per_tensor_quantized_tensor %1118, %1121, %1122 : !torch.vtensor<[120,40,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,40,1,1],!torch.qint8> | |
| %1124 = torch.aten.dequantize.self %1123 : !torch.vtensor<[120,40,1,1],!torch.qint8> -> !torch.vtensor<[120,40,1,1],f32> | |
| %1125 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1126 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_266 = torch.constant.int 12 | |
| %1127 = torch.aten.item %1125 : !torch.vtensor<[],f32> -> !torch.float | |
| %1128 = torch.aten.item %1126 : !torch.vtensor<[],si8> -> !torch.int | |
| %1129 = torch.aten.quantize_per_tensor %39, %1127, %1128, %int12_266 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1130 = torch.aten.int_repr %1129 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %1131 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1132 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1133 = torch.aten.item %1131 : !torch.vtensor<[],f32> -> !torch.float | |
| %1134 = torch.aten.item %1132 : !torch.vtensor<[],si8> -> !torch.int | |
| %1135 = torch.aten._make_per_tensor_quantized_tensor %1130, %1133, %1134 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1136 = torch.aten.dequantize.self %1135 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int0_267 = torch.constant.int 0 | |
| %int0_268 = torch.constant.int 0 | |
| %int1_269 = torch.constant.int 1 | |
| %int1_270 = torch.constant.int 1 | |
| %int1_271 = torch.constant.int 1 | |
| %int1_272 = torch.constant.int 1 | |
| %int0_273 = torch.constant.int 0 | |
| %1137 = torch.prim.ListConstruct %int0_267, %int0_268 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1138 = torch.prim.ListConstruct %int1_269, %int1_270 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1139 = torch.prim.ListConstruct %int1_271, %int1_272 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1140 = torch.prim.ListConstruct %int0_273, %int0_273 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_274 = torch.constant.bool false | |
| %int1_275 = torch.constant.int 1 | |
| %1141 = torch.aten.convolution %1112, %1124, %1136, %1139, %1137, %1138, %false_274, %1140, %int1_275 : !torch.vtensor<[1,40,28,28],f32>, !torch.vtensor<[120,40,1,1],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,28,28],f32> | |
| %1142 = torch.aten.relu %1141 : !torch.vtensor<[1,120,28,28],f32> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1143 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1144 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_276 = torch.constant.int 12 | |
| %1145 = torch.aten.item %1143 : !torch.vtensor<[],f32> -> !torch.float | |
| %1146 = torch.aten.item %1144 : !torch.vtensor<[],si8> -> !torch.int | |
| %1147 = torch.aten.quantize_per_tensor %1142, %1145, %1146, %int12_276 : !torch.vtensor<[1,120,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1148 = torch.aten.int_repr %1147 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],si8> | |
| %1149 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1150 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1151 = torch.aten.item %1149 : !torch.vtensor<[],f32> -> !torch.float | |
| %1152 = torch.aten.item %1150 : !torch.vtensor<[],si8> -> !torch.int | |
| %1153 = torch.aten._make_per_tensor_quantized_tensor %1148, %1151, %1152 : !torch.vtensor<[1,120,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1154 = torch.aten.dequantize.self %1153 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1155 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1156 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_277 = torch.constant.int 12 | |
| %1157 = torch.aten.item %1155 : !torch.vtensor<[],f32> -> !torch.float | |
| %1158 = torch.aten.item %1156 : !torch.vtensor<[],si8> -> !torch.int | |
| %1159 = torch.aten.quantize_per_tensor %40, %1157, %1158, %int12_277 : !torch.vtensor<[120,1,5,5],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,1,5,5],!torch.qint8> | |
| %1160 = torch.aten.int_repr %1159 : !torch.vtensor<[120,1,5,5],!torch.qint8> -> !torch.vtensor<[120,1,5,5],si8> | |
| %1161 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1162 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1163 = torch.aten.item %1161 : !torch.vtensor<[],f32> -> !torch.float | |
| %1164 = torch.aten.item %1162 : !torch.vtensor<[],si8> -> !torch.int | |
| %1165 = torch.aten._make_per_tensor_quantized_tensor %1160, %1163, %1164 : !torch.vtensor<[120,1,5,5],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,1,5,5],!torch.qint8> | |
| %1166 = torch.aten.dequantize.self %1165 : !torch.vtensor<[120,1,5,5],!torch.qint8> -> !torch.vtensor<[120,1,5,5],f32> | |
| %1167 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1168 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_278 = torch.constant.int 12 | |
| %1169 = torch.aten.item %1167 : !torch.vtensor<[],f32> -> !torch.float | |
| %1170 = torch.aten.item %1168 : !torch.vtensor<[],si8> -> !torch.int | |
| %1171 = torch.aten.quantize_per_tensor %41, %1169, %1170, %int12_278 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1172 = torch.aten.int_repr %1171 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %1173 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1174 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1175 = torch.aten.item %1173 : !torch.vtensor<[],f32> -> !torch.float | |
| %1176 = torch.aten.item %1174 : !torch.vtensor<[],si8> -> !torch.int | |
| %1177 = torch.aten._make_per_tensor_quantized_tensor %1172, %1175, %1176 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1178 = torch.aten.dequantize.self %1177 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int2_279 = torch.constant.int 2 | |
| %int2_280 = torch.constant.int 2 | |
| %int1_281 = torch.constant.int 1 | |
| %int1_282 = torch.constant.int 1 | |
| %int1_283 = torch.constant.int 1 | |
| %int1_284 = torch.constant.int 1 | |
| %int0_285 = torch.constant.int 0 | |
| %1179 = torch.prim.ListConstruct %int2_279, %int2_280 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1180 = torch.prim.ListConstruct %int1_281, %int1_282 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1181 = torch.prim.ListConstruct %int1_283, %int1_284 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1182 = torch.prim.ListConstruct %int0_285, %int0_285 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_286 = torch.constant.bool false | |
| %int120_287 = torch.constant.int 120 | |
| %1183 = torch.aten.convolution %1154, %1166, %1178, %1181, %1179, %1180, %false_286, %1182, %int120_287 : !torch.vtensor<[1,120,28,28],f32>, !torch.vtensor<[120,1,5,5],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,28,28],f32> | |
| %1184 = torch.aten.relu %1183 : !torch.vtensor<[1,120,28,28],f32> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1185 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1186 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_288 = torch.constant.int 12 | |
| %1187 = torch.aten.item %1185 : !torch.vtensor<[],f32> -> !torch.float | |
| %1188 = torch.aten.item %1186 : !torch.vtensor<[],si8> -> !torch.int | |
| %1189 = torch.aten.quantize_per_tensor %1184, %1187, %1188, %int12_288 : !torch.vtensor<[1,120,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1190 = torch.aten.int_repr %1189 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],si8> | |
| %1191 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1192 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1193 = torch.aten.item %1191 : !torch.vtensor<[],f32> -> !torch.float | |
| %1194 = torch.aten.item %1192 : !torch.vtensor<[],si8> -> !torch.int | |
| %1195 = torch.aten._make_per_tensor_quantized_tensor %1190, %1193, %1194 : !torch.vtensor<[1,120,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1196 = torch.aten.dequantize.self %1195 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],f32> | |
| %int0_289 = torch.constant.int 0 | |
| %int1_290 = torch.constant.int 1 | |
| %int28_291 = torch.constant.int 28 | |
| %int28_292 = torch.constant.int 28 | |
| %1197 = torch.prim.ListConstruct %int28_291, %int28_292 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1198 = torch.prim.ListConstruct %int0_289, %int0_289 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1199 = torch.prim.ListConstruct %int1_290, %int1_290 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_293 = torch.constant.bool false | |
| %none_294 = torch.constant.none | |
| %1200 = torch.aten.avg_pool2d %1196, %1197, %1199, %1198, %false_293, %false_293, %none_294 : !torch.vtensor<[1,120,28,28],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,120,1,1],f32> | |
| %1201 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1202 = torch.aten.mul.Tensor %1200, %1201 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1203 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1204 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_295 = torch.constant.int 12 | |
| %1205 = torch.aten.item %1203 : !torch.vtensor<[],f32> -> !torch.float | |
| %1206 = torch.aten.item %1204 : !torch.vtensor<[],si8> -> !torch.int | |
| %1207 = torch.aten.quantize_per_tensor %1202, %1205, %1206, %int12_295 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1208 = torch.aten.int_repr %1207 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %1209 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1210 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1211 = torch.aten.item %1209 : !torch.vtensor<[],f32> -> !torch.float | |
| %1212 = torch.aten.item %1210 : !torch.vtensor<[],si8> -> !torch.int | |
| %1213 = torch.aten._make_per_tensor_quantized_tensor %1208, %1211, %1212 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1214 = torch.aten.dequantize.self %1213 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1215 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1216 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_296 = torch.constant.int 12 | |
| %1217 = torch.aten.item %1215 : !torch.vtensor<[],f32> -> !torch.float | |
| %1218 = torch.aten.item %1216 : !torch.vtensor<[],si8> -> !torch.int | |
| %1219 = torch.aten.quantize_per_tensor %42, %1217, %1218, %int12_296 : !torch.vtensor<[32,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32,120,1,1],!torch.qint8> | |
| %1220 = torch.aten.int_repr %1219 : !torch.vtensor<[32,120,1,1],!torch.qint8> -> !torch.vtensor<[32,120,1,1],si8> | |
| %1221 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1222 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1223 = torch.aten.item %1221 : !torch.vtensor<[],f32> -> !torch.float | |
| %1224 = torch.aten.item %1222 : !torch.vtensor<[],si8> -> !torch.int | |
| %1225 = torch.aten._make_per_tensor_quantized_tensor %1220, %1223, %1224 : !torch.vtensor<[32,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[32,120,1,1],!torch.qint8> | |
| %1226 = torch.aten.dequantize.self %1225 : !torch.vtensor<[32,120,1,1],!torch.qint8> -> !torch.vtensor<[32,120,1,1],f32> | |
| %1227 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1228 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_297 = torch.constant.int 12 | |
| %1229 = torch.aten.item %1227 : !torch.vtensor<[],f32> -> !torch.float | |
| %1230 = torch.aten.item %1228 : !torch.vtensor<[],si8> -> !torch.int | |
| %1231 = torch.aten.quantize_per_tensor %43, %1229, %1230, %int12_297 : !torch.vtensor<[32],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %1232 = torch.aten.int_repr %1231 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],si8> | |
| %1233 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1234 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1235 = torch.aten.item %1233 : !torch.vtensor<[],f32> -> !torch.float | |
| %1236 = torch.aten.item %1234 : !torch.vtensor<[],si8> -> !torch.int | |
| %1237 = torch.aten._make_per_tensor_quantized_tensor %1232, %1235, %1236 : !torch.vtensor<[32],si8>, !torch.float, !torch.int -> !torch.vtensor<[32],!torch.qint8> | |
| %1238 = torch.aten.dequantize.self %1237 : !torch.vtensor<[32],!torch.qint8> -> !torch.vtensor<[32],f32> | |
| %int0_298 = torch.constant.int 0 | |
| %int0_299 = torch.constant.int 0 | |
| %int1_300 = torch.constant.int 1 | |
| %int1_301 = torch.constant.int 1 | |
| %int1_302 = torch.constant.int 1 | |
| %int1_303 = torch.constant.int 1 | |
| %int0_304 = torch.constant.int 0 | |
| %1239 = torch.prim.ListConstruct %int0_298, %int0_299 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1240 = torch.prim.ListConstruct %int1_300, %int1_301 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1241 = torch.prim.ListConstruct %int1_302, %int1_303 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1242 = torch.prim.ListConstruct %int0_304, %int0_304 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_305 = torch.constant.bool false | |
| %int1_306 = torch.constant.int 1 | |
| %1243 = torch.aten.convolution %1214, %1226, %1238, %1241, %1239, %1240, %false_305, %1242, %int1_306 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[32,120,1,1],f32>, !torch.vtensor<[32],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,32,1,1],f32> | |
| %1244 = torch.aten.relu %1243 : !torch.vtensor<[1,32,1,1],f32> -> !torch.vtensor<[1,32,1,1],f32> | |
| %1245 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1246 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_307 = torch.constant.int 12 | |
| %1247 = torch.aten.item %1245 : !torch.vtensor<[],f32> -> !torch.float | |
| %1248 = torch.aten.item %1246 : !torch.vtensor<[],si8> -> !torch.int | |
| %1249 = torch.aten.quantize_per_tensor %1244, %1247, %1248, %int12_307 : !torch.vtensor<[1,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,32,1,1],!torch.qint8> | |
| %1250 = torch.aten.int_repr %1249 : !torch.vtensor<[1,32,1,1],!torch.qint8> -> !torch.vtensor<[1,32,1,1],si8> | |
| %1251 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1252 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1253 = torch.aten.item %1251 : !torch.vtensor<[],f32> -> !torch.float | |
| %1254 = torch.aten.item %1252 : !torch.vtensor<[],si8> -> !torch.int | |
| %1255 = torch.aten._make_per_tensor_quantized_tensor %1250, %1253, %1254 : !torch.vtensor<[1,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,32,1,1],!torch.qint8> | |
| %1256 = torch.aten.dequantize.self %1255 : !torch.vtensor<[1,32,1,1],!torch.qint8> -> !torch.vtensor<[1,32,1,1],f32> | |
| %1257 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1258 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_308 = torch.constant.int 12 | |
| %1259 = torch.aten.item %1257 : !torch.vtensor<[],f32> -> !torch.float | |
| %1260 = torch.aten.item %1258 : !torch.vtensor<[],si8> -> !torch.int | |
| %1261 = torch.aten.quantize_per_tensor %44, %1259, %1260, %int12_308 : !torch.vtensor<[120,32,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,32,1,1],!torch.qint8> | |
| %1262 = torch.aten.int_repr %1261 : !torch.vtensor<[120,32,1,1],!torch.qint8> -> !torch.vtensor<[120,32,1,1],si8> | |
| %1263 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1264 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1265 = torch.aten.item %1263 : !torch.vtensor<[],f32> -> !torch.float | |
| %1266 = torch.aten.item %1264 : !torch.vtensor<[],si8> -> !torch.int | |
| %1267 = torch.aten._make_per_tensor_quantized_tensor %1262, %1265, %1266 : !torch.vtensor<[120,32,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,32,1,1],!torch.qint8> | |
| %1268 = torch.aten.dequantize.self %1267 : !torch.vtensor<[120,32,1,1],!torch.qint8> -> !torch.vtensor<[120,32,1,1],f32> | |
| %1269 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1270 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_309 = torch.constant.int 12 | |
| %1271 = torch.aten.item %1269 : !torch.vtensor<[],f32> -> !torch.float | |
| %1272 = torch.aten.item %1270 : !torch.vtensor<[],si8> -> !torch.int | |
| %1273 = torch.aten.quantize_per_tensor %45, %1271, %1272, %int12_309 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1274 = torch.aten.int_repr %1273 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %1275 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1276 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1277 = torch.aten.item %1275 : !torch.vtensor<[],f32> -> !torch.float | |
| %1278 = torch.aten.item %1276 : !torch.vtensor<[],si8> -> !torch.int | |
| %1279 = torch.aten._make_per_tensor_quantized_tensor %1274, %1277, %1278 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %1280 = torch.aten.dequantize.self %1279 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int0_310 = torch.constant.int 0 | |
| %int0_311 = torch.constant.int 0 | |
| %int1_312 = torch.constant.int 1 | |
| %int1_313 = torch.constant.int 1 | |
| %int1_314 = torch.constant.int 1 | |
| %int1_315 = torch.constant.int 1 | |
| %int0_316 = torch.constant.int 0 | |
| %1281 = torch.prim.ListConstruct %int0_310, %int0_311 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1282 = torch.prim.ListConstruct %int1_312, %int1_313 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1283 = torch.prim.ListConstruct %int1_314, %int1_315 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1284 = torch.prim.ListConstruct %int0_316, %int0_316 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_317 = torch.constant.bool false | |
| %int1_318 = torch.constant.int 1 | |
| %1285 = torch.aten.convolution %1256, %1268, %1280, %1283, %1281, %1282, %false_317, %1284, %int1_318 : !torch.vtensor<[1,32,1,1],f32>, !torch.vtensor<[120,32,1,1],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,1,1],f32> | |
| %1286 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1287 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_319 = torch.constant.int 12 | |
| %1288 = torch.aten.item %1286 : !torch.vtensor<[],f32> -> !torch.float | |
| %1289 = torch.aten.item %1287 : !torch.vtensor<[],si8> -> !torch.int | |
| %1290 = torch.aten.quantize_per_tensor %1285, %1288, %1289, %int12_319 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1291 = torch.aten.int_repr %1290 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %1292 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1293 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1294 = torch.aten.item %1292 : !torch.vtensor<[],f32> -> !torch.float | |
| %1295 = torch.aten.item %1293 : !torch.vtensor<[],si8> -> !torch.int | |
| %1296 = torch.aten._make_per_tensor_quantized_tensor %1291, %1294, %1295 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1297 = torch.aten.dequantize.self %1296 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1298 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_320 = torch.constant.int 1 | |
| %1299 = torch.aten.add.Tensor %1297, %1298, %int1_320 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,120,1,1],f32> | |
| %1300 = torch.aten.relu %1299 : !torch.vtensor<[1,120,1,1],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1301 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1302 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_321 = torch.constant.int 6 | |
| %none_322 = torch.constant.none | |
| %false_323 = torch.constant.bool false | |
| %1303 = torch.aten.to.dtype %1301, %int6_321, %false_323, %false_323, %none_322 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_324 = torch.constant.int 6 | |
| %none_325 = torch.constant.none | |
| %false_326 = torch.constant.bool false | |
| %1304 = torch.aten.to.dtype %1302, %int6_324, %false_326, %false_326, %none_325 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1305 = torch.aten.clamp.Tensor %1300, %1303, %1304 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1306 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1307 = torch.aten.mul.Tensor %1305, %1306 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1308 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1309 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_327 = torch.constant.int 12 | |
| %1310 = torch.aten.item %1308 : !torch.vtensor<[],f32> -> !torch.float | |
| %1311 = torch.aten.item %1309 : !torch.vtensor<[],si8> -> !torch.int | |
| %1312 = torch.aten.quantize_per_tensor %1307, %1310, %1311, %int12_327 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1313 = torch.aten.int_repr %1312 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %1314 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1315 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1316 = torch.aten.item %1314 : !torch.vtensor<[],f32> -> !torch.float | |
| %1317 = torch.aten.item %1315 : !torch.vtensor<[],si8> -> !torch.int | |
| %1318 = torch.aten._make_per_tensor_quantized_tensor %1313, %1316, %1317 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %1319 = torch.aten.dequantize.self %1318 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %1320 = torch.aten.mul.Tensor %1319, %1196 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[1,120,28,28],f32> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1321 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1322 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_328 = torch.constant.int 12 | |
| %1323 = torch.aten.item %1321 : !torch.vtensor<[],f32> -> !torch.float | |
| %1324 = torch.aten.item %1322 : !torch.vtensor<[],si8> -> !torch.int | |
| %1325 = torch.aten.quantize_per_tensor %1320, %1323, %1324, %int12_328 : !torch.vtensor<[1,120,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1326 = torch.aten.int_repr %1325 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],si8> | |
| %1327 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1328 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1329 = torch.aten.item %1327 : !torch.vtensor<[],f32> -> !torch.float | |
| %1330 = torch.aten.item %1328 : !torch.vtensor<[],si8> -> !torch.int | |
| %1331 = torch.aten._make_per_tensor_quantized_tensor %1326, %1329, %1330 : !torch.vtensor<[1,120,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,28,28],!torch.qint8> | |
| %1332 = torch.aten.dequantize.self %1331 : !torch.vtensor<[1,120,28,28],!torch.qint8> -> !torch.vtensor<[1,120,28,28],f32> | |
| %1333 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1334 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_329 = torch.constant.int 12 | |
| %1335 = torch.aten.item %1333 : !torch.vtensor<[],f32> -> !torch.float | |
| %1336 = torch.aten.item %1334 : !torch.vtensor<[],si8> -> !torch.int | |
| %1337 = torch.aten.quantize_per_tensor %46, %1335, %1336, %int12_329 : !torch.vtensor<[40,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[40,120,1,1],!torch.qint8> | |
| %1338 = torch.aten.int_repr %1337 : !torch.vtensor<[40,120,1,1],!torch.qint8> -> !torch.vtensor<[40,120,1,1],si8> | |
| %1339 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1340 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1341 = torch.aten.item %1339 : !torch.vtensor<[],f32> -> !torch.float | |
| %1342 = torch.aten.item %1340 : !torch.vtensor<[],si8> -> !torch.int | |
| %1343 = torch.aten._make_per_tensor_quantized_tensor %1338, %1341, %1342 : !torch.vtensor<[40,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[40,120,1,1],!torch.qint8> | |
| %1344 = torch.aten.dequantize.self %1343 : !torch.vtensor<[40,120,1,1],!torch.qint8> -> !torch.vtensor<[40,120,1,1],f32> | |
| %1345 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1346 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_330 = torch.constant.int 12 | |
| %1347 = torch.aten.item %1345 : !torch.vtensor<[],f32> -> !torch.float | |
| %1348 = torch.aten.item %1346 : !torch.vtensor<[],si8> -> !torch.int | |
| %1349 = torch.aten.quantize_per_tensor %47, %1347, %1348, %int12_330 : !torch.vtensor<[40],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[40],!torch.qint8> | |
| %1350 = torch.aten.int_repr %1349 : !torch.vtensor<[40],!torch.qint8> -> !torch.vtensor<[40],si8> | |
| %1351 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1352 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1353 = torch.aten.item %1351 : !torch.vtensor<[],f32> -> !torch.float | |
| %1354 = torch.aten.item %1352 : !torch.vtensor<[],si8> -> !torch.int | |
| %1355 = torch.aten._make_per_tensor_quantized_tensor %1350, %1353, %1354 : !torch.vtensor<[40],si8>, !torch.float, !torch.int -> !torch.vtensor<[40],!torch.qint8> | |
| %1356 = torch.aten.dequantize.self %1355 : !torch.vtensor<[40],!torch.qint8> -> !torch.vtensor<[40],f32> | |
| %int0_331 = torch.constant.int 0 | |
| %int0_332 = torch.constant.int 0 | |
| %int1_333 = torch.constant.int 1 | |
| %int1_334 = torch.constant.int 1 | |
| %int1_335 = torch.constant.int 1 | |
| %int1_336 = torch.constant.int 1 | |
| %int0_337 = torch.constant.int 0 | |
| %1357 = torch.prim.ListConstruct %int0_331, %int0_332 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1358 = torch.prim.ListConstruct %int1_333, %int1_334 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1359 = torch.prim.ListConstruct %int1_335, %int1_336 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1360 = torch.prim.ListConstruct %int0_337, %int0_337 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_338 = torch.constant.bool false | |
| %int1_339 = torch.constant.int 1 | |
| %1361 = torch.aten.convolution %1332, %1344, %1356, %1359, %1357, %1358, %false_338, %1360, %int1_339 : !torch.vtensor<[1,120,28,28],f32>, !torch.vtensor<[40,120,1,1],f32>, !torch.vtensor<[40],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,40,28,28],f32> | |
| %1362 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1363 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_340 = torch.constant.int 12 | |
| %1364 = torch.aten.item %1362 : !torch.vtensor<[],f32> -> !torch.float | |
| %1365 = torch.aten.item %1363 : !torch.vtensor<[],si8> -> !torch.int | |
| %1366 = torch.aten.quantize_per_tensor %1361, %1364, %1365, %int12_340 : !torch.vtensor<[1,40,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1367 = torch.aten.int_repr %1366 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],si8> | |
| %1368 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1369 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1370 = torch.aten.item %1368 : !torch.vtensor<[],f32> -> !torch.float | |
| %1371 = torch.aten.item %1369 : !torch.vtensor<[],si8> -> !torch.int | |
| %1372 = torch.aten._make_per_tensor_quantized_tensor %1367, %1370, %1371 : !torch.vtensor<[1,40,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1373 = torch.aten.dequantize.self %1372 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],f32> | |
| %int1_341 = torch.constant.int 1 | |
| %1374 = torch.aten.add.Tensor %1373, %1112, %int1_341 : !torch.vtensor<[1,40,28,28],f32>, !torch.vtensor<[1,40,28,28],f32>, !torch.int -> !torch.vtensor<[1,40,28,28],f32> | |
| %1375 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1376 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_342 = torch.constant.int 12 | |
| %1377 = torch.aten.item %1375 : !torch.vtensor<[],f32> -> !torch.float | |
| %1378 = torch.aten.item %1376 : !torch.vtensor<[],si8> -> !torch.int | |
| %1379 = torch.aten.quantize_per_tensor %1374, %1377, %1378, %int12_342 : !torch.vtensor<[1,40,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1380 = torch.aten.int_repr %1379 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],si8> | |
| %1381 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1382 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1383 = torch.aten.item %1381 : !torch.vtensor<[],f32> -> !torch.float | |
| %1384 = torch.aten.item %1382 : !torch.vtensor<[],si8> -> !torch.int | |
| %1385 = torch.aten._make_per_tensor_quantized_tensor %1380, %1383, %1384 : !torch.vtensor<[1,40,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,40,28,28],!torch.qint8> | |
| %1386 = torch.aten.dequantize.self %1385 : !torch.vtensor<[1,40,28,28],!torch.qint8> -> !torch.vtensor<[1,40,28,28],f32> | |
| %1387 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1388 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_343 = torch.constant.int 12 | |
| %1389 = torch.aten.item %1387 : !torch.vtensor<[],f32> -> !torch.float | |
| %1390 = torch.aten.item %1388 : !torch.vtensor<[],si8> -> !torch.int | |
| %1391 = torch.aten.quantize_per_tensor %48, %1389, %1390, %int12_343 : !torch.vtensor<[240,40,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240,40,1,1],!torch.qint8> | |
| %1392 = torch.aten.int_repr %1391 : !torch.vtensor<[240,40,1,1],!torch.qint8> -> !torch.vtensor<[240,40,1,1],si8> | |
| %1393 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1394 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1395 = torch.aten.item %1393 : !torch.vtensor<[],f32> -> !torch.float | |
| %1396 = torch.aten.item %1394 : !torch.vtensor<[],si8> -> !torch.int | |
| %1397 = torch.aten._make_per_tensor_quantized_tensor %1392, %1395, %1396 : !torch.vtensor<[240,40,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[240,40,1,1],!torch.qint8> | |
| %1398 = torch.aten.dequantize.self %1397 : !torch.vtensor<[240,40,1,1],!torch.qint8> -> !torch.vtensor<[240,40,1,1],f32> | |
| %1399 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1400 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_344 = torch.constant.int 12 | |
| %1401 = torch.aten.item %1399 : !torch.vtensor<[],f32> -> !torch.float | |
| %1402 = torch.aten.item %1400 : !torch.vtensor<[],si8> -> !torch.int | |
| %1403 = torch.aten.quantize_per_tensor %49, %1401, %1402, %int12_344 : !torch.vtensor<[240],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %1404 = torch.aten.int_repr %1403 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],si8> | |
| %1405 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1406 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1407 = torch.aten.item %1405 : !torch.vtensor<[],f32> -> !torch.float | |
| %1408 = torch.aten.item %1406 : !torch.vtensor<[],si8> -> !torch.int | |
| %1409 = torch.aten._make_per_tensor_quantized_tensor %1404, %1407, %1408 : !torch.vtensor<[240],si8>, !torch.float, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %1410 = torch.aten.dequantize.self %1409 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],f32> | |
| %int0_345 = torch.constant.int 0 | |
| %int0_346 = torch.constant.int 0 | |
| %int1_347 = torch.constant.int 1 | |
| %int1_348 = torch.constant.int 1 | |
| %int1_349 = torch.constant.int 1 | |
| %int1_350 = torch.constant.int 1 | |
| %int0_351 = torch.constant.int 0 | |
| %1411 = torch.prim.ListConstruct %int0_345, %int0_346 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1412 = torch.prim.ListConstruct %int1_347, %int1_348 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1413 = torch.prim.ListConstruct %int1_349, %int1_350 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1414 = torch.prim.ListConstruct %int0_351, %int0_351 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_352 = torch.constant.bool false | |
| %int1_353 = torch.constant.int 1 | |
| %1415 = torch.aten.convolution %1386, %1398, %1410, %1413, %1411, %1412, %false_352, %1414, %int1_353 : !torch.vtensor<[1,40,28,28],f32>, !torch.vtensor<[240,40,1,1],f32>, !torch.vtensor<[240],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,240,28,28],f32> | |
| %1416 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1417 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_354 = torch.constant.int 12 | |
| %1418 = torch.aten.item %1416 : !torch.vtensor<[],f32> -> !torch.float | |
| %1419 = torch.aten.item %1417 : !torch.vtensor<[],si8> -> !torch.int | |
| %1420 = torch.aten.quantize_per_tensor %1415, %1418, %1419, %int12_354 : !torch.vtensor<[1,240,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,28,28],!torch.qint8> | |
| %1421 = torch.aten.int_repr %1420 : !torch.vtensor<[1,240,28,28],!torch.qint8> -> !torch.vtensor<[1,240,28,28],si8> | |
| %1422 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1423 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1424 = torch.aten.item %1422 : !torch.vtensor<[],f32> -> !torch.float | |
| %1425 = torch.aten.item %1423 : !torch.vtensor<[],si8> -> !torch.int | |
| %1426 = torch.aten._make_per_tensor_quantized_tensor %1421, %1424, %1425 : !torch.vtensor<[1,240,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,28,28],!torch.qint8> | |
| %1427 = torch.aten.dequantize.self %1426 : !torch.vtensor<[1,240,28,28],!torch.qint8> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1428 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_355 = torch.constant.int 1 | |
| %1429 = torch.aten.add.Tensor %1427, %1428, %int1_355 : !torch.vtensor<[1,240,28,28],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,240,28,28],f32> | |
| %1430 = torch.aten.relu %1429 : !torch.vtensor<[1,240,28,28],f32> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1431 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1432 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_356 = torch.constant.int 6 | |
| %none_357 = torch.constant.none | |
| %false_358 = torch.constant.bool false | |
| %1433 = torch.aten.to.dtype %1431, %int6_356, %false_358, %false_358, %none_357 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_359 = torch.constant.int 6 | |
| %none_360 = torch.constant.none | |
| %false_361 = torch.constant.bool false | |
| %1434 = torch.aten.to.dtype %1432, %int6_359, %false_361, %false_361, %none_360 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1435 = torch.aten.clamp.Tensor %1430, %1433, %1434 : !torch.vtensor<[1,240,28,28],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1436 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1437 = torch.aten.mul.Tensor %1435, %1436 : !torch.vtensor<[1,240,28,28],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1438 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1439 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_362 = torch.constant.int 12 | |
| %1440 = torch.aten.item %1438 : !torch.vtensor<[],f32> -> !torch.float | |
| %1441 = torch.aten.item %1439 : !torch.vtensor<[],si8> -> !torch.int | |
| %1442 = torch.aten.quantize_per_tensor %1437, %1440, %1441, %int12_362 : !torch.vtensor<[1,240,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,28,28],!torch.qint8> | |
| %1443 = torch.aten.int_repr %1442 : !torch.vtensor<[1,240,28,28],!torch.qint8> -> !torch.vtensor<[1,240,28,28],si8> | |
| %1444 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1445 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1446 = torch.aten.item %1444 : !torch.vtensor<[],f32> -> !torch.float | |
| %1447 = torch.aten.item %1445 : !torch.vtensor<[],si8> -> !torch.int | |
| %1448 = torch.aten._make_per_tensor_quantized_tensor %1443, %1446, %1447 : !torch.vtensor<[1,240,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,28,28],!torch.qint8> | |
| %1449 = torch.aten.dequantize.self %1448 : !torch.vtensor<[1,240,28,28],!torch.qint8> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1450 = torch.aten.mul.Tensor %1427, %1449 : !torch.vtensor<[1,240,28,28],f32>, !torch.vtensor<[1,240,28,28],f32> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1451 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1452 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_363 = torch.constant.int 12 | |
| %1453 = torch.aten.item %1451 : !torch.vtensor<[],f32> -> !torch.float | |
| %1454 = torch.aten.item %1452 : !torch.vtensor<[],si8> -> !torch.int | |
| %1455 = torch.aten.quantize_per_tensor %1450, %1453, %1454, %int12_363 : !torch.vtensor<[1,240,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,28,28],!torch.qint8> | |
| %1456 = torch.aten.int_repr %1455 : !torch.vtensor<[1,240,28,28],!torch.qint8> -> !torch.vtensor<[1,240,28,28],si8> | |
| %1457 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1458 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1459 = torch.aten.item %1457 : !torch.vtensor<[],f32> -> !torch.float | |
| %1460 = torch.aten.item %1458 : !torch.vtensor<[],si8> -> !torch.int | |
| %1461 = torch.aten._make_per_tensor_quantized_tensor %1456, %1459, %1460 : !torch.vtensor<[1,240,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,28,28],!torch.qint8> | |
| %1462 = torch.aten.dequantize.self %1461 : !torch.vtensor<[1,240,28,28],!torch.qint8> -> !torch.vtensor<[1,240,28,28],f32> | |
| %1463 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1464 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_364 = torch.constant.int 12 | |
| %1465 = torch.aten.item %1463 : !torch.vtensor<[],f32> -> !torch.float | |
| %1466 = torch.aten.item %1464 : !torch.vtensor<[],si8> -> !torch.int | |
| %1467 = torch.aten.quantize_per_tensor %50, %1465, %1466, %int12_364 : !torch.vtensor<[240,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240,1,3,3],!torch.qint8> | |
| %1468 = torch.aten.int_repr %1467 : !torch.vtensor<[240,1,3,3],!torch.qint8> -> !torch.vtensor<[240,1,3,3],si8> | |
| %1469 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1470 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1471 = torch.aten.item %1469 : !torch.vtensor<[],f32> -> !torch.float | |
| %1472 = torch.aten.item %1470 : !torch.vtensor<[],si8> -> !torch.int | |
| %1473 = torch.aten._make_per_tensor_quantized_tensor %1468, %1471, %1472 : !torch.vtensor<[240,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[240,1,3,3],!torch.qint8> | |
| %1474 = torch.aten.dequantize.self %1473 : !torch.vtensor<[240,1,3,3],!torch.qint8> -> !torch.vtensor<[240,1,3,3],f32> | |
| %1475 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1476 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_365 = torch.constant.int 12 | |
| %1477 = torch.aten.item %1475 : !torch.vtensor<[],f32> -> !torch.float | |
| %1478 = torch.aten.item %1476 : !torch.vtensor<[],si8> -> !torch.int | |
| %1479 = torch.aten.quantize_per_tensor %51, %1477, %1478, %int12_365 : !torch.vtensor<[240],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %1480 = torch.aten.int_repr %1479 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],si8> | |
| %1481 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1482 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1483 = torch.aten.item %1481 : !torch.vtensor<[],f32> -> !torch.float | |
| %1484 = torch.aten.item %1482 : !torch.vtensor<[],si8> -> !torch.int | |
| %1485 = torch.aten._make_per_tensor_quantized_tensor %1480, %1483, %1484 : !torch.vtensor<[240],si8>, !torch.float, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %1486 = torch.aten.dequantize.self %1485 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],f32> | |
| %int1_366 = torch.constant.int 1 | |
| %int1_367 = torch.constant.int 1 | |
| %int1_368 = torch.constant.int 1 | |
| %int1_369 = torch.constant.int 1 | |
| %int2_370 = torch.constant.int 2 | |
| %int2_371 = torch.constant.int 2 | |
| %int0_372 = torch.constant.int 0 | |
| %1487 = torch.prim.ListConstruct %int1_366, %int1_367 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1488 = torch.prim.ListConstruct %int1_368, %int1_369 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1489 = torch.prim.ListConstruct %int2_370, %int2_371 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1490 = torch.prim.ListConstruct %int0_372, %int0_372 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_373 = torch.constant.bool false | |
| %int240 = torch.constant.int 240 | |
| %1491 = torch.aten.convolution %1462, %1474, %1486, %1489, %1487, %1488, %false_373, %1490, %int240 : !torch.vtensor<[1,240,28,28],f32>, !torch.vtensor<[240,1,3,3],f32>, !torch.vtensor<[240],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,240,14,14],f32> | |
| %1492 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1493 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_374 = torch.constant.int 12 | |
| %1494 = torch.aten.item %1492 : !torch.vtensor<[],f32> -> !torch.float | |
| %1495 = torch.aten.item %1493 : !torch.vtensor<[],si8> -> !torch.int | |
| %1496 = torch.aten.quantize_per_tensor %1491, %1494, %1495, %int12_374 : !torch.vtensor<[1,240,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,14,14],!torch.qint8> | |
| %1497 = torch.aten.int_repr %1496 : !torch.vtensor<[1,240,14,14],!torch.qint8> -> !torch.vtensor<[1,240,14,14],si8> | |
| %1498 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1499 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1500 = torch.aten.item %1498 : !torch.vtensor<[],f32> -> !torch.float | |
| %1501 = torch.aten.item %1499 : !torch.vtensor<[],si8> -> !torch.int | |
| %1502 = torch.aten._make_per_tensor_quantized_tensor %1497, %1500, %1501 : !torch.vtensor<[1,240,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,14,14],!torch.qint8> | |
| %1503 = torch.aten.dequantize.self %1502 : !torch.vtensor<[1,240,14,14],!torch.qint8> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1504 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_375 = torch.constant.int 1 | |
| %1505 = torch.aten.add.Tensor %1503, %1504, %int1_375 : !torch.vtensor<[1,240,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,240,14,14],f32> | |
| %1506 = torch.aten.relu %1505 : !torch.vtensor<[1,240,14,14],f32> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1507 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1508 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_376 = torch.constant.int 6 | |
| %none_377 = torch.constant.none | |
| %false_378 = torch.constant.bool false | |
| %1509 = torch.aten.to.dtype %1507, %int6_376, %false_378, %false_378, %none_377 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_379 = torch.constant.int 6 | |
| %none_380 = torch.constant.none | |
| %false_381 = torch.constant.bool false | |
| %1510 = torch.aten.to.dtype %1508, %int6_379, %false_381, %false_381, %none_380 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1511 = torch.aten.clamp.Tensor %1506, %1509, %1510 : !torch.vtensor<[1,240,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1512 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1513 = torch.aten.mul.Tensor %1511, %1512 : !torch.vtensor<[1,240,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1514 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1515 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_382 = torch.constant.int 12 | |
| %1516 = torch.aten.item %1514 : !torch.vtensor<[],f32> -> !torch.float | |
| %1517 = torch.aten.item %1515 : !torch.vtensor<[],si8> -> !torch.int | |
| %1518 = torch.aten.quantize_per_tensor %1513, %1516, %1517, %int12_382 : !torch.vtensor<[1,240,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,14,14],!torch.qint8> | |
| %1519 = torch.aten.int_repr %1518 : !torch.vtensor<[1,240,14,14],!torch.qint8> -> !torch.vtensor<[1,240,14,14],si8> | |
| %1520 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1521 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1522 = torch.aten.item %1520 : !torch.vtensor<[],f32> -> !torch.float | |
| %1523 = torch.aten.item %1521 : !torch.vtensor<[],si8> -> !torch.int | |
| %1524 = torch.aten._make_per_tensor_quantized_tensor %1519, %1522, %1523 : !torch.vtensor<[1,240,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,14,14],!torch.qint8> | |
| %1525 = torch.aten.dequantize.self %1524 : !torch.vtensor<[1,240,14,14],!torch.qint8> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1526 = torch.aten.mul.Tensor %1503, %1525 : !torch.vtensor<[1,240,14,14],f32>, !torch.vtensor<[1,240,14,14],f32> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1527 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1528 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_383 = torch.constant.int 12 | |
| %1529 = torch.aten.item %1527 : !torch.vtensor<[],f32> -> !torch.float | |
| %1530 = torch.aten.item %1528 : !torch.vtensor<[],si8> -> !torch.int | |
| %1531 = torch.aten.quantize_per_tensor %1526, %1529, %1530, %int12_383 : !torch.vtensor<[1,240,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,14,14],!torch.qint8> | |
| %1532 = torch.aten.int_repr %1531 : !torch.vtensor<[1,240,14,14],!torch.qint8> -> !torch.vtensor<[1,240,14,14],si8> | |
| %1533 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1534 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1535 = torch.aten.item %1533 : !torch.vtensor<[],f32> -> !torch.float | |
| %1536 = torch.aten.item %1534 : !torch.vtensor<[],si8> -> !torch.int | |
| %1537 = torch.aten._make_per_tensor_quantized_tensor %1532, %1535, %1536 : !torch.vtensor<[1,240,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,14,14],!torch.qint8> | |
| %1538 = torch.aten.dequantize.self %1537 : !torch.vtensor<[1,240,14,14],!torch.qint8> -> !torch.vtensor<[1,240,14,14],f32> | |
| %1539 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1540 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_384 = torch.constant.int 12 | |
| %1541 = torch.aten.item %1539 : !torch.vtensor<[],f32> -> !torch.float | |
| %1542 = torch.aten.item %1540 : !torch.vtensor<[],si8> -> !torch.int | |
| %1543 = torch.aten.quantize_per_tensor %52, %1541, %1542, %int12_384 : !torch.vtensor<[80,240,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,240,1,1],!torch.qint8> | |
| %1544 = torch.aten.int_repr %1543 : !torch.vtensor<[80,240,1,1],!torch.qint8> -> !torch.vtensor<[80,240,1,1],si8> | |
| %1545 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1546 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1547 = torch.aten.item %1545 : !torch.vtensor<[],f32> -> !torch.float | |
| %1548 = torch.aten.item %1546 : !torch.vtensor<[],si8> -> !torch.int | |
| %1549 = torch.aten._make_per_tensor_quantized_tensor %1544, %1547, %1548 : !torch.vtensor<[80,240,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,240,1,1],!torch.qint8> | |
| %1550 = torch.aten.dequantize.self %1549 : !torch.vtensor<[80,240,1,1],!torch.qint8> -> !torch.vtensor<[80,240,1,1],f32> | |
| %1551 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1552 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_385 = torch.constant.int 12 | |
| %1553 = torch.aten.item %1551 : !torch.vtensor<[],f32> -> !torch.float | |
| %1554 = torch.aten.item %1552 : !torch.vtensor<[],si8> -> !torch.int | |
| %1555 = torch.aten.quantize_per_tensor %53, %1553, %1554, %int12_385 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %1556 = torch.aten.int_repr %1555 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %1557 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1558 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1559 = torch.aten.item %1557 : !torch.vtensor<[],f32> -> !torch.float | |
| %1560 = torch.aten.item %1558 : !torch.vtensor<[],si8> -> !torch.int | |
| %1561 = torch.aten._make_per_tensor_quantized_tensor %1556, %1559, %1560 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %1562 = torch.aten.dequantize.self %1561 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_386 = torch.constant.int 0 | |
| %int0_387 = torch.constant.int 0 | |
| %int1_388 = torch.constant.int 1 | |
| %int1_389 = torch.constant.int 1 | |
| %int1_390 = torch.constant.int 1 | |
| %int1_391 = torch.constant.int 1 | |
| %int0_392 = torch.constant.int 0 | |
| %1563 = torch.prim.ListConstruct %int0_386, %int0_387 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1564 = torch.prim.ListConstruct %int1_388, %int1_389 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1565 = torch.prim.ListConstruct %int1_390, %int1_391 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1566 = torch.prim.ListConstruct %int0_392, %int0_392 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_393 = torch.constant.bool false | |
| %int1_394 = torch.constant.int 1 | |
| %1567 = torch.aten.convolution %1538, %1550, %1562, %1565, %1563, %1564, %false_393, %1566, %int1_394 : !torch.vtensor<[1,240,14,14],f32>, !torch.vtensor<[80,240,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %1568 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1569 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_395 = torch.constant.int 12 | |
| %1570 = torch.aten.item %1568 : !torch.vtensor<[],f32> -> !torch.float | |
| %1571 = torch.aten.item %1569 : !torch.vtensor<[],si8> -> !torch.int | |
| %1572 = torch.aten.quantize_per_tensor %1567, %1570, %1571, %int12_395 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1573 = torch.aten.int_repr %1572 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %1574 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1575 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1576 = torch.aten.item %1574 : !torch.vtensor<[],f32> -> !torch.float | |
| %1577 = torch.aten.item %1575 : !torch.vtensor<[],si8> -> !torch.int | |
| %1578 = torch.aten._make_per_tensor_quantized_tensor %1573, %1576, %1577 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1579 = torch.aten.dequantize.self %1578 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %1580 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1581 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_396 = torch.constant.int 12 | |
| %1582 = torch.aten.item %1580 : !torch.vtensor<[],f32> -> !torch.float | |
| %1583 = torch.aten.item %1581 : !torch.vtensor<[],si8> -> !torch.int | |
| %1584 = torch.aten.quantize_per_tensor %54, %1582, %1583, %int12_396 : !torch.vtensor<[200,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[200,80,1,1],!torch.qint8> | |
| %1585 = torch.aten.int_repr %1584 : !torch.vtensor<[200,80,1,1],!torch.qint8> -> !torch.vtensor<[200,80,1,1],si8> | |
| %1586 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1587 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1588 = torch.aten.item %1586 : !torch.vtensor<[],f32> -> !torch.float | |
| %1589 = torch.aten.item %1587 : !torch.vtensor<[],si8> -> !torch.int | |
| %1590 = torch.aten._make_per_tensor_quantized_tensor %1585, %1588, %1589 : !torch.vtensor<[200,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[200,80,1,1],!torch.qint8> | |
| %1591 = torch.aten.dequantize.self %1590 : !torch.vtensor<[200,80,1,1],!torch.qint8> -> !torch.vtensor<[200,80,1,1],f32> | |
| %1592 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1593 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_397 = torch.constant.int 12 | |
| %1594 = torch.aten.item %1592 : !torch.vtensor<[],f32> -> !torch.float | |
| %1595 = torch.aten.item %1593 : !torch.vtensor<[],si8> -> !torch.int | |
| %1596 = torch.aten.quantize_per_tensor %55, %1594, %1595, %int12_397 : !torch.vtensor<[200],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[200],!torch.qint8> | |
| %1597 = torch.aten.int_repr %1596 : !torch.vtensor<[200],!torch.qint8> -> !torch.vtensor<[200],si8> | |
| %1598 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1599 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1600 = torch.aten.item %1598 : !torch.vtensor<[],f32> -> !torch.float | |
| %1601 = torch.aten.item %1599 : !torch.vtensor<[],si8> -> !torch.int | |
| %1602 = torch.aten._make_per_tensor_quantized_tensor %1597, %1600, %1601 : !torch.vtensor<[200],si8>, !torch.float, !torch.int -> !torch.vtensor<[200],!torch.qint8> | |
| %1603 = torch.aten.dequantize.self %1602 : !torch.vtensor<[200],!torch.qint8> -> !torch.vtensor<[200],f32> | |
| %int0_398 = torch.constant.int 0 | |
| %int0_399 = torch.constant.int 0 | |
| %int1_400 = torch.constant.int 1 | |
| %int1_401 = torch.constant.int 1 | |
| %int1_402 = torch.constant.int 1 | |
| %int1_403 = torch.constant.int 1 | |
| %int0_404 = torch.constant.int 0 | |
| %1604 = torch.prim.ListConstruct %int0_398, %int0_399 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1605 = torch.prim.ListConstruct %int1_400, %int1_401 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1606 = torch.prim.ListConstruct %int1_402, %int1_403 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1607 = torch.prim.ListConstruct %int0_404, %int0_404 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_405 = torch.constant.bool false | |
| %int1_406 = torch.constant.int 1 | |
| %1608 = torch.aten.convolution %1579, %1591, %1603, %1606, %1604, %1605, %false_405, %1607, %int1_406 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[200,80,1,1],f32>, !torch.vtensor<[200],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,200,14,14],f32> | |
| %1609 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1610 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_407 = torch.constant.int 12 | |
| %1611 = torch.aten.item %1609 : !torch.vtensor<[],f32> -> !torch.float | |
| %1612 = torch.aten.item %1610 : !torch.vtensor<[],si8> -> !torch.int | |
| %1613 = torch.aten.quantize_per_tensor %1608, %1611, %1612, %int12_407 : !torch.vtensor<[1,200,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1614 = torch.aten.int_repr %1613 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],si8> | |
| %1615 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1616 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1617 = torch.aten.item %1615 : !torch.vtensor<[],f32> -> !torch.float | |
| %1618 = torch.aten.item %1616 : !torch.vtensor<[],si8> -> !torch.int | |
| %1619 = torch.aten._make_per_tensor_quantized_tensor %1614, %1617, %1618 : !torch.vtensor<[1,200,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1620 = torch.aten.dequantize.self %1619 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1621 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_408 = torch.constant.int 1 | |
| %1622 = torch.aten.add.Tensor %1620, %1621, %int1_408 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,200,14,14],f32> | |
| %1623 = torch.aten.relu %1622 : !torch.vtensor<[1,200,14,14],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1624 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1625 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_409 = torch.constant.int 6 | |
| %none_410 = torch.constant.none | |
| %false_411 = torch.constant.bool false | |
| %1626 = torch.aten.to.dtype %1624, %int6_409, %false_411, %false_411, %none_410 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_412 = torch.constant.int 6 | |
| %none_413 = torch.constant.none | |
| %false_414 = torch.constant.bool false | |
| %1627 = torch.aten.to.dtype %1625, %int6_412, %false_414, %false_414, %none_413 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1628 = torch.aten.clamp.Tensor %1623, %1626, %1627 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1629 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1630 = torch.aten.mul.Tensor %1628, %1629 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1631 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1632 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_415 = torch.constant.int 12 | |
| %1633 = torch.aten.item %1631 : !torch.vtensor<[],f32> -> !torch.float | |
| %1634 = torch.aten.item %1632 : !torch.vtensor<[],si8> -> !torch.int | |
| %1635 = torch.aten.quantize_per_tensor %1630, %1633, %1634, %int12_415 : !torch.vtensor<[1,200,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1636 = torch.aten.int_repr %1635 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],si8> | |
| %1637 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1638 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1639 = torch.aten.item %1637 : !torch.vtensor<[],f32> -> !torch.float | |
| %1640 = torch.aten.item %1638 : !torch.vtensor<[],si8> -> !torch.int | |
| %1641 = torch.aten._make_per_tensor_quantized_tensor %1636, %1639, %1640 : !torch.vtensor<[1,200,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1642 = torch.aten.dequantize.self %1641 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1643 = torch.aten.mul.Tensor %1620, %1642 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[1,200,14,14],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1644 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1645 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_416 = torch.constant.int 12 | |
| %1646 = torch.aten.item %1644 : !torch.vtensor<[],f32> -> !torch.float | |
| %1647 = torch.aten.item %1645 : !torch.vtensor<[],si8> -> !torch.int | |
| %1648 = torch.aten.quantize_per_tensor %1643, %1646, %1647, %int12_416 : !torch.vtensor<[1,200,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1649 = torch.aten.int_repr %1648 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],si8> | |
| %1650 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1651 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1652 = torch.aten.item %1650 : !torch.vtensor<[],f32> -> !torch.float | |
| %1653 = torch.aten.item %1651 : !torch.vtensor<[],si8> -> !torch.int | |
| %1654 = torch.aten._make_per_tensor_quantized_tensor %1649, %1652, %1653 : !torch.vtensor<[1,200,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1655 = torch.aten.dequantize.self %1654 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1656 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1657 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_417 = torch.constant.int 12 | |
| %1658 = torch.aten.item %1656 : !torch.vtensor<[],f32> -> !torch.float | |
| %1659 = torch.aten.item %1657 : !torch.vtensor<[],si8> -> !torch.int | |
| %1660 = torch.aten.quantize_per_tensor %56, %1658, %1659, %int12_417 : !torch.vtensor<[200,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[200,1,3,3],!torch.qint8> | |
| %1661 = torch.aten.int_repr %1660 : !torch.vtensor<[200,1,3,3],!torch.qint8> -> !torch.vtensor<[200,1,3,3],si8> | |
| %1662 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1663 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1664 = torch.aten.item %1662 : !torch.vtensor<[],f32> -> !torch.float | |
| %1665 = torch.aten.item %1663 : !torch.vtensor<[],si8> -> !torch.int | |
| %1666 = torch.aten._make_per_tensor_quantized_tensor %1661, %1664, %1665 : !torch.vtensor<[200,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[200,1,3,3],!torch.qint8> | |
| %1667 = torch.aten.dequantize.self %1666 : !torch.vtensor<[200,1,3,3],!torch.qint8> -> !torch.vtensor<[200,1,3,3],f32> | |
| %1668 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1669 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_418 = torch.constant.int 12 | |
| %1670 = torch.aten.item %1668 : !torch.vtensor<[],f32> -> !torch.float | |
| %1671 = torch.aten.item %1669 : !torch.vtensor<[],si8> -> !torch.int | |
| %1672 = torch.aten.quantize_per_tensor %57, %1670, %1671, %int12_418 : !torch.vtensor<[200],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[200],!torch.qint8> | |
| %1673 = torch.aten.int_repr %1672 : !torch.vtensor<[200],!torch.qint8> -> !torch.vtensor<[200],si8> | |
| %1674 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1675 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1676 = torch.aten.item %1674 : !torch.vtensor<[],f32> -> !torch.float | |
| %1677 = torch.aten.item %1675 : !torch.vtensor<[],si8> -> !torch.int | |
| %1678 = torch.aten._make_per_tensor_quantized_tensor %1673, %1676, %1677 : !torch.vtensor<[200],si8>, !torch.float, !torch.int -> !torch.vtensor<[200],!torch.qint8> | |
| %1679 = torch.aten.dequantize.self %1678 : !torch.vtensor<[200],!torch.qint8> -> !torch.vtensor<[200],f32> | |
| %int1_419 = torch.constant.int 1 | |
| %int1_420 = torch.constant.int 1 | |
| %int1_421 = torch.constant.int 1 | |
| %int1_422 = torch.constant.int 1 | |
| %int1_423 = torch.constant.int 1 | |
| %int1_424 = torch.constant.int 1 | |
| %int0_425 = torch.constant.int 0 | |
| %1680 = torch.prim.ListConstruct %int1_419, %int1_420 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1681 = torch.prim.ListConstruct %int1_421, %int1_422 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1682 = torch.prim.ListConstruct %int1_423, %int1_424 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1683 = torch.prim.ListConstruct %int0_425, %int0_425 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_426 = torch.constant.bool false | |
| %int200 = torch.constant.int 200 | |
| %1684 = torch.aten.convolution %1655, %1667, %1679, %1682, %1680, %1681, %false_426, %1683, %int200 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[200,1,3,3],f32>, !torch.vtensor<[200],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,200,14,14],f32> | |
| %1685 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1686 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_427 = torch.constant.int 12 | |
| %1687 = torch.aten.item %1685 : !torch.vtensor<[],f32> -> !torch.float | |
| %1688 = torch.aten.item %1686 : !torch.vtensor<[],si8> -> !torch.int | |
| %1689 = torch.aten.quantize_per_tensor %1684, %1687, %1688, %int12_427 : !torch.vtensor<[1,200,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1690 = torch.aten.int_repr %1689 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],si8> | |
| %1691 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1692 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1693 = torch.aten.item %1691 : !torch.vtensor<[],f32> -> !torch.float | |
| %1694 = torch.aten.item %1692 : !torch.vtensor<[],si8> -> !torch.int | |
| %1695 = torch.aten._make_per_tensor_quantized_tensor %1690, %1693, %1694 : !torch.vtensor<[1,200,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1696 = torch.aten.dequantize.self %1695 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1697 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_428 = torch.constant.int 1 | |
| %1698 = torch.aten.add.Tensor %1696, %1697, %int1_428 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,200,14,14],f32> | |
| %1699 = torch.aten.relu %1698 : !torch.vtensor<[1,200,14,14],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1700 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1701 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_429 = torch.constant.int 6 | |
| %none_430 = torch.constant.none | |
| %false_431 = torch.constant.bool false | |
| %1702 = torch.aten.to.dtype %1700, %int6_429, %false_431, %false_431, %none_430 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_432 = torch.constant.int 6 | |
| %none_433 = torch.constant.none | |
| %false_434 = torch.constant.bool false | |
| %1703 = torch.aten.to.dtype %1701, %int6_432, %false_434, %false_434, %none_433 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1704 = torch.aten.clamp.Tensor %1699, %1702, %1703 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1705 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1706 = torch.aten.mul.Tensor %1704, %1705 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1707 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1708 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_435 = torch.constant.int 12 | |
| %1709 = torch.aten.item %1707 : !torch.vtensor<[],f32> -> !torch.float | |
| %1710 = torch.aten.item %1708 : !torch.vtensor<[],si8> -> !torch.int | |
| %1711 = torch.aten.quantize_per_tensor %1706, %1709, %1710, %int12_435 : !torch.vtensor<[1,200,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1712 = torch.aten.int_repr %1711 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],si8> | |
| %1713 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1714 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1715 = torch.aten.item %1713 : !torch.vtensor<[],f32> -> !torch.float | |
| %1716 = torch.aten.item %1714 : !torch.vtensor<[],si8> -> !torch.int | |
| %1717 = torch.aten._make_per_tensor_quantized_tensor %1712, %1715, %1716 : !torch.vtensor<[1,200,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1718 = torch.aten.dequantize.self %1717 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1719 = torch.aten.mul.Tensor %1696, %1718 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[1,200,14,14],f32> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1720 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1721 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_436 = torch.constant.int 12 | |
| %1722 = torch.aten.item %1720 : !torch.vtensor<[],f32> -> !torch.float | |
| %1723 = torch.aten.item %1721 : !torch.vtensor<[],si8> -> !torch.int | |
| %1724 = torch.aten.quantize_per_tensor %1719, %1722, %1723, %int12_436 : !torch.vtensor<[1,200,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1725 = torch.aten.int_repr %1724 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],si8> | |
| %1726 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1727 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1728 = torch.aten.item %1726 : !torch.vtensor<[],f32> -> !torch.float | |
| %1729 = torch.aten.item %1727 : !torch.vtensor<[],si8> -> !torch.int | |
| %1730 = torch.aten._make_per_tensor_quantized_tensor %1725, %1728, %1729 : !torch.vtensor<[1,200,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,200,14,14],!torch.qint8> | |
| %1731 = torch.aten.dequantize.self %1730 : !torch.vtensor<[1,200,14,14],!torch.qint8> -> !torch.vtensor<[1,200,14,14],f32> | |
| %1732 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1733 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_437 = torch.constant.int 12 | |
| %1734 = torch.aten.item %1732 : !torch.vtensor<[],f32> -> !torch.float | |
| %1735 = torch.aten.item %1733 : !torch.vtensor<[],si8> -> !torch.int | |
| %1736 = torch.aten.quantize_per_tensor %58, %1734, %1735, %int12_437 : !torch.vtensor<[80,200,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,200,1,1],!torch.qint8> | |
| %1737 = torch.aten.int_repr %1736 : !torch.vtensor<[80,200,1,1],!torch.qint8> -> !torch.vtensor<[80,200,1,1],si8> | |
| %1738 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1739 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1740 = torch.aten.item %1738 : !torch.vtensor<[],f32> -> !torch.float | |
| %1741 = torch.aten.item %1739 : !torch.vtensor<[],si8> -> !torch.int | |
| %1742 = torch.aten._make_per_tensor_quantized_tensor %1737, %1740, %1741 : !torch.vtensor<[80,200,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,200,1,1],!torch.qint8> | |
| %1743 = torch.aten.dequantize.self %1742 : !torch.vtensor<[80,200,1,1],!torch.qint8> -> !torch.vtensor<[80,200,1,1],f32> | |
| %1744 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1745 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_438 = torch.constant.int 12 | |
| %1746 = torch.aten.item %1744 : !torch.vtensor<[],f32> -> !torch.float | |
| %1747 = torch.aten.item %1745 : !torch.vtensor<[],si8> -> !torch.int | |
| %1748 = torch.aten.quantize_per_tensor %59, %1746, %1747, %int12_438 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %1749 = torch.aten.int_repr %1748 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %1750 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1751 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1752 = torch.aten.item %1750 : !torch.vtensor<[],f32> -> !torch.float | |
| %1753 = torch.aten.item %1751 : !torch.vtensor<[],si8> -> !torch.int | |
| %1754 = torch.aten._make_per_tensor_quantized_tensor %1749, %1752, %1753 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %1755 = torch.aten.dequantize.self %1754 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_439 = torch.constant.int 0 | |
| %int0_440 = torch.constant.int 0 | |
| %int1_441 = torch.constant.int 1 | |
| %int1_442 = torch.constant.int 1 | |
| %int1_443 = torch.constant.int 1 | |
| %int1_444 = torch.constant.int 1 | |
| %int0_445 = torch.constant.int 0 | |
| %1756 = torch.prim.ListConstruct %int0_439, %int0_440 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1757 = torch.prim.ListConstruct %int1_441, %int1_442 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1758 = torch.prim.ListConstruct %int1_443, %int1_444 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1759 = torch.prim.ListConstruct %int0_445, %int0_445 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_446 = torch.constant.bool false | |
| %int1_447 = torch.constant.int 1 | |
| %1760 = torch.aten.convolution %1731, %1743, %1755, %1758, %1756, %1757, %false_446, %1759, %int1_447 : !torch.vtensor<[1,200,14,14],f32>, !torch.vtensor<[80,200,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %1761 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1762 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_448 = torch.constant.int 12 | |
| %1763 = torch.aten.item %1761 : !torch.vtensor<[],f32> -> !torch.float | |
| %1764 = torch.aten.item %1762 : !torch.vtensor<[],si8> -> !torch.int | |
| %1765 = torch.aten.quantize_per_tensor %1760, %1763, %1764, %int12_448 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1766 = torch.aten.int_repr %1765 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %1767 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1768 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1769 = torch.aten.item %1767 : !torch.vtensor<[],f32> -> !torch.float | |
| %1770 = torch.aten.item %1768 : !torch.vtensor<[],si8> -> !torch.int | |
| %1771 = torch.aten._make_per_tensor_quantized_tensor %1766, %1769, %1770 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1772 = torch.aten.dequantize.self %1771 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %int1_449 = torch.constant.int 1 | |
| %1773 = torch.aten.add.Tensor %1772, %1579, %int1_449 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[1,80,14,14],f32>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %1774 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1775 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_450 = torch.constant.int 12 | |
| %1776 = torch.aten.item %1774 : !torch.vtensor<[],f32> -> !torch.float | |
| %1777 = torch.aten.item %1775 : !torch.vtensor<[],si8> -> !torch.int | |
| %1778 = torch.aten.quantize_per_tensor %1773, %1776, %1777, %int12_450 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1779 = torch.aten.int_repr %1778 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %1780 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1781 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1782 = torch.aten.item %1780 : !torch.vtensor<[],f32> -> !torch.float | |
| %1783 = torch.aten.item %1781 : !torch.vtensor<[],si8> -> !torch.int | |
| %1784 = torch.aten._make_per_tensor_quantized_tensor %1779, %1782, %1783 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1785 = torch.aten.dequantize.self %1784 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %1786 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1787 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_451 = torch.constant.int 12 | |
| %1788 = torch.aten.item %1786 : !torch.vtensor<[],f32> -> !torch.float | |
| %1789 = torch.aten.item %1787 : !torch.vtensor<[],si8> -> !torch.int | |
| %1790 = torch.aten.quantize_per_tensor %60, %1788, %1789, %int12_451 : !torch.vtensor<[184,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184,80,1,1],!torch.qint8> | |
| %1791 = torch.aten.int_repr %1790 : !torch.vtensor<[184,80,1,1],!torch.qint8> -> !torch.vtensor<[184,80,1,1],si8> | |
| %1792 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1793 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1794 = torch.aten.item %1792 : !torch.vtensor<[],f32> -> !torch.float | |
| %1795 = torch.aten.item %1793 : !torch.vtensor<[],si8> -> !torch.int | |
| %1796 = torch.aten._make_per_tensor_quantized_tensor %1791, %1794, %1795 : !torch.vtensor<[184,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[184,80,1,1],!torch.qint8> | |
| %1797 = torch.aten.dequantize.self %1796 : !torch.vtensor<[184,80,1,1],!torch.qint8> -> !torch.vtensor<[184,80,1,1],f32> | |
| %1798 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1799 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_452 = torch.constant.int 12 | |
| %1800 = torch.aten.item %1798 : !torch.vtensor<[],f32> -> !torch.float | |
| %1801 = torch.aten.item %1799 : !torch.vtensor<[],si8> -> !torch.int | |
| %1802 = torch.aten.quantize_per_tensor %61, %1800, %1801, %int12_452 : !torch.vtensor<[184],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %1803 = torch.aten.int_repr %1802 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],si8> | |
| %1804 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1805 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1806 = torch.aten.item %1804 : !torch.vtensor<[],f32> -> !torch.float | |
| %1807 = torch.aten.item %1805 : !torch.vtensor<[],si8> -> !torch.int | |
| %1808 = torch.aten._make_per_tensor_quantized_tensor %1803, %1806, %1807 : !torch.vtensor<[184],si8>, !torch.float, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %1809 = torch.aten.dequantize.self %1808 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],f32> | |
| %int0_453 = torch.constant.int 0 | |
| %int0_454 = torch.constant.int 0 | |
| %int1_455 = torch.constant.int 1 | |
| %int1_456 = torch.constant.int 1 | |
| %int1_457 = torch.constant.int 1 | |
| %int1_458 = torch.constant.int 1 | |
| %int0_459 = torch.constant.int 0 | |
| %1810 = torch.prim.ListConstruct %int0_453, %int0_454 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1811 = torch.prim.ListConstruct %int1_455, %int1_456 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1812 = torch.prim.ListConstruct %int1_457, %int1_458 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1813 = torch.prim.ListConstruct %int0_459, %int0_459 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_460 = torch.constant.bool false | |
| %int1_461 = torch.constant.int 1 | |
| %1814 = torch.aten.convolution %1785, %1797, %1809, %1812, %1810, %1811, %false_460, %1813, %int1_461 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[184,80,1,1],f32>, !torch.vtensor<[184],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %1815 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1816 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_462 = torch.constant.int 12 | |
| %1817 = torch.aten.item %1815 : !torch.vtensor<[],f32> -> !torch.float | |
| %1818 = torch.aten.item %1816 : !torch.vtensor<[],si8> -> !torch.int | |
| %1819 = torch.aten.quantize_per_tensor %1814, %1817, %1818, %int12_462 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1820 = torch.aten.int_repr %1819 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %1821 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1822 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1823 = torch.aten.item %1821 : !torch.vtensor<[],f32> -> !torch.float | |
| %1824 = torch.aten.item %1822 : !torch.vtensor<[],si8> -> !torch.int | |
| %1825 = torch.aten._make_per_tensor_quantized_tensor %1820, %1823, %1824 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1826 = torch.aten.dequantize.self %1825 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1827 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_463 = torch.constant.int 1 | |
| %1828 = torch.aten.add.Tensor %1826, %1827, %int1_463 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %1829 = torch.aten.relu %1828 : !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1830 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1831 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_464 = torch.constant.int 6 | |
| %none_465 = torch.constant.none | |
| %false_466 = torch.constant.bool false | |
| %1832 = torch.aten.to.dtype %1830, %int6_464, %false_466, %false_466, %none_465 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_467 = torch.constant.int 6 | |
| %none_468 = torch.constant.none | |
| %false_469 = torch.constant.bool false | |
| %1833 = torch.aten.to.dtype %1831, %int6_467, %false_469, %false_469, %none_468 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1834 = torch.aten.clamp.Tensor %1829, %1832, %1833 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1835 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1836 = torch.aten.mul.Tensor %1834, %1835 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1837 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1838 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_470 = torch.constant.int 12 | |
| %1839 = torch.aten.item %1837 : !torch.vtensor<[],f32> -> !torch.float | |
| %1840 = torch.aten.item %1838 : !torch.vtensor<[],si8> -> !torch.int | |
| %1841 = torch.aten.quantize_per_tensor %1836, %1839, %1840, %int12_470 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1842 = torch.aten.int_repr %1841 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %1843 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1844 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1845 = torch.aten.item %1843 : !torch.vtensor<[],f32> -> !torch.float | |
| %1846 = torch.aten.item %1844 : !torch.vtensor<[],si8> -> !torch.int | |
| %1847 = torch.aten._make_per_tensor_quantized_tensor %1842, %1845, %1846 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1848 = torch.aten.dequantize.self %1847 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1849 = torch.aten.mul.Tensor %1826, %1848 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1850 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1851 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_471 = torch.constant.int 12 | |
| %1852 = torch.aten.item %1850 : !torch.vtensor<[],f32> -> !torch.float | |
| %1853 = torch.aten.item %1851 : !torch.vtensor<[],si8> -> !torch.int | |
| %1854 = torch.aten.quantize_per_tensor %1849, %1852, %1853, %int12_471 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1855 = torch.aten.int_repr %1854 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %1856 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1857 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1858 = torch.aten.item %1856 : !torch.vtensor<[],f32> -> !torch.float | |
| %1859 = torch.aten.item %1857 : !torch.vtensor<[],si8> -> !torch.int | |
| %1860 = torch.aten._make_per_tensor_quantized_tensor %1855, %1858, %1859 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1861 = torch.aten.dequantize.self %1860 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1862 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1863 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_472 = torch.constant.int 12 | |
| %1864 = torch.aten.item %1862 : !torch.vtensor<[],f32> -> !torch.float | |
| %1865 = torch.aten.item %1863 : !torch.vtensor<[],si8> -> !torch.int | |
| %1866 = torch.aten.quantize_per_tensor %62, %1864, %1865, %int12_472 : !torch.vtensor<[184,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184,1,3,3],!torch.qint8> | |
| %1867 = torch.aten.int_repr %1866 : !torch.vtensor<[184,1,3,3],!torch.qint8> -> !torch.vtensor<[184,1,3,3],si8> | |
| %1868 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1869 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1870 = torch.aten.item %1868 : !torch.vtensor<[],f32> -> !torch.float | |
| %1871 = torch.aten.item %1869 : !torch.vtensor<[],si8> -> !torch.int | |
| %1872 = torch.aten._make_per_tensor_quantized_tensor %1867, %1870, %1871 : !torch.vtensor<[184,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[184,1,3,3],!torch.qint8> | |
| %1873 = torch.aten.dequantize.self %1872 : !torch.vtensor<[184,1,3,3],!torch.qint8> -> !torch.vtensor<[184,1,3,3],f32> | |
| %1874 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1875 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_473 = torch.constant.int 12 | |
| %1876 = torch.aten.item %1874 : !torch.vtensor<[],f32> -> !torch.float | |
| %1877 = torch.aten.item %1875 : !torch.vtensor<[],si8> -> !torch.int | |
| %1878 = torch.aten.quantize_per_tensor %63, %1876, %1877, %int12_473 : !torch.vtensor<[184],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %1879 = torch.aten.int_repr %1878 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],si8> | |
| %1880 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1881 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1882 = torch.aten.item %1880 : !torch.vtensor<[],f32> -> !torch.float | |
| %1883 = torch.aten.item %1881 : !torch.vtensor<[],si8> -> !torch.int | |
| %1884 = torch.aten._make_per_tensor_quantized_tensor %1879, %1882, %1883 : !torch.vtensor<[184],si8>, !torch.float, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %1885 = torch.aten.dequantize.self %1884 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],f32> | |
| %int1_474 = torch.constant.int 1 | |
| %int1_475 = torch.constant.int 1 | |
| %int1_476 = torch.constant.int 1 | |
| %int1_477 = torch.constant.int 1 | |
| %int1_478 = torch.constant.int 1 | |
| %int1_479 = torch.constant.int 1 | |
| %int0_480 = torch.constant.int 0 | |
| %1886 = torch.prim.ListConstruct %int1_474, %int1_475 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1887 = torch.prim.ListConstruct %int1_476, %int1_477 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1888 = torch.prim.ListConstruct %int1_478, %int1_479 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1889 = torch.prim.ListConstruct %int0_480, %int0_480 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_481 = torch.constant.bool false | |
| %int184 = torch.constant.int 184 | |
| %1890 = torch.aten.convolution %1861, %1873, %1885, %1888, %1886, %1887, %false_481, %1889, %int184 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[184,1,3,3],f32>, !torch.vtensor<[184],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %1891 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1892 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_482 = torch.constant.int 12 | |
| %1893 = torch.aten.item %1891 : !torch.vtensor<[],f32> -> !torch.float | |
| %1894 = torch.aten.item %1892 : !torch.vtensor<[],si8> -> !torch.int | |
| %1895 = torch.aten.quantize_per_tensor %1890, %1893, %1894, %int12_482 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1896 = torch.aten.int_repr %1895 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %1897 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1898 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1899 = torch.aten.item %1897 : !torch.vtensor<[],f32> -> !torch.float | |
| %1900 = torch.aten.item %1898 : !torch.vtensor<[],si8> -> !torch.int | |
| %1901 = torch.aten._make_per_tensor_quantized_tensor %1896, %1899, %1900 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1902 = torch.aten.dequantize.self %1901 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1903 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_483 = torch.constant.int 1 | |
| %1904 = torch.aten.add.Tensor %1902, %1903, %int1_483 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %1905 = torch.aten.relu %1904 : !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1906 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1907 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_484 = torch.constant.int 6 | |
| %none_485 = torch.constant.none | |
| %false_486 = torch.constant.bool false | |
| %1908 = torch.aten.to.dtype %1906, %int6_484, %false_486, %false_486, %none_485 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_487 = torch.constant.int 6 | |
| %none_488 = torch.constant.none | |
| %false_489 = torch.constant.bool false | |
| %1909 = torch.aten.to.dtype %1907, %int6_487, %false_489, %false_489, %none_488 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %1910 = torch.aten.clamp.Tensor %1905, %1908, %1909 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1911 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1912 = torch.aten.mul.Tensor %1910, %1911 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1913 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1914 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_490 = torch.constant.int 12 | |
| %1915 = torch.aten.item %1913 : !torch.vtensor<[],f32> -> !torch.float | |
| %1916 = torch.aten.item %1914 : !torch.vtensor<[],si8> -> !torch.int | |
| %1917 = torch.aten.quantize_per_tensor %1912, %1915, %1916, %int12_490 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1918 = torch.aten.int_repr %1917 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %1919 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1920 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1921 = torch.aten.item %1919 : !torch.vtensor<[],f32> -> !torch.float | |
| %1922 = torch.aten.item %1920 : !torch.vtensor<[],si8> -> !torch.int | |
| %1923 = torch.aten._make_per_tensor_quantized_tensor %1918, %1921, %1922 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1924 = torch.aten.dequantize.self %1923 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1925 = torch.aten.mul.Tensor %1902, %1924 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1926 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1927 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_491 = torch.constant.int 12 | |
| %1928 = torch.aten.item %1926 : !torch.vtensor<[],f32> -> !torch.float | |
| %1929 = torch.aten.item %1927 : !torch.vtensor<[],si8> -> !torch.int | |
| %1930 = torch.aten.quantize_per_tensor %1925, %1928, %1929, %int12_491 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1931 = torch.aten.int_repr %1930 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %1932 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1933 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1934 = torch.aten.item %1932 : !torch.vtensor<[],f32> -> !torch.float | |
| %1935 = torch.aten.item %1933 : !torch.vtensor<[],si8> -> !torch.int | |
| %1936 = torch.aten._make_per_tensor_quantized_tensor %1931, %1934, %1935 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %1937 = torch.aten.dequantize.self %1936 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %1938 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1939 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_492 = torch.constant.int 12 | |
| %1940 = torch.aten.item %1938 : !torch.vtensor<[],f32> -> !torch.float | |
| %1941 = torch.aten.item %1939 : !torch.vtensor<[],si8> -> !torch.int | |
| %1942 = torch.aten.quantize_per_tensor %64, %1940, %1941, %int12_492 : !torch.vtensor<[80,184,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,184,1,1],!torch.qint8> | |
| %1943 = torch.aten.int_repr %1942 : !torch.vtensor<[80,184,1,1],!torch.qint8> -> !torch.vtensor<[80,184,1,1],si8> | |
| %1944 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1945 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1946 = torch.aten.item %1944 : !torch.vtensor<[],f32> -> !torch.float | |
| %1947 = torch.aten.item %1945 : !torch.vtensor<[],si8> -> !torch.int | |
| %1948 = torch.aten._make_per_tensor_quantized_tensor %1943, %1946, %1947 : !torch.vtensor<[80,184,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,184,1,1],!torch.qint8> | |
| %1949 = torch.aten.dequantize.self %1948 : !torch.vtensor<[80,184,1,1],!torch.qint8> -> !torch.vtensor<[80,184,1,1],f32> | |
| %1950 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1951 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_493 = torch.constant.int 12 | |
| %1952 = torch.aten.item %1950 : !torch.vtensor<[],f32> -> !torch.float | |
| %1953 = torch.aten.item %1951 : !torch.vtensor<[],si8> -> !torch.int | |
| %1954 = torch.aten.quantize_per_tensor %65, %1952, %1953, %int12_493 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %1955 = torch.aten.int_repr %1954 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %1956 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1957 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1958 = torch.aten.item %1956 : !torch.vtensor<[],f32> -> !torch.float | |
| %1959 = torch.aten.item %1957 : !torch.vtensor<[],si8> -> !torch.int | |
| %1960 = torch.aten._make_per_tensor_quantized_tensor %1955, %1958, %1959 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %1961 = torch.aten.dequantize.self %1960 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_494 = torch.constant.int 0 | |
| %int0_495 = torch.constant.int 0 | |
| %int1_496 = torch.constant.int 1 | |
| %int1_497 = torch.constant.int 1 | |
| %int1_498 = torch.constant.int 1 | |
| %int1_499 = torch.constant.int 1 | |
| %int0_500 = torch.constant.int 0 | |
| %1962 = torch.prim.ListConstruct %int0_494, %int0_495 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1963 = torch.prim.ListConstruct %int1_496, %int1_497 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1964 = torch.prim.ListConstruct %int1_498, %int1_499 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %1965 = torch.prim.ListConstruct %int0_500, %int0_500 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_501 = torch.constant.bool false | |
| %int1_502 = torch.constant.int 1 | |
| %1966 = torch.aten.convolution %1937, %1949, %1961, %1964, %1962, %1963, %false_501, %1965, %int1_502 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[80,184,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %1967 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1968 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_503 = torch.constant.int 12 | |
| %1969 = torch.aten.item %1967 : !torch.vtensor<[],f32> -> !torch.float | |
| %1970 = torch.aten.item %1968 : !torch.vtensor<[],si8> -> !torch.int | |
| %1971 = torch.aten.quantize_per_tensor %1966, %1969, %1970, %int12_503 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1972 = torch.aten.int_repr %1971 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %1973 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1974 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1975 = torch.aten.item %1973 : !torch.vtensor<[],f32> -> !torch.float | |
| %1976 = torch.aten.item %1974 : !torch.vtensor<[],si8> -> !torch.int | |
| %1977 = torch.aten._make_per_tensor_quantized_tensor %1972, %1975, %1976 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1978 = torch.aten.dequantize.self %1977 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %int1_504 = torch.constant.int 1 | |
| %1979 = torch.aten.add.Tensor %1978, %1785, %int1_504 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[1,80,14,14],f32>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %1980 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1981 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_505 = torch.constant.int 12 | |
| %1982 = torch.aten.item %1980 : !torch.vtensor<[],f32> -> !torch.float | |
| %1983 = torch.aten.item %1981 : !torch.vtensor<[],si8> -> !torch.int | |
| %1984 = torch.aten.quantize_per_tensor %1979, %1982, %1983, %int12_505 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1985 = torch.aten.int_repr %1984 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %1986 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1987 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %1988 = torch.aten.item %1986 : !torch.vtensor<[],f32> -> !torch.float | |
| %1989 = torch.aten.item %1987 : !torch.vtensor<[],si8> -> !torch.int | |
| %1990 = torch.aten._make_per_tensor_quantized_tensor %1985, %1988, %1989 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %1991 = torch.aten.dequantize.self %1990 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %1992 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1993 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_506 = torch.constant.int 12 | |
| %1994 = torch.aten.item %1992 : !torch.vtensor<[],f32> -> !torch.float | |
| %1995 = torch.aten.item %1993 : !torch.vtensor<[],si8> -> !torch.int | |
| %1996 = torch.aten.quantize_per_tensor %66, %1994, %1995, %int12_506 : !torch.vtensor<[184,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184,80,1,1],!torch.qint8> | |
| %1997 = torch.aten.int_repr %1996 : !torch.vtensor<[184,80,1,1],!torch.qint8> -> !torch.vtensor<[184,80,1,1],si8> | |
| %1998 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %1999 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2000 = torch.aten.item %1998 : !torch.vtensor<[],f32> -> !torch.float | |
| %2001 = torch.aten.item %1999 : !torch.vtensor<[],si8> -> !torch.int | |
| %2002 = torch.aten._make_per_tensor_quantized_tensor %1997, %2000, %2001 : !torch.vtensor<[184,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[184,80,1,1],!torch.qint8> | |
| %2003 = torch.aten.dequantize.self %2002 : !torch.vtensor<[184,80,1,1],!torch.qint8> -> !torch.vtensor<[184,80,1,1],f32> | |
| %2004 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2005 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_507 = torch.constant.int 12 | |
| %2006 = torch.aten.item %2004 : !torch.vtensor<[],f32> -> !torch.float | |
| %2007 = torch.aten.item %2005 : !torch.vtensor<[],si8> -> !torch.int | |
| %2008 = torch.aten.quantize_per_tensor %67, %2006, %2007, %int12_507 : !torch.vtensor<[184],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %2009 = torch.aten.int_repr %2008 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],si8> | |
| %2010 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2011 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2012 = torch.aten.item %2010 : !torch.vtensor<[],f32> -> !torch.float | |
| %2013 = torch.aten.item %2011 : !torch.vtensor<[],si8> -> !torch.int | |
| %2014 = torch.aten._make_per_tensor_quantized_tensor %2009, %2012, %2013 : !torch.vtensor<[184],si8>, !torch.float, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %2015 = torch.aten.dequantize.self %2014 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],f32> | |
| %int0_508 = torch.constant.int 0 | |
| %int0_509 = torch.constant.int 0 | |
| %int1_510 = torch.constant.int 1 | |
| %int1_511 = torch.constant.int 1 | |
| %int1_512 = torch.constant.int 1 | |
| %int1_513 = torch.constant.int 1 | |
| %int0_514 = torch.constant.int 0 | |
| %2016 = torch.prim.ListConstruct %int0_508, %int0_509 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2017 = torch.prim.ListConstruct %int1_510, %int1_511 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2018 = torch.prim.ListConstruct %int1_512, %int1_513 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2019 = torch.prim.ListConstruct %int0_514, %int0_514 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_515 = torch.constant.bool false | |
| %int1_516 = torch.constant.int 1 | |
| %2020 = torch.aten.convolution %1991, %2003, %2015, %2018, %2016, %2017, %false_515, %2019, %int1_516 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[184,80,1,1],f32>, !torch.vtensor<[184],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %2021 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2022 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_517 = torch.constant.int 12 | |
| %2023 = torch.aten.item %2021 : !torch.vtensor<[],f32> -> !torch.float | |
| %2024 = torch.aten.item %2022 : !torch.vtensor<[],si8> -> !torch.int | |
| %2025 = torch.aten.quantize_per_tensor %2020, %2023, %2024, %int12_517 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2026 = torch.aten.int_repr %2025 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %2027 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2028 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2029 = torch.aten.item %2027 : !torch.vtensor<[],f32> -> !torch.float | |
| %2030 = torch.aten.item %2028 : !torch.vtensor<[],si8> -> !torch.int | |
| %2031 = torch.aten._make_per_tensor_quantized_tensor %2026, %2029, %2030 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2032 = torch.aten.dequantize.self %2031 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2033 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_518 = torch.constant.int 1 | |
| %2034 = torch.aten.add.Tensor %2032, %2033, %int1_518 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %2035 = torch.aten.relu %2034 : !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2036 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2037 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_519 = torch.constant.int 6 | |
| %none_520 = torch.constant.none | |
| %false_521 = torch.constant.bool false | |
| %2038 = torch.aten.to.dtype %2036, %int6_519, %false_521, %false_521, %none_520 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_522 = torch.constant.int 6 | |
| %none_523 = torch.constant.none | |
| %false_524 = torch.constant.bool false | |
| %2039 = torch.aten.to.dtype %2037, %int6_522, %false_524, %false_524, %none_523 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2040 = torch.aten.clamp.Tensor %2035, %2038, %2039 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2041 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2042 = torch.aten.mul.Tensor %2040, %2041 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2043 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2044 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_525 = torch.constant.int 12 | |
| %2045 = torch.aten.item %2043 : !torch.vtensor<[],f32> -> !torch.float | |
| %2046 = torch.aten.item %2044 : !torch.vtensor<[],si8> -> !torch.int | |
| %2047 = torch.aten.quantize_per_tensor %2042, %2045, %2046, %int12_525 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2048 = torch.aten.int_repr %2047 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %2049 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2050 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2051 = torch.aten.item %2049 : !torch.vtensor<[],f32> -> !torch.float | |
| %2052 = torch.aten.item %2050 : !torch.vtensor<[],si8> -> !torch.int | |
| %2053 = torch.aten._make_per_tensor_quantized_tensor %2048, %2051, %2052 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2054 = torch.aten.dequantize.self %2053 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2055 = torch.aten.mul.Tensor %2032, %2054 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2056 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2057 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_526 = torch.constant.int 12 | |
| %2058 = torch.aten.item %2056 : !torch.vtensor<[],f32> -> !torch.float | |
| %2059 = torch.aten.item %2057 : !torch.vtensor<[],si8> -> !torch.int | |
| %2060 = torch.aten.quantize_per_tensor %2055, %2058, %2059, %int12_526 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2061 = torch.aten.int_repr %2060 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %2062 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2063 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2064 = torch.aten.item %2062 : !torch.vtensor<[],f32> -> !torch.float | |
| %2065 = torch.aten.item %2063 : !torch.vtensor<[],si8> -> !torch.int | |
| %2066 = torch.aten._make_per_tensor_quantized_tensor %2061, %2064, %2065 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2067 = torch.aten.dequantize.self %2066 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2068 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2069 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_527 = torch.constant.int 12 | |
| %2070 = torch.aten.item %2068 : !torch.vtensor<[],f32> -> !torch.float | |
| %2071 = torch.aten.item %2069 : !torch.vtensor<[],si8> -> !torch.int | |
| %2072 = torch.aten.quantize_per_tensor %68, %2070, %2071, %int12_527 : !torch.vtensor<[184,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184,1,3,3],!torch.qint8> | |
| %2073 = torch.aten.int_repr %2072 : !torch.vtensor<[184,1,3,3],!torch.qint8> -> !torch.vtensor<[184,1,3,3],si8> | |
| %2074 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2075 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2076 = torch.aten.item %2074 : !torch.vtensor<[],f32> -> !torch.float | |
| %2077 = torch.aten.item %2075 : !torch.vtensor<[],si8> -> !torch.int | |
| %2078 = torch.aten._make_per_tensor_quantized_tensor %2073, %2076, %2077 : !torch.vtensor<[184,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[184,1,3,3],!torch.qint8> | |
| %2079 = torch.aten.dequantize.self %2078 : !torch.vtensor<[184,1,3,3],!torch.qint8> -> !torch.vtensor<[184,1,3,3],f32> | |
| %2080 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2081 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_528 = torch.constant.int 12 | |
| %2082 = torch.aten.item %2080 : !torch.vtensor<[],f32> -> !torch.float | |
| %2083 = torch.aten.item %2081 : !torch.vtensor<[],si8> -> !torch.int | |
| %2084 = torch.aten.quantize_per_tensor %69, %2082, %2083, %int12_528 : !torch.vtensor<[184],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %2085 = torch.aten.int_repr %2084 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],si8> | |
| %2086 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2087 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2088 = torch.aten.item %2086 : !torch.vtensor<[],f32> -> !torch.float | |
| %2089 = torch.aten.item %2087 : !torch.vtensor<[],si8> -> !torch.int | |
| %2090 = torch.aten._make_per_tensor_quantized_tensor %2085, %2088, %2089 : !torch.vtensor<[184],si8>, !torch.float, !torch.int -> !torch.vtensor<[184],!torch.qint8> | |
| %2091 = torch.aten.dequantize.self %2090 : !torch.vtensor<[184],!torch.qint8> -> !torch.vtensor<[184],f32> | |
| %int1_529 = torch.constant.int 1 | |
| %int1_530 = torch.constant.int 1 | |
| %int1_531 = torch.constant.int 1 | |
| %int1_532 = torch.constant.int 1 | |
| %int1_533 = torch.constant.int 1 | |
| %int1_534 = torch.constant.int 1 | |
| %int0_535 = torch.constant.int 0 | |
| %2092 = torch.prim.ListConstruct %int1_529, %int1_530 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2093 = torch.prim.ListConstruct %int1_531, %int1_532 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2094 = torch.prim.ListConstruct %int1_533, %int1_534 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2095 = torch.prim.ListConstruct %int0_535, %int0_535 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_536 = torch.constant.bool false | |
| %int184_537 = torch.constant.int 184 | |
| %2096 = torch.aten.convolution %2067, %2079, %2091, %2094, %2092, %2093, %false_536, %2095, %int184_537 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[184,1,3,3],f32>, !torch.vtensor<[184],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %2097 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2098 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_538 = torch.constant.int 12 | |
| %2099 = torch.aten.item %2097 : !torch.vtensor<[],f32> -> !torch.float | |
| %2100 = torch.aten.item %2098 : !torch.vtensor<[],si8> -> !torch.int | |
| %2101 = torch.aten.quantize_per_tensor %2096, %2099, %2100, %int12_538 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2102 = torch.aten.int_repr %2101 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %2103 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2104 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2105 = torch.aten.item %2103 : !torch.vtensor<[],f32> -> !torch.float | |
| %2106 = torch.aten.item %2104 : !torch.vtensor<[],si8> -> !torch.int | |
| %2107 = torch.aten._make_per_tensor_quantized_tensor %2102, %2105, %2106 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2108 = torch.aten.dequantize.self %2107 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2109 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_539 = torch.constant.int 1 | |
| %2110 = torch.aten.add.Tensor %2108, %2109, %int1_539 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,184,14,14],f32> | |
| %2111 = torch.aten.relu %2110 : !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2112 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2113 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_540 = torch.constant.int 6 | |
| %none_541 = torch.constant.none | |
| %false_542 = torch.constant.bool false | |
| %2114 = torch.aten.to.dtype %2112, %int6_540, %false_542, %false_542, %none_541 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_543 = torch.constant.int 6 | |
| %none_544 = torch.constant.none | |
| %false_545 = torch.constant.bool false | |
| %2115 = torch.aten.to.dtype %2113, %int6_543, %false_545, %false_545, %none_544 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2116 = torch.aten.clamp.Tensor %2111, %2114, %2115 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2117 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2118 = torch.aten.mul.Tensor %2116, %2117 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2119 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2120 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_546 = torch.constant.int 12 | |
| %2121 = torch.aten.item %2119 : !torch.vtensor<[],f32> -> !torch.float | |
| %2122 = torch.aten.item %2120 : !torch.vtensor<[],si8> -> !torch.int | |
| %2123 = torch.aten.quantize_per_tensor %2118, %2121, %2122, %int12_546 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2124 = torch.aten.int_repr %2123 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %2125 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2126 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2127 = torch.aten.item %2125 : !torch.vtensor<[],f32> -> !torch.float | |
| %2128 = torch.aten.item %2126 : !torch.vtensor<[],si8> -> !torch.int | |
| %2129 = torch.aten._make_per_tensor_quantized_tensor %2124, %2127, %2128 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2130 = torch.aten.dequantize.self %2129 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2131 = torch.aten.mul.Tensor %2108, %2130 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[1,184,14,14],f32> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2132 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2133 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_547 = torch.constant.int 12 | |
| %2134 = torch.aten.item %2132 : !torch.vtensor<[],f32> -> !torch.float | |
| %2135 = torch.aten.item %2133 : !torch.vtensor<[],si8> -> !torch.int | |
| %2136 = torch.aten.quantize_per_tensor %2131, %2134, %2135, %int12_547 : !torch.vtensor<[1,184,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2137 = torch.aten.int_repr %2136 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],si8> | |
| %2138 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2139 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2140 = torch.aten.item %2138 : !torch.vtensor<[],f32> -> !torch.float | |
| %2141 = torch.aten.item %2139 : !torch.vtensor<[],si8> -> !torch.int | |
| %2142 = torch.aten._make_per_tensor_quantized_tensor %2137, %2140, %2141 : !torch.vtensor<[1,184,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,184,14,14],!torch.qint8> | |
| %2143 = torch.aten.dequantize.self %2142 : !torch.vtensor<[1,184,14,14],!torch.qint8> -> !torch.vtensor<[1,184,14,14],f32> | |
| %2144 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2145 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_548 = torch.constant.int 12 | |
| %2146 = torch.aten.item %2144 : !torch.vtensor<[],f32> -> !torch.float | |
| %2147 = torch.aten.item %2145 : !torch.vtensor<[],si8> -> !torch.int | |
| %2148 = torch.aten.quantize_per_tensor %70, %2146, %2147, %int12_548 : !torch.vtensor<[80,184,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80,184,1,1],!torch.qint8> | |
| %2149 = torch.aten.int_repr %2148 : !torch.vtensor<[80,184,1,1],!torch.qint8> -> !torch.vtensor<[80,184,1,1],si8> | |
| %2150 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2151 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2152 = torch.aten.item %2150 : !torch.vtensor<[],f32> -> !torch.float | |
| %2153 = torch.aten.item %2151 : !torch.vtensor<[],si8> -> !torch.int | |
| %2154 = torch.aten._make_per_tensor_quantized_tensor %2149, %2152, %2153 : !torch.vtensor<[80,184,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[80,184,1,1],!torch.qint8> | |
| %2155 = torch.aten.dequantize.self %2154 : !torch.vtensor<[80,184,1,1],!torch.qint8> -> !torch.vtensor<[80,184,1,1],f32> | |
| %2156 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2157 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_549 = torch.constant.int 12 | |
| %2158 = torch.aten.item %2156 : !torch.vtensor<[],f32> -> !torch.float | |
| %2159 = torch.aten.item %2157 : !torch.vtensor<[],si8> -> !torch.int | |
| %2160 = torch.aten.quantize_per_tensor %71, %2158, %2159, %int12_549 : !torch.vtensor<[80],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %2161 = torch.aten.int_repr %2160 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],si8> | |
| %2162 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2163 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2164 = torch.aten.item %2162 : !torch.vtensor<[],f32> -> !torch.float | |
| %2165 = torch.aten.item %2163 : !torch.vtensor<[],si8> -> !torch.int | |
| %2166 = torch.aten._make_per_tensor_quantized_tensor %2161, %2164, %2165 : !torch.vtensor<[80],si8>, !torch.float, !torch.int -> !torch.vtensor<[80],!torch.qint8> | |
| %2167 = torch.aten.dequantize.self %2166 : !torch.vtensor<[80],!torch.qint8> -> !torch.vtensor<[80],f32> | |
| %int0_550 = torch.constant.int 0 | |
| %int0_551 = torch.constant.int 0 | |
| %int1_552 = torch.constant.int 1 | |
| %int1_553 = torch.constant.int 1 | |
| %int1_554 = torch.constant.int 1 | |
| %int1_555 = torch.constant.int 1 | |
| %int0_556 = torch.constant.int 0 | |
| %2168 = torch.prim.ListConstruct %int0_550, %int0_551 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2169 = torch.prim.ListConstruct %int1_552, %int1_553 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2170 = torch.prim.ListConstruct %int1_554, %int1_555 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2171 = torch.prim.ListConstruct %int0_556, %int0_556 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_557 = torch.constant.bool false | |
| %int1_558 = torch.constant.int 1 | |
| %2172 = torch.aten.convolution %2143, %2155, %2167, %2170, %2168, %2169, %false_557, %2171, %int1_558 : !torch.vtensor<[1,184,14,14],f32>, !torch.vtensor<[80,184,1,1],f32>, !torch.vtensor<[80],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %2173 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2174 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_559 = torch.constant.int 12 | |
| %2175 = torch.aten.item %2173 : !torch.vtensor<[],f32> -> !torch.float | |
| %2176 = torch.aten.item %2174 : !torch.vtensor<[],si8> -> !torch.int | |
| %2177 = torch.aten.quantize_per_tensor %2172, %2175, %2176, %int12_559 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %2178 = torch.aten.int_repr %2177 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %2179 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2180 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2181 = torch.aten.item %2179 : !torch.vtensor<[],f32> -> !torch.float | |
| %2182 = torch.aten.item %2180 : !torch.vtensor<[],si8> -> !torch.int | |
| %2183 = torch.aten._make_per_tensor_quantized_tensor %2178, %2181, %2182 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %2184 = torch.aten.dequantize.self %2183 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %int1_560 = torch.constant.int 1 | |
| %2185 = torch.aten.add.Tensor %2184, %1991, %int1_560 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[1,80,14,14],f32>, !torch.int -> !torch.vtensor<[1,80,14,14],f32> | |
| %2186 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2187 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_561 = torch.constant.int 12 | |
| %2188 = torch.aten.item %2186 : !torch.vtensor<[],f32> -> !torch.float | |
| %2189 = torch.aten.item %2187 : !torch.vtensor<[],si8> -> !torch.int | |
| %2190 = torch.aten.quantize_per_tensor %2185, %2188, %2189, %int12_561 : !torch.vtensor<[1,80,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %2191 = torch.aten.int_repr %2190 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],si8> | |
| %2192 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2193 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2194 = torch.aten.item %2192 : !torch.vtensor<[],f32> -> !torch.float | |
| %2195 = torch.aten.item %2193 : !torch.vtensor<[],si8> -> !torch.int | |
| %2196 = torch.aten._make_per_tensor_quantized_tensor %2191, %2194, %2195 : !torch.vtensor<[1,80,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,80,14,14],!torch.qint8> | |
| %2197 = torch.aten.dequantize.self %2196 : !torch.vtensor<[1,80,14,14],!torch.qint8> -> !torch.vtensor<[1,80,14,14],f32> | |
| %2198 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2199 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_562 = torch.constant.int 12 | |
| %2200 = torch.aten.item %2198 : !torch.vtensor<[],f32> -> !torch.float | |
| %2201 = torch.aten.item %2199 : !torch.vtensor<[],si8> -> !torch.int | |
| %2202 = torch.aten.quantize_per_tensor %72, %2200, %2201, %int12_562 : !torch.vtensor<[480,80,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[480,80,1,1],!torch.qint8> | |
| %2203 = torch.aten.int_repr %2202 : !torch.vtensor<[480,80,1,1],!torch.qint8> -> !torch.vtensor<[480,80,1,1],si8> | |
| %2204 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2205 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2206 = torch.aten.item %2204 : !torch.vtensor<[],f32> -> !torch.float | |
| %2207 = torch.aten.item %2205 : !torch.vtensor<[],si8> -> !torch.int | |
| %2208 = torch.aten._make_per_tensor_quantized_tensor %2203, %2206, %2207 : !torch.vtensor<[480,80,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[480,80,1,1],!torch.qint8> | |
| %2209 = torch.aten.dequantize.self %2208 : !torch.vtensor<[480,80,1,1],!torch.qint8> -> !torch.vtensor<[480,80,1,1],f32> | |
| %2210 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2211 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_563 = torch.constant.int 12 | |
| %2212 = torch.aten.item %2210 : !torch.vtensor<[],f32> -> !torch.float | |
| %2213 = torch.aten.item %2211 : !torch.vtensor<[],si8> -> !torch.int | |
| %2214 = torch.aten.quantize_per_tensor %73, %2212, %2213, %int12_563 : !torch.vtensor<[480],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[480],!torch.qint8> | |
| %2215 = torch.aten.int_repr %2214 : !torch.vtensor<[480],!torch.qint8> -> !torch.vtensor<[480],si8> | |
| %2216 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2217 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2218 = torch.aten.item %2216 : !torch.vtensor<[],f32> -> !torch.float | |
| %2219 = torch.aten.item %2217 : !torch.vtensor<[],si8> -> !torch.int | |
| %2220 = torch.aten._make_per_tensor_quantized_tensor %2215, %2218, %2219 : !torch.vtensor<[480],si8>, !torch.float, !torch.int -> !torch.vtensor<[480],!torch.qint8> | |
| %2221 = torch.aten.dequantize.self %2220 : !torch.vtensor<[480],!torch.qint8> -> !torch.vtensor<[480],f32> | |
| %int0_564 = torch.constant.int 0 | |
| %int0_565 = torch.constant.int 0 | |
| %int1_566 = torch.constant.int 1 | |
| %int1_567 = torch.constant.int 1 | |
| %int1_568 = torch.constant.int 1 | |
| %int1_569 = torch.constant.int 1 | |
| %int0_570 = torch.constant.int 0 | |
| %2222 = torch.prim.ListConstruct %int0_564, %int0_565 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2223 = torch.prim.ListConstruct %int1_566, %int1_567 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2224 = torch.prim.ListConstruct %int1_568, %int1_569 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2225 = torch.prim.ListConstruct %int0_570, %int0_570 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_571 = torch.constant.bool false | |
| %int1_572 = torch.constant.int 1 | |
| %2226 = torch.aten.convolution %2197, %2209, %2221, %2224, %2222, %2223, %false_571, %2225, %int1_572 : !torch.vtensor<[1,80,14,14],f32>, !torch.vtensor<[480,80,1,1],f32>, !torch.vtensor<[480],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,480,14,14],f32> | |
| %2227 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2228 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_573 = torch.constant.int 12 | |
| %2229 = torch.aten.item %2227 : !torch.vtensor<[],f32> -> !torch.float | |
| %2230 = torch.aten.item %2228 : !torch.vtensor<[],si8> -> !torch.int | |
| %2231 = torch.aten.quantize_per_tensor %2226, %2229, %2230, %int12_573 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2232 = torch.aten.int_repr %2231 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2233 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2234 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2235 = torch.aten.item %2233 : !torch.vtensor<[],f32> -> !torch.float | |
| %2236 = torch.aten.item %2234 : !torch.vtensor<[],si8> -> !torch.int | |
| %2237 = torch.aten._make_per_tensor_quantized_tensor %2232, %2235, %2236 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2238 = torch.aten.dequantize.self %2237 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2239 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_574 = torch.constant.int 1 | |
| %2240 = torch.aten.add.Tensor %2238, %2239, %int1_574 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,480,14,14],f32> | |
| %2241 = torch.aten.relu %2240 : !torch.vtensor<[1,480,14,14],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2242 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2243 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_575 = torch.constant.int 6 | |
| %none_576 = torch.constant.none | |
| %false_577 = torch.constant.bool false | |
| %2244 = torch.aten.to.dtype %2242, %int6_575, %false_577, %false_577, %none_576 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_578 = torch.constant.int 6 | |
| %none_579 = torch.constant.none | |
| %false_580 = torch.constant.bool false | |
| %2245 = torch.aten.to.dtype %2243, %int6_578, %false_580, %false_580, %none_579 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2246 = torch.aten.clamp.Tensor %2241, %2244, %2245 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2247 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2248 = torch.aten.mul.Tensor %2246, %2247 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2249 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2250 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_581 = torch.constant.int 12 | |
| %2251 = torch.aten.item %2249 : !torch.vtensor<[],f32> -> !torch.float | |
| %2252 = torch.aten.item %2250 : !torch.vtensor<[],si8> -> !torch.int | |
| %2253 = torch.aten.quantize_per_tensor %2248, %2251, %2252, %int12_581 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2254 = torch.aten.int_repr %2253 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2255 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2256 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2257 = torch.aten.item %2255 : !torch.vtensor<[],f32> -> !torch.float | |
| %2258 = torch.aten.item %2256 : !torch.vtensor<[],si8> -> !torch.int | |
| %2259 = torch.aten._make_per_tensor_quantized_tensor %2254, %2257, %2258 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2260 = torch.aten.dequantize.self %2259 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2261 = torch.aten.mul.Tensor %2238, %2260 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[1,480,14,14],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2262 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2263 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_582 = torch.constant.int 12 | |
| %2264 = torch.aten.item %2262 : !torch.vtensor<[],f32> -> !torch.float | |
| %2265 = torch.aten.item %2263 : !torch.vtensor<[],si8> -> !torch.int | |
| %2266 = torch.aten.quantize_per_tensor %2261, %2264, %2265, %int12_582 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2267 = torch.aten.int_repr %2266 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2268 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2269 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2270 = torch.aten.item %2268 : !torch.vtensor<[],f32> -> !torch.float | |
| %2271 = torch.aten.item %2269 : !torch.vtensor<[],si8> -> !torch.int | |
| %2272 = torch.aten._make_per_tensor_quantized_tensor %2267, %2270, %2271 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2273 = torch.aten.dequantize.self %2272 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2274 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2275 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_583 = torch.constant.int 12 | |
| %2276 = torch.aten.item %2274 : !torch.vtensor<[],f32> -> !torch.float | |
| %2277 = torch.aten.item %2275 : !torch.vtensor<[],si8> -> !torch.int | |
| %2278 = torch.aten.quantize_per_tensor %74, %2276, %2277, %int12_583 : !torch.vtensor<[480,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[480,1,3,3],!torch.qint8> | |
| %2279 = torch.aten.int_repr %2278 : !torch.vtensor<[480,1,3,3],!torch.qint8> -> !torch.vtensor<[480,1,3,3],si8> | |
| %2280 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2281 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2282 = torch.aten.item %2280 : !torch.vtensor<[],f32> -> !torch.float | |
| %2283 = torch.aten.item %2281 : !torch.vtensor<[],si8> -> !torch.int | |
| %2284 = torch.aten._make_per_tensor_quantized_tensor %2279, %2282, %2283 : !torch.vtensor<[480,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[480,1,3,3],!torch.qint8> | |
| %2285 = torch.aten.dequantize.self %2284 : !torch.vtensor<[480,1,3,3],!torch.qint8> -> !torch.vtensor<[480,1,3,3],f32> | |
| %2286 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2287 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_584 = torch.constant.int 12 | |
| %2288 = torch.aten.item %2286 : !torch.vtensor<[],f32> -> !torch.float | |
| %2289 = torch.aten.item %2287 : !torch.vtensor<[],si8> -> !torch.int | |
| %2290 = torch.aten.quantize_per_tensor %75, %2288, %2289, %int12_584 : !torch.vtensor<[480],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[480],!torch.qint8> | |
| %2291 = torch.aten.int_repr %2290 : !torch.vtensor<[480],!torch.qint8> -> !torch.vtensor<[480],si8> | |
| %2292 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2293 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2294 = torch.aten.item %2292 : !torch.vtensor<[],f32> -> !torch.float | |
| %2295 = torch.aten.item %2293 : !torch.vtensor<[],si8> -> !torch.int | |
| %2296 = torch.aten._make_per_tensor_quantized_tensor %2291, %2294, %2295 : !torch.vtensor<[480],si8>, !torch.float, !torch.int -> !torch.vtensor<[480],!torch.qint8> | |
| %2297 = torch.aten.dequantize.self %2296 : !torch.vtensor<[480],!torch.qint8> -> !torch.vtensor<[480],f32> | |
| %int1_585 = torch.constant.int 1 | |
| %int1_586 = torch.constant.int 1 | |
| %int1_587 = torch.constant.int 1 | |
| %int1_588 = torch.constant.int 1 | |
| %int1_589 = torch.constant.int 1 | |
| %int1_590 = torch.constant.int 1 | |
| %int0_591 = torch.constant.int 0 | |
| %2298 = torch.prim.ListConstruct %int1_585, %int1_586 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2299 = torch.prim.ListConstruct %int1_587, %int1_588 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2300 = torch.prim.ListConstruct %int1_589, %int1_590 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2301 = torch.prim.ListConstruct %int0_591, %int0_591 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_592 = torch.constant.bool false | |
| %int480 = torch.constant.int 480 | |
| %2302 = torch.aten.convolution %2273, %2285, %2297, %2300, %2298, %2299, %false_592, %2301, %int480 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[480,1,3,3],f32>, !torch.vtensor<[480],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,480,14,14],f32> | |
| %2303 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2304 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_593 = torch.constant.int 12 | |
| %2305 = torch.aten.item %2303 : !torch.vtensor<[],f32> -> !torch.float | |
| %2306 = torch.aten.item %2304 : !torch.vtensor<[],si8> -> !torch.int | |
| %2307 = torch.aten.quantize_per_tensor %2302, %2305, %2306, %int12_593 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2308 = torch.aten.int_repr %2307 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2309 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2310 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2311 = torch.aten.item %2309 : !torch.vtensor<[],f32> -> !torch.float | |
| %2312 = torch.aten.item %2310 : !torch.vtensor<[],si8> -> !torch.int | |
| %2313 = torch.aten._make_per_tensor_quantized_tensor %2308, %2311, %2312 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2314 = torch.aten.dequantize.self %2313 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2315 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_594 = torch.constant.int 1 | |
| %2316 = torch.aten.add.Tensor %2314, %2315, %int1_594 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,480,14,14],f32> | |
| %2317 = torch.aten.relu %2316 : !torch.vtensor<[1,480,14,14],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2318 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2319 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_595 = torch.constant.int 6 | |
| %none_596 = torch.constant.none | |
| %false_597 = torch.constant.bool false | |
| %2320 = torch.aten.to.dtype %2318, %int6_595, %false_597, %false_597, %none_596 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_598 = torch.constant.int 6 | |
| %none_599 = torch.constant.none | |
| %false_600 = torch.constant.bool false | |
| %2321 = torch.aten.to.dtype %2319, %int6_598, %false_600, %false_600, %none_599 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2322 = torch.aten.clamp.Tensor %2317, %2320, %2321 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2323 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2324 = torch.aten.mul.Tensor %2322, %2323 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2325 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2326 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_601 = torch.constant.int 12 | |
| %2327 = torch.aten.item %2325 : !torch.vtensor<[],f32> -> !torch.float | |
| %2328 = torch.aten.item %2326 : !torch.vtensor<[],si8> -> !torch.int | |
| %2329 = torch.aten.quantize_per_tensor %2324, %2327, %2328, %int12_601 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2330 = torch.aten.int_repr %2329 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2331 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2332 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2333 = torch.aten.item %2331 : !torch.vtensor<[],f32> -> !torch.float | |
| %2334 = torch.aten.item %2332 : !torch.vtensor<[],si8> -> !torch.int | |
| %2335 = torch.aten._make_per_tensor_quantized_tensor %2330, %2333, %2334 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2336 = torch.aten.dequantize.self %2335 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2337 = torch.aten.mul.Tensor %2314, %2336 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[1,480,14,14],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2338 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2339 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_602 = torch.constant.int 12 | |
| %2340 = torch.aten.item %2338 : !torch.vtensor<[],f32> -> !torch.float | |
| %2341 = torch.aten.item %2339 : !torch.vtensor<[],si8> -> !torch.int | |
| %2342 = torch.aten.quantize_per_tensor %2337, %2340, %2341, %int12_602 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2343 = torch.aten.int_repr %2342 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2344 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2345 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2346 = torch.aten.item %2344 : !torch.vtensor<[],f32> -> !torch.float | |
| %2347 = torch.aten.item %2345 : !torch.vtensor<[],si8> -> !torch.int | |
| %2348 = torch.aten._make_per_tensor_quantized_tensor %2343, %2346, %2347 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2349 = torch.aten.dequantize.self %2348 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %int0_603 = torch.constant.int 0 | |
| %int1_604 = torch.constant.int 1 | |
| %int14 = torch.constant.int 14 | |
| %int14_605 = torch.constant.int 14 | |
| %2350 = torch.prim.ListConstruct %int14, %int14_605 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2351 = torch.prim.ListConstruct %int0_603, %int0_603 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2352 = torch.prim.ListConstruct %int1_604, %int1_604 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_606 = torch.constant.bool false | |
| %none_607 = torch.constant.none | |
| %2353 = torch.aten.avg_pool2d %2349, %2350, %2352, %2351, %false_606, %false_606, %none_607 : !torch.vtensor<[1,480,14,14],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,480,1,1],f32> | |
| %2354 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2355 = torch.aten.mul.Tensor %2353, %2354 : !torch.vtensor<[1,480,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2356 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2357 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_608 = torch.constant.int 12 | |
| %2358 = torch.aten.item %2356 : !torch.vtensor<[],f32> -> !torch.float | |
| %2359 = torch.aten.item %2357 : !torch.vtensor<[],si8> -> !torch.int | |
| %2360 = torch.aten.quantize_per_tensor %2355, %2358, %2359, %int12_608 : !torch.vtensor<[1,480,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,1,1],!torch.qint8> | |
| %2361 = torch.aten.int_repr %2360 : !torch.vtensor<[1,480,1,1],!torch.qint8> -> !torch.vtensor<[1,480,1,1],si8> | |
| %2362 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2363 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2364 = torch.aten.item %2362 : !torch.vtensor<[],f32> -> !torch.float | |
| %2365 = torch.aten.item %2363 : !torch.vtensor<[],si8> -> !torch.int | |
| %2366 = torch.aten._make_per_tensor_quantized_tensor %2361, %2364, %2365 : !torch.vtensor<[1,480,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,1,1],!torch.qint8> | |
| %2367 = torch.aten.dequantize.self %2366 : !torch.vtensor<[1,480,1,1],!torch.qint8> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2368 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2369 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_609 = torch.constant.int 12 | |
| %2370 = torch.aten.item %2368 : !torch.vtensor<[],f32> -> !torch.float | |
| %2371 = torch.aten.item %2369 : !torch.vtensor<[],si8> -> !torch.int | |
| %2372 = torch.aten.quantize_per_tensor %76, %2370, %2371, %int12_609 : !torch.vtensor<[120,480,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120,480,1,1],!torch.qint8> | |
| %2373 = torch.aten.int_repr %2372 : !torch.vtensor<[120,480,1,1],!torch.qint8> -> !torch.vtensor<[120,480,1,1],si8> | |
| %2374 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2375 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2376 = torch.aten.item %2374 : !torch.vtensor<[],f32> -> !torch.float | |
| %2377 = torch.aten.item %2375 : !torch.vtensor<[],si8> -> !torch.int | |
| %2378 = torch.aten._make_per_tensor_quantized_tensor %2373, %2376, %2377 : !torch.vtensor<[120,480,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[120,480,1,1],!torch.qint8> | |
| %2379 = torch.aten.dequantize.self %2378 : !torch.vtensor<[120,480,1,1],!torch.qint8> -> !torch.vtensor<[120,480,1,1],f32> | |
| %2380 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2381 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_610 = torch.constant.int 12 | |
| %2382 = torch.aten.item %2380 : !torch.vtensor<[],f32> -> !torch.float | |
| %2383 = torch.aten.item %2381 : !torch.vtensor<[],si8> -> !torch.int | |
| %2384 = torch.aten.quantize_per_tensor %77, %2382, %2383, %int12_610 : !torch.vtensor<[120],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %2385 = torch.aten.int_repr %2384 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],si8> | |
| %2386 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2387 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2388 = torch.aten.item %2386 : !torch.vtensor<[],f32> -> !torch.float | |
| %2389 = torch.aten.item %2387 : !torch.vtensor<[],si8> -> !torch.int | |
| %2390 = torch.aten._make_per_tensor_quantized_tensor %2385, %2388, %2389 : !torch.vtensor<[120],si8>, !torch.float, !torch.int -> !torch.vtensor<[120],!torch.qint8> | |
| %2391 = torch.aten.dequantize.self %2390 : !torch.vtensor<[120],!torch.qint8> -> !torch.vtensor<[120],f32> | |
| %int0_611 = torch.constant.int 0 | |
| %int0_612 = torch.constant.int 0 | |
| %int1_613 = torch.constant.int 1 | |
| %int1_614 = torch.constant.int 1 | |
| %int1_615 = torch.constant.int 1 | |
| %int1_616 = torch.constant.int 1 | |
| %int0_617 = torch.constant.int 0 | |
| %2392 = torch.prim.ListConstruct %int0_611, %int0_612 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2393 = torch.prim.ListConstruct %int1_613, %int1_614 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2394 = torch.prim.ListConstruct %int1_615, %int1_616 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2395 = torch.prim.ListConstruct %int0_617, %int0_617 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_618 = torch.constant.bool false | |
| %int1_619 = torch.constant.int 1 | |
| %2396 = torch.aten.convolution %2367, %2379, %2391, %2394, %2392, %2393, %false_618, %2395, %int1_619 : !torch.vtensor<[1,480,1,1],f32>, !torch.vtensor<[120,480,1,1],f32>, !torch.vtensor<[120],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,120,1,1],f32> | |
| %2397 = torch.aten.relu %2396 : !torch.vtensor<[1,120,1,1],f32> -> !torch.vtensor<[1,120,1,1],f32> | |
| %2398 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2399 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_620 = torch.constant.int 12 | |
| %2400 = torch.aten.item %2398 : !torch.vtensor<[],f32> -> !torch.float | |
| %2401 = torch.aten.item %2399 : !torch.vtensor<[],si8> -> !torch.int | |
| %2402 = torch.aten.quantize_per_tensor %2397, %2400, %2401, %int12_620 : !torch.vtensor<[1,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %2403 = torch.aten.int_repr %2402 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],si8> | |
| %2404 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2405 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2406 = torch.aten.item %2404 : !torch.vtensor<[],f32> -> !torch.float | |
| %2407 = torch.aten.item %2405 : !torch.vtensor<[],si8> -> !torch.int | |
| %2408 = torch.aten._make_per_tensor_quantized_tensor %2403, %2406, %2407 : !torch.vtensor<[1,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,120,1,1],!torch.qint8> | |
| %2409 = torch.aten.dequantize.self %2408 : !torch.vtensor<[1,120,1,1],!torch.qint8> -> !torch.vtensor<[1,120,1,1],f32> | |
| %2410 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2411 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_621 = torch.constant.int 12 | |
| %2412 = torch.aten.item %2410 : !torch.vtensor<[],f32> -> !torch.float | |
| %2413 = torch.aten.item %2411 : !torch.vtensor<[],si8> -> !torch.int | |
| %2414 = torch.aten.quantize_per_tensor %78, %2412, %2413, %int12_621 : !torch.vtensor<[480,120,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[480,120,1,1],!torch.qint8> | |
| %2415 = torch.aten.int_repr %2414 : !torch.vtensor<[480,120,1,1],!torch.qint8> -> !torch.vtensor<[480,120,1,1],si8> | |
| %2416 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2417 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2418 = torch.aten.item %2416 : !torch.vtensor<[],f32> -> !torch.float | |
| %2419 = torch.aten.item %2417 : !torch.vtensor<[],si8> -> !torch.int | |
| %2420 = torch.aten._make_per_tensor_quantized_tensor %2415, %2418, %2419 : !torch.vtensor<[480,120,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[480,120,1,1],!torch.qint8> | |
| %2421 = torch.aten.dequantize.self %2420 : !torch.vtensor<[480,120,1,1],!torch.qint8> -> !torch.vtensor<[480,120,1,1],f32> | |
| %2422 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2423 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_622 = torch.constant.int 12 | |
| %2424 = torch.aten.item %2422 : !torch.vtensor<[],f32> -> !torch.float | |
| %2425 = torch.aten.item %2423 : !torch.vtensor<[],si8> -> !torch.int | |
| %2426 = torch.aten.quantize_per_tensor %79, %2424, %2425, %int12_622 : !torch.vtensor<[480],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[480],!torch.qint8> | |
| %2427 = torch.aten.int_repr %2426 : !torch.vtensor<[480],!torch.qint8> -> !torch.vtensor<[480],si8> | |
| %2428 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2429 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2430 = torch.aten.item %2428 : !torch.vtensor<[],f32> -> !torch.float | |
| %2431 = torch.aten.item %2429 : !torch.vtensor<[],si8> -> !torch.int | |
| %2432 = torch.aten._make_per_tensor_quantized_tensor %2427, %2430, %2431 : !torch.vtensor<[480],si8>, !torch.float, !torch.int -> !torch.vtensor<[480],!torch.qint8> | |
| %2433 = torch.aten.dequantize.self %2432 : !torch.vtensor<[480],!torch.qint8> -> !torch.vtensor<[480],f32> | |
| %int0_623 = torch.constant.int 0 | |
| %int0_624 = torch.constant.int 0 | |
| %int1_625 = torch.constant.int 1 | |
| %int1_626 = torch.constant.int 1 | |
| %int1_627 = torch.constant.int 1 | |
| %int1_628 = torch.constant.int 1 | |
| %int0_629 = torch.constant.int 0 | |
| %2434 = torch.prim.ListConstruct %int0_623, %int0_624 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2435 = torch.prim.ListConstruct %int1_625, %int1_626 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2436 = torch.prim.ListConstruct %int1_627, %int1_628 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2437 = torch.prim.ListConstruct %int0_629, %int0_629 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_630 = torch.constant.bool false | |
| %int1_631 = torch.constant.int 1 | |
| %2438 = torch.aten.convolution %2409, %2421, %2433, %2436, %2434, %2435, %false_630, %2437, %int1_631 : !torch.vtensor<[1,120,1,1],f32>, !torch.vtensor<[480,120,1,1],f32>, !torch.vtensor<[480],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,480,1,1],f32> | |
| %2439 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2440 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_632 = torch.constant.int 12 | |
| %2441 = torch.aten.item %2439 : !torch.vtensor<[],f32> -> !torch.float | |
| %2442 = torch.aten.item %2440 : !torch.vtensor<[],si8> -> !torch.int | |
| %2443 = torch.aten.quantize_per_tensor %2438, %2441, %2442, %int12_632 : !torch.vtensor<[1,480,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,1,1],!torch.qint8> | |
| %2444 = torch.aten.int_repr %2443 : !torch.vtensor<[1,480,1,1],!torch.qint8> -> !torch.vtensor<[1,480,1,1],si8> | |
| %2445 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2446 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2447 = torch.aten.item %2445 : !torch.vtensor<[],f32> -> !torch.float | |
| %2448 = torch.aten.item %2446 : !torch.vtensor<[],si8> -> !torch.int | |
| %2449 = torch.aten._make_per_tensor_quantized_tensor %2444, %2447, %2448 : !torch.vtensor<[1,480,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,1,1],!torch.qint8> | |
| %2450 = torch.aten.dequantize.self %2449 : !torch.vtensor<[1,480,1,1],!torch.qint8> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2451 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_633 = torch.constant.int 1 | |
| %2452 = torch.aten.add.Tensor %2450, %2451, %int1_633 : !torch.vtensor<[1,480,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,480,1,1],f32> | |
| %2453 = torch.aten.relu %2452 : !torch.vtensor<[1,480,1,1],f32> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2454 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2455 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_634 = torch.constant.int 6 | |
| %none_635 = torch.constant.none | |
| %false_636 = torch.constant.bool false | |
| %2456 = torch.aten.to.dtype %2454, %int6_634, %false_636, %false_636, %none_635 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_637 = torch.constant.int 6 | |
| %none_638 = torch.constant.none | |
| %false_639 = torch.constant.bool false | |
| %2457 = torch.aten.to.dtype %2455, %int6_637, %false_639, %false_639, %none_638 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2458 = torch.aten.clamp.Tensor %2453, %2456, %2457 : !torch.vtensor<[1,480,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2459 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2460 = torch.aten.mul.Tensor %2458, %2459 : !torch.vtensor<[1,480,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2461 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2462 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_640 = torch.constant.int 12 | |
| %2463 = torch.aten.item %2461 : !torch.vtensor<[],f32> -> !torch.float | |
| %2464 = torch.aten.item %2462 : !torch.vtensor<[],si8> -> !torch.int | |
| %2465 = torch.aten.quantize_per_tensor %2460, %2463, %2464, %int12_640 : !torch.vtensor<[1,480,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,1,1],!torch.qint8> | |
| %2466 = torch.aten.int_repr %2465 : !torch.vtensor<[1,480,1,1],!torch.qint8> -> !torch.vtensor<[1,480,1,1],si8> | |
| %2467 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2468 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2469 = torch.aten.item %2467 : !torch.vtensor<[],f32> -> !torch.float | |
| %2470 = torch.aten.item %2468 : !torch.vtensor<[],si8> -> !torch.int | |
| %2471 = torch.aten._make_per_tensor_quantized_tensor %2466, %2469, %2470 : !torch.vtensor<[1,480,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,1,1],!torch.qint8> | |
| %2472 = torch.aten.dequantize.self %2471 : !torch.vtensor<[1,480,1,1],!torch.qint8> -> !torch.vtensor<[1,480,1,1],f32> | |
| %2473 = torch.aten.mul.Tensor %2472, %2349 : !torch.vtensor<[1,480,1,1],f32>, !torch.vtensor<[1,480,14,14],f32> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2474 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2475 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_641 = torch.constant.int 12 | |
| %2476 = torch.aten.item %2474 : !torch.vtensor<[],f32> -> !torch.float | |
| %2477 = torch.aten.item %2475 : !torch.vtensor<[],si8> -> !torch.int | |
| %2478 = torch.aten.quantize_per_tensor %2473, %2476, %2477, %int12_641 : !torch.vtensor<[1,480,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2479 = torch.aten.int_repr %2478 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],si8> | |
| %2480 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2481 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2482 = torch.aten.item %2480 : !torch.vtensor<[],f32> -> !torch.float | |
| %2483 = torch.aten.item %2481 : !torch.vtensor<[],si8> -> !torch.int | |
| %2484 = torch.aten._make_per_tensor_quantized_tensor %2479, %2482, %2483 : !torch.vtensor<[1,480,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,480,14,14],!torch.qint8> | |
| %2485 = torch.aten.dequantize.self %2484 : !torch.vtensor<[1,480,14,14],!torch.qint8> -> !torch.vtensor<[1,480,14,14],f32> | |
| %2486 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2487 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_642 = torch.constant.int 12 | |
| %2488 = torch.aten.item %2486 : !torch.vtensor<[],f32> -> !torch.float | |
| %2489 = torch.aten.item %2487 : !torch.vtensor<[],si8> -> !torch.int | |
| %2490 = torch.aten.quantize_per_tensor %80, %2488, %2489, %int12_642 : !torch.vtensor<[112,480,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[112,480,1,1],!torch.qint8> | |
| %2491 = torch.aten.int_repr %2490 : !torch.vtensor<[112,480,1,1],!torch.qint8> -> !torch.vtensor<[112,480,1,1],si8> | |
| %2492 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2493 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2494 = torch.aten.item %2492 : !torch.vtensor<[],f32> -> !torch.float | |
| %2495 = torch.aten.item %2493 : !torch.vtensor<[],si8> -> !torch.int | |
| %2496 = torch.aten._make_per_tensor_quantized_tensor %2491, %2494, %2495 : !torch.vtensor<[112,480,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[112,480,1,1],!torch.qint8> | |
| %2497 = torch.aten.dequantize.self %2496 : !torch.vtensor<[112,480,1,1],!torch.qint8> -> !torch.vtensor<[112,480,1,1],f32> | |
| %2498 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2499 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_643 = torch.constant.int 12 | |
| %2500 = torch.aten.item %2498 : !torch.vtensor<[],f32> -> !torch.float | |
| %2501 = torch.aten.item %2499 : !torch.vtensor<[],si8> -> !torch.int | |
| %2502 = torch.aten.quantize_per_tensor %81, %2500, %2501, %int12_643 : !torch.vtensor<[112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[112],!torch.qint8> | |
| %2503 = torch.aten.int_repr %2502 : !torch.vtensor<[112],!torch.qint8> -> !torch.vtensor<[112],si8> | |
| %2504 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2505 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2506 = torch.aten.item %2504 : !torch.vtensor<[],f32> -> !torch.float | |
| %2507 = torch.aten.item %2505 : !torch.vtensor<[],si8> -> !torch.int | |
| %2508 = torch.aten._make_per_tensor_quantized_tensor %2503, %2506, %2507 : !torch.vtensor<[112],si8>, !torch.float, !torch.int -> !torch.vtensor<[112],!torch.qint8> | |
| %2509 = torch.aten.dequantize.self %2508 : !torch.vtensor<[112],!torch.qint8> -> !torch.vtensor<[112],f32> | |
| %int0_644 = torch.constant.int 0 | |
| %int0_645 = torch.constant.int 0 | |
| %int1_646 = torch.constant.int 1 | |
| %int1_647 = torch.constant.int 1 | |
| %int1_648 = torch.constant.int 1 | |
| %int1_649 = torch.constant.int 1 | |
| %int0_650 = torch.constant.int 0 | |
| %2510 = torch.prim.ListConstruct %int0_644, %int0_645 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2511 = torch.prim.ListConstruct %int1_646, %int1_647 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2512 = torch.prim.ListConstruct %int1_648, %int1_649 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2513 = torch.prim.ListConstruct %int0_650, %int0_650 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_651 = torch.constant.bool false | |
| %int1_652 = torch.constant.int 1 | |
| %2514 = torch.aten.convolution %2485, %2497, %2509, %2512, %2510, %2511, %false_651, %2513, %int1_652 : !torch.vtensor<[1,480,14,14],f32>, !torch.vtensor<[112,480,1,1],f32>, !torch.vtensor<[112],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,112,14,14],f32> | |
| %2515 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2516 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_653 = torch.constant.int 12 | |
| %2517 = torch.aten.item %2515 : !torch.vtensor<[],f32> -> !torch.float | |
| %2518 = torch.aten.item %2516 : !torch.vtensor<[],si8> -> !torch.int | |
| %2519 = torch.aten.quantize_per_tensor %2514, %2517, %2518, %int12_653 : !torch.vtensor<[1,112,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,112,14,14],!torch.qint8> | |
| %2520 = torch.aten.int_repr %2519 : !torch.vtensor<[1,112,14,14],!torch.qint8> -> !torch.vtensor<[1,112,14,14],si8> | |
| %2521 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2522 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2523 = torch.aten.item %2521 : !torch.vtensor<[],f32> -> !torch.float | |
| %2524 = torch.aten.item %2522 : !torch.vtensor<[],si8> -> !torch.int | |
| %2525 = torch.aten._make_per_tensor_quantized_tensor %2520, %2523, %2524 : !torch.vtensor<[1,112,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,112,14,14],!torch.qint8> | |
| %2526 = torch.aten.dequantize.self %2525 : !torch.vtensor<[1,112,14,14],!torch.qint8> -> !torch.vtensor<[1,112,14,14],f32> | |
| %2527 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2528 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_654 = torch.constant.int 12 | |
| %2529 = torch.aten.item %2527 : !torch.vtensor<[],f32> -> !torch.float | |
| %2530 = torch.aten.item %2528 : !torch.vtensor<[],si8> -> !torch.int | |
| %2531 = torch.aten.quantize_per_tensor %82, %2529, %2530, %int12_654 : !torch.vtensor<[672,112,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672,112,1,1],!torch.qint8> | |
| %2532 = torch.aten.int_repr %2531 : !torch.vtensor<[672,112,1,1],!torch.qint8> -> !torch.vtensor<[672,112,1,1],si8> | |
| %2533 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2534 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2535 = torch.aten.item %2533 : !torch.vtensor<[],f32> -> !torch.float | |
| %2536 = torch.aten.item %2534 : !torch.vtensor<[],si8> -> !torch.int | |
| %2537 = torch.aten._make_per_tensor_quantized_tensor %2532, %2535, %2536 : !torch.vtensor<[672,112,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[672,112,1,1],!torch.qint8> | |
| %2538 = torch.aten.dequantize.self %2537 : !torch.vtensor<[672,112,1,1],!torch.qint8> -> !torch.vtensor<[672,112,1,1],f32> | |
| %2539 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2540 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_655 = torch.constant.int 12 | |
| %2541 = torch.aten.item %2539 : !torch.vtensor<[],f32> -> !torch.float | |
| %2542 = torch.aten.item %2540 : !torch.vtensor<[],si8> -> !torch.int | |
| %2543 = torch.aten.quantize_per_tensor %83, %2541, %2542, %int12_655 : !torch.vtensor<[672],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2544 = torch.aten.int_repr %2543 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],si8> | |
| %2545 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2546 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2547 = torch.aten.item %2545 : !torch.vtensor<[],f32> -> !torch.float | |
| %2548 = torch.aten.item %2546 : !torch.vtensor<[],si8> -> !torch.int | |
| %2549 = torch.aten._make_per_tensor_quantized_tensor %2544, %2547, %2548 : !torch.vtensor<[672],si8>, !torch.float, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2550 = torch.aten.dequantize.self %2549 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],f32> | |
| %int0_656 = torch.constant.int 0 | |
| %int0_657 = torch.constant.int 0 | |
| %int1_658 = torch.constant.int 1 | |
| %int1_659 = torch.constant.int 1 | |
| %int1_660 = torch.constant.int 1 | |
| %int1_661 = torch.constant.int 1 | |
| %int0_662 = torch.constant.int 0 | |
| %2551 = torch.prim.ListConstruct %int0_656, %int0_657 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2552 = torch.prim.ListConstruct %int1_658, %int1_659 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2553 = torch.prim.ListConstruct %int1_660, %int1_661 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2554 = torch.prim.ListConstruct %int0_662, %int0_662 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_663 = torch.constant.bool false | |
| %int1_664 = torch.constant.int 1 | |
| %2555 = torch.aten.convolution %2526, %2538, %2550, %2553, %2551, %2552, %false_663, %2554, %int1_664 : !torch.vtensor<[1,112,14,14],f32>, !torch.vtensor<[672,112,1,1],f32>, !torch.vtensor<[672],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2556 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2557 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_665 = torch.constant.int 12 | |
| %2558 = torch.aten.item %2556 : !torch.vtensor<[],f32> -> !torch.float | |
| %2559 = torch.aten.item %2557 : !torch.vtensor<[],si8> -> !torch.int | |
| %2560 = torch.aten.quantize_per_tensor %2555, %2558, %2559, %int12_665 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2561 = torch.aten.int_repr %2560 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2562 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2563 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2564 = torch.aten.item %2562 : !torch.vtensor<[],f32> -> !torch.float | |
| %2565 = torch.aten.item %2563 : !torch.vtensor<[],si8> -> !torch.int | |
| %2566 = torch.aten._make_per_tensor_quantized_tensor %2561, %2564, %2565 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2567 = torch.aten.dequantize.self %2566 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2568 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_666 = torch.constant.int 1 | |
| %2569 = torch.aten.add.Tensor %2567, %2568, %int1_666 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2570 = torch.aten.relu %2569 : !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2571 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2572 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_667 = torch.constant.int 6 | |
| %none_668 = torch.constant.none | |
| %false_669 = torch.constant.bool false | |
| %2573 = torch.aten.to.dtype %2571, %int6_667, %false_669, %false_669, %none_668 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_670 = torch.constant.int 6 | |
| %none_671 = torch.constant.none | |
| %false_672 = torch.constant.bool false | |
| %2574 = torch.aten.to.dtype %2572, %int6_670, %false_672, %false_672, %none_671 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2575 = torch.aten.clamp.Tensor %2570, %2573, %2574 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2576 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2577 = torch.aten.mul.Tensor %2575, %2576 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2578 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2579 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_673 = torch.constant.int 12 | |
| %2580 = torch.aten.item %2578 : !torch.vtensor<[],f32> -> !torch.float | |
| %2581 = torch.aten.item %2579 : !torch.vtensor<[],si8> -> !torch.int | |
| %2582 = torch.aten.quantize_per_tensor %2577, %2580, %2581, %int12_673 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2583 = torch.aten.int_repr %2582 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2584 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2585 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2586 = torch.aten.item %2584 : !torch.vtensor<[],f32> -> !torch.float | |
| %2587 = torch.aten.item %2585 : !torch.vtensor<[],si8> -> !torch.int | |
| %2588 = torch.aten._make_per_tensor_quantized_tensor %2583, %2586, %2587 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2589 = torch.aten.dequantize.self %2588 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2590 = torch.aten.mul.Tensor %2567, %2589 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2591 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2592 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_674 = torch.constant.int 12 | |
| %2593 = torch.aten.item %2591 : !torch.vtensor<[],f32> -> !torch.float | |
| %2594 = torch.aten.item %2592 : !torch.vtensor<[],si8> -> !torch.int | |
| %2595 = torch.aten.quantize_per_tensor %2590, %2593, %2594, %int12_674 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2596 = torch.aten.int_repr %2595 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2597 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2598 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2599 = torch.aten.item %2597 : !torch.vtensor<[],f32> -> !torch.float | |
| %2600 = torch.aten.item %2598 : !torch.vtensor<[],si8> -> !torch.int | |
| %2601 = torch.aten._make_per_tensor_quantized_tensor %2596, %2599, %2600 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2602 = torch.aten.dequantize.self %2601 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2603 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2604 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_675 = torch.constant.int 12 | |
| %2605 = torch.aten.item %2603 : !torch.vtensor<[],f32> -> !torch.float | |
| %2606 = torch.aten.item %2604 : !torch.vtensor<[],si8> -> !torch.int | |
| %2607 = torch.aten.quantize_per_tensor %84, %2605, %2606, %int12_675 : !torch.vtensor<[672,1,3,3],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672,1,3,3],!torch.qint8> | |
| %2608 = torch.aten.int_repr %2607 : !torch.vtensor<[672,1,3,3],!torch.qint8> -> !torch.vtensor<[672,1,3,3],si8> | |
| %2609 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2610 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2611 = torch.aten.item %2609 : !torch.vtensor<[],f32> -> !torch.float | |
| %2612 = torch.aten.item %2610 : !torch.vtensor<[],si8> -> !torch.int | |
| %2613 = torch.aten._make_per_tensor_quantized_tensor %2608, %2611, %2612 : !torch.vtensor<[672,1,3,3],si8>, !torch.float, !torch.int -> !torch.vtensor<[672,1,3,3],!torch.qint8> | |
| %2614 = torch.aten.dequantize.self %2613 : !torch.vtensor<[672,1,3,3],!torch.qint8> -> !torch.vtensor<[672,1,3,3],f32> | |
| %2615 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2616 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_676 = torch.constant.int 12 | |
| %2617 = torch.aten.item %2615 : !torch.vtensor<[],f32> -> !torch.float | |
| %2618 = torch.aten.item %2616 : !torch.vtensor<[],si8> -> !torch.int | |
| %2619 = torch.aten.quantize_per_tensor %85, %2617, %2618, %int12_676 : !torch.vtensor<[672],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2620 = torch.aten.int_repr %2619 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],si8> | |
| %2621 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2622 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2623 = torch.aten.item %2621 : !torch.vtensor<[],f32> -> !torch.float | |
| %2624 = torch.aten.item %2622 : !torch.vtensor<[],si8> -> !torch.int | |
| %2625 = torch.aten._make_per_tensor_quantized_tensor %2620, %2623, %2624 : !torch.vtensor<[672],si8>, !torch.float, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2626 = torch.aten.dequantize.self %2625 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],f32> | |
| %int1_677 = torch.constant.int 1 | |
| %int1_678 = torch.constant.int 1 | |
| %int1_679 = torch.constant.int 1 | |
| %int1_680 = torch.constant.int 1 | |
| %int1_681 = torch.constant.int 1 | |
| %int1_682 = torch.constant.int 1 | |
| %int0_683 = torch.constant.int 0 | |
| %2627 = torch.prim.ListConstruct %int1_677, %int1_678 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2628 = torch.prim.ListConstruct %int1_679, %int1_680 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2629 = torch.prim.ListConstruct %int1_681, %int1_682 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2630 = torch.prim.ListConstruct %int0_683, %int0_683 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_684 = torch.constant.bool false | |
| %int672 = torch.constant.int 672 | |
| %2631 = torch.aten.convolution %2602, %2614, %2626, %2629, %2627, %2628, %false_684, %2630, %int672 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[672,1,3,3],f32>, !torch.vtensor<[672],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2632 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2633 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_685 = torch.constant.int 12 | |
| %2634 = torch.aten.item %2632 : !torch.vtensor<[],f32> -> !torch.float | |
| %2635 = torch.aten.item %2633 : !torch.vtensor<[],si8> -> !torch.int | |
| %2636 = torch.aten.quantize_per_tensor %2631, %2634, %2635, %int12_685 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2637 = torch.aten.int_repr %2636 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2638 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2639 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2640 = torch.aten.item %2638 : !torch.vtensor<[],f32> -> !torch.float | |
| %2641 = torch.aten.item %2639 : !torch.vtensor<[],si8> -> !torch.int | |
| %2642 = torch.aten._make_per_tensor_quantized_tensor %2637, %2640, %2641 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2643 = torch.aten.dequantize.self %2642 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2644 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_686 = torch.constant.int 1 | |
| %2645 = torch.aten.add.Tensor %2643, %2644, %int1_686 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2646 = torch.aten.relu %2645 : !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2647 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2648 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_687 = torch.constant.int 6 | |
| %none_688 = torch.constant.none | |
| %false_689 = torch.constant.bool false | |
| %2649 = torch.aten.to.dtype %2647, %int6_687, %false_689, %false_689, %none_688 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_690 = torch.constant.int 6 | |
| %none_691 = torch.constant.none | |
| %false_692 = torch.constant.bool false | |
| %2650 = torch.aten.to.dtype %2648, %int6_690, %false_692, %false_692, %none_691 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2651 = torch.aten.clamp.Tensor %2646, %2649, %2650 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2652 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2653 = torch.aten.mul.Tensor %2651, %2652 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2654 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2655 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_693 = torch.constant.int 12 | |
| %2656 = torch.aten.item %2654 : !torch.vtensor<[],f32> -> !torch.float | |
| %2657 = torch.aten.item %2655 : !torch.vtensor<[],si8> -> !torch.int | |
| %2658 = torch.aten.quantize_per_tensor %2653, %2656, %2657, %int12_693 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2659 = torch.aten.int_repr %2658 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2660 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2661 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2662 = torch.aten.item %2660 : !torch.vtensor<[],f32> -> !torch.float | |
| %2663 = torch.aten.item %2661 : !torch.vtensor<[],si8> -> !torch.int | |
| %2664 = torch.aten._make_per_tensor_quantized_tensor %2659, %2662, %2663 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2665 = torch.aten.dequantize.self %2664 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2666 = torch.aten.mul.Tensor %2643, %2665 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2667 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2668 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_694 = torch.constant.int 12 | |
| %2669 = torch.aten.item %2667 : !torch.vtensor<[],f32> -> !torch.float | |
| %2670 = torch.aten.item %2668 : !torch.vtensor<[],si8> -> !torch.int | |
| %2671 = torch.aten.quantize_per_tensor %2666, %2669, %2670, %int12_694 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2672 = torch.aten.int_repr %2671 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2673 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2674 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2675 = torch.aten.item %2673 : !torch.vtensor<[],f32> -> !torch.float | |
| %2676 = torch.aten.item %2674 : !torch.vtensor<[],si8> -> !torch.int | |
| %2677 = torch.aten._make_per_tensor_quantized_tensor %2672, %2675, %2676 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2678 = torch.aten.dequantize.self %2677 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %int0_695 = torch.constant.int 0 | |
| %int1_696 = torch.constant.int 1 | |
| %int14_697 = torch.constant.int 14 | |
| %int14_698 = torch.constant.int 14 | |
| %2679 = torch.prim.ListConstruct %int14_697, %int14_698 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2680 = torch.prim.ListConstruct %int0_695, %int0_695 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2681 = torch.prim.ListConstruct %int1_696, %int1_696 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_699 = torch.constant.bool false | |
| %none_700 = torch.constant.none | |
| %2682 = torch.aten.avg_pool2d %2678, %2679, %2681, %2680, %false_699, %false_699, %none_700 : !torch.vtensor<[1,672,14,14],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,672,1,1],f32> | |
| %2683 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2684 = torch.aten.mul.Tensor %2682, %2683 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2685 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2686 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_701 = torch.constant.int 12 | |
| %2687 = torch.aten.item %2685 : !torch.vtensor<[],f32> -> !torch.float | |
| %2688 = torch.aten.item %2686 : !torch.vtensor<[],si8> -> !torch.int | |
| %2689 = torch.aten.quantize_per_tensor %2684, %2687, %2688, %int12_701 : !torch.vtensor<[1,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %2690 = torch.aten.int_repr %2689 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],si8> | |
| %2691 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2692 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2693 = torch.aten.item %2691 : !torch.vtensor<[],f32> -> !torch.float | |
| %2694 = torch.aten.item %2692 : !torch.vtensor<[],si8> -> !torch.int | |
| %2695 = torch.aten._make_per_tensor_quantized_tensor %2690, %2693, %2694 : !torch.vtensor<[1,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %2696 = torch.aten.dequantize.self %2695 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2697 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2698 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_702 = torch.constant.int 12 | |
| %2699 = torch.aten.item %2697 : !torch.vtensor<[],f32> -> !torch.float | |
| %2700 = torch.aten.item %2698 : !torch.vtensor<[],si8> -> !torch.int | |
| %2701 = torch.aten.quantize_per_tensor %86, %2699, %2700, %int12_702 : !torch.vtensor<[168,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[168,672,1,1],!torch.qint8> | |
| %2702 = torch.aten.int_repr %2701 : !torch.vtensor<[168,672,1,1],!torch.qint8> -> !torch.vtensor<[168,672,1,1],si8> | |
| %2703 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2704 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2705 = torch.aten.item %2703 : !torch.vtensor<[],f32> -> !torch.float | |
| %2706 = torch.aten.item %2704 : !torch.vtensor<[],si8> -> !torch.int | |
| %2707 = torch.aten._make_per_tensor_quantized_tensor %2702, %2705, %2706 : !torch.vtensor<[168,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[168,672,1,1],!torch.qint8> | |
| %2708 = torch.aten.dequantize.self %2707 : !torch.vtensor<[168,672,1,1],!torch.qint8> -> !torch.vtensor<[168,672,1,1],f32> | |
| %2709 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2710 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_703 = torch.constant.int 12 | |
| %2711 = torch.aten.item %2709 : !torch.vtensor<[],f32> -> !torch.float | |
| %2712 = torch.aten.item %2710 : !torch.vtensor<[],si8> -> !torch.int | |
| %2713 = torch.aten.quantize_per_tensor %87, %2711, %2712, %int12_703 : !torch.vtensor<[168],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[168],!torch.qint8> | |
| %2714 = torch.aten.int_repr %2713 : !torch.vtensor<[168],!torch.qint8> -> !torch.vtensor<[168],si8> | |
| %2715 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2716 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2717 = torch.aten.item %2715 : !torch.vtensor<[],f32> -> !torch.float | |
| %2718 = torch.aten.item %2716 : !torch.vtensor<[],si8> -> !torch.int | |
| %2719 = torch.aten._make_per_tensor_quantized_tensor %2714, %2717, %2718 : !torch.vtensor<[168],si8>, !torch.float, !torch.int -> !torch.vtensor<[168],!torch.qint8> | |
| %2720 = torch.aten.dequantize.self %2719 : !torch.vtensor<[168],!torch.qint8> -> !torch.vtensor<[168],f32> | |
| %int0_704 = torch.constant.int 0 | |
| %int0_705 = torch.constant.int 0 | |
| %int1_706 = torch.constant.int 1 | |
| %int1_707 = torch.constant.int 1 | |
| %int1_708 = torch.constant.int 1 | |
| %int1_709 = torch.constant.int 1 | |
| %int0_710 = torch.constant.int 0 | |
| %2721 = torch.prim.ListConstruct %int0_704, %int0_705 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2722 = torch.prim.ListConstruct %int1_706, %int1_707 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2723 = torch.prim.ListConstruct %int1_708, %int1_709 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2724 = torch.prim.ListConstruct %int0_710, %int0_710 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_711 = torch.constant.bool false | |
| %int1_712 = torch.constant.int 1 | |
| %2725 = torch.aten.convolution %2696, %2708, %2720, %2723, %2721, %2722, %false_711, %2724, %int1_712 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[168,672,1,1],f32>, !torch.vtensor<[168],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,168,1,1],f32> | |
| %2726 = torch.aten.relu %2725 : !torch.vtensor<[1,168,1,1],f32> -> !torch.vtensor<[1,168,1,1],f32> | |
| %2727 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2728 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_713 = torch.constant.int 12 | |
| %2729 = torch.aten.item %2727 : !torch.vtensor<[],f32> -> !torch.float | |
| %2730 = torch.aten.item %2728 : !torch.vtensor<[],si8> -> !torch.int | |
| %2731 = torch.aten.quantize_per_tensor %2726, %2729, %2730, %int12_713 : !torch.vtensor<[1,168,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,168,1,1],!torch.qint8> | |
| %2732 = torch.aten.int_repr %2731 : !torch.vtensor<[1,168,1,1],!torch.qint8> -> !torch.vtensor<[1,168,1,1],si8> | |
| %2733 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2734 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2735 = torch.aten.item %2733 : !torch.vtensor<[],f32> -> !torch.float | |
| %2736 = torch.aten.item %2734 : !torch.vtensor<[],si8> -> !torch.int | |
| %2737 = torch.aten._make_per_tensor_quantized_tensor %2732, %2735, %2736 : !torch.vtensor<[1,168,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,168,1,1],!torch.qint8> | |
| %2738 = torch.aten.dequantize.self %2737 : !torch.vtensor<[1,168,1,1],!torch.qint8> -> !torch.vtensor<[1,168,1,1],f32> | |
| %2739 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2740 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_714 = torch.constant.int 12 | |
| %2741 = torch.aten.item %2739 : !torch.vtensor<[],f32> -> !torch.float | |
| %2742 = torch.aten.item %2740 : !torch.vtensor<[],si8> -> !torch.int | |
| %2743 = torch.aten.quantize_per_tensor %88, %2741, %2742, %int12_714 : !torch.vtensor<[672,168,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672,168,1,1],!torch.qint8> | |
| %2744 = torch.aten.int_repr %2743 : !torch.vtensor<[672,168,1,1],!torch.qint8> -> !torch.vtensor<[672,168,1,1],si8> | |
| %2745 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2746 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2747 = torch.aten.item %2745 : !torch.vtensor<[],f32> -> !torch.float | |
| %2748 = torch.aten.item %2746 : !torch.vtensor<[],si8> -> !torch.int | |
| %2749 = torch.aten._make_per_tensor_quantized_tensor %2744, %2747, %2748 : !torch.vtensor<[672,168,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[672,168,1,1],!torch.qint8> | |
| %2750 = torch.aten.dequantize.self %2749 : !torch.vtensor<[672,168,1,1],!torch.qint8> -> !torch.vtensor<[672,168,1,1],f32> | |
| %2751 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2752 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_715 = torch.constant.int 12 | |
| %2753 = torch.aten.item %2751 : !torch.vtensor<[],f32> -> !torch.float | |
| %2754 = torch.aten.item %2752 : !torch.vtensor<[],si8> -> !torch.int | |
| %2755 = torch.aten.quantize_per_tensor %89, %2753, %2754, %int12_715 : !torch.vtensor<[672],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2756 = torch.aten.int_repr %2755 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],si8> | |
| %2757 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2758 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2759 = torch.aten.item %2757 : !torch.vtensor<[],f32> -> !torch.float | |
| %2760 = torch.aten.item %2758 : !torch.vtensor<[],si8> -> !torch.int | |
| %2761 = torch.aten._make_per_tensor_quantized_tensor %2756, %2759, %2760 : !torch.vtensor<[672],si8>, !torch.float, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2762 = torch.aten.dequantize.self %2761 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],f32> | |
| %int0_716 = torch.constant.int 0 | |
| %int0_717 = torch.constant.int 0 | |
| %int1_718 = torch.constant.int 1 | |
| %int1_719 = torch.constant.int 1 | |
| %int1_720 = torch.constant.int 1 | |
| %int1_721 = torch.constant.int 1 | |
| %int0_722 = torch.constant.int 0 | |
| %2763 = torch.prim.ListConstruct %int0_716, %int0_717 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2764 = torch.prim.ListConstruct %int1_718, %int1_719 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2765 = torch.prim.ListConstruct %int1_720, %int1_721 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2766 = torch.prim.ListConstruct %int0_722, %int0_722 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_723 = torch.constant.bool false | |
| %int1_724 = torch.constant.int 1 | |
| %2767 = torch.aten.convolution %2738, %2750, %2762, %2765, %2763, %2764, %false_723, %2766, %int1_724 : !torch.vtensor<[1,168,1,1],f32>, !torch.vtensor<[672,168,1,1],f32>, !torch.vtensor<[672],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,672,1,1],f32> | |
| %2768 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2769 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_725 = torch.constant.int 12 | |
| %2770 = torch.aten.item %2768 : !torch.vtensor<[],f32> -> !torch.float | |
| %2771 = torch.aten.item %2769 : !torch.vtensor<[],si8> -> !torch.int | |
| %2772 = torch.aten.quantize_per_tensor %2767, %2770, %2771, %int12_725 : !torch.vtensor<[1,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %2773 = torch.aten.int_repr %2772 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],si8> | |
| %2774 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2775 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2776 = torch.aten.item %2774 : !torch.vtensor<[],f32> -> !torch.float | |
| %2777 = torch.aten.item %2775 : !torch.vtensor<[],si8> -> !torch.int | |
| %2778 = torch.aten._make_per_tensor_quantized_tensor %2773, %2776, %2777 : !torch.vtensor<[1,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %2779 = torch.aten.dequantize.self %2778 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2780 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_726 = torch.constant.int 1 | |
| %2781 = torch.aten.add.Tensor %2779, %2780, %int1_726 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,672,1,1],f32> | |
| %2782 = torch.aten.relu %2781 : !torch.vtensor<[1,672,1,1],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2783 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2784 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_727 = torch.constant.int 6 | |
| %none_728 = torch.constant.none | |
| %false_729 = torch.constant.bool false | |
| %2785 = torch.aten.to.dtype %2783, %int6_727, %false_729, %false_729, %none_728 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_730 = torch.constant.int 6 | |
| %none_731 = torch.constant.none | |
| %false_732 = torch.constant.bool false | |
| %2786 = torch.aten.to.dtype %2784, %int6_730, %false_732, %false_732, %none_731 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2787 = torch.aten.clamp.Tensor %2782, %2785, %2786 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2788 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2789 = torch.aten.mul.Tensor %2787, %2788 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2790 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2791 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_733 = torch.constant.int 12 | |
| %2792 = torch.aten.item %2790 : !torch.vtensor<[],f32> -> !torch.float | |
| %2793 = torch.aten.item %2791 : !torch.vtensor<[],si8> -> !torch.int | |
| %2794 = torch.aten.quantize_per_tensor %2789, %2792, %2793, %int12_733 : !torch.vtensor<[1,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %2795 = torch.aten.int_repr %2794 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],si8> | |
| %2796 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2797 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2798 = torch.aten.item %2796 : !torch.vtensor<[],f32> -> !torch.float | |
| %2799 = torch.aten.item %2797 : !torch.vtensor<[],si8> -> !torch.int | |
| %2800 = torch.aten._make_per_tensor_quantized_tensor %2795, %2798, %2799 : !torch.vtensor<[1,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %2801 = torch.aten.dequantize.self %2800 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],f32> | |
| %2802 = torch.aten.mul.Tensor %2801, %2678 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2803 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2804 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_734 = torch.constant.int 12 | |
| %2805 = torch.aten.item %2803 : !torch.vtensor<[],f32> -> !torch.float | |
| %2806 = torch.aten.item %2804 : !torch.vtensor<[],si8> -> !torch.int | |
| %2807 = torch.aten.quantize_per_tensor %2802, %2805, %2806, %int12_734 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2808 = torch.aten.int_repr %2807 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2809 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2810 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2811 = torch.aten.item %2809 : !torch.vtensor<[],f32> -> !torch.float | |
| %2812 = torch.aten.item %2810 : !torch.vtensor<[],si8> -> !torch.int | |
| %2813 = torch.aten._make_per_tensor_quantized_tensor %2808, %2811, %2812 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2814 = torch.aten.dequantize.self %2813 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2815 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2816 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_735 = torch.constant.int 12 | |
| %2817 = torch.aten.item %2815 : !torch.vtensor<[],f32> -> !torch.float | |
| %2818 = torch.aten.item %2816 : !torch.vtensor<[],si8> -> !torch.int | |
| %2819 = torch.aten.quantize_per_tensor %90, %2817, %2818, %int12_735 : !torch.vtensor<[112,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[112,672,1,1],!torch.qint8> | |
| %2820 = torch.aten.int_repr %2819 : !torch.vtensor<[112,672,1,1],!torch.qint8> -> !torch.vtensor<[112,672,1,1],si8> | |
| %2821 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2822 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2823 = torch.aten.item %2821 : !torch.vtensor<[],f32> -> !torch.float | |
| %2824 = torch.aten.item %2822 : !torch.vtensor<[],si8> -> !torch.int | |
| %2825 = torch.aten._make_per_tensor_quantized_tensor %2820, %2823, %2824 : !torch.vtensor<[112,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[112,672,1,1],!torch.qint8> | |
| %2826 = torch.aten.dequantize.self %2825 : !torch.vtensor<[112,672,1,1],!torch.qint8> -> !torch.vtensor<[112,672,1,1],f32> | |
| %2827 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2828 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_736 = torch.constant.int 12 | |
| %2829 = torch.aten.item %2827 : !torch.vtensor<[],f32> -> !torch.float | |
| %2830 = torch.aten.item %2828 : !torch.vtensor<[],si8> -> !torch.int | |
| %2831 = torch.aten.quantize_per_tensor %91, %2829, %2830, %int12_736 : !torch.vtensor<[112],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[112],!torch.qint8> | |
| %2832 = torch.aten.int_repr %2831 : !torch.vtensor<[112],!torch.qint8> -> !torch.vtensor<[112],si8> | |
| %2833 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2834 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2835 = torch.aten.item %2833 : !torch.vtensor<[],f32> -> !torch.float | |
| %2836 = torch.aten.item %2834 : !torch.vtensor<[],si8> -> !torch.int | |
| %2837 = torch.aten._make_per_tensor_quantized_tensor %2832, %2835, %2836 : !torch.vtensor<[112],si8>, !torch.float, !torch.int -> !torch.vtensor<[112],!torch.qint8> | |
| %2838 = torch.aten.dequantize.self %2837 : !torch.vtensor<[112],!torch.qint8> -> !torch.vtensor<[112],f32> | |
| %int0_737 = torch.constant.int 0 | |
| %int0_738 = torch.constant.int 0 | |
| %int1_739 = torch.constant.int 1 | |
| %int1_740 = torch.constant.int 1 | |
| %int1_741 = torch.constant.int 1 | |
| %int1_742 = torch.constant.int 1 | |
| %int0_743 = torch.constant.int 0 | |
| %2839 = torch.prim.ListConstruct %int0_737, %int0_738 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2840 = torch.prim.ListConstruct %int1_739, %int1_740 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2841 = torch.prim.ListConstruct %int1_741, %int1_742 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2842 = torch.prim.ListConstruct %int0_743, %int0_743 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_744 = torch.constant.bool false | |
| %int1_745 = torch.constant.int 1 | |
| %2843 = torch.aten.convolution %2814, %2826, %2838, %2841, %2839, %2840, %false_744, %2842, %int1_745 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[112,672,1,1],f32>, !torch.vtensor<[112],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,112,14,14],f32> | |
| %2844 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2845 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_746 = torch.constant.int 12 | |
| %2846 = torch.aten.item %2844 : !torch.vtensor<[],f32> -> !torch.float | |
| %2847 = torch.aten.item %2845 : !torch.vtensor<[],si8> -> !torch.int | |
| %2848 = torch.aten.quantize_per_tensor %2843, %2846, %2847, %int12_746 : !torch.vtensor<[1,112,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,112,14,14],!torch.qint8> | |
| %2849 = torch.aten.int_repr %2848 : !torch.vtensor<[1,112,14,14],!torch.qint8> -> !torch.vtensor<[1,112,14,14],si8> | |
| %2850 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2851 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2852 = torch.aten.item %2850 : !torch.vtensor<[],f32> -> !torch.float | |
| %2853 = torch.aten.item %2851 : !torch.vtensor<[],si8> -> !torch.int | |
| %2854 = torch.aten._make_per_tensor_quantized_tensor %2849, %2852, %2853 : !torch.vtensor<[1,112,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,112,14,14],!torch.qint8> | |
| %2855 = torch.aten.dequantize.self %2854 : !torch.vtensor<[1,112,14,14],!torch.qint8> -> !torch.vtensor<[1,112,14,14],f32> | |
| %int1_747 = torch.constant.int 1 | |
| %2856 = torch.aten.add.Tensor %2855, %2526, %int1_747 : !torch.vtensor<[1,112,14,14],f32>, !torch.vtensor<[1,112,14,14],f32>, !torch.int -> !torch.vtensor<[1,112,14,14],f32> | |
| %2857 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2858 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_748 = torch.constant.int 12 | |
| %2859 = torch.aten.item %2857 : !torch.vtensor<[],f32> -> !torch.float | |
| %2860 = torch.aten.item %2858 : !torch.vtensor<[],si8> -> !torch.int | |
| %2861 = torch.aten.quantize_per_tensor %2856, %2859, %2860, %int12_748 : !torch.vtensor<[1,112,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,112,14,14],!torch.qint8> | |
| %2862 = torch.aten.int_repr %2861 : !torch.vtensor<[1,112,14,14],!torch.qint8> -> !torch.vtensor<[1,112,14,14],si8> | |
| %2863 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2864 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2865 = torch.aten.item %2863 : !torch.vtensor<[],f32> -> !torch.float | |
| %2866 = torch.aten.item %2864 : !torch.vtensor<[],si8> -> !torch.int | |
| %2867 = torch.aten._make_per_tensor_quantized_tensor %2862, %2865, %2866 : !torch.vtensor<[1,112,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,112,14,14],!torch.qint8> | |
| %2868 = torch.aten.dequantize.self %2867 : !torch.vtensor<[1,112,14,14],!torch.qint8> -> !torch.vtensor<[1,112,14,14],f32> | |
| %2869 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2870 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_749 = torch.constant.int 12 | |
| %2871 = torch.aten.item %2869 : !torch.vtensor<[],f32> -> !torch.float | |
| %2872 = torch.aten.item %2870 : !torch.vtensor<[],si8> -> !torch.int | |
| %2873 = torch.aten.quantize_per_tensor %92, %2871, %2872, %int12_749 : !torch.vtensor<[672,112,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672,112,1,1],!torch.qint8> | |
| %2874 = torch.aten.int_repr %2873 : !torch.vtensor<[672,112,1,1],!torch.qint8> -> !torch.vtensor<[672,112,1,1],si8> | |
| %2875 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2876 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2877 = torch.aten.item %2875 : !torch.vtensor<[],f32> -> !torch.float | |
| %2878 = torch.aten.item %2876 : !torch.vtensor<[],si8> -> !torch.int | |
| %2879 = torch.aten._make_per_tensor_quantized_tensor %2874, %2877, %2878 : !torch.vtensor<[672,112,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[672,112,1,1],!torch.qint8> | |
| %2880 = torch.aten.dequantize.self %2879 : !torch.vtensor<[672,112,1,1],!torch.qint8> -> !torch.vtensor<[672,112,1,1],f32> | |
| %2881 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2882 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_750 = torch.constant.int 12 | |
| %2883 = torch.aten.item %2881 : !torch.vtensor<[],f32> -> !torch.float | |
| %2884 = torch.aten.item %2882 : !torch.vtensor<[],si8> -> !torch.int | |
| %2885 = torch.aten.quantize_per_tensor %93, %2883, %2884, %int12_750 : !torch.vtensor<[672],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2886 = torch.aten.int_repr %2885 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],si8> | |
| %2887 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2888 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2889 = torch.aten.item %2887 : !torch.vtensor<[],f32> -> !torch.float | |
| %2890 = torch.aten.item %2888 : !torch.vtensor<[],si8> -> !torch.int | |
| %2891 = torch.aten._make_per_tensor_quantized_tensor %2886, %2889, %2890 : !torch.vtensor<[672],si8>, !torch.float, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2892 = torch.aten.dequantize.self %2891 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],f32> | |
| %int0_751 = torch.constant.int 0 | |
| %int0_752 = torch.constant.int 0 | |
| %int1_753 = torch.constant.int 1 | |
| %int1_754 = torch.constant.int 1 | |
| %int1_755 = torch.constant.int 1 | |
| %int1_756 = torch.constant.int 1 | |
| %int0_757 = torch.constant.int 0 | |
| %2893 = torch.prim.ListConstruct %int0_751, %int0_752 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2894 = torch.prim.ListConstruct %int1_753, %int1_754 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2895 = torch.prim.ListConstruct %int1_755, %int1_756 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2896 = torch.prim.ListConstruct %int0_757, %int0_757 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_758 = torch.constant.bool false | |
| %int1_759 = torch.constant.int 1 | |
| %2897 = torch.aten.convolution %2868, %2880, %2892, %2895, %2893, %2894, %false_758, %2896, %int1_759 : !torch.vtensor<[1,112,14,14],f32>, !torch.vtensor<[672,112,1,1],f32>, !torch.vtensor<[672],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2898 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2899 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_760 = torch.constant.int 12 | |
| %2900 = torch.aten.item %2898 : !torch.vtensor<[],f32> -> !torch.float | |
| %2901 = torch.aten.item %2899 : !torch.vtensor<[],si8> -> !torch.int | |
| %2902 = torch.aten.quantize_per_tensor %2897, %2900, %2901, %int12_760 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2903 = torch.aten.int_repr %2902 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2904 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2905 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2906 = torch.aten.item %2904 : !torch.vtensor<[],f32> -> !torch.float | |
| %2907 = torch.aten.item %2905 : !torch.vtensor<[],si8> -> !torch.int | |
| %2908 = torch.aten._make_per_tensor_quantized_tensor %2903, %2906, %2907 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2909 = torch.aten.dequantize.self %2908 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2910 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_761 = torch.constant.int 1 | |
| %2911 = torch.aten.add.Tensor %2909, %2910, %int1_761 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2912 = torch.aten.relu %2911 : !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2913 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2914 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_762 = torch.constant.int 6 | |
| %none_763 = torch.constant.none | |
| %false_764 = torch.constant.bool false | |
| %2915 = torch.aten.to.dtype %2913, %int6_762, %false_764, %false_764, %none_763 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_765 = torch.constant.int 6 | |
| %none_766 = torch.constant.none | |
| %false_767 = torch.constant.bool false | |
| %2916 = torch.aten.to.dtype %2914, %int6_765, %false_767, %false_767, %none_766 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2917 = torch.aten.clamp.Tensor %2912, %2915, %2916 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2918 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2919 = torch.aten.mul.Tensor %2917, %2918 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2920 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2921 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_768 = torch.constant.int 12 | |
| %2922 = torch.aten.item %2920 : !torch.vtensor<[],f32> -> !torch.float | |
| %2923 = torch.aten.item %2921 : !torch.vtensor<[],si8> -> !torch.int | |
| %2924 = torch.aten.quantize_per_tensor %2919, %2922, %2923, %int12_768 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2925 = torch.aten.int_repr %2924 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2926 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2927 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2928 = torch.aten.item %2926 : !torch.vtensor<[],f32> -> !torch.float | |
| %2929 = torch.aten.item %2927 : !torch.vtensor<[],si8> -> !torch.int | |
| %2930 = torch.aten._make_per_tensor_quantized_tensor %2925, %2928, %2929 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2931 = torch.aten.dequantize.self %2930 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2932 = torch.aten.mul.Tensor %2909, %2931 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2933 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2934 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_769 = torch.constant.int 12 | |
| %2935 = torch.aten.item %2933 : !torch.vtensor<[],f32> -> !torch.float | |
| %2936 = torch.aten.item %2934 : !torch.vtensor<[],si8> -> !torch.int | |
| %2937 = torch.aten.quantize_per_tensor %2932, %2935, %2936, %int12_769 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2938 = torch.aten.int_repr %2937 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2939 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2940 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2941 = torch.aten.item %2939 : !torch.vtensor<[],f32> -> !torch.float | |
| %2942 = torch.aten.item %2940 : !torch.vtensor<[],si8> -> !torch.int | |
| %2943 = torch.aten._make_per_tensor_quantized_tensor %2938, %2941, %2942 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2944 = torch.aten.dequantize.self %2943 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2945 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2946 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_770 = torch.constant.int 12 | |
| %2947 = torch.aten.item %2945 : !torch.vtensor<[],f32> -> !torch.float | |
| %2948 = torch.aten.item %2946 : !torch.vtensor<[],si8> -> !torch.int | |
| %2949 = torch.aten.quantize_per_tensor %94, %2947, %2948, %int12_770 : !torch.vtensor<[672,1,5,5],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672,1,5,5],!torch.qint8> | |
| %2950 = torch.aten.int_repr %2949 : !torch.vtensor<[672,1,5,5],!torch.qint8> -> !torch.vtensor<[672,1,5,5],si8> | |
| %2951 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2952 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2953 = torch.aten.item %2951 : !torch.vtensor<[],f32> -> !torch.float | |
| %2954 = torch.aten.item %2952 : !torch.vtensor<[],si8> -> !torch.int | |
| %2955 = torch.aten._make_per_tensor_quantized_tensor %2950, %2953, %2954 : !torch.vtensor<[672,1,5,5],si8>, !torch.float, !torch.int -> !torch.vtensor<[672,1,5,5],!torch.qint8> | |
| %2956 = torch.aten.dequantize.self %2955 : !torch.vtensor<[672,1,5,5],!torch.qint8> -> !torch.vtensor<[672,1,5,5],f32> | |
| %2957 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2958 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_771 = torch.constant.int 12 | |
| %2959 = torch.aten.item %2957 : !torch.vtensor<[],f32> -> !torch.float | |
| %2960 = torch.aten.item %2958 : !torch.vtensor<[],si8> -> !torch.int | |
| %2961 = torch.aten.quantize_per_tensor %95, %2959, %2960, %int12_771 : !torch.vtensor<[672],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2962 = torch.aten.int_repr %2961 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],si8> | |
| %2963 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2964 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2965 = torch.aten.item %2963 : !torch.vtensor<[],f32> -> !torch.float | |
| %2966 = torch.aten.item %2964 : !torch.vtensor<[],si8> -> !torch.int | |
| %2967 = torch.aten._make_per_tensor_quantized_tensor %2962, %2965, %2966 : !torch.vtensor<[672],si8>, !torch.float, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %2968 = torch.aten.dequantize.self %2967 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],f32> | |
| %int4 = torch.constant.int 4 | |
| %int4_772 = torch.constant.int 4 | |
| %int2_773 = torch.constant.int 2 | |
| %int2_774 = torch.constant.int 2 | |
| %int1_775 = torch.constant.int 1 | |
| %int1_776 = torch.constant.int 1 | |
| %int0_777 = torch.constant.int 0 | |
| %2969 = torch.prim.ListConstruct %int4, %int4_772 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2970 = torch.prim.ListConstruct %int2_773, %int2_774 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2971 = torch.prim.ListConstruct %int1_775, %int1_776 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %2972 = torch.prim.ListConstruct %int0_777, %int0_777 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_778 = torch.constant.bool false | |
| %int672_779 = torch.constant.int 672 | |
| %2973 = torch.aten.convolution %2944, %2956, %2968, %2971, %2969, %2970, %false_778, %2972, %int672_779 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[672,1,5,5],f32>, !torch.vtensor<[672],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2974 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2975 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_780 = torch.constant.int 12 | |
| %2976 = torch.aten.item %2974 : !torch.vtensor<[],f32> -> !torch.float | |
| %2977 = torch.aten.item %2975 : !torch.vtensor<[],si8> -> !torch.int | |
| %2978 = torch.aten.quantize_per_tensor %2973, %2976, %2977, %int12_780 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2979 = torch.aten.int_repr %2978 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %2980 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2981 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %2982 = torch.aten.item %2980 : !torch.vtensor<[],f32> -> !torch.float | |
| %2983 = torch.aten.item %2981 : !torch.vtensor<[],si8> -> !torch.int | |
| %2984 = torch.aten._make_per_tensor_quantized_tensor %2979, %2982, %2983 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %2985 = torch.aten.dequantize.self %2984 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2986 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_781 = torch.constant.int 1 | |
| %2987 = torch.aten.add.Tensor %2985, %2986, %int1_781 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,672,14,14],f32> | |
| %2988 = torch.aten.relu %2987 : !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2989 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2990 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_782 = torch.constant.int 6 | |
| %none_783 = torch.constant.none | |
| %false_784 = torch.constant.bool false | |
| %2991 = torch.aten.to.dtype %2989, %int6_782, %false_784, %false_784, %none_783 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_785 = torch.constant.int 6 | |
| %none_786 = torch.constant.none | |
| %false_787 = torch.constant.bool false | |
| %2992 = torch.aten.to.dtype %2990, %int6_785, %false_787, %false_787, %none_786 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %2993 = torch.aten.clamp.Tensor %2988, %2991, %2992 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2994 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2995 = torch.aten.mul.Tensor %2993, %2994 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %2996 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %2997 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_788 = torch.constant.int 12 | |
| %2998 = torch.aten.item %2996 : !torch.vtensor<[],f32> -> !torch.float | |
| %2999 = torch.aten.item %2997 : !torch.vtensor<[],si8> -> !torch.int | |
| %3000 = torch.aten.quantize_per_tensor %2995, %2998, %2999, %int12_788 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %3001 = torch.aten.int_repr %3000 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %3002 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3003 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3004 = torch.aten.item %3002 : !torch.vtensor<[],f32> -> !torch.float | |
| %3005 = torch.aten.item %3003 : !torch.vtensor<[],si8> -> !torch.int | |
| %3006 = torch.aten._make_per_tensor_quantized_tensor %3001, %3004, %3005 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %3007 = torch.aten.dequantize.self %3006 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %3008 = torch.aten.mul.Tensor %2985, %3007 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %3009 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3010 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_789 = torch.constant.int 12 | |
| %3011 = torch.aten.item %3009 : !torch.vtensor<[],f32> -> !torch.float | |
| %3012 = torch.aten.item %3010 : !torch.vtensor<[],si8> -> !torch.int | |
| %3013 = torch.aten.quantize_per_tensor %3008, %3011, %3012, %int12_789 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %3014 = torch.aten.int_repr %3013 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %3015 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3016 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3017 = torch.aten.item %3015 : !torch.vtensor<[],f32> -> !torch.float | |
| %3018 = torch.aten.item %3016 : !torch.vtensor<[],si8> -> !torch.int | |
| %3019 = torch.aten._make_per_tensor_quantized_tensor %3014, %3017, %3018 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %3020 = torch.aten.dequantize.self %3019 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %int0_790 = torch.constant.int 0 | |
| %int1_791 = torch.constant.int 1 | |
| %int14_792 = torch.constant.int 14 | |
| %int14_793 = torch.constant.int 14 | |
| %3021 = torch.prim.ListConstruct %int14_792, %int14_793 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3022 = torch.prim.ListConstruct %int0_790, %int0_790 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3023 = torch.prim.ListConstruct %int1_791, %int1_791 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_794 = torch.constant.bool false | |
| %none_795 = torch.constant.none | |
| %3024 = torch.aten.avg_pool2d %3020, %3021, %3023, %3022, %false_794, %false_794, %none_795 : !torch.vtensor<[1,672,14,14],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,672,1,1],f32> | |
| %3025 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3026 = torch.aten.mul.Tensor %3024, %3025 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3027 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3028 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_796 = torch.constant.int 12 | |
| %3029 = torch.aten.item %3027 : !torch.vtensor<[],f32> -> !torch.float | |
| %3030 = torch.aten.item %3028 : !torch.vtensor<[],si8> -> !torch.int | |
| %3031 = torch.aten.quantize_per_tensor %3026, %3029, %3030, %int12_796 : !torch.vtensor<[1,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %3032 = torch.aten.int_repr %3031 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],si8> | |
| %3033 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3034 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3035 = torch.aten.item %3033 : !torch.vtensor<[],f32> -> !torch.float | |
| %3036 = torch.aten.item %3034 : !torch.vtensor<[],si8> -> !torch.int | |
| %3037 = torch.aten._make_per_tensor_quantized_tensor %3032, %3035, %3036 : !torch.vtensor<[1,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %3038 = torch.aten.dequantize.self %3037 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3039 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3040 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_797 = torch.constant.int 12 | |
| %3041 = torch.aten.item %3039 : !torch.vtensor<[],f32> -> !torch.float | |
| %3042 = torch.aten.item %3040 : !torch.vtensor<[],si8> -> !torch.int | |
| %3043 = torch.aten.quantize_per_tensor %96, %3041, %3042, %int12_797 : !torch.vtensor<[168,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[168,672,1,1],!torch.qint8> | |
| %3044 = torch.aten.int_repr %3043 : !torch.vtensor<[168,672,1,1],!torch.qint8> -> !torch.vtensor<[168,672,1,1],si8> | |
| %3045 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3046 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3047 = torch.aten.item %3045 : !torch.vtensor<[],f32> -> !torch.float | |
| %3048 = torch.aten.item %3046 : !torch.vtensor<[],si8> -> !torch.int | |
| %3049 = torch.aten._make_per_tensor_quantized_tensor %3044, %3047, %3048 : !torch.vtensor<[168,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[168,672,1,1],!torch.qint8> | |
| %3050 = torch.aten.dequantize.self %3049 : !torch.vtensor<[168,672,1,1],!torch.qint8> -> !torch.vtensor<[168,672,1,1],f32> | |
| %3051 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3052 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_798 = torch.constant.int 12 | |
| %3053 = torch.aten.item %3051 : !torch.vtensor<[],f32> -> !torch.float | |
| %3054 = torch.aten.item %3052 : !torch.vtensor<[],si8> -> !torch.int | |
| %3055 = torch.aten.quantize_per_tensor %97, %3053, %3054, %int12_798 : !torch.vtensor<[168],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[168],!torch.qint8> | |
| %3056 = torch.aten.int_repr %3055 : !torch.vtensor<[168],!torch.qint8> -> !torch.vtensor<[168],si8> | |
| %3057 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3058 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3059 = torch.aten.item %3057 : !torch.vtensor<[],f32> -> !torch.float | |
| %3060 = torch.aten.item %3058 : !torch.vtensor<[],si8> -> !torch.int | |
| %3061 = torch.aten._make_per_tensor_quantized_tensor %3056, %3059, %3060 : !torch.vtensor<[168],si8>, !torch.float, !torch.int -> !torch.vtensor<[168],!torch.qint8> | |
| %3062 = torch.aten.dequantize.self %3061 : !torch.vtensor<[168],!torch.qint8> -> !torch.vtensor<[168],f32> | |
| %int0_799 = torch.constant.int 0 | |
| %int0_800 = torch.constant.int 0 | |
| %int1_801 = torch.constant.int 1 | |
| %int1_802 = torch.constant.int 1 | |
| %int1_803 = torch.constant.int 1 | |
| %int1_804 = torch.constant.int 1 | |
| %int0_805 = torch.constant.int 0 | |
| %3063 = torch.prim.ListConstruct %int0_799, %int0_800 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3064 = torch.prim.ListConstruct %int1_801, %int1_802 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3065 = torch.prim.ListConstruct %int1_803, %int1_804 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3066 = torch.prim.ListConstruct %int0_805, %int0_805 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_806 = torch.constant.bool false | |
| %int1_807 = torch.constant.int 1 | |
| %3067 = torch.aten.convolution %3038, %3050, %3062, %3065, %3063, %3064, %false_806, %3066, %int1_807 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[168,672,1,1],f32>, !torch.vtensor<[168],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,168,1,1],f32> | |
| %3068 = torch.aten.relu %3067 : !torch.vtensor<[1,168,1,1],f32> -> !torch.vtensor<[1,168,1,1],f32> | |
| %3069 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3070 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_808 = torch.constant.int 12 | |
| %3071 = torch.aten.item %3069 : !torch.vtensor<[],f32> -> !torch.float | |
| %3072 = torch.aten.item %3070 : !torch.vtensor<[],si8> -> !torch.int | |
| %3073 = torch.aten.quantize_per_tensor %3068, %3071, %3072, %int12_808 : !torch.vtensor<[1,168,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,168,1,1],!torch.qint8> | |
| %3074 = torch.aten.int_repr %3073 : !torch.vtensor<[1,168,1,1],!torch.qint8> -> !torch.vtensor<[1,168,1,1],si8> | |
| %3075 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3076 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3077 = torch.aten.item %3075 : !torch.vtensor<[],f32> -> !torch.float | |
| %3078 = torch.aten.item %3076 : !torch.vtensor<[],si8> -> !torch.int | |
| %3079 = torch.aten._make_per_tensor_quantized_tensor %3074, %3077, %3078 : !torch.vtensor<[1,168,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,168,1,1],!torch.qint8> | |
| %3080 = torch.aten.dequantize.self %3079 : !torch.vtensor<[1,168,1,1],!torch.qint8> -> !torch.vtensor<[1,168,1,1],f32> | |
| %3081 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3082 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_809 = torch.constant.int 12 | |
| %3083 = torch.aten.item %3081 : !torch.vtensor<[],f32> -> !torch.float | |
| %3084 = torch.aten.item %3082 : !torch.vtensor<[],si8> -> !torch.int | |
| %3085 = torch.aten.quantize_per_tensor %98, %3083, %3084, %int12_809 : !torch.vtensor<[672,168,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672,168,1,1],!torch.qint8> | |
| %3086 = torch.aten.int_repr %3085 : !torch.vtensor<[672,168,1,1],!torch.qint8> -> !torch.vtensor<[672,168,1,1],si8> | |
| %3087 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3088 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3089 = torch.aten.item %3087 : !torch.vtensor<[],f32> -> !torch.float | |
| %3090 = torch.aten.item %3088 : !torch.vtensor<[],si8> -> !torch.int | |
| %3091 = torch.aten._make_per_tensor_quantized_tensor %3086, %3089, %3090 : !torch.vtensor<[672,168,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[672,168,1,1],!torch.qint8> | |
| %3092 = torch.aten.dequantize.self %3091 : !torch.vtensor<[672,168,1,1],!torch.qint8> -> !torch.vtensor<[672,168,1,1],f32> | |
| %3093 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3094 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_810 = torch.constant.int 12 | |
| %3095 = torch.aten.item %3093 : !torch.vtensor<[],f32> -> !torch.float | |
| %3096 = torch.aten.item %3094 : !torch.vtensor<[],si8> -> !torch.int | |
| %3097 = torch.aten.quantize_per_tensor %99, %3095, %3096, %int12_810 : !torch.vtensor<[672],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %3098 = torch.aten.int_repr %3097 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],si8> | |
| %3099 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3100 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3101 = torch.aten.item %3099 : !torch.vtensor<[],f32> -> !torch.float | |
| %3102 = torch.aten.item %3100 : !torch.vtensor<[],si8> -> !torch.int | |
| %3103 = torch.aten._make_per_tensor_quantized_tensor %3098, %3101, %3102 : !torch.vtensor<[672],si8>, !torch.float, !torch.int -> !torch.vtensor<[672],!torch.qint8> | |
| %3104 = torch.aten.dequantize.self %3103 : !torch.vtensor<[672],!torch.qint8> -> !torch.vtensor<[672],f32> | |
| %int0_811 = torch.constant.int 0 | |
| %int0_812 = torch.constant.int 0 | |
| %int1_813 = torch.constant.int 1 | |
| %int1_814 = torch.constant.int 1 | |
| %int1_815 = torch.constant.int 1 | |
| %int1_816 = torch.constant.int 1 | |
| %int0_817 = torch.constant.int 0 | |
| %3105 = torch.prim.ListConstruct %int0_811, %int0_812 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3106 = torch.prim.ListConstruct %int1_813, %int1_814 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3107 = torch.prim.ListConstruct %int1_815, %int1_816 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3108 = torch.prim.ListConstruct %int0_817, %int0_817 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_818 = torch.constant.bool false | |
| %int1_819 = torch.constant.int 1 | |
| %3109 = torch.aten.convolution %3080, %3092, %3104, %3107, %3105, %3106, %false_818, %3108, %int1_819 : !torch.vtensor<[1,168,1,1],f32>, !torch.vtensor<[672,168,1,1],f32>, !torch.vtensor<[672],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,672,1,1],f32> | |
| %3110 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3111 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_820 = torch.constant.int 12 | |
| %3112 = torch.aten.item %3110 : !torch.vtensor<[],f32> -> !torch.float | |
| %3113 = torch.aten.item %3111 : !torch.vtensor<[],si8> -> !torch.int | |
| %3114 = torch.aten.quantize_per_tensor %3109, %3112, %3113, %int12_820 : !torch.vtensor<[1,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %3115 = torch.aten.int_repr %3114 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],si8> | |
| %3116 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3117 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3118 = torch.aten.item %3116 : !torch.vtensor<[],f32> -> !torch.float | |
| %3119 = torch.aten.item %3117 : !torch.vtensor<[],si8> -> !torch.int | |
| %3120 = torch.aten._make_per_tensor_quantized_tensor %3115, %3118, %3119 : !torch.vtensor<[1,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %3121 = torch.aten.dequantize.self %3120 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3122 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_821 = torch.constant.int 1 | |
| %3123 = torch.aten.add.Tensor %3121, %3122, %int1_821 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,672,1,1],f32> | |
| %3124 = torch.aten.relu %3123 : !torch.vtensor<[1,672,1,1],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3125 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3126 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_822 = torch.constant.int 6 | |
| %none_823 = torch.constant.none | |
| %false_824 = torch.constant.bool false | |
| %3127 = torch.aten.to.dtype %3125, %int6_822, %false_824, %false_824, %none_823 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_825 = torch.constant.int 6 | |
| %none_826 = torch.constant.none | |
| %false_827 = torch.constant.bool false | |
| %3128 = torch.aten.to.dtype %3126, %int6_825, %false_827, %false_827, %none_826 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3129 = torch.aten.clamp.Tensor %3124, %3127, %3128 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3130 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3131 = torch.aten.mul.Tensor %3129, %3130 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3132 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3133 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_828 = torch.constant.int 12 | |
| %3134 = torch.aten.item %3132 : !torch.vtensor<[],f32> -> !torch.float | |
| %3135 = torch.aten.item %3133 : !torch.vtensor<[],si8> -> !torch.int | |
| %3136 = torch.aten.quantize_per_tensor %3131, %3134, %3135, %int12_828 : !torch.vtensor<[1,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %3137 = torch.aten.int_repr %3136 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],si8> | |
| %3138 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3139 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3140 = torch.aten.item %3138 : !torch.vtensor<[],f32> -> !torch.float | |
| %3141 = torch.aten.item %3139 : !torch.vtensor<[],si8> -> !torch.int | |
| %3142 = torch.aten._make_per_tensor_quantized_tensor %3137, %3140, %3141 : !torch.vtensor<[1,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,1,1],!torch.qint8> | |
| %3143 = torch.aten.dequantize.self %3142 : !torch.vtensor<[1,672,1,1],!torch.qint8> -> !torch.vtensor<[1,672,1,1],f32> | |
| %3144 = torch.aten.mul.Tensor %3143, %3020 : !torch.vtensor<[1,672,1,1],f32>, !torch.vtensor<[1,672,14,14],f32> -> !torch.vtensor<[1,672,14,14],f32> | |
| %3145 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3146 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_829 = torch.constant.int 12 | |
| %3147 = torch.aten.item %3145 : !torch.vtensor<[],f32> -> !torch.float | |
| %3148 = torch.aten.item %3146 : !torch.vtensor<[],si8> -> !torch.int | |
| %3149 = torch.aten.quantize_per_tensor %3144, %3147, %3148, %int12_829 : !torch.vtensor<[1,672,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %3150 = torch.aten.int_repr %3149 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],si8> | |
| %3151 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3152 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3153 = torch.aten.item %3151 : !torch.vtensor<[],f32> -> !torch.float | |
| %3154 = torch.aten.item %3152 : !torch.vtensor<[],si8> -> !torch.int | |
| %3155 = torch.aten._make_per_tensor_quantized_tensor %3150, %3153, %3154 : !torch.vtensor<[1,672,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,672,14,14],!torch.qint8> | |
| %3156 = torch.aten.dequantize.self %3155 : !torch.vtensor<[1,672,14,14],!torch.qint8> -> !torch.vtensor<[1,672,14,14],f32> | |
| %3157 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3158 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_830 = torch.constant.int 12 | |
| %3159 = torch.aten.item %3157 : !torch.vtensor<[],f32> -> !torch.float | |
| %3160 = torch.aten.item %3158 : !torch.vtensor<[],si8> -> !torch.int | |
| %3161 = torch.aten.quantize_per_tensor %100, %3159, %3160, %int12_830 : !torch.vtensor<[160,672,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[160,672,1,1],!torch.qint8> | |
| %3162 = torch.aten.int_repr %3161 : !torch.vtensor<[160,672,1,1],!torch.qint8> -> !torch.vtensor<[160,672,1,1],si8> | |
| %3163 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3164 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3165 = torch.aten.item %3163 : !torch.vtensor<[],f32> -> !torch.float | |
| %3166 = torch.aten.item %3164 : !torch.vtensor<[],si8> -> !torch.int | |
| %3167 = torch.aten._make_per_tensor_quantized_tensor %3162, %3165, %3166 : !torch.vtensor<[160,672,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[160,672,1,1],!torch.qint8> | |
| %3168 = torch.aten.dequantize.self %3167 : !torch.vtensor<[160,672,1,1],!torch.qint8> -> !torch.vtensor<[160,672,1,1],f32> | |
| %3169 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3170 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_831 = torch.constant.int 12 | |
| %3171 = torch.aten.item %3169 : !torch.vtensor<[],f32> -> !torch.float | |
| %3172 = torch.aten.item %3170 : !torch.vtensor<[],si8> -> !torch.int | |
| %3173 = torch.aten.quantize_per_tensor %101, %3171, %3172, %int12_831 : !torch.vtensor<[160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[160],!torch.qint8> | |
| %3174 = torch.aten.int_repr %3173 : !torch.vtensor<[160],!torch.qint8> -> !torch.vtensor<[160],si8> | |
| %3175 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3176 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3177 = torch.aten.item %3175 : !torch.vtensor<[],f32> -> !torch.float | |
| %3178 = torch.aten.item %3176 : !torch.vtensor<[],si8> -> !torch.int | |
| %3179 = torch.aten._make_per_tensor_quantized_tensor %3174, %3177, %3178 : !torch.vtensor<[160],si8>, !torch.float, !torch.int -> !torch.vtensor<[160],!torch.qint8> | |
| %3180 = torch.aten.dequantize.self %3179 : !torch.vtensor<[160],!torch.qint8> -> !torch.vtensor<[160],f32> | |
| %int0_832 = torch.constant.int 0 | |
| %int0_833 = torch.constant.int 0 | |
| %int1_834 = torch.constant.int 1 | |
| %int1_835 = torch.constant.int 1 | |
| %int1_836 = torch.constant.int 1 | |
| %int1_837 = torch.constant.int 1 | |
| %int0_838 = torch.constant.int 0 | |
| %3181 = torch.prim.ListConstruct %int0_832, %int0_833 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3182 = torch.prim.ListConstruct %int1_834, %int1_835 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3183 = torch.prim.ListConstruct %int1_836, %int1_837 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3184 = torch.prim.ListConstruct %int0_838, %int0_838 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_839 = torch.constant.bool false | |
| %int1_840 = torch.constant.int 1 | |
| %3185 = torch.aten.convolution %3156, %3168, %3180, %3183, %3181, %3182, %false_839, %3184, %int1_840 : !torch.vtensor<[1,672,14,14],f32>, !torch.vtensor<[160,672,1,1],f32>, !torch.vtensor<[160],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,160,14,14],f32> | |
| %3186 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3187 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_841 = torch.constant.int 12 | |
| %3188 = torch.aten.item %3186 : !torch.vtensor<[],f32> -> !torch.float | |
| %3189 = torch.aten.item %3187 : !torch.vtensor<[],si8> -> !torch.int | |
| %3190 = torch.aten.quantize_per_tensor %3185, %3188, %3189, %int12_841 : !torch.vtensor<[1,160,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3191 = torch.aten.int_repr %3190 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],si8> | |
| %3192 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3193 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3194 = torch.aten.item %3192 : !torch.vtensor<[],f32> -> !torch.float | |
| %3195 = torch.aten.item %3193 : !torch.vtensor<[],si8> -> !torch.int | |
| %3196 = torch.aten._make_per_tensor_quantized_tensor %3191, %3194, %3195 : !torch.vtensor<[1,160,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3197 = torch.aten.dequantize.self %3196 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],f32> | |
| %3198 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3199 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_842 = torch.constant.int 12 | |
| %3200 = torch.aten.item %3198 : !torch.vtensor<[],f32> -> !torch.float | |
| %3201 = torch.aten.item %3199 : !torch.vtensor<[],si8> -> !torch.int | |
| %3202 = torch.aten.quantize_per_tensor %102, %3200, %3201, %int12_842 : !torch.vtensor<[960,160,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,160,1,1],!torch.qint8> | |
| %3203 = torch.aten.int_repr %3202 : !torch.vtensor<[960,160,1,1],!torch.qint8> -> !torch.vtensor<[960,160,1,1],si8> | |
| %3204 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3205 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3206 = torch.aten.item %3204 : !torch.vtensor<[],f32> -> !torch.float | |
| %3207 = torch.aten.item %3205 : !torch.vtensor<[],si8> -> !torch.int | |
| %3208 = torch.aten._make_per_tensor_quantized_tensor %3203, %3206, %3207 : !torch.vtensor<[960,160,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,160,1,1],!torch.qint8> | |
| %3209 = torch.aten.dequantize.self %3208 : !torch.vtensor<[960,160,1,1],!torch.qint8> -> !torch.vtensor<[960,160,1,1],f32> | |
| %3210 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3211 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_843 = torch.constant.int 12 | |
| %3212 = torch.aten.item %3210 : !torch.vtensor<[],f32> -> !torch.float | |
| %3213 = torch.aten.item %3211 : !torch.vtensor<[],si8> -> !torch.int | |
| %3214 = torch.aten.quantize_per_tensor %103, %3212, %3213, %int12_843 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3215 = torch.aten.int_repr %3214 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3216 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3217 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3218 = torch.aten.item %3216 : !torch.vtensor<[],f32> -> !torch.float | |
| %3219 = torch.aten.item %3217 : !torch.vtensor<[],si8> -> !torch.int | |
| %3220 = torch.aten._make_per_tensor_quantized_tensor %3215, %3218, %3219 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3221 = torch.aten.dequantize.self %3220 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int0_844 = torch.constant.int 0 | |
| %int0_845 = torch.constant.int 0 | |
| %int1_846 = torch.constant.int 1 | |
| %int1_847 = torch.constant.int 1 | |
| %int1_848 = torch.constant.int 1 | |
| %int1_849 = torch.constant.int 1 | |
| %int0_850 = torch.constant.int 0 | |
| %3222 = torch.prim.ListConstruct %int0_844, %int0_845 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3223 = torch.prim.ListConstruct %int1_846, %int1_847 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3224 = torch.prim.ListConstruct %int1_848, %int1_849 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3225 = torch.prim.ListConstruct %int0_850, %int0_850 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_851 = torch.constant.bool false | |
| %int1_852 = torch.constant.int 1 | |
| %3226 = torch.aten.convolution %3197, %3209, %3221, %3224, %3222, %3223, %false_851, %3225, %int1_852 : !torch.vtensor<[1,160,14,14],f32>, !torch.vtensor<[960,160,1,1],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3227 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3228 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_853 = torch.constant.int 12 | |
| %3229 = torch.aten.item %3227 : !torch.vtensor<[],f32> -> !torch.float | |
| %3230 = torch.aten.item %3228 : !torch.vtensor<[],si8> -> !torch.int | |
| %3231 = torch.aten.quantize_per_tensor %3226, %3229, %3230, %int12_853 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3232 = torch.aten.int_repr %3231 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3233 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3234 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3235 = torch.aten.item %3233 : !torch.vtensor<[],f32> -> !torch.float | |
| %3236 = torch.aten.item %3234 : !torch.vtensor<[],si8> -> !torch.int | |
| %3237 = torch.aten._make_per_tensor_quantized_tensor %3232, %3235, %3236 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3238 = torch.aten.dequantize.self %3237 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3239 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_854 = torch.constant.int 1 | |
| %3240 = torch.aten.add.Tensor %3238, %3239, %int1_854 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3241 = torch.aten.relu %3240 : !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3242 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3243 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_855 = torch.constant.int 6 | |
| %none_856 = torch.constant.none | |
| %false_857 = torch.constant.bool false | |
| %3244 = torch.aten.to.dtype %3242, %int6_855, %false_857, %false_857, %none_856 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_858 = torch.constant.int 6 | |
| %none_859 = torch.constant.none | |
| %false_860 = torch.constant.bool false | |
| %3245 = torch.aten.to.dtype %3243, %int6_858, %false_860, %false_860, %none_859 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3246 = torch.aten.clamp.Tensor %3241, %3244, %3245 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3247 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3248 = torch.aten.mul.Tensor %3246, %3247 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3249 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3250 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_861 = torch.constant.int 12 | |
| %3251 = torch.aten.item %3249 : !torch.vtensor<[],f32> -> !torch.float | |
| %3252 = torch.aten.item %3250 : !torch.vtensor<[],si8> -> !torch.int | |
| %3253 = torch.aten.quantize_per_tensor %3248, %3251, %3252, %int12_861 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3254 = torch.aten.int_repr %3253 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3255 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3256 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3257 = torch.aten.item %3255 : !torch.vtensor<[],f32> -> !torch.float | |
| %3258 = torch.aten.item %3256 : !torch.vtensor<[],si8> -> !torch.int | |
| %3259 = torch.aten._make_per_tensor_quantized_tensor %3254, %3257, %3258 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3260 = torch.aten.dequantize.self %3259 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3261 = torch.aten.mul.Tensor %3238, %3260 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3262 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3263 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_862 = torch.constant.int 12 | |
| %3264 = torch.aten.item %3262 : !torch.vtensor<[],f32> -> !torch.float | |
| %3265 = torch.aten.item %3263 : !torch.vtensor<[],si8> -> !torch.int | |
| %3266 = torch.aten.quantize_per_tensor %3261, %3264, %3265, %int12_862 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3267 = torch.aten.int_repr %3266 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3268 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3269 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3270 = torch.aten.item %3268 : !torch.vtensor<[],f32> -> !torch.float | |
| %3271 = torch.aten.item %3269 : !torch.vtensor<[],si8> -> !torch.int | |
| %3272 = torch.aten._make_per_tensor_quantized_tensor %3267, %3270, %3271 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3273 = torch.aten.dequantize.self %3272 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3274 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3275 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_863 = torch.constant.int 12 | |
| %3276 = torch.aten.item %3274 : !torch.vtensor<[],f32> -> !torch.float | |
| %3277 = torch.aten.item %3275 : !torch.vtensor<[],si8> -> !torch.int | |
| %3278 = torch.aten.quantize_per_tensor %104, %3276, %3277, %int12_863 : !torch.vtensor<[960,1,5,5],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,1,5,5],!torch.qint8> | |
| %3279 = torch.aten.int_repr %3278 : !torch.vtensor<[960,1,5,5],!torch.qint8> -> !torch.vtensor<[960,1,5,5],si8> | |
| %3280 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3281 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3282 = torch.aten.item %3280 : !torch.vtensor<[],f32> -> !torch.float | |
| %3283 = torch.aten.item %3281 : !torch.vtensor<[],si8> -> !torch.int | |
| %3284 = torch.aten._make_per_tensor_quantized_tensor %3279, %3282, %3283 : !torch.vtensor<[960,1,5,5],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,1,5,5],!torch.qint8> | |
| %3285 = torch.aten.dequantize.self %3284 : !torch.vtensor<[960,1,5,5],!torch.qint8> -> !torch.vtensor<[960,1,5,5],f32> | |
| %3286 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3287 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_864 = torch.constant.int 12 | |
| %3288 = torch.aten.item %3286 : !torch.vtensor<[],f32> -> !torch.float | |
| %3289 = torch.aten.item %3287 : !torch.vtensor<[],si8> -> !torch.int | |
| %3290 = torch.aten.quantize_per_tensor %105, %3288, %3289, %int12_864 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3291 = torch.aten.int_repr %3290 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3292 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3293 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3294 = torch.aten.item %3292 : !torch.vtensor<[],f32> -> !torch.float | |
| %3295 = torch.aten.item %3293 : !torch.vtensor<[],si8> -> !torch.int | |
| %3296 = torch.aten._make_per_tensor_quantized_tensor %3291, %3294, %3295 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3297 = torch.aten.dequantize.self %3296 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int4_865 = torch.constant.int 4 | |
| %int4_866 = torch.constant.int 4 | |
| %int2_867 = torch.constant.int 2 | |
| %int2_868 = torch.constant.int 2 | |
| %int1_869 = torch.constant.int 1 | |
| %int1_870 = torch.constant.int 1 | |
| %int0_871 = torch.constant.int 0 | |
| %3298 = torch.prim.ListConstruct %int4_865, %int4_866 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3299 = torch.prim.ListConstruct %int2_867, %int2_868 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3300 = torch.prim.ListConstruct %int1_869, %int1_870 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3301 = torch.prim.ListConstruct %int0_871, %int0_871 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_872 = torch.constant.bool false | |
| %int960 = torch.constant.int 960 | |
| %3302 = torch.aten.convolution %3273, %3285, %3297, %3300, %3298, %3299, %false_872, %3301, %int960 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[960,1,5,5],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3303 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3304 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_873 = torch.constant.int 12 | |
| %3305 = torch.aten.item %3303 : !torch.vtensor<[],f32> -> !torch.float | |
| %3306 = torch.aten.item %3304 : !torch.vtensor<[],si8> -> !torch.int | |
| %3307 = torch.aten.quantize_per_tensor %3302, %3305, %3306, %int12_873 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3308 = torch.aten.int_repr %3307 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3309 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3310 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3311 = torch.aten.item %3309 : !torch.vtensor<[],f32> -> !torch.float | |
| %3312 = torch.aten.item %3310 : !torch.vtensor<[],si8> -> !torch.int | |
| %3313 = torch.aten._make_per_tensor_quantized_tensor %3308, %3311, %3312 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3314 = torch.aten.dequantize.self %3313 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3315 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_874 = torch.constant.int 1 | |
| %3316 = torch.aten.add.Tensor %3314, %3315, %int1_874 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3317 = torch.aten.relu %3316 : !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3318 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3319 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_875 = torch.constant.int 6 | |
| %none_876 = torch.constant.none | |
| %false_877 = torch.constant.bool false | |
| %3320 = torch.aten.to.dtype %3318, %int6_875, %false_877, %false_877, %none_876 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_878 = torch.constant.int 6 | |
| %none_879 = torch.constant.none | |
| %false_880 = torch.constant.bool false | |
| %3321 = torch.aten.to.dtype %3319, %int6_878, %false_880, %false_880, %none_879 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3322 = torch.aten.clamp.Tensor %3317, %3320, %3321 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3323 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3324 = torch.aten.mul.Tensor %3322, %3323 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3325 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3326 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_881 = torch.constant.int 12 | |
| %3327 = torch.aten.item %3325 : !torch.vtensor<[],f32> -> !torch.float | |
| %3328 = torch.aten.item %3326 : !torch.vtensor<[],si8> -> !torch.int | |
| %3329 = torch.aten.quantize_per_tensor %3324, %3327, %3328, %int12_881 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3330 = torch.aten.int_repr %3329 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3331 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3332 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3333 = torch.aten.item %3331 : !torch.vtensor<[],f32> -> !torch.float | |
| %3334 = torch.aten.item %3332 : !torch.vtensor<[],si8> -> !torch.int | |
| %3335 = torch.aten._make_per_tensor_quantized_tensor %3330, %3333, %3334 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3336 = torch.aten.dequantize.self %3335 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3337 = torch.aten.mul.Tensor %3314, %3336 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3338 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3339 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_882 = torch.constant.int 12 | |
| %3340 = torch.aten.item %3338 : !torch.vtensor<[],f32> -> !torch.float | |
| %3341 = torch.aten.item %3339 : !torch.vtensor<[],si8> -> !torch.int | |
| %3342 = torch.aten.quantize_per_tensor %3337, %3340, %3341, %int12_882 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3343 = torch.aten.int_repr %3342 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3344 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3345 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3346 = torch.aten.item %3344 : !torch.vtensor<[],f32> -> !torch.float | |
| %3347 = torch.aten.item %3345 : !torch.vtensor<[],si8> -> !torch.int | |
| %3348 = torch.aten._make_per_tensor_quantized_tensor %3343, %3346, %3347 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3349 = torch.aten.dequantize.self %3348 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %int0_883 = torch.constant.int 0 | |
| %int1_884 = torch.constant.int 1 | |
| %int14_885 = torch.constant.int 14 | |
| %int14_886 = torch.constant.int 14 | |
| %3350 = torch.prim.ListConstruct %int14_885, %int14_886 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3351 = torch.prim.ListConstruct %int0_883, %int0_883 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3352 = torch.prim.ListConstruct %int1_884, %int1_884 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_887 = torch.constant.bool false | |
| %none_888 = torch.constant.none | |
| %3353 = torch.aten.avg_pool2d %3349, %3350, %3352, %3351, %false_887, %false_887, %none_888 : !torch.vtensor<[1,960,14,14],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,960,1,1],f32> | |
| %3354 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3355 = torch.aten.mul.Tensor %3353, %3354 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3356 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3357 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_889 = torch.constant.int 12 | |
| %3358 = torch.aten.item %3356 : !torch.vtensor<[],f32> -> !torch.float | |
| %3359 = torch.aten.item %3357 : !torch.vtensor<[],si8> -> !torch.int | |
| %3360 = torch.aten.quantize_per_tensor %3355, %3358, %3359, %int12_889 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3361 = torch.aten.int_repr %3360 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %3362 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3363 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3364 = torch.aten.item %3362 : !torch.vtensor<[],f32> -> !torch.float | |
| %3365 = torch.aten.item %3363 : !torch.vtensor<[],si8> -> !torch.int | |
| %3366 = torch.aten._make_per_tensor_quantized_tensor %3361, %3364, %3365 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3367 = torch.aten.dequantize.self %3366 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3368 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3369 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_890 = torch.constant.int 12 | |
| %3370 = torch.aten.item %3368 : !torch.vtensor<[],f32> -> !torch.float | |
| %3371 = torch.aten.item %3369 : !torch.vtensor<[],si8> -> !torch.int | |
| %3372 = torch.aten.quantize_per_tensor %106, %3370, %3371, %int12_890 : !torch.vtensor<[240,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240,960,1,1],!torch.qint8> | |
| %3373 = torch.aten.int_repr %3372 : !torch.vtensor<[240,960,1,1],!torch.qint8> -> !torch.vtensor<[240,960,1,1],si8> | |
| %3374 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3375 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3376 = torch.aten.item %3374 : !torch.vtensor<[],f32> -> !torch.float | |
| %3377 = torch.aten.item %3375 : !torch.vtensor<[],si8> -> !torch.int | |
| %3378 = torch.aten._make_per_tensor_quantized_tensor %3373, %3376, %3377 : !torch.vtensor<[240,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[240,960,1,1],!torch.qint8> | |
| %3379 = torch.aten.dequantize.self %3378 : !torch.vtensor<[240,960,1,1],!torch.qint8> -> !torch.vtensor<[240,960,1,1],f32> | |
| %3380 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3381 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_891 = torch.constant.int 12 | |
| %3382 = torch.aten.item %3380 : !torch.vtensor<[],f32> -> !torch.float | |
| %3383 = torch.aten.item %3381 : !torch.vtensor<[],si8> -> !torch.int | |
| %3384 = torch.aten.quantize_per_tensor %107, %3382, %3383, %int12_891 : !torch.vtensor<[240],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %3385 = torch.aten.int_repr %3384 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],si8> | |
| %3386 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3387 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3388 = torch.aten.item %3386 : !torch.vtensor<[],f32> -> !torch.float | |
| %3389 = torch.aten.item %3387 : !torch.vtensor<[],si8> -> !torch.int | |
| %3390 = torch.aten._make_per_tensor_quantized_tensor %3385, %3388, %3389 : !torch.vtensor<[240],si8>, !torch.float, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %3391 = torch.aten.dequantize.self %3390 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],f32> | |
| %int0_892 = torch.constant.int 0 | |
| %int0_893 = torch.constant.int 0 | |
| %int1_894 = torch.constant.int 1 | |
| %int1_895 = torch.constant.int 1 | |
| %int1_896 = torch.constant.int 1 | |
| %int1_897 = torch.constant.int 1 | |
| %int0_898 = torch.constant.int 0 | |
| %3392 = torch.prim.ListConstruct %int0_892, %int0_893 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3393 = torch.prim.ListConstruct %int1_894, %int1_895 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3394 = torch.prim.ListConstruct %int1_896, %int1_897 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3395 = torch.prim.ListConstruct %int0_898, %int0_898 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_899 = torch.constant.bool false | |
| %int1_900 = torch.constant.int 1 | |
| %3396 = torch.aten.convolution %3367, %3379, %3391, %3394, %3392, %3393, %false_899, %3395, %int1_900 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[240,960,1,1],f32>, !torch.vtensor<[240],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,240,1,1],f32> | |
| %3397 = torch.aten.relu %3396 : !torch.vtensor<[1,240,1,1],f32> -> !torch.vtensor<[1,240,1,1],f32> | |
| %3398 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3399 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_901 = torch.constant.int 12 | |
| %3400 = torch.aten.item %3398 : !torch.vtensor<[],f32> -> !torch.float | |
| %3401 = torch.aten.item %3399 : !torch.vtensor<[],si8> -> !torch.int | |
| %3402 = torch.aten.quantize_per_tensor %3397, %3400, %3401, %int12_901 : !torch.vtensor<[1,240,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,1,1],!torch.qint8> | |
| %3403 = torch.aten.int_repr %3402 : !torch.vtensor<[1,240,1,1],!torch.qint8> -> !torch.vtensor<[1,240,1,1],si8> | |
| %3404 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3405 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3406 = torch.aten.item %3404 : !torch.vtensor<[],f32> -> !torch.float | |
| %3407 = torch.aten.item %3405 : !torch.vtensor<[],si8> -> !torch.int | |
| %3408 = torch.aten._make_per_tensor_quantized_tensor %3403, %3406, %3407 : !torch.vtensor<[1,240,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,1,1],!torch.qint8> | |
| %3409 = torch.aten.dequantize.self %3408 : !torch.vtensor<[1,240,1,1],!torch.qint8> -> !torch.vtensor<[1,240,1,1],f32> | |
| %3410 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3411 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_902 = torch.constant.int 12 | |
| %3412 = torch.aten.item %3410 : !torch.vtensor<[],f32> -> !torch.float | |
| %3413 = torch.aten.item %3411 : !torch.vtensor<[],si8> -> !torch.int | |
| %3414 = torch.aten.quantize_per_tensor %108, %3412, %3413, %int12_902 : !torch.vtensor<[960,240,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,240,1,1],!torch.qint8> | |
| %3415 = torch.aten.int_repr %3414 : !torch.vtensor<[960,240,1,1],!torch.qint8> -> !torch.vtensor<[960,240,1,1],si8> | |
| %3416 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3417 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3418 = torch.aten.item %3416 : !torch.vtensor<[],f32> -> !torch.float | |
| %3419 = torch.aten.item %3417 : !torch.vtensor<[],si8> -> !torch.int | |
| %3420 = torch.aten._make_per_tensor_quantized_tensor %3415, %3418, %3419 : !torch.vtensor<[960,240,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,240,1,1],!torch.qint8> | |
| %3421 = torch.aten.dequantize.self %3420 : !torch.vtensor<[960,240,1,1],!torch.qint8> -> !torch.vtensor<[960,240,1,1],f32> | |
| %3422 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3423 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_903 = torch.constant.int 12 | |
| %3424 = torch.aten.item %3422 : !torch.vtensor<[],f32> -> !torch.float | |
| %3425 = torch.aten.item %3423 : !torch.vtensor<[],si8> -> !torch.int | |
| %3426 = torch.aten.quantize_per_tensor %109, %3424, %3425, %int12_903 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3427 = torch.aten.int_repr %3426 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3428 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3429 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3430 = torch.aten.item %3428 : !torch.vtensor<[],f32> -> !torch.float | |
| %3431 = torch.aten.item %3429 : !torch.vtensor<[],si8> -> !torch.int | |
| %3432 = torch.aten._make_per_tensor_quantized_tensor %3427, %3430, %3431 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3433 = torch.aten.dequantize.self %3432 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int0_904 = torch.constant.int 0 | |
| %int0_905 = torch.constant.int 0 | |
| %int1_906 = torch.constant.int 1 | |
| %int1_907 = torch.constant.int 1 | |
| %int1_908 = torch.constant.int 1 | |
| %int1_909 = torch.constant.int 1 | |
| %int0_910 = torch.constant.int 0 | |
| %3434 = torch.prim.ListConstruct %int0_904, %int0_905 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3435 = torch.prim.ListConstruct %int1_906, %int1_907 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3436 = torch.prim.ListConstruct %int1_908, %int1_909 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3437 = torch.prim.ListConstruct %int0_910, %int0_910 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_911 = torch.constant.bool false | |
| %int1_912 = torch.constant.int 1 | |
| %3438 = torch.aten.convolution %3409, %3421, %3433, %3436, %3434, %3435, %false_911, %3437, %int1_912 : !torch.vtensor<[1,240,1,1],f32>, !torch.vtensor<[960,240,1,1],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,1,1],f32> | |
| %3439 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3440 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_913 = torch.constant.int 12 | |
| %3441 = torch.aten.item %3439 : !torch.vtensor<[],f32> -> !torch.float | |
| %3442 = torch.aten.item %3440 : !torch.vtensor<[],si8> -> !torch.int | |
| %3443 = torch.aten.quantize_per_tensor %3438, %3441, %3442, %int12_913 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3444 = torch.aten.int_repr %3443 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %3445 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3446 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3447 = torch.aten.item %3445 : !torch.vtensor<[],f32> -> !torch.float | |
| %3448 = torch.aten.item %3446 : !torch.vtensor<[],si8> -> !torch.int | |
| %3449 = torch.aten._make_per_tensor_quantized_tensor %3444, %3447, %3448 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3450 = torch.aten.dequantize.self %3449 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3451 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_914 = torch.constant.int 1 | |
| %3452 = torch.aten.add.Tensor %3450, %3451, %int1_914 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,1,1],f32> | |
| %3453 = torch.aten.relu %3452 : !torch.vtensor<[1,960,1,1],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3454 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3455 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_915 = torch.constant.int 6 | |
| %none_916 = torch.constant.none | |
| %false_917 = torch.constant.bool false | |
| %3456 = torch.aten.to.dtype %3454, %int6_915, %false_917, %false_917, %none_916 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_918 = torch.constant.int 6 | |
| %none_919 = torch.constant.none | |
| %false_920 = torch.constant.bool false | |
| %3457 = torch.aten.to.dtype %3455, %int6_918, %false_920, %false_920, %none_919 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3458 = torch.aten.clamp.Tensor %3453, %3456, %3457 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3459 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3460 = torch.aten.mul.Tensor %3458, %3459 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3461 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3462 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_921 = torch.constant.int 12 | |
| %3463 = torch.aten.item %3461 : !torch.vtensor<[],f32> -> !torch.float | |
| %3464 = torch.aten.item %3462 : !torch.vtensor<[],si8> -> !torch.int | |
| %3465 = torch.aten.quantize_per_tensor %3460, %3463, %3464, %int12_921 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3466 = torch.aten.int_repr %3465 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %3467 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3468 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3469 = torch.aten.item %3467 : !torch.vtensor<[],f32> -> !torch.float | |
| %3470 = torch.aten.item %3468 : !torch.vtensor<[],si8> -> !torch.int | |
| %3471 = torch.aten._make_per_tensor_quantized_tensor %3466, %3469, %3470 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3472 = torch.aten.dequantize.self %3471 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3473 = torch.aten.mul.Tensor %3472, %3349 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3474 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3475 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_922 = torch.constant.int 12 | |
| %3476 = torch.aten.item %3474 : !torch.vtensor<[],f32> -> !torch.float | |
| %3477 = torch.aten.item %3475 : !torch.vtensor<[],si8> -> !torch.int | |
| %3478 = torch.aten.quantize_per_tensor %3473, %3476, %3477, %int12_922 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3479 = torch.aten.int_repr %3478 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3480 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3481 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3482 = torch.aten.item %3480 : !torch.vtensor<[],f32> -> !torch.float | |
| %3483 = torch.aten.item %3481 : !torch.vtensor<[],si8> -> !torch.int | |
| %3484 = torch.aten._make_per_tensor_quantized_tensor %3479, %3482, %3483 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3485 = torch.aten.dequantize.self %3484 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3486 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3487 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_923 = torch.constant.int 12 | |
| %3488 = torch.aten.item %3486 : !torch.vtensor<[],f32> -> !torch.float | |
| %3489 = torch.aten.item %3487 : !torch.vtensor<[],si8> -> !torch.int | |
| %3490 = torch.aten.quantize_per_tensor %110, %3488, %3489, %int12_923 : !torch.vtensor<[160,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[160,960,1,1],!torch.qint8> | |
| %3491 = torch.aten.int_repr %3490 : !torch.vtensor<[160,960,1,1],!torch.qint8> -> !torch.vtensor<[160,960,1,1],si8> | |
| %3492 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3493 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3494 = torch.aten.item %3492 : !torch.vtensor<[],f32> -> !torch.float | |
| %3495 = torch.aten.item %3493 : !torch.vtensor<[],si8> -> !torch.int | |
| %3496 = torch.aten._make_per_tensor_quantized_tensor %3491, %3494, %3495 : !torch.vtensor<[160,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[160,960,1,1],!torch.qint8> | |
| %3497 = torch.aten.dequantize.self %3496 : !torch.vtensor<[160,960,1,1],!torch.qint8> -> !torch.vtensor<[160,960,1,1],f32> | |
| %3498 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3499 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_924 = torch.constant.int 12 | |
| %3500 = torch.aten.item %3498 : !torch.vtensor<[],f32> -> !torch.float | |
| %3501 = torch.aten.item %3499 : !torch.vtensor<[],si8> -> !torch.int | |
| %3502 = torch.aten.quantize_per_tensor %111, %3500, %3501, %int12_924 : !torch.vtensor<[160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[160],!torch.qint8> | |
| %3503 = torch.aten.int_repr %3502 : !torch.vtensor<[160],!torch.qint8> -> !torch.vtensor<[160],si8> | |
| %3504 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3505 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3506 = torch.aten.item %3504 : !torch.vtensor<[],f32> -> !torch.float | |
| %3507 = torch.aten.item %3505 : !torch.vtensor<[],si8> -> !torch.int | |
| %3508 = torch.aten._make_per_tensor_quantized_tensor %3503, %3506, %3507 : !torch.vtensor<[160],si8>, !torch.float, !torch.int -> !torch.vtensor<[160],!torch.qint8> | |
| %3509 = torch.aten.dequantize.self %3508 : !torch.vtensor<[160],!torch.qint8> -> !torch.vtensor<[160],f32> | |
| %int0_925 = torch.constant.int 0 | |
| %int0_926 = torch.constant.int 0 | |
| %int1_927 = torch.constant.int 1 | |
| %int1_928 = torch.constant.int 1 | |
| %int1_929 = torch.constant.int 1 | |
| %int1_930 = torch.constant.int 1 | |
| %int0_931 = torch.constant.int 0 | |
| %3510 = torch.prim.ListConstruct %int0_925, %int0_926 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3511 = torch.prim.ListConstruct %int1_927, %int1_928 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3512 = torch.prim.ListConstruct %int1_929, %int1_930 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3513 = torch.prim.ListConstruct %int0_931, %int0_931 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_932 = torch.constant.bool false | |
| %int1_933 = torch.constant.int 1 | |
| %3514 = torch.aten.convolution %3485, %3497, %3509, %3512, %3510, %3511, %false_932, %3513, %int1_933 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[160,960,1,1],f32>, !torch.vtensor<[160],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,160,14,14],f32> | |
| %3515 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3516 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_934 = torch.constant.int 12 | |
| %3517 = torch.aten.item %3515 : !torch.vtensor<[],f32> -> !torch.float | |
| %3518 = torch.aten.item %3516 : !torch.vtensor<[],si8> -> !torch.int | |
| %3519 = torch.aten.quantize_per_tensor %3514, %3517, %3518, %int12_934 : !torch.vtensor<[1,160,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3520 = torch.aten.int_repr %3519 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],si8> | |
| %3521 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3522 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3523 = torch.aten.item %3521 : !torch.vtensor<[],f32> -> !torch.float | |
| %3524 = torch.aten.item %3522 : !torch.vtensor<[],si8> -> !torch.int | |
| %3525 = torch.aten._make_per_tensor_quantized_tensor %3520, %3523, %3524 : !torch.vtensor<[1,160,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3526 = torch.aten.dequantize.self %3525 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],f32> | |
| %int1_935 = torch.constant.int 1 | |
| %3527 = torch.aten.add.Tensor %3526, %3197, %int1_935 : !torch.vtensor<[1,160,14,14],f32>, !torch.vtensor<[1,160,14,14],f32>, !torch.int -> !torch.vtensor<[1,160,14,14],f32> | |
| %3528 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3529 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_936 = torch.constant.int 12 | |
| %3530 = torch.aten.item %3528 : !torch.vtensor<[],f32> -> !torch.float | |
| %3531 = torch.aten.item %3529 : !torch.vtensor<[],si8> -> !torch.int | |
| %3532 = torch.aten.quantize_per_tensor %3527, %3530, %3531, %int12_936 : !torch.vtensor<[1,160,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3533 = torch.aten.int_repr %3532 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],si8> | |
| %3534 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3535 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3536 = torch.aten.item %3534 : !torch.vtensor<[],f32> -> !torch.float | |
| %3537 = torch.aten.item %3535 : !torch.vtensor<[],si8> -> !torch.int | |
| %3538 = torch.aten._make_per_tensor_quantized_tensor %3533, %3536, %3537 : !torch.vtensor<[1,160,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3539 = torch.aten.dequantize.self %3538 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],f32> | |
| %3540 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3541 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_937 = torch.constant.int 12 | |
| %3542 = torch.aten.item %3540 : !torch.vtensor<[],f32> -> !torch.float | |
| %3543 = torch.aten.item %3541 : !torch.vtensor<[],si8> -> !torch.int | |
| %3544 = torch.aten.quantize_per_tensor %112, %3542, %3543, %int12_937 : !torch.vtensor<[960,160,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,160,1,1],!torch.qint8> | |
| %3545 = torch.aten.int_repr %3544 : !torch.vtensor<[960,160,1,1],!torch.qint8> -> !torch.vtensor<[960,160,1,1],si8> | |
| %3546 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3547 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3548 = torch.aten.item %3546 : !torch.vtensor<[],f32> -> !torch.float | |
| %3549 = torch.aten.item %3547 : !torch.vtensor<[],si8> -> !torch.int | |
| %3550 = torch.aten._make_per_tensor_quantized_tensor %3545, %3548, %3549 : !torch.vtensor<[960,160,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,160,1,1],!torch.qint8> | |
| %3551 = torch.aten.dequantize.self %3550 : !torch.vtensor<[960,160,1,1],!torch.qint8> -> !torch.vtensor<[960,160,1,1],f32> | |
| %3552 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3553 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_938 = torch.constant.int 12 | |
| %3554 = torch.aten.item %3552 : !torch.vtensor<[],f32> -> !torch.float | |
| %3555 = torch.aten.item %3553 : !torch.vtensor<[],si8> -> !torch.int | |
| %3556 = torch.aten.quantize_per_tensor %113, %3554, %3555, %int12_938 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3557 = torch.aten.int_repr %3556 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3558 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3559 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3560 = torch.aten.item %3558 : !torch.vtensor<[],f32> -> !torch.float | |
| %3561 = torch.aten.item %3559 : !torch.vtensor<[],si8> -> !torch.int | |
| %3562 = torch.aten._make_per_tensor_quantized_tensor %3557, %3560, %3561 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3563 = torch.aten.dequantize.self %3562 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int0_939 = torch.constant.int 0 | |
| %int0_940 = torch.constant.int 0 | |
| %int1_941 = torch.constant.int 1 | |
| %int1_942 = torch.constant.int 1 | |
| %int1_943 = torch.constant.int 1 | |
| %int1_944 = torch.constant.int 1 | |
| %int0_945 = torch.constant.int 0 | |
| %3564 = torch.prim.ListConstruct %int0_939, %int0_940 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3565 = torch.prim.ListConstruct %int1_941, %int1_942 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3566 = torch.prim.ListConstruct %int1_943, %int1_944 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3567 = torch.prim.ListConstruct %int0_945, %int0_945 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_946 = torch.constant.bool false | |
| %int1_947 = torch.constant.int 1 | |
| %3568 = torch.aten.convolution %3539, %3551, %3563, %3566, %3564, %3565, %false_946, %3567, %int1_947 : !torch.vtensor<[1,160,14,14],f32>, !torch.vtensor<[960,160,1,1],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3569 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3570 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_948 = torch.constant.int 12 | |
| %3571 = torch.aten.item %3569 : !torch.vtensor<[],f32> -> !torch.float | |
| %3572 = torch.aten.item %3570 : !torch.vtensor<[],si8> -> !torch.int | |
| %3573 = torch.aten.quantize_per_tensor %3568, %3571, %3572, %int12_948 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3574 = torch.aten.int_repr %3573 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3575 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3576 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3577 = torch.aten.item %3575 : !torch.vtensor<[],f32> -> !torch.float | |
| %3578 = torch.aten.item %3576 : !torch.vtensor<[],si8> -> !torch.int | |
| %3579 = torch.aten._make_per_tensor_quantized_tensor %3574, %3577, %3578 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3580 = torch.aten.dequantize.self %3579 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3581 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_949 = torch.constant.int 1 | |
| %3582 = torch.aten.add.Tensor %3580, %3581, %int1_949 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3583 = torch.aten.relu %3582 : !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3584 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3585 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_950 = torch.constant.int 6 | |
| %none_951 = torch.constant.none | |
| %false_952 = torch.constant.bool false | |
| %3586 = torch.aten.to.dtype %3584, %int6_950, %false_952, %false_952, %none_951 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_953 = torch.constant.int 6 | |
| %none_954 = torch.constant.none | |
| %false_955 = torch.constant.bool false | |
| %3587 = torch.aten.to.dtype %3585, %int6_953, %false_955, %false_955, %none_954 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3588 = torch.aten.clamp.Tensor %3583, %3586, %3587 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3589 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3590 = torch.aten.mul.Tensor %3588, %3589 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3591 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3592 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_956 = torch.constant.int 12 | |
| %3593 = torch.aten.item %3591 : !torch.vtensor<[],f32> -> !torch.float | |
| %3594 = torch.aten.item %3592 : !torch.vtensor<[],si8> -> !torch.int | |
| %3595 = torch.aten.quantize_per_tensor %3590, %3593, %3594, %int12_956 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3596 = torch.aten.int_repr %3595 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3597 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3598 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3599 = torch.aten.item %3597 : !torch.vtensor<[],f32> -> !torch.float | |
| %3600 = torch.aten.item %3598 : !torch.vtensor<[],si8> -> !torch.int | |
| %3601 = torch.aten._make_per_tensor_quantized_tensor %3596, %3599, %3600 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3602 = torch.aten.dequantize.self %3601 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3603 = torch.aten.mul.Tensor %3580, %3602 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3604 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3605 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_957 = torch.constant.int 12 | |
| %3606 = torch.aten.item %3604 : !torch.vtensor<[],f32> -> !torch.float | |
| %3607 = torch.aten.item %3605 : !torch.vtensor<[],si8> -> !torch.int | |
| %3608 = torch.aten.quantize_per_tensor %3603, %3606, %3607, %int12_957 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3609 = torch.aten.int_repr %3608 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3610 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3611 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3612 = torch.aten.item %3610 : !torch.vtensor<[],f32> -> !torch.float | |
| %3613 = torch.aten.item %3611 : !torch.vtensor<[],si8> -> !torch.int | |
| %3614 = torch.aten._make_per_tensor_quantized_tensor %3609, %3612, %3613 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3615 = torch.aten.dequantize.self %3614 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3616 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3617 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_958 = torch.constant.int 12 | |
| %3618 = torch.aten.item %3616 : !torch.vtensor<[],f32> -> !torch.float | |
| %3619 = torch.aten.item %3617 : !torch.vtensor<[],si8> -> !torch.int | |
| %3620 = torch.aten.quantize_per_tensor %114, %3618, %3619, %int12_958 : !torch.vtensor<[960,1,5,5],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,1,5,5],!torch.qint8> | |
| %3621 = torch.aten.int_repr %3620 : !torch.vtensor<[960,1,5,5],!torch.qint8> -> !torch.vtensor<[960,1,5,5],si8> | |
| %3622 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3623 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3624 = torch.aten.item %3622 : !torch.vtensor<[],f32> -> !torch.float | |
| %3625 = torch.aten.item %3623 : !torch.vtensor<[],si8> -> !torch.int | |
| %3626 = torch.aten._make_per_tensor_quantized_tensor %3621, %3624, %3625 : !torch.vtensor<[960,1,5,5],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,1,5,5],!torch.qint8> | |
| %3627 = torch.aten.dequantize.self %3626 : !torch.vtensor<[960,1,5,5],!torch.qint8> -> !torch.vtensor<[960,1,5,5],f32> | |
| %3628 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3629 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_959 = torch.constant.int 12 | |
| %3630 = torch.aten.item %3628 : !torch.vtensor<[],f32> -> !torch.float | |
| %3631 = torch.aten.item %3629 : !torch.vtensor<[],si8> -> !torch.int | |
| %3632 = torch.aten.quantize_per_tensor %115, %3630, %3631, %int12_959 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3633 = torch.aten.int_repr %3632 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3634 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3635 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3636 = torch.aten.item %3634 : !torch.vtensor<[],f32> -> !torch.float | |
| %3637 = torch.aten.item %3635 : !torch.vtensor<[],si8> -> !torch.int | |
| %3638 = torch.aten._make_per_tensor_quantized_tensor %3633, %3636, %3637 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3639 = torch.aten.dequantize.self %3638 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int4_960 = torch.constant.int 4 | |
| %int4_961 = torch.constant.int 4 | |
| %int2_962 = torch.constant.int 2 | |
| %int2_963 = torch.constant.int 2 | |
| %int1_964 = torch.constant.int 1 | |
| %int1_965 = torch.constant.int 1 | |
| %int0_966 = torch.constant.int 0 | |
| %3640 = torch.prim.ListConstruct %int4_960, %int4_961 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3641 = torch.prim.ListConstruct %int2_962, %int2_963 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3642 = torch.prim.ListConstruct %int1_964, %int1_965 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3643 = torch.prim.ListConstruct %int0_966, %int0_966 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_967 = torch.constant.bool false | |
| %int960_968 = torch.constant.int 960 | |
| %3644 = torch.aten.convolution %3615, %3627, %3639, %3642, %3640, %3641, %false_967, %3643, %int960_968 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[960,1,5,5],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3645 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3646 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_969 = torch.constant.int 12 | |
| %3647 = torch.aten.item %3645 : !torch.vtensor<[],f32> -> !torch.float | |
| %3648 = torch.aten.item %3646 : !torch.vtensor<[],si8> -> !torch.int | |
| %3649 = torch.aten.quantize_per_tensor %3644, %3647, %3648, %int12_969 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3650 = torch.aten.int_repr %3649 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3651 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3652 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3653 = torch.aten.item %3651 : !torch.vtensor<[],f32> -> !torch.float | |
| %3654 = torch.aten.item %3652 : !torch.vtensor<[],si8> -> !torch.int | |
| %3655 = torch.aten._make_per_tensor_quantized_tensor %3650, %3653, %3654 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3656 = torch.aten.dequantize.self %3655 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3657 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_970 = torch.constant.int 1 | |
| %3658 = torch.aten.add.Tensor %3656, %3657, %int1_970 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3659 = torch.aten.relu %3658 : !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3660 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3661 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_971 = torch.constant.int 6 | |
| %none_972 = torch.constant.none | |
| %false_973 = torch.constant.bool false | |
| %3662 = torch.aten.to.dtype %3660, %int6_971, %false_973, %false_973, %none_972 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_974 = torch.constant.int 6 | |
| %none_975 = torch.constant.none | |
| %false_976 = torch.constant.bool false | |
| %3663 = torch.aten.to.dtype %3661, %int6_974, %false_976, %false_976, %none_975 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3664 = torch.aten.clamp.Tensor %3659, %3662, %3663 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3665 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3666 = torch.aten.mul.Tensor %3664, %3665 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3667 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3668 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_977 = torch.constant.int 12 | |
| %3669 = torch.aten.item %3667 : !torch.vtensor<[],f32> -> !torch.float | |
| %3670 = torch.aten.item %3668 : !torch.vtensor<[],si8> -> !torch.int | |
| %3671 = torch.aten.quantize_per_tensor %3666, %3669, %3670, %int12_977 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3672 = torch.aten.int_repr %3671 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3673 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3674 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3675 = torch.aten.item %3673 : !torch.vtensor<[],f32> -> !torch.float | |
| %3676 = torch.aten.item %3674 : !torch.vtensor<[],si8> -> !torch.int | |
| %3677 = torch.aten._make_per_tensor_quantized_tensor %3672, %3675, %3676 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3678 = torch.aten.dequantize.self %3677 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3679 = torch.aten.mul.Tensor %3656, %3678 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3680 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3681 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_978 = torch.constant.int 12 | |
| %3682 = torch.aten.item %3680 : !torch.vtensor<[],f32> -> !torch.float | |
| %3683 = torch.aten.item %3681 : !torch.vtensor<[],si8> -> !torch.int | |
| %3684 = torch.aten.quantize_per_tensor %3679, %3682, %3683, %int12_978 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3685 = torch.aten.int_repr %3684 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3686 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3687 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3688 = torch.aten.item %3686 : !torch.vtensor<[],f32> -> !torch.float | |
| %3689 = torch.aten.item %3687 : !torch.vtensor<[],si8> -> !torch.int | |
| %3690 = torch.aten._make_per_tensor_quantized_tensor %3685, %3688, %3689 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3691 = torch.aten.dequantize.self %3690 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %int0_979 = torch.constant.int 0 | |
| %int1_980 = torch.constant.int 1 | |
| %int14_981 = torch.constant.int 14 | |
| %int14_982 = torch.constant.int 14 | |
| %3692 = torch.prim.ListConstruct %int14_981, %int14_982 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3693 = torch.prim.ListConstruct %int0_979, %int0_979 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3694 = torch.prim.ListConstruct %int1_980, %int1_980 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_983 = torch.constant.bool false | |
| %none_984 = torch.constant.none | |
| %3695 = torch.aten.avg_pool2d %3691, %3692, %3694, %3693, %false_983, %false_983, %none_984 : !torch.vtensor<[1,960,14,14],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,960,1,1],f32> | |
| %3696 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3697 = torch.aten.mul.Tensor %3695, %3696 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3698 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3699 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_985 = torch.constant.int 12 | |
| %3700 = torch.aten.item %3698 : !torch.vtensor<[],f32> -> !torch.float | |
| %3701 = torch.aten.item %3699 : !torch.vtensor<[],si8> -> !torch.int | |
| %3702 = torch.aten.quantize_per_tensor %3697, %3700, %3701, %int12_985 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3703 = torch.aten.int_repr %3702 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %3704 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3705 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3706 = torch.aten.item %3704 : !torch.vtensor<[],f32> -> !torch.float | |
| %3707 = torch.aten.item %3705 : !torch.vtensor<[],si8> -> !torch.int | |
| %3708 = torch.aten._make_per_tensor_quantized_tensor %3703, %3706, %3707 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3709 = torch.aten.dequantize.self %3708 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3710 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3711 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_986 = torch.constant.int 12 | |
| %3712 = torch.aten.item %3710 : !torch.vtensor<[],f32> -> !torch.float | |
| %3713 = torch.aten.item %3711 : !torch.vtensor<[],si8> -> !torch.int | |
| %3714 = torch.aten.quantize_per_tensor %116, %3712, %3713, %int12_986 : !torch.vtensor<[240,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240,960,1,1],!torch.qint8> | |
| %3715 = torch.aten.int_repr %3714 : !torch.vtensor<[240,960,1,1],!torch.qint8> -> !torch.vtensor<[240,960,1,1],si8> | |
| %3716 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3717 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3718 = torch.aten.item %3716 : !torch.vtensor<[],f32> -> !torch.float | |
| %3719 = torch.aten.item %3717 : !torch.vtensor<[],si8> -> !torch.int | |
| %3720 = torch.aten._make_per_tensor_quantized_tensor %3715, %3718, %3719 : !torch.vtensor<[240,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[240,960,1,1],!torch.qint8> | |
| %3721 = torch.aten.dequantize.self %3720 : !torch.vtensor<[240,960,1,1],!torch.qint8> -> !torch.vtensor<[240,960,1,1],f32> | |
| %3722 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3723 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_987 = torch.constant.int 12 | |
| %3724 = torch.aten.item %3722 : !torch.vtensor<[],f32> -> !torch.float | |
| %3725 = torch.aten.item %3723 : !torch.vtensor<[],si8> -> !torch.int | |
| %3726 = torch.aten.quantize_per_tensor %117, %3724, %3725, %int12_987 : !torch.vtensor<[240],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %3727 = torch.aten.int_repr %3726 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],si8> | |
| %3728 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3729 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3730 = torch.aten.item %3728 : !torch.vtensor<[],f32> -> !torch.float | |
| %3731 = torch.aten.item %3729 : !torch.vtensor<[],si8> -> !torch.int | |
| %3732 = torch.aten._make_per_tensor_quantized_tensor %3727, %3730, %3731 : !torch.vtensor<[240],si8>, !torch.float, !torch.int -> !torch.vtensor<[240],!torch.qint8> | |
| %3733 = torch.aten.dequantize.self %3732 : !torch.vtensor<[240],!torch.qint8> -> !torch.vtensor<[240],f32> | |
| %int0_988 = torch.constant.int 0 | |
| %int0_989 = torch.constant.int 0 | |
| %int1_990 = torch.constant.int 1 | |
| %int1_991 = torch.constant.int 1 | |
| %int1_992 = torch.constant.int 1 | |
| %int1_993 = torch.constant.int 1 | |
| %int0_994 = torch.constant.int 0 | |
| %3734 = torch.prim.ListConstruct %int0_988, %int0_989 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3735 = torch.prim.ListConstruct %int1_990, %int1_991 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3736 = torch.prim.ListConstruct %int1_992, %int1_993 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3737 = torch.prim.ListConstruct %int0_994, %int0_994 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_995 = torch.constant.bool false | |
| %int1_996 = torch.constant.int 1 | |
| %3738 = torch.aten.convolution %3709, %3721, %3733, %3736, %3734, %3735, %false_995, %3737, %int1_996 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[240,960,1,1],f32>, !torch.vtensor<[240],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,240,1,1],f32> | |
| %3739 = torch.aten.relu %3738 : !torch.vtensor<[1,240,1,1],f32> -> !torch.vtensor<[1,240,1,1],f32> | |
| %3740 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3741 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_997 = torch.constant.int 12 | |
| %3742 = torch.aten.item %3740 : !torch.vtensor<[],f32> -> !torch.float | |
| %3743 = torch.aten.item %3741 : !torch.vtensor<[],si8> -> !torch.int | |
| %3744 = torch.aten.quantize_per_tensor %3739, %3742, %3743, %int12_997 : !torch.vtensor<[1,240,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,240,1,1],!torch.qint8> | |
| %3745 = torch.aten.int_repr %3744 : !torch.vtensor<[1,240,1,1],!torch.qint8> -> !torch.vtensor<[1,240,1,1],si8> | |
| %3746 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3747 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3748 = torch.aten.item %3746 : !torch.vtensor<[],f32> -> !torch.float | |
| %3749 = torch.aten.item %3747 : !torch.vtensor<[],si8> -> !torch.int | |
| %3750 = torch.aten._make_per_tensor_quantized_tensor %3745, %3748, %3749 : !torch.vtensor<[1,240,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,240,1,1],!torch.qint8> | |
| %3751 = torch.aten.dequantize.self %3750 : !torch.vtensor<[1,240,1,1],!torch.qint8> -> !torch.vtensor<[1,240,1,1],f32> | |
| %3752 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3753 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_998 = torch.constant.int 12 | |
| %3754 = torch.aten.item %3752 : !torch.vtensor<[],f32> -> !torch.float | |
| %3755 = torch.aten.item %3753 : !torch.vtensor<[],si8> -> !torch.int | |
| %3756 = torch.aten.quantize_per_tensor %118, %3754, %3755, %int12_998 : !torch.vtensor<[960,240,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,240,1,1],!torch.qint8> | |
| %3757 = torch.aten.int_repr %3756 : !torch.vtensor<[960,240,1,1],!torch.qint8> -> !torch.vtensor<[960,240,1,1],si8> | |
| %3758 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3759 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3760 = torch.aten.item %3758 : !torch.vtensor<[],f32> -> !torch.float | |
| %3761 = torch.aten.item %3759 : !torch.vtensor<[],si8> -> !torch.int | |
| %3762 = torch.aten._make_per_tensor_quantized_tensor %3757, %3760, %3761 : !torch.vtensor<[960,240,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,240,1,1],!torch.qint8> | |
| %3763 = torch.aten.dequantize.self %3762 : !torch.vtensor<[960,240,1,1],!torch.qint8> -> !torch.vtensor<[960,240,1,1],f32> | |
| %3764 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3765 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_999 = torch.constant.int 12 | |
| %3766 = torch.aten.item %3764 : !torch.vtensor<[],f32> -> !torch.float | |
| %3767 = torch.aten.item %3765 : !torch.vtensor<[],si8> -> !torch.int | |
| %3768 = torch.aten.quantize_per_tensor %119, %3766, %3767, %int12_999 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3769 = torch.aten.int_repr %3768 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3770 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3771 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3772 = torch.aten.item %3770 : !torch.vtensor<[],f32> -> !torch.float | |
| %3773 = torch.aten.item %3771 : !torch.vtensor<[],si8> -> !torch.int | |
| %3774 = torch.aten._make_per_tensor_quantized_tensor %3769, %3772, %3773 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3775 = torch.aten.dequantize.self %3774 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int0_1000 = torch.constant.int 0 | |
| %int0_1001 = torch.constant.int 0 | |
| %int1_1002 = torch.constant.int 1 | |
| %int1_1003 = torch.constant.int 1 | |
| %int1_1004 = torch.constant.int 1 | |
| %int1_1005 = torch.constant.int 1 | |
| %int0_1006 = torch.constant.int 0 | |
| %3776 = torch.prim.ListConstruct %int0_1000, %int0_1001 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3777 = torch.prim.ListConstruct %int1_1002, %int1_1003 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3778 = torch.prim.ListConstruct %int1_1004, %int1_1005 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3779 = torch.prim.ListConstruct %int0_1006, %int0_1006 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1007 = torch.constant.bool false | |
| %int1_1008 = torch.constant.int 1 | |
| %3780 = torch.aten.convolution %3751, %3763, %3775, %3778, %3776, %3777, %false_1007, %3779, %int1_1008 : !torch.vtensor<[1,240,1,1],f32>, !torch.vtensor<[960,240,1,1],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,1,1],f32> | |
| %3781 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3782 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1009 = torch.constant.int 12 | |
| %3783 = torch.aten.item %3781 : !torch.vtensor<[],f32> -> !torch.float | |
| %3784 = torch.aten.item %3782 : !torch.vtensor<[],si8> -> !torch.int | |
| %3785 = torch.aten.quantize_per_tensor %3780, %3783, %3784, %int12_1009 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3786 = torch.aten.int_repr %3785 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %3787 = torch.vtensor.literal(dense<6.250000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3788 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3789 = torch.aten.item %3787 : !torch.vtensor<[],f32> -> !torch.float | |
| %3790 = torch.aten.item %3788 : !torch.vtensor<[],si8> -> !torch.int | |
| %3791 = torch.aten._make_per_tensor_quantized_tensor %3786, %3789, %3790 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3792 = torch.aten.dequantize.self %3791 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3793 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_1010 = torch.constant.int 1 | |
| %3794 = torch.aten.add.Tensor %3792, %3793, %int1_1010 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,1,1],f32> | |
| %3795 = torch.aten.relu %3794 : !torch.vtensor<[1,960,1,1],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3796 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3797 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_1011 = torch.constant.int 6 | |
| %none_1012 = torch.constant.none | |
| %false_1013 = torch.constant.bool false | |
| %3798 = torch.aten.to.dtype %3796, %int6_1011, %false_1013, %false_1013, %none_1012 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_1014 = torch.constant.int 6 | |
| %none_1015 = torch.constant.none | |
| %false_1016 = torch.constant.bool false | |
| %3799 = torch.aten.to.dtype %3797, %int6_1014, %false_1016, %false_1016, %none_1015 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3800 = torch.aten.clamp.Tensor %3795, %3798, %3799 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3801 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3802 = torch.aten.mul.Tensor %3800, %3801 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3803 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3804 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1017 = torch.constant.int 12 | |
| %3805 = torch.aten.item %3803 : !torch.vtensor<[],f32> -> !torch.float | |
| %3806 = torch.aten.item %3804 : !torch.vtensor<[],si8> -> !torch.int | |
| %3807 = torch.aten.quantize_per_tensor %3802, %3805, %3806, %int12_1017 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3808 = torch.aten.int_repr %3807 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %3809 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3810 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3811 = torch.aten.item %3809 : !torch.vtensor<[],f32> -> !torch.float | |
| %3812 = torch.aten.item %3810 : !torch.vtensor<[],si8> -> !torch.int | |
| %3813 = torch.aten._make_per_tensor_quantized_tensor %3808, %3811, %3812 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %3814 = torch.aten.dequantize.self %3813 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %3815 = torch.aten.mul.Tensor %3814, %3691 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3816 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3817 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1018 = torch.constant.int 12 | |
| %3818 = torch.aten.item %3816 : !torch.vtensor<[],f32> -> !torch.float | |
| %3819 = torch.aten.item %3817 : !torch.vtensor<[],si8> -> !torch.int | |
| %3820 = torch.aten.quantize_per_tensor %3815, %3818, %3819, %int12_1018 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3821 = torch.aten.int_repr %3820 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3822 = torch.vtensor.literal(dense<3.906250e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3823 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3824 = torch.aten.item %3822 : !torch.vtensor<[],f32> -> !torch.float | |
| %3825 = torch.aten.item %3823 : !torch.vtensor<[],si8> -> !torch.int | |
| %3826 = torch.aten._make_per_tensor_quantized_tensor %3821, %3824, %3825 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3827 = torch.aten.dequantize.self %3826 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3828 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3829 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1019 = torch.constant.int 12 | |
| %3830 = torch.aten.item %3828 : !torch.vtensor<[],f32> -> !torch.float | |
| %3831 = torch.aten.item %3829 : !torch.vtensor<[],si8> -> !torch.int | |
| %3832 = torch.aten.quantize_per_tensor %120, %3830, %3831, %int12_1019 : !torch.vtensor<[160,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[160,960,1,1],!torch.qint8> | |
| %3833 = torch.aten.int_repr %3832 : !torch.vtensor<[160,960,1,1],!torch.qint8> -> !torch.vtensor<[160,960,1,1],si8> | |
| %3834 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3835 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3836 = torch.aten.item %3834 : !torch.vtensor<[],f32> -> !torch.float | |
| %3837 = torch.aten.item %3835 : !torch.vtensor<[],si8> -> !torch.int | |
| %3838 = torch.aten._make_per_tensor_quantized_tensor %3833, %3836, %3837 : !torch.vtensor<[160,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[160,960,1,1],!torch.qint8> | |
| %3839 = torch.aten.dequantize.self %3838 : !torch.vtensor<[160,960,1,1],!torch.qint8> -> !torch.vtensor<[160,960,1,1],f32> | |
| %3840 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3841 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1020 = torch.constant.int 12 | |
| %3842 = torch.aten.item %3840 : !torch.vtensor<[],f32> -> !torch.float | |
| %3843 = torch.aten.item %3841 : !torch.vtensor<[],si8> -> !torch.int | |
| %3844 = torch.aten.quantize_per_tensor %121, %3842, %3843, %int12_1020 : !torch.vtensor<[160],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[160],!torch.qint8> | |
| %3845 = torch.aten.int_repr %3844 : !torch.vtensor<[160],!torch.qint8> -> !torch.vtensor<[160],si8> | |
| %3846 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3847 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3848 = torch.aten.item %3846 : !torch.vtensor<[],f32> -> !torch.float | |
| %3849 = torch.aten.item %3847 : !torch.vtensor<[],si8> -> !torch.int | |
| %3850 = torch.aten._make_per_tensor_quantized_tensor %3845, %3848, %3849 : !torch.vtensor<[160],si8>, !torch.float, !torch.int -> !torch.vtensor<[160],!torch.qint8> | |
| %3851 = torch.aten.dequantize.self %3850 : !torch.vtensor<[160],!torch.qint8> -> !torch.vtensor<[160],f32> | |
| %int0_1021 = torch.constant.int 0 | |
| %int0_1022 = torch.constant.int 0 | |
| %int1_1023 = torch.constant.int 1 | |
| %int1_1024 = torch.constant.int 1 | |
| %int1_1025 = torch.constant.int 1 | |
| %int1_1026 = torch.constant.int 1 | |
| %int0_1027 = torch.constant.int 0 | |
| %3852 = torch.prim.ListConstruct %int0_1021, %int0_1022 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3853 = torch.prim.ListConstruct %int1_1023, %int1_1024 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3854 = torch.prim.ListConstruct %int1_1025, %int1_1026 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3855 = torch.prim.ListConstruct %int0_1027, %int0_1027 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1028 = torch.constant.bool false | |
| %int1_1029 = torch.constant.int 1 | |
| %3856 = torch.aten.convolution %3827, %3839, %3851, %3854, %3852, %3853, %false_1028, %3855, %int1_1029 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[160,960,1,1],f32>, !torch.vtensor<[160],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,160,14,14],f32> | |
| %3857 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3858 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1030 = torch.constant.int 12 | |
| %3859 = torch.aten.item %3857 : !torch.vtensor<[],f32> -> !torch.float | |
| %3860 = torch.aten.item %3858 : !torch.vtensor<[],si8> -> !torch.int | |
| %3861 = torch.aten.quantize_per_tensor %3856, %3859, %3860, %int12_1030 : !torch.vtensor<[1,160,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3862 = torch.aten.int_repr %3861 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],si8> | |
| %3863 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3864 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3865 = torch.aten.item %3863 : !torch.vtensor<[],f32> -> !torch.float | |
| %3866 = torch.aten.item %3864 : !torch.vtensor<[],si8> -> !torch.int | |
| %3867 = torch.aten._make_per_tensor_quantized_tensor %3862, %3865, %3866 : !torch.vtensor<[1,160,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3868 = torch.aten.dequantize.self %3867 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],f32> | |
| %int1_1031 = torch.constant.int 1 | |
| %3869 = torch.aten.add.Tensor %3868, %3539, %int1_1031 : !torch.vtensor<[1,160,14,14],f32>, !torch.vtensor<[1,160,14,14],f32>, !torch.int -> !torch.vtensor<[1,160,14,14],f32> | |
| %3870 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3871 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1032 = torch.constant.int 12 | |
| %3872 = torch.aten.item %3870 : !torch.vtensor<[],f32> -> !torch.float | |
| %3873 = torch.aten.item %3871 : !torch.vtensor<[],si8> -> !torch.int | |
| %3874 = torch.aten.quantize_per_tensor %3869, %3872, %3873, %int12_1032 : !torch.vtensor<[1,160,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3875 = torch.aten.int_repr %3874 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],si8> | |
| %3876 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3877 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3878 = torch.aten.item %3876 : !torch.vtensor<[],f32> -> !torch.float | |
| %3879 = torch.aten.item %3877 : !torch.vtensor<[],si8> -> !torch.int | |
| %3880 = torch.aten._make_per_tensor_quantized_tensor %3875, %3878, %3879 : !torch.vtensor<[1,160,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,160,14,14],!torch.qint8> | |
| %3881 = torch.aten.dequantize.self %3880 : !torch.vtensor<[1,160,14,14],!torch.qint8> -> !torch.vtensor<[1,160,14,14],f32> | |
| %3882 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3883 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1033 = torch.constant.int 12 | |
| %3884 = torch.aten.item %3882 : !torch.vtensor<[],f32> -> !torch.float | |
| %3885 = torch.aten.item %3883 : !torch.vtensor<[],si8> -> !torch.int | |
| %3886 = torch.aten.quantize_per_tensor %122, %3884, %3885, %int12_1033 : !torch.vtensor<[960,160,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960,160,1,1],!torch.qint8> | |
| %3887 = torch.aten.int_repr %3886 : !torch.vtensor<[960,160,1,1],!torch.qint8> -> !torch.vtensor<[960,160,1,1],si8> | |
| %3888 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3889 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3890 = torch.aten.item %3888 : !torch.vtensor<[],f32> -> !torch.float | |
| %3891 = torch.aten.item %3889 : !torch.vtensor<[],si8> -> !torch.int | |
| %3892 = torch.aten._make_per_tensor_quantized_tensor %3887, %3890, %3891 : !torch.vtensor<[960,160,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[960,160,1,1],!torch.qint8> | |
| %3893 = torch.aten.dequantize.self %3892 : !torch.vtensor<[960,160,1,1],!torch.qint8> -> !torch.vtensor<[960,160,1,1],f32> | |
| %3894 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3895 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1034 = torch.constant.int 12 | |
| %3896 = torch.aten.item %3894 : !torch.vtensor<[],f32> -> !torch.float | |
| %3897 = torch.aten.item %3895 : !torch.vtensor<[],si8> -> !torch.int | |
| %3898 = torch.aten.quantize_per_tensor %123, %3896, %3897, %int12_1034 : !torch.vtensor<[960],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3899 = torch.aten.int_repr %3898 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],si8> | |
| %3900 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3901 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3902 = torch.aten.item %3900 : !torch.vtensor<[],f32> -> !torch.float | |
| %3903 = torch.aten.item %3901 : !torch.vtensor<[],si8> -> !torch.int | |
| %3904 = torch.aten._make_per_tensor_quantized_tensor %3899, %3902, %3903 : !torch.vtensor<[960],si8>, !torch.float, !torch.int -> !torch.vtensor<[960],!torch.qint8> | |
| %3905 = torch.aten.dequantize.self %3904 : !torch.vtensor<[960],!torch.qint8> -> !torch.vtensor<[960],f32> | |
| %int0_1035 = torch.constant.int 0 | |
| %int0_1036 = torch.constant.int 0 | |
| %int1_1037 = torch.constant.int 1 | |
| %int1_1038 = torch.constant.int 1 | |
| %int1_1039 = torch.constant.int 1 | |
| %int1_1040 = torch.constant.int 1 | |
| %int0_1041 = torch.constant.int 0 | |
| %3906 = torch.prim.ListConstruct %int0_1035, %int0_1036 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3907 = torch.prim.ListConstruct %int1_1037, %int1_1038 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3908 = torch.prim.ListConstruct %int1_1039, %int1_1040 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3909 = torch.prim.ListConstruct %int0_1041, %int0_1041 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1042 = torch.constant.bool false | |
| %int1_1043 = torch.constant.int 1 | |
| %3910 = torch.aten.convolution %3881, %3893, %3905, %3908, %3906, %3907, %false_1042, %3909, %int1_1043 : !torch.vtensor<[1,160,14,14],f32>, !torch.vtensor<[960,160,1,1],f32>, !torch.vtensor<[960],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3911 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3912 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1044 = torch.constant.int 12 | |
| %3913 = torch.aten.item %3911 : !torch.vtensor<[],f32> -> !torch.float | |
| %3914 = torch.aten.item %3912 : !torch.vtensor<[],si8> -> !torch.int | |
| %3915 = torch.aten.quantize_per_tensor %3910, %3913, %3914, %int12_1044 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3916 = torch.aten.int_repr %3915 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3917 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3918 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3919 = torch.aten.item %3917 : !torch.vtensor<[],f32> -> !torch.float | |
| %3920 = torch.aten.item %3918 : !torch.vtensor<[],si8> -> !torch.int | |
| %3921 = torch.aten._make_per_tensor_quantized_tensor %3916, %3919, %3920 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3922 = torch.aten.dequantize.self %3921 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3923 = torch.vtensor.literal(dense<3.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int1_1045 = torch.constant.int 1 | |
| %3924 = torch.aten.add.Tensor %3922, %3923, %int1_1045 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.int -> !torch.vtensor<[1,960,14,14],f32> | |
| %3925 = torch.aten.relu %3924 : !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3926 = torch.vtensor.literal(dense<0.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3927 = torch.vtensor.literal(dense<6.000000e+00> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %int6_1046 = torch.constant.int 6 | |
| %none_1047 = torch.constant.none | |
| %false_1048 = torch.constant.bool false | |
| %3928 = torch.aten.to.dtype %3926, %int6_1046, %false_1048, %false_1048, %none_1047 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %int6_1049 = torch.constant.int 6 | |
| %none_1050 = torch.constant.none | |
| %false_1051 = torch.constant.bool false | |
| %3929 = torch.aten.to.dtype %3927, %int6_1049, %false_1051, %false_1051, %none_1050 : !torch.vtensor<[],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> | |
| %3930 = torch.aten.clamp.Tensor %3925, %3928, %3929 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3931 = torch.vtensor.literal(dense<0.166687012> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3932 = torch.aten.mul.Tensor %3930, %3931 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3933 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3934 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1052 = torch.constant.int 12 | |
| %3935 = torch.aten.item %3933 : !torch.vtensor<[],f32> -> !torch.float | |
| %3936 = torch.aten.item %3934 : !torch.vtensor<[],si8> -> !torch.int | |
| %3937 = torch.aten.quantize_per_tensor %3932, %3935, %3936, %int12_1052 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3938 = torch.aten.int_repr %3937 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3939 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3940 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3941 = torch.aten.item %3939 : !torch.vtensor<[],f32> -> !torch.float | |
| %3942 = torch.aten.item %3940 : !torch.vtensor<[],si8> -> !torch.int | |
| %3943 = torch.aten._make_per_tensor_quantized_tensor %3938, %3941, %3942 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3944 = torch.aten.dequantize.self %3943 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3945 = torch.aten.mul.Tensor %3922, %3944 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[1,960,14,14],f32> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3946 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3947 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1053 = torch.constant.int 12 | |
| %3948 = torch.aten.item %3946 : !torch.vtensor<[],f32> -> !torch.float | |
| %3949 = torch.aten.item %3947 : !torch.vtensor<[],si8> -> !torch.int | |
| %3950 = torch.aten.quantize_per_tensor %3945, %3948, %3949, %int12_1053 : !torch.vtensor<[1,960,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3951 = torch.aten.int_repr %3950 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],si8> | |
| %3952 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3953 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3954 = torch.aten.item %3952 : !torch.vtensor<[],f32> -> !torch.float | |
| %3955 = torch.aten.item %3953 : !torch.vtensor<[],si8> -> !torch.int | |
| %3956 = torch.aten._make_per_tensor_quantized_tensor %3951, %3954, %3955 : !torch.vtensor<[1,960,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,14,14],!torch.qint8> | |
| %3957 = torch.aten.dequantize.self %3956 : !torch.vtensor<[1,960,14,14],!torch.qint8> -> !torch.vtensor<[1,960,14,14],f32> | |
| %3958 = torch.vtensor.literal(dense<9.765625E-4> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3959 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1054 = torch.constant.int 12 | |
| %3960 = torch.aten.item %3958 : !torch.vtensor<[],f32> -> !torch.float | |
| %3961 = torch.aten.item %3959 : !torch.vtensor<[],si8> -> !torch.int | |
| %3962 = torch.aten.quantize_per_tensor %124, %3960, %3961, %int12_1054 : !torch.vtensor<[128,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,960,1,1],!torch.qint8> | |
| %3963 = torch.aten.int_repr %3962 : !torch.vtensor<[128,960,1,1],!torch.qint8> -> !torch.vtensor<[128,960,1,1],si8> | |
| %3964 = torch.vtensor.literal(dense<9.765625E-4> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3965 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3966 = torch.aten.item %3964 : !torch.vtensor<[],f32> -> !torch.float | |
| %3967 = torch.aten.item %3965 : !torch.vtensor<[],si8> -> !torch.int | |
| %3968 = torch.aten._make_per_tensor_quantized_tensor %3963, %3966, %3967 : !torch.vtensor<[128,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,960,1,1],!torch.qint8> | |
| %3969 = torch.aten.dequantize.self %3968 : !torch.vtensor<[128,960,1,1],!torch.qint8> -> !torch.vtensor<[128,960,1,1],f32> | |
| %3970 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3971 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1055 = torch.constant.int 12 | |
| %3972 = torch.aten.item %3970 : !torch.vtensor<[],f32> -> !torch.float | |
| %3973 = torch.aten.item %3971 : !torch.vtensor<[],si8> -> !torch.int | |
| %3974 = torch.aten.quantize_per_tensor %125, %3972, %3973, %int12_1055 : !torch.vtensor<[128],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %3975 = torch.aten.int_repr %3974 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],si8> | |
| %3976 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3977 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3978 = torch.aten.item %3976 : !torch.vtensor<[],f32> -> !torch.float | |
| %3979 = torch.aten.item %3977 : !torch.vtensor<[],si8> -> !torch.int | |
| %3980 = torch.aten._make_per_tensor_quantized_tensor %3975, %3978, %3979 : !torch.vtensor<[128],si8>, !torch.float, !torch.int -> !torch.vtensor<[128],!torch.qint8> | |
| %3981 = torch.aten.dequantize.self %3980 : !torch.vtensor<[128],!torch.qint8> -> !torch.vtensor<[128],f32> | |
| %int0_1056 = torch.constant.int 0 | |
| %int0_1057 = torch.constant.int 0 | |
| %int1_1058 = torch.constant.int 1 | |
| %int1_1059 = torch.constant.int 1 | |
| %int1_1060 = torch.constant.int 1 | |
| %int1_1061 = torch.constant.int 1 | |
| %int0_1062 = torch.constant.int 0 | |
| %3982 = torch.prim.ListConstruct %int0_1056, %int0_1057 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3983 = torch.prim.ListConstruct %int1_1058, %int1_1059 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3984 = torch.prim.ListConstruct %int1_1060, %int1_1061 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %3985 = torch.prim.ListConstruct %int0_1062, %int0_1062 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1063 = torch.constant.bool false | |
| %int1_1064 = torch.constant.int 1 | |
| %3986 = torch.aten.convolution %3957, %3969, %3981, %3984, %3982, %3983, %false_1063, %3985, %int1_1064 : !torch.vtensor<[1,960,14,14],f32>, !torch.vtensor<[128,960,1,1],f32>, !torch.vtensor<[128],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,14,14],f32> | |
| %3987 = torch.aten.relu %3986 : !torch.vtensor<[1,128,14,14],f32> -> !torch.vtensor<[1,128,14,14],f32> | |
| %3988 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3989 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1065 = torch.constant.int 12 | |
| %3990 = torch.aten.item %3988 : !torch.vtensor<[],f32> -> !torch.float | |
| %3991 = torch.aten.item %3989 : !torch.vtensor<[],si8> -> !torch.int | |
| %3992 = torch.aten.quantize_per_tensor %3987, %3990, %3991, %int12_1065 : !torch.vtensor<[1,128,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,14,14],!torch.qint8> | |
| %3993 = torch.aten.int_repr %3992 : !torch.vtensor<[1,128,14,14],!torch.qint8> -> !torch.vtensor<[1,128,14,14],si8> | |
| %3994 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %3995 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %3996 = torch.aten.item %3994 : !torch.vtensor<[],f32> -> !torch.float | |
| %3997 = torch.aten.item %3995 : !torch.vtensor<[],si8> -> !torch.int | |
| %3998 = torch.aten._make_per_tensor_quantized_tensor %3993, %3996, %3997 : !torch.vtensor<[1,128,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,14,14],!torch.qint8> | |
| %3999 = torch.aten.dequantize.self %3998 : !torch.vtensor<[1,128,14,14],!torch.qint8> -> !torch.vtensor<[1,128,14,14],f32> | |
| %int0_1066 = torch.constant.int 0 | |
| %int1_1067 = torch.constant.int 1 | |
| %int14_1068 = torch.constant.int 14 | |
| %int14_1069 = torch.constant.int 14 | |
| %4000 = torch.prim.ListConstruct %int14_1068, %int14_1069 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4001 = torch.prim.ListConstruct %int0_1066, %int0_1066 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4002 = torch.prim.ListConstruct %int1_1067, %int1_1067 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1070 = torch.constant.bool false | |
| %none_1071 = torch.constant.none | |
| %4003 = torch.aten.avg_pool2d %3957, %4000, %4002, %4001, %false_1070, %false_1070, %none_1071 : !torch.vtensor<[1,960,14,14],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,960,1,1],f32> | |
| %4004 = torch.vtensor.literal(dense<1.00488281> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4005 = torch.aten.mul.Tensor %4003, %4004 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,960,1,1],f32> | |
| %4006 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4007 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1072 = torch.constant.int 12 | |
| %4008 = torch.aten.item %4006 : !torch.vtensor<[],f32> -> !torch.float | |
| %4009 = torch.aten.item %4007 : !torch.vtensor<[],si8> -> !torch.int | |
| %4010 = torch.aten.quantize_per_tensor %4005, %4008, %4009, %int12_1072 : !torch.vtensor<[1,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %4011 = torch.aten.int_repr %4010 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],si8> | |
| %4012 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4013 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4014 = torch.aten.item %4012 : !torch.vtensor<[],f32> -> !torch.float | |
| %4015 = torch.aten.item %4013 : !torch.vtensor<[],si8> -> !torch.int | |
| %4016 = torch.aten._make_per_tensor_quantized_tensor %4011, %4014, %4015 : !torch.vtensor<[1,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,960,1,1],!torch.qint8> | |
| %4017 = torch.aten.dequantize.self %4016 : !torch.vtensor<[1,960,1,1],!torch.qint8> -> !torch.vtensor<[1,960,1,1],f32> | |
| %4018 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4019 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1073 = torch.constant.int 12 | |
| %4020 = torch.aten.item %4018 : !torch.vtensor<[],f32> -> !torch.float | |
| %4021 = torch.aten.item %4019 : !torch.vtensor<[],si8> -> !torch.int | |
| %4022 = torch.aten.quantize_per_tensor %126, %4020, %4021, %int12_1073 : !torch.vtensor<[128,960,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[128,960,1,1],!torch.qint8> | |
| %4023 = torch.aten.int_repr %4022 : !torch.vtensor<[128,960,1,1],!torch.qint8> -> !torch.vtensor<[128,960,1,1],si8> | |
| %4024 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4025 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4026 = torch.aten.item %4024 : !torch.vtensor<[],f32> -> !torch.float | |
| %4027 = torch.aten.item %4025 : !torch.vtensor<[],si8> -> !torch.int | |
| %4028 = torch.aten._make_per_tensor_quantized_tensor %4023, %4026, %4027 : !torch.vtensor<[128,960,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[128,960,1,1],!torch.qint8> | |
| %4029 = torch.aten.dequantize.self %4028 : !torch.vtensor<[128,960,1,1],!torch.qint8> -> !torch.vtensor<[128,960,1,1],f32> | |
| %int0_1074 = torch.constant.int 0 | |
| %int0_1075 = torch.constant.int 0 | |
| %int1_1076 = torch.constant.int 1 | |
| %int1_1077 = torch.constant.int 1 | |
| %int1_1078 = torch.constant.int 1 | |
| %int1_1079 = torch.constant.int 1 | |
| %int0_1080 = torch.constant.int 0 | |
| %4030 = torch.prim.ListConstruct %int0_1074, %int0_1075 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4031 = torch.prim.ListConstruct %int1_1076, %int1_1077 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4032 = torch.prim.ListConstruct %int1_1078, %int1_1079 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4033 = torch.prim.ListConstruct %int0_1080, %int0_1080 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1081 = torch.constant.bool false | |
| %none_1082 = torch.constant.none | |
| %int1_1083 = torch.constant.int 1 | |
| %4034 = torch.aten.convolution %4017, %4029, %none_1082, %4032, %4030, %4031, %false_1081, %4033, %int1_1083 : !torch.vtensor<[1,960,1,1],f32>, !torch.vtensor<[128,960,1,1],f32>, !torch.none, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,128,1,1],f32> | |
| %4035 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4036 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1084 = torch.constant.int 12 | |
| %4037 = torch.aten.item %4035 : !torch.vtensor<[],f32> -> !torch.float | |
| %4038 = torch.aten.item %4036 : !torch.vtensor<[],si8> -> !torch.int | |
| %4039 = torch.aten.quantize_per_tensor %4034, %4037, %4038, %int12_1084 : !torch.vtensor<[1,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,1,1],!torch.qint8> | |
| %4040 = torch.aten.int_repr %4039 : !torch.vtensor<[1,128,1,1],!torch.qint8> -> !torch.vtensor<[1,128,1,1],si8> | |
| %4041 = torch.vtensor.literal(dense<3.125000e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4042 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4043 = torch.aten.item %4041 : !torch.vtensor<[],f32> -> !torch.float | |
| %4044 = torch.aten.item %4042 : !torch.vtensor<[],si8> -> !torch.int | |
| %4045 = torch.aten._make_per_tensor_quantized_tensor %4040, %4043, %4044 : !torch.vtensor<[1,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1,1],!torch.qint8> | |
| %4046 = torch.aten.dequantize.self %4045 : !torch.vtensor<[1,128,1,1],!torch.qint8> -> !torch.vtensor<[1,128,1,1],f32> | |
| %4047 = torch.aten.sigmoid %4046 : !torch.vtensor<[1,128,1,1],f32> -> !torch.vtensor<[1,128,1,1],f32> | |
| %4048 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4049 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1085 = torch.constant.int 12 | |
| %4050 = torch.aten.item %4048 : !torch.vtensor<[],f32> -> !torch.float | |
| %4051 = torch.aten.item %4049 : !torch.vtensor<[],si8> -> !torch.int | |
| %4052 = torch.aten.quantize_per_tensor %4047, %4050, %4051, %int12_1085 : !torch.vtensor<[1,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,1,1],!torch.qint8> | |
| %4053 = torch.aten.int_repr %4052 : !torch.vtensor<[1,128,1,1],!torch.qint8> -> !torch.vtensor<[1,128,1,1],si8> | |
| %4054 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4055 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4056 = torch.aten.item %4054 : !torch.vtensor<[],f32> -> !torch.float | |
| %4057 = torch.aten.item %4055 : !torch.vtensor<[],si8> -> !torch.int | |
| %4058 = torch.aten._make_per_tensor_quantized_tensor %4053, %4056, %4057 : !torch.vtensor<[1,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1,1],!torch.qint8> | |
| %4059 = torch.aten.dequantize.self %4058 : !torch.vtensor<[1,128,1,1],!torch.qint8> -> !torch.vtensor<[1,128,1,1],f32> | |
| %4060 = torch.aten.mul.Tensor %3999, %4059 : !torch.vtensor<[1,128,14,14],f32>, !torch.vtensor<[1,128,1,1],f32> -> !torch.vtensor<[1,128,14,14],f32> | |
| %4061 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4062 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1086 = torch.constant.int 12 | |
| %4063 = torch.aten.item %4061 : !torch.vtensor<[],f32> -> !torch.float | |
| %4064 = torch.aten.item %4062 : !torch.vtensor<[],si8> -> !torch.int | |
| %4065 = torch.aten.quantize_per_tensor %4060, %4063, %4064, %int12_1086 : !torch.vtensor<[1,128,14,14],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,128,14,14],!torch.qint8> | |
| %4066 = torch.aten.int_repr %4065 : !torch.vtensor<[1,128,14,14],!torch.qint8> -> !torch.vtensor<[1,128,14,14],si8> | |
| %4067 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4068 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4069 = torch.aten.item %4067 : !torch.vtensor<[],f32> -> !torch.float | |
| %4070 = torch.aten.item %4068 : !torch.vtensor<[],si8> -> !torch.int | |
| %4071 = torch.aten._make_per_tensor_quantized_tensor %4066, %4069, %4070 : !torch.vtensor<[1,128,14,14],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,128,14,14],!torch.qint8> | |
| %4072 = torch.aten.dequantize.self %4071 : !torch.vtensor<[1,128,14,14],!torch.qint8> -> !torch.vtensor<[1,128,14,14],f32> | |
| %4073 = torch.vtensor.literal(dense<28> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4074 = torch.vtensor.literal(dense<28> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4075 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1087 = torch.constant.int 0 | |
| %int0_1088 = torch.constant.int 0 | |
| %int0_1089 = torch.constant.int 0 | |
| %4076 = torch.aten.select.int %4075, %int0_1087, %int0_1089 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4077 = torch.aten.item %4076 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4078 = torch.aten.lt.int %4077, %int0_1087 : !torch.int, !torch.int -> !torch.bool | |
| %4079 = torch.aten.Int.bool %4078 : !torch.bool -> !torch.int | |
| %4080 = torch.aten.mul.int %4079, %int0_1088 : !torch.int, !torch.int -> !torch.int | |
| %4081 = torch.aten.add.int %4077, %4080 : !torch.int, !torch.int -> !torch.int | |
| %4082 = torch.prim.ListConstruct %4081 : (!torch.int) -> !torch.list<int> | |
| %false_1090 = torch.constant.bool false | |
| %none_1091 = torch.constant.none | |
| %4083 = torch.aten.tensor %4082, %none_1091, %none_1091, %false_1090 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values, %indices = torch.aten.sort %4083, %int0_1087, %false_1090 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1092 = torch.constant.int 0 | |
| %4084 = torch.aten.select.int %values, %int0_1087, %int0_1092 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4085 = torch.aten.item %4084 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4086 = torch.aten.unsqueeze %4073, %4085 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4087 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1093 = torch.constant.int 0 | |
| %int0_1094 = torch.constant.int 0 | |
| %int0_1095 = torch.constant.int 0 | |
| %4088 = torch.aten.select.int %4087, %int0_1093, %int0_1095 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4089 = torch.aten.item %4088 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4090 = torch.aten.lt.int %4089, %int0_1093 : !torch.int, !torch.int -> !torch.bool | |
| %4091 = torch.aten.Int.bool %4090 : !torch.bool -> !torch.int | |
| %4092 = torch.aten.mul.int %4091, %int0_1094 : !torch.int, !torch.int -> !torch.int | |
| %4093 = torch.aten.add.int %4089, %4092 : !torch.int, !torch.int -> !torch.int | |
| %4094 = torch.prim.ListConstruct %4093 : (!torch.int) -> !torch.list<int> | |
| %false_1096 = torch.constant.bool false | |
| %none_1097 = torch.constant.none | |
| %4095 = torch.aten.tensor %4094, %none_1097, %none_1097, %false_1096 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1098, %indices_1099 = torch.aten.sort %4095, %int0_1093, %false_1096 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1100 = torch.constant.int 0 | |
| %4096 = torch.aten.select.int %values_1098, %int0_1093, %int0_1100 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4097 = torch.aten.item %4096 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4098 = torch.aten.unsqueeze %4074, %4097 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4099 = torch.prim.ListConstruct %4086, %4098 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_1101 = torch.constant.int 0 | |
| %4100 = torch.aten.cat %4099, %int0_1101 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[2],si64> | |
| %4101 = torch.aten._shape_as_tensor %4072 : !torch.vtensor<[1,128,14,14],f32> -> !torch.vtensor<[4],si64> | |
| %4102 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4103 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4104 = torch.vtensor.literal(dense<2> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %none_1102 = torch.constant.none | |
| %int1_1103 = torch.constant.int 1 | |
| %4105 = torch.prim.ListConstruct %int1_1103 : (!torch.int) -> !torch.list<int> | |
| %4106 = torch.aten.ones %4105, %none_1102, %none_1102, %none_1102, %none_1102 : !torch.list<int>, !torch.none, !torch.none, !torch.none, !torch.none -> !torch.vtensor<[1],si64> | |
| %int0_1104 = torch.constant.int 0 | |
| %int0_1105 = torch.constant.int 0 | |
| %4107 = torch.prim.NumToTensor.Scalar %int0_1105 : !torch.int -> !torch.vtensor<[1],si64> | |
| %4108 = torch.aten.index_select %4103, %int0_1104, %4107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4109 = torch.aten.item %4108 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4110 = torch.aten.index_select %4104, %int0_1104, %4107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4111 = torch.aten.item %4110 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4112 = torch.aten.index_select %4102, %int0_1104, %4107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4113 = torch.aten.item %4112 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4114 = torch.aten.index_select %4106, %int0_1104, %4107 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4115 = torch.aten.item %4114 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4116 = torch.aten.slice.Tensor %4101, %4113, %4109, %4111, %4115 : !torch.vtensor<[4],si64>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[2],si64> | |
| %int4_1106 = torch.constant.int 4 | |
| %none_1107 = torch.constant.none | |
| %false_1108 = torch.constant.bool false | |
| %4117 = torch.aten.to.dtype %4100, %int4_1106, %false_1108, %false_1108, %none_1107 : !torch.vtensor<[2],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[2],si64> | |
| %4118 = torch.prim.ListConstruct %4116, %4117 : (!torch.vtensor<[2],si64>, !torch.vtensor<[2],si64>) -> !torch.list<vtensor> | |
| %int0_1109 = torch.constant.int 0 | |
| %4119 = torch.aten.cat %4118, %int0_1109 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],si64> | |
| %4120 = torch.operator "onnx.Resize"(%4072, %none, %none, %4119) {torch.onnx.coordinate_transformation_mode = "half_pixel", torch.onnx.cubic_coeff_a = -7.500000e-01 : f32, torch.onnx.mode = "linear", torch.onnx.nearest_mode = "floor"} : (!torch.vtensor<[1,128,14,14],f32>, !torch.none, !torch.none, !torch.vtensor<[4],si64>) -> !torch.vtensor<[?,?,?,?],f32> | |
| %4121 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4122 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1110 = torch.constant.int 12 | |
| %4123 = torch.aten.item %4121 : !torch.vtensor<[],f32> -> !torch.float | |
| %4124 = torch.aten.item %4122 : !torch.vtensor<[],si8> -> !torch.int | |
| %4125 = torch.aten.quantize_per_tensor %4120, %4123, %4124, %int12_1110 : !torch.vtensor<[?,?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4126 = torch.aten.int_repr %4125 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],si8> | |
| %4127 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4128 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4129 = torch.aten.item %4127 : !torch.vtensor<[],f32> -> !torch.float | |
| %4130 = torch.aten.item %4128 : !torch.vtensor<[],si8> -> !torch.int | |
| %4131 = torch.aten._make_per_tensor_quantized_tensor %4126, %4129, %4130 : !torch.vtensor<[?,?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4132 = torch.aten.dequantize.self %4131 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],f32> | |
| %4133 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4134 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1111 = torch.constant.int 12 | |
| %4135 = torch.aten.item %4133 : !torch.vtensor<[],f32> -> !torch.float | |
| %4136 = torch.aten.item %4134 : !torch.vtensor<[],si8> -> !torch.int | |
| %4137 = torch.aten.quantize_per_tensor %127, %4135, %4136, %int12_1111 : !torch.vtensor<[21,40,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[21,40,1,1],!torch.qint8> | |
| %4138 = torch.aten.int_repr %4137 : !torch.vtensor<[21,40,1,1],!torch.qint8> -> !torch.vtensor<[21,40,1,1],si8> | |
| %4139 = torch.vtensor.literal(dense<0.001953125> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4140 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4141 = torch.aten.item %4139 : !torch.vtensor<[],f32> -> !torch.float | |
| %4142 = torch.aten.item %4140 : !torch.vtensor<[],si8> -> !torch.int | |
| %4143 = torch.aten._make_per_tensor_quantized_tensor %4138, %4141, %4142 : !torch.vtensor<[21,40,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[21,40,1,1],!torch.qint8> | |
| %4144 = torch.aten.dequantize.self %4143 : !torch.vtensor<[21,40,1,1],!torch.qint8> -> !torch.vtensor<[21,40,1,1],f32> | |
| %4145 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4146 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1112 = torch.constant.int 12 | |
| %4147 = torch.aten.item %4145 : !torch.vtensor<[],f32> -> !torch.float | |
| %4148 = torch.aten.item %4146 : !torch.vtensor<[],si8> -> !torch.int | |
| %4149 = torch.aten.quantize_per_tensor %128, %4147, %4148, %int12_1112 : !torch.vtensor<[21],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[21],!torch.qint8> | |
| %4150 = torch.aten.int_repr %4149 : !torch.vtensor<[21],!torch.qint8> -> !torch.vtensor<[21],si8> | |
| %4151 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4152 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4153 = torch.aten.item %4151 : !torch.vtensor<[],f32> -> !torch.float | |
| %4154 = torch.aten.item %4152 : !torch.vtensor<[],si8> -> !torch.int | |
| %4155 = torch.aten._make_per_tensor_quantized_tensor %4150, %4153, %4154 : !torch.vtensor<[21],si8>, !torch.float, !torch.int -> !torch.vtensor<[21],!torch.qint8> | |
| %4156 = torch.aten.dequantize.self %4155 : !torch.vtensor<[21],!torch.qint8> -> !torch.vtensor<[21],f32> | |
| %int0_1113 = torch.constant.int 0 | |
| %int0_1114 = torch.constant.int 0 | |
| %int1_1115 = torch.constant.int 1 | |
| %int1_1116 = torch.constant.int 1 | |
| %int1_1117 = torch.constant.int 1 | |
| %int1_1118 = torch.constant.int 1 | |
| %int0_1119 = torch.constant.int 0 | |
| %4157 = torch.prim.ListConstruct %int0_1113, %int0_1114 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4158 = torch.prim.ListConstruct %int1_1115, %int1_1116 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4159 = torch.prim.ListConstruct %int1_1117, %int1_1118 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4160 = torch.prim.ListConstruct %int0_1119, %int0_1119 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1120 = torch.constant.bool false | |
| %int1_1121 = torch.constant.int 1 | |
| %4161 = torch.aten.convolution %838, %4144, %4156, %4159, %4157, %4158, %false_1120, %4160, %int1_1121 : !torch.vtensor<[1,40,28,28],f32>, !torch.vtensor<[21,40,1,1],f32>, !torch.vtensor<[21],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[1,21,28,28],f32> | |
| %4162 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4163 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1122 = torch.constant.int 12 | |
| %4164 = torch.aten.item %4162 : !torch.vtensor<[],f32> -> !torch.float | |
| %4165 = torch.aten.item %4163 : !torch.vtensor<[],si8> -> !torch.int | |
| %4166 = torch.aten.quantize_per_tensor %4161, %4164, %4165, %int12_1122 : !torch.vtensor<[1,21,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[1,21,28,28],!torch.qint8> | |
| %4167 = torch.aten.int_repr %4166 : !torch.vtensor<[1,21,28,28],!torch.qint8> -> !torch.vtensor<[1,21,28,28],si8> | |
| %4168 = torch.vtensor.literal(dense<1.562500e-02> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4169 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4170 = torch.aten.item %4168 : !torch.vtensor<[],f32> -> !torch.float | |
| %4171 = torch.aten.item %4169 : !torch.vtensor<[],si8> -> !torch.int | |
| %4172 = torch.aten._make_per_tensor_quantized_tensor %4167, %4170, %4171 : !torch.vtensor<[1,21,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,21,28,28],!torch.qint8> | |
| %4173 = torch.aten.dequantize.self %4172 : !torch.vtensor<[1,21,28,28],!torch.qint8> -> !torch.vtensor<[1,21,28,28],f32> | |
| %4174 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4175 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1123 = torch.constant.int 12 | |
| %4176 = torch.aten.item %4174 : !torch.vtensor<[],f32> -> !torch.float | |
| %4177 = torch.aten.item %4175 : !torch.vtensor<[],si8> -> !torch.int | |
| %4178 = torch.aten.quantize_per_tensor %129, %4176, %4177, %int12_1123 : !torch.vtensor<[21,128,1,1],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[21,128,1,1],!torch.qint8> | |
| %4179 = torch.aten.int_repr %4178 : !torch.vtensor<[21,128,1,1],!torch.qint8> -> !torch.vtensor<[21,128,1,1],si8> | |
| %4180 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4181 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4182 = torch.aten.item %4180 : !torch.vtensor<[],f32> -> !torch.float | |
| %4183 = torch.aten.item %4181 : !torch.vtensor<[],si8> -> !torch.int | |
| %4184 = torch.aten._make_per_tensor_quantized_tensor %4179, %4182, %4183 : !torch.vtensor<[21,128,1,1],si8>, !torch.float, !torch.int -> !torch.vtensor<[21,128,1,1],!torch.qint8> | |
| %4185 = torch.aten.dequantize.self %4184 : !torch.vtensor<[21,128,1,1],!torch.qint8> -> !torch.vtensor<[21,128,1,1],f32> | |
| %4186 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4187 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1124 = torch.constant.int 12 | |
| %4188 = torch.aten.item %4186 : !torch.vtensor<[],f32> -> !torch.float | |
| %4189 = torch.aten.item %4187 : !torch.vtensor<[],si8> -> !torch.int | |
| %4190 = torch.aten.quantize_per_tensor %130, %4188, %4189, %int12_1124 : !torch.vtensor<[21],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[21],!torch.qint8> | |
| %4191 = torch.aten.int_repr %4190 : !torch.vtensor<[21],!torch.qint8> -> !torch.vtensor<[21],si8> | |
| %4192 = torch.vtensor.literal(dense<7.812500e-03> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4193 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4194 = torch.aten.item %4192 : !torch.vtensor<[],f32> -> !torch.float | |
| %4195 = torch.aten.item %4193 : !torch.vtensor<[],si8> -> !torch.int | |
| %4196 = torch.aten._make_per_tensor_quantized_tensor %4191, %4194, %4195 : !torch.vtensor<[21],si8>, !torch.float, !torch.int -> !torch.vtensor<[21],!torch.qint8> | |
| %4197 = torch.aten.dequantize.self %4196 : !torch.vtensor<[21],!torch.qint8> -> !torch.vtensor<[21],f32> | |
| %int0_1125 = torch.constant.int 0 | |
| %int0_1126 = torch.constant.int 0 | |
| %int1_1127 = torch.constant.int 1 | |
| %int1_1128 = torch.constant.int 1 | |
| %int1_1129 = torch.constant.int 1 | |
| %int1_1130 = torch.constant.int 1 | |
| %int0_1131 = torch.constant.int 0 | |
| %4198 = torch.prim.ListConstruct %int0_1125, %int0_1126 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4199 = torch.prim.ListConstruct %int1_1127, %int1_1128 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4200 = torch.prim.ListConstruct %int1_1129, %int1_1130 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %4201 = torch.prim.ListConstruct %int0_1131, %int0_1131 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %false_1132 = torch.constant.bool false | |
| %int1_1133 = torch.constant.int 1 | |
| %4202 = torch.aten.convolution %4132, %4185, %4197, %4200, %4198, %4199, %false_1132, %4201, %int1_1133 : !torch.vtensor<[?,?,?,?],f32>, !torch.vtensor<[21,128,1,1],f32>, !torch.vtensor<[21],f32>, !torch.list<int>, !torch.list<int>, !torch.list<int>, !torch.bool, !torch.list<int>, !torch.int -> !torch.vtensor<[?,21,?,?],f32> | |
| %4203 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4204 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1134 = torch.constant.int 12 | |
| %4205 = torch.aten.item %4203 : !torch.vtensor<[],f32> -> !torch.float | |
| %4206 = torch.aten.item %4204 : !torch.vtensor<[],si8> -> !torch.int | |
| %4207 = torch.aten.quantize_per_tensor %4202, %4205, %4206, %int12_1134 : !torch.vtensor<[?,21,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,21,?,?],!torch.qint8> | |
| %4208 = torch.aten.int_repr %4207 : !torch.vtensor<[?,21,?,?],!torch.qint8> -> !torch.vtensor<[?,21,?,?],si8> | |
| %4209 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4210 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4211 = torch.aten.item %4209 : !torch.vtensor<[],f32> -> !torch.float | |
| %4212 = torch.aten.item %4210 : !torch.vtensor<[],si8> -> !torch.int | |
| %4213 = torch.aten._make_per_tensor_quantized_tensor %4208, %4211, %4212 : !torch.vtensor<[?,21,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,21,?,?],!torch.qint8> | |
| %4214 = torch.aten.dequantize.self %4213 : !torch.vtensor<[?,21,?,?],!torch.qint8> -> !torch.vtensor<[?,21,?,?],f32> | |
| %int1_1135 = torch.constant.int 1 | |
| %4215 = torch.aten.add.Tensor %4173, %4214, %int1_1135 : !torch.vtensor<[1,21,28,28],f32>, !torch.vtensor<[?,21,?,?],f32>, !torch.int -> !torch.vtensor<[?,21,28,28],f32> | |
| %4216 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4217 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1136 = torch.constant.int 12 | |
| %4218 = torch.aten.item %4216 : !torch.vtensor<[],f32> -> !torch.float | |
| %4219 = torch.aten.item %4217 : !torch.vtensor<[],si8> -> !torch.int | |
| %4220 = torch.aten.quantize_per_tensor %4215, %4218, %4219, %int12_1136 : !torch.vtensor<[?,21,28,28],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,21,28,28],!torch.qint8> | |
| %4221 = torch.aten.int_repr %4220 : !torch.vtensor<[?,21,28,28],!torch.qint8> -> !torch.vtensor<[?,21,28,28],si8> | |
| %4222 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4223 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4224 = torch.aten.item %4222 : !torch.vtensor<[],f32> -> !torch.float | |
| %4225 = torch.aten.item %4223 : !torch.vtensor<[],si8> -> !torch.int | |
| %4226 = torch.aten._make_per_tensor_quantized_tensor %4221, %4224, %4225 : !torch.vtensor<[?,21,28,28],si8>, !torch.float, !torch.int -> !torch.vtensor<[?,21,28,28],!torch.qint8> | |
| %4227 = torch.aten.dequantize.self %4226 : !torch.vtensor<[?,21,28,28],!torch.qint8> -> !torch.vtensor<[?,21,28,28],f32> | |
| %4228 = torch.vtensor.literal(dense<224> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4229 = torch.vtensor.literal(dense<224> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %4230 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1137 = torch.constant.int 0 | |
| %int0_1138 = torch.constant.int 0 | |
| %int0_1139 = torch.constant.int 0 | |
| %4231 = torch.aten.select.int %4230, %int0_1137, %int0_1139 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4232 = torch.aten.item %4231 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4233 = torch.aten.lt.int %4232, %int0_1137 : !torch.int, !torch.int -> !torch.bool | |
| %4234 = torch.aten.Int.bool %4233 : !torch.bool -> !torch.int | |
| %4235 = torch.aten.mul.int %4234, %int0_1138 : !torch.int, !torch.int -> !torch.int | |
| %4236 = torch.aten.add.int %4232, %4235 : !torch.int, !torch.int -> !torch.int | |
| %4237 = torch.prim.ListConstruct %4236 : (!torch.int) -> !torch.list<int> | |
| %false_1140 = torch.constant.bool false | |
| %none_1141 = torch.constant.none | |
| %4238 = torch.aten.tensor %4237, %none_1141, %none_1141, %false_1140 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1142, %indices_1143 = torch.aten.sort %4238, %int0_1137, %false_1140 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1144 = torch.constant.int 0 | |
| %4239 = torch.aten.select.int %values_1142, %int0_1137, %int0_1144 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4240 = torch.aten.item %4239 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4241 = torch.aten.unsqueeze %4228, %4240 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4242 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %int0_1145 = torch.constant.int 0 | |
| %int0_1146 = torch.constant.int 0 | |
| %int0_1147 = torch.constant.int 0 | |
| %4243 = torch.aten.select.int %4242, %int0_1145, %int0_1147 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4244 = torch.aten.item %4243 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4245 = torch.aten.lt.int %4244, %int0_1145 : !torch.int, !torch.int -> !torch.bool | |
| %4246 = torch.aten.Int.bool %4245 : !torch.bool -> !torch.int | |
| %4247 = torch.aten.mul.int %4246, %int0_1146 : !torch.int, !torch.int -> !torch.int | |
| %4248 = torch.aten.add.int %4244, %4247 : !torch.int, !torch.int -> !torch.int | |
| %4249 = torch.prim.ListConstruct %4248 : (!torch.int) -> !torch.list<int> | |
| %false_1148 = torch.constant.bool false | |
| %none_1149 = torch.constant.none | |
| %4250 = torch.aten.tensor %4249, %none_1149, %none_1149, %false_1148 : !torch.list<int>, !torch.none, !torch.none, !torch.bool -> !torch.vtensor<[1],si64> | |
| %values_1150, %indices_1151 = torch.aten.sort %4250, %int0_1145, %false_1148 : !torch.vtensor<[1],si64>, !torch.int, !torch.bool -> !torch.vtensor<[1],si64>, !torch.vtensor<[1],si64> | |
| %int0_1152 = torch.constant.int 0 | |
| %4251 = torch.aten.select.int %values_1150, %int0_1145, %int0_1152 : !torch.vtensor<[1],si64>, !torch.int, !torch.int -> !torch.vtensor<[1],si64> | |
| %4252 = torch.aten.item %4251 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4253 = torch.aten.unsqueeze %4229, %4252 : !torch.vtensor<[],si64>, !torch.int -> !torch.vtensor<[1],si64> | |
| %4254 = torch.prim.ListConstruct %4241, %4253 : (!torch.vtensor<[1],si64>, !torch.vtensor<[1],si64>) -> !torch.list<vtensor> | |
| %int0_1153 = torch.constant.int 0 | |
| %4255 = torch.aten.cat %4254, %int0_1153 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[2],si64> | |
| %4256 = torch.aten._shape_as_tensor %4227 : !torch.vtensor<[?,21,28,28],f32> -> !torch.vtensor<[4],si64> | |
| %4257 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4258 = torch.vtensor.literal(dense<0> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %4259 = torch.vtensor.literal(dense<2> : tensor<1xsi64>) : !torch.vtensor<[1],si64> | |
| %none_1154 = torch.constant.none | |
| %int1_1155 = torch.constant.int 1 | |
| %4260 = torch.prim.ListConstruct %int1_1155 : (!torch.int) -> !torch.list<int> | |
| %4261 = torch.aten.ones %4260, %none_1154, %none_1154, %none_1154, %none_1154 : !torch.list<int>, !torch.none, !torch.none, !torch.none, !torch.none -> !torch.vtensor<[1],si64> | |
| %int0_1156 = torch.constant.int 0 | |
| %int0_1157 = torch.constant.int 0 | |
| %4262 = torch.prim.NumToTensor.Scalar %int0_1157 : !torch.int -> !torch.vtensor<[1],si64> | |
| %4263 = torch.aten.index_select %4258, %int0_1156, %4262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4264 = torch.aten.item %4263 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4265 = torch.aten.index_select %4259, %int0_1156, %4262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4266 = torch.aten.item %4265 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4267 = torch.aten.index_select %4257, %int0_1156, %4262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4268 = torch.aten.item %4267 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4269 = torch.aten.index_select %4261, %int0_1156, %4262 : !torch.vtensor<[1],si64>, !torch.int, !torch.vtensor<[1],si64> -> !torch.vtensor<[1],si64> | |
| %4270 = torch.aten.item %4269 : !torch.vtensor<[1],si64> -> !torch.int | |
| %4271 = torch.aten.slice.Tensor %4256, %4268, %4264, %4266, %4270 : !torch.vtensor<[4],si64>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[2],si64> | |
| %int4_1158 = torch.constant.int 4 | |
| %none_1159 = torch.constant.none | |
| %false_1160 = torch.constant.bool false | |
| %4272 = torch.aten.to.dtype %4255, %int4_1158, %false_1160, %false_1160, %none_1159 : !torch.vtensor<[2],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[2],si64> | |
| %4273 = torch.prim.ListConstruct %4271, %4272 : (!torch.vtensor<[2],si64>, !torch.vtensor<[2],si64>) -> !torch.list<vtensor> | |
| %int0_1161 = torch.constant.int 0 | |
| %4274 = torch.aten.cat %4273, %int0_1161 : !torch.list<vtensor>, !torch.int -> !torch.vtensor<[4],si64> | |
| %4275 = torch.operator "onnx.Resize"(%4227, %none, %none, %4274) {torch.onnx.coordinate_transformation_mode = "half_pixel", torch.onnx.cubic_coeff_a = -7.500000e-01 : f32, torch.onnx.mode = "linear", torch.onnx.nearest_mode = "floor"} : (!torch.vtensor<[?,21,28,28],f32>, !torch.none, !torch.none, !torch.vtensor<[4],si64>) -> !torch.vtensor<[?,?,?,?],f32> | |
| %4276 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4277 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %int12_1162 = torch.constant.int 12 | |
| %4278 = torch.aten.item %4276 : !torch.vtensor<[],f32> -> !torch.float | |
| %4279 = torch.aten.item %4277 : !torch.vtensor<[],si8> -> !torch.int | |
| %4280 = torch.aten.quantize_per_tensor %4275, %4278, %4279, %int12_1162 : !torch.vtensor<[?,?,?,?],f32>, !torch.float, !torch.int, !torch.int -> !torch.vtensor<[?,?,?,?],!torch.qint8> | |
| %4281 = torch.aten.int_repr %4280 : !torch.vtensor<[?,?,?,?],!torch.qint8> -> !torch.vtensor<[?,?,?,?],si8> | |
| %4282 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f32>) : !torch.vtensor<[],f32> | |
| %4283 = torch.vtensor.literal(dense<0> : tensor<si8>) : !torch.vtensor<[],si8> | |
| %4284 = torch.aten.item %4282 : !torch.vtensor<[],f32> -> !torch.float | |
| %4285 = torch.aten.item %4283 : !torch.vtensor<[],si8> -> !torch.int | |
| %4286 = torch.aten._make_per_tensor_quantized_tensor %4281, %4284, %4285 : !torch.vtensor<[?,?,?,?],si8>, !torch.float, !torch.int -> !torch.vtensor<[1,21,224,224],!torch.qint8> | |
| %4287 = torch.aten.dequantize.self %4286 : !torch.vtensor<[1,21,224,224],!torch.qint8> -> !torch.vtensor<[1,21,224,224],f32> | |
| return %4287 : !torch.vtensor<[1,21,224,224],f32> | |
| } | |
| } | |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment