Created
February 15, 2023 00:42
-
-
Save AmosLewis/7da724d4fa77eae18d6de02ff82105b0 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| ➜ ~ torch-mlir-opt -pass-pipeline='builtin.module(torchscript-module-to-torch-backend-pipeline{backend-legal-ops=torch.aten.flatten.using_ints,torch.aten.native_layer_norm,torch.aten.linear})' /tmp/_lambda.mlir --mlir-print-ir-after-failure -mlir-disable-threading | |
| <eval_with_key>.2:5:16: error: unsupported by backend contract: tensor with unknown rank | |
| <eval_with_key>.2:5:16: note: see current operation: %36 = "torch.tensor_static_info_cast"(%35) : (!torch.vtensor<[1,4],si64>) -> !torch.vtensor<*,si64> | |
| <eval_with_key>.2:5:16: note: this is likely due to a missing transfer function in abstract_interp_lib_gen.py | |
| // -----// IR Dump After LowerToBackendContract Failed (torch-lower-to-backend-contract) //----- // | |
| module attributes {torch.debug_module_name = "_lambda"} { | |
| func.func @forward(%arg0: !torch.vtensor<[1,15],si64>, %arg1: !torch.vtensor<[1,4],si64>) -> !torch.vtensor<[1,4,32128],f32> { | |
| %int512 = torch.constant.int 512 | |
| %int1 = torch.constant.int 1 | |
| %int3 = torch.constant.int 3 | |
| %int0 = torch.constant.int 0 | |
| %int-1 = torch.constant.int -1 | |
| %true = torch.constant.bool true | |
| %int4 = torch.constant.int 4 | |
| %false = torch.constant.bool false | |
| %none = torch.constant.none | |
| %int15 = torch.constant.int 15 | |
| %0 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.vtensor<[512,2048],f32> | |
| %1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.vtensor<[2048,512],f32> | |
| %2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32x8xf32>) : !torch.vtensor<[32,8],f32> | |
| %3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.vtensor<[512,512],f32> | |
| %4 = torch.vtensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.vtensor<[512],f32> | |
| %5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<32128x512xf32>) : !torch.vtensor<[32128,512],f32> | |
| %6 = torch.vtensor.literal(dense<0> : tensor<si64>) : !torch.vtensor<[],si64> | |
| %int-100 = torch.constant.int -100 | |
| %float-3.402820e38 = torch.constant.float -3.4028234663852886E+38 | |
| %int6 = torch.constant.int 6 | |
| %int9223372036854775807 = torch.constant.int 9223372036854775807 | |
| %int2 = torch.constant.int 2 | |
| %float1.000000e00 = torch.constant.float 1.000000e+00 | |
| %float9.999990e-07 = torch.constant.float 9.9999999999999995E-7 | |
| %int8 = torch.constant.int 8 | |
| %int64 = torch.constant.int 64 | |
| %int16 = torch.constant.int 16 | |
| %float2.772590e00 = torch.constant.float 2.7725887222397811 | |
| %int2048 = torch.constant.int 2048 | |
| %float2.079440e00 = torch.constant.float 2.0794415416798357 | |
| %int31 = torch.constant.int 31 | |
| %float4.419420e-02 = torch.constant.float 0.044194173824159223 | |
| %int32128 = torch.constant.int 32128 | |
| %cpu = torch.constant.device "cpu" | |
| %7 = torch.prim.ListConstruct %int1, %int4 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %8 = torch.aten.zeros %7, %int4, %int0, %cpu, %false : !torch.list<int>, !torch.int, !torch.int, !torch.Device, !torch.bool -> !torch.vtensor<[1,4],si64> | |
| %9 = torch.tensor_static_info_cast %8 : !torch.vtensor<[1,4],si64> to !torch.vtensor<*,si64> | |
| %10 = torch.copy.to_tensor %9 : !torch.tensor<*,si64> | |
| %11 = torch.aten.slice.Tensor %arg1, %int1, %int0, %int-1, %int1 : !torch.vtensor<[1,4],si64>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,3],si64> | |
| %12 = torch.aten.clone %11, %none : !torch.vtensor<[1,3],si64>, !torch.none -> !torch.vtensor<[1,3],si64> | |
| %13 = torch.aten.slice.Tensor %10, %int1, %int1, %int9223372036854775807, %int1 : !torch.tensor<*,si64>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor<[1,3],si64> | |
| %14 = torch.tensor_static_info_cast %13 : !torch.tensor<[1,3],si64> to !torch.tensor<*,si64> | |
| %15 = torch.prim.ListConstruct %int1, %int3 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %16 = torch.aten.broadcast_to %12, %15 : !torch.vtensor<[1,3],si64>, !torch.list<int> -> !torch.vtensor<[1,3],si64> | |
| %17 = torch.tensor_static_info_cast %16 : !torch.vtensor<[1,3],si64> to !torch.vtensor<*,si64> | |
| torch.overwrite.tensor.contents %17 overwrites %14 : !torch.vtensor<*,si64>, !torch.tensor<*,si64> | |
| %18 = torch.aten.clone %6, %none : !torch.vtensor<[],si64>, !torch.none -> !torch.vtensor<[],si64> | |
| %19 = torch.aten.slice.Tensor %10, %int1, %int0, %int1, %int1 : !torch.tensor<*,si64>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor<[1,1],si64> | |
| %20 = torch.aten.squeeze.dim %19, %int1 : !torch.tensor<[1,1],si64>, !torch.int -> !torch.tensor<[1],si64> | |
| %21 = torch.tensor_static_info_cast %20 : !torch.tensor<[1],si64> to !torch.tensor<*,si64> | |
| %22 = torch.copy.to_vtensor %21 : !torch.vtensor<*,si64> | |
| %23 = torch.aten.fill.Tensor %22, %18 : !torch.vtensor<*,si64>, !torch.vtensor<[],si64> -> !torch.vtensor<[1],si64> | |
| %24 = torch.tensor_static_info_cast %23 : !torch.vtensor<[1],si64> to !torch.vtensor<*,si64> | |
| torch.overwrite.tensor.contents %24 overwrites %21 : !torch.vtensor<*,si64>, !torch.tensor<*,si64> | |
| %25 = torch.copy.to_vtensor %10 : !torch.vtensor<*,si64> | |
| %26 = torch.aten.eq.Scalar %25, %int-100 : !torch.vtensor<*,si64>, !torch.int -> !torch.vtensor<[1,4],i1> | |
| %27 = torch.copy.to_vtensor %10 : !torch.vtensor<*,si64> | |
| %28 = torch.prim.ListConstruct : () -> !torch.list<int> | |
| %29 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> | |
| %30 = torch.aten.broadcast_to %29, %28 : !torch.vtensor<[],si64>, !torch.list<int> -> !torch.vtensor<[],si64> | |
| %31 = torch.aten.where.self %26, %30, %27 : !torch.vtensor<[1,4],i1>, !torch.vtensor<[],si64>, !torch.vtensor<*,si64> -> !torch.vtensor<[1,4],si64> | |
| %32 = torch.tensor_static_info_cast %31 : !torch.vtensor<[1,4],si64> to !torch.vtensor<*,si64> | |
| torch.overwrite.tensor.contents %32 overwrites %10 : !torch.vtensor<*,si64>, !torch.tensor<*,si64> | |
| %33 = torch.prim.ListConstruct %int-1, %int15 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %34 = torch.aten.view %arg0, %33 : !torch.vtensor<[1,15],si64>, !torch.list<int> -> !torch.vtensor<[1,15],si64> | |
| %35 = torch.aten.embedding %5, %34, %int-1, %false, %false : !torch.vtensor<[32128,512],f32>, !torch.vtensor<[1,15],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[1,15,512],f32> | |
| %36 = torch.prim.ListConstruct %int1, %int15 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %37 = torch.aten.ones %36, %none, %none, %cpu, %false : !torch.list<int>, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[1,15],f32> | |
| %38 = torch.aten.unsqueeze %37, %int1 : !torch.vtensor<[1,15],f32>, !torch.int -> !torch.vtensor<[1,1,15],f32> | |
| %39 = torch.aten.unsqueeze %38, %int2 : !torch.vtensor<[1,1,15],f32>, !torch.int -> !torch.vtensor<[1,1,1,15],f32> | |
| %40 = torch.aten.rsub.Scalar %39, %float1.000000e00, %int1 : !torch.vtensor<[1,1,1,15],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,1,1,15],f32> | |
| %41 = torch.aten.mul.Scalar %40, %float-3.402820e38 : !torch.vtensor<[1,1,1,15],f32>, !torch.float -> !torch.vtensor<[1,1,1,15],f32> | |
| %42 = torch.aten.pow.Tensor_Scalar %35, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %43 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> | |
| %44 = torch.aten.sum.dim_IntList %42, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %45 = torch.aten.div.Scalar %44, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %46 = torch.aten.add.Scalar %45, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %47 = torch.aten.rsqrt %46 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %48 = torch.aten.mul.Tensor %35, %47 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %49 = torch.aten.mul.Tensor %4, %48 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %50 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %51 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %52 = torch.aten.view %49, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %53 = torch.aten.mm %52, %50 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %54 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %55 = torch.aten.view %53, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %56 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %57 = torch.aten.view %55, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %58 = torch.aten.transpose.int %57, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %59 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %60 = torch.aten.view %49, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %61 = torch.aten.mm %60, %59 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %62 = torch.aten.view %61, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %63 = torch.aten.view %62, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %64 = torch.aten.transpose.int %63, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %65 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %66 = torch.aten.view %49, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %67 = torch.aten.mm %66, %65 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %68 = torch.aten.view %67, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %69 = torch.aten.view %68, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %70 = torch.aten.transpose.int %69, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %71 = torch.aten.transpose.int %64, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %72 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %73 = torch.aten.broadcast_to %58, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %74 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %75 = torch.aten.view %73, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %76 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %77 = torch.aten.broadcast_to %71, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %78 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %79 = torch.aten.view %77, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %80 = torch.aten.bmm %75, %79 : !torch.vtensor<[8,15,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,15,15],f32> | |
| %81 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %82 = torch.aten.view %80, %81 : !torch.vtensor<[8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %83 = torch.aten.arange.start_step %int0, %int15, %int1, %int4, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[15],si64> | |
| %84 = torch.aten.unsqueeze %83, %int1 : !torch.vtensor<[15],si64>, !torch.int -> !torch.vtensor<[15,1],si64> | |
| %85 = torch.aten.arange.start_step %int0, %int15, %int1, %int4, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[15],si64> | |
| %86 = torch.aten.unsqueeze %85, %int0 : !torch.vtensor<[15],si64>, !torch.int -> !torch.vtensor<[1,15],si64> | |
| %87 = torch.aten.sub.Tensor %86, %84, %int1 : !torch.vtensor<[1,15],si64>, !torch.vtensor<[15,1],si64>, !torch.int -> !torch.vtensor<[15,15],si64> | |
| %88 = torch.aten.gt.Scalar %87, %int0 : !torch.vtensor<[15,15],si64>, !torch.int -> !torch.vtensor<[15,15],i1> | |
| %89 = torch.aten.to.dtype %88, %int4, %false, %false, %none : !torch.vtensor<[15,15],i1>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[15,15],si64> | |
| %90 = torch.aten.mul.Scalar %89, %int16 : !torch.vtensor<[15,15],si64>, !torch.int -> !torch.vtensor<[15,15],si64> | |
| %91 = torch.aten.add.Scalar %90, %int0, %int1 : !torch.vtensor<[15,15],si64>, !torch.int, !torch.int -> !torch.vtensor<[15,15],si64> | |
| %92 = torch.aten.abs %87 : !torch.vtensor<[15,15],si64> -> !torch.vtensor<[15,15],si64> | |
| %93 = torch.aten.lt.Scalar %92, %int8 : !torch.vtensor<[15,15],si64>, !torch.int -> !torch.vtensor<[15,15],i1> | |
| %94 = torch.aten.to.dtype %92, %int6, %false, %false, %none : !torch.vtensor<[15,15],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[15,15],f32> | |
| %95 = torch.aten.div.Scalar %94, %int8 : !torch.vtensor<[15,15],f32>, !torch.int -> !torch.vtensor<[15,15],f32> | |
| %96 = torch.aten.log %95 : !torch.vtensor<[15,15],f32> -> !torch.vtensor<[15,15],f32> | |
| %97 = torch.aten.div.Scalar %96, %float2.772590e00 : !torch.vtensor<[15,15],f32>, !torch.float -> !torch.vtensor<[15,15],f32> | |
| %98 = torch.aten.mul.Scalar %97, %int8 : !torch.vtensor<[15,15],f32>, !torch.int -> !torch.vtensor<[15,15],f32> | |
| %99 = torch.aten.to.dtype %98, %int4, %false, %false, %none : !torch.vtensor<[15,15],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[15,15],si64> | |
| %100 = torch.aten.add.Scalar %99, %int8, %int1 : !torch.vtensor<[15,15],si64>, !torch.int, !torch.int -> !torch.vtensor<[15,15],si64> | |
| %101 = torch.prim.NumToTensor.Scalar %int15 : !torch.int -> !torch.vtensor<[],si64> | |
| %102 = torch.prim.ListConstruct %int15, %int15 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %103 = torch.aten.broadcast_to %101, %102 : !torch.vtensor<[],si64>, !torch.list<int> -> !torch.vtensor<[15,15],si64> | |
| %104 = torch.aten.minimum %100, %103 : !torch.vtensor<[15,15],si64>, !torch.vtensor<[15,15],si64> -> !torch.vtensor<[15,15],si64> | |
| %105 = torch.aten.where.self %93, %92, %104 : !torch.vtensor<[15,15],i1>, !torch.vtensor<[15,15],si64>, !torch.vtensor<[15,15],si64> -> !torch.vtensor<[15,15],si64> | |
| %106 = torch.aten.add.Tensor %91, %105, %int1 : !torch.vtensor<[15,15],si64>, !torch.vtensor<[15,15],si64>, !torch.int -> !torch.vtensor<[15,15],si64> | |
| %107 = torch.aten.embedding %2, %106, %int-1, %false, %false : !torch.vtensor<[32,8],f32>, !torch.vtensor<[15,15],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[15,15,8],f32> | |
| %108 = torch.prim.ListConstruct %int2, %int0, %int1 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %109 = torch.aten.permute %107, %108 : !torch.vtensor<[15,15,8],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %110 = torch.aten.unsqueeze %109, %int0 : !torch.vtensor<[8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %111 = torch.aten.add.Tensor %110, %41, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,1,1,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %112 = torch.aten.add.Tensor %82, %111, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %values, %indices = torch.aten.max.dim %112, %int-1, %true : !torch.vtensor<[1,8,15,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,15,1],f32>, !torch.vtensor<[1,8,15,1],si64> | |
| %113 = torch.aten.sub.Tensor %112, %values, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %114 = torch.aten.exp %113 : !torch.vtensor<[1,8,15,15],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %115 = torch.aten.sum.dim_IntList %114, %43, %true, %none : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,15,1],f32> | |
| %116 = torch.aten.div.Tensor %114, %115 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %117 = torch.aten.broadcast_to %116, %81 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %118 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %119 = torch.aten.view %117, %118 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %120 = torch.aten.broadcast_to %70, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %121 = torch.aten.view %120, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %122 = torch.aten.bmm %119, %121 : !torch.vtensor<[8,15,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,15,64],f32> | |
| %123 = torch.aten.view %122, %72 : !torch.vtensor<[8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %124 = torch.aten.transpose.int %123, %int1, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %125 = torch.aten.clone %124, %int0 : !torch.vtensor<[1,15,8,64],f32>, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %126 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %127 = torch.aten.view %125, %126 : !torch.vtensor<[1,15,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %128 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %129 = torch.aten.view %127, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %130 = torch.aten.mm %129, %128 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %131 = torch.aten.view %130, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %132 = torch.aten.add.Tensor %35, %131, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %133 = torch.aten.pow.Tensor_Scalar %132, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %134 = torch.aten.sum.dim_IntList %133, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %135 = torch.aten.div.Scalar %134, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %136 = torch.aten.add.Scalar %135, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %137 = torch.aten.rsqrt %136 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %138 = torch.aten.mul.Tensor %132, %137 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %139 = torch.aten.mul.Tensor %4, %138 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %140 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %141 = torch.aten.view %139, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %142 = torch.aten.mm %141, %140 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[15,2048],f32> | |
| %143 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %144 = torch.aten.view %142, %143 : !torch.vtensor<[15,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,15,2048],f32> | |
| %145 = torch.aten.relu %144 : !torch.vtensor<[1,15,2048],f32> -> !torch.vtensor<[1,15,2048],f32> | |
| %146 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %147 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %148 = torch.aten.view %145, %147 : !torch.vtensor<[1,15,2048],f32>, !torch.list<int> -> !torch.vtensor<[15,2048],f32> | |
| %149 = torch.aten.mm %148, %146 : !torch.vtensor<[15,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %150 = torch.aten.view %149, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %151 = torch.aten.add.Tensor %132, %150, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %152 = torch.aten.pow.Tensor_Scalar %151, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %153 = torch.aten.sum.dim_IntList %152, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %154 = torch.aten.div.Scalar %153, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %155 = torch.aten.add.Scalar %154, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %156 = torch.aten.rsqrt %155 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %157 = torch.aten.mul.Tensor %151, %156 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %158 = torch.aten.mul.Tensor %4, %157 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %159 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %160 = torch.aten.view %158, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %161 = torch.aten.mm %160, %159 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %162 = torch.aten.view %161, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %163 = torch.aten.view %162, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %164 = torch.aten.transpose.int %163, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %165 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %166 = torch.aten.view %158, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %167 = torch.aten.mm %166, %165 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %168 = torch.aten.view %167, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %169 = torch.aten.view %168, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %170 = torch.aten.transpose.int %169, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %171 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %172 = torch.aten.view %158, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %173 = torch.aten.mm %172, %171 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %174 = torch.aten.view %173, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %175 = torch.aten.view %174, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %176 = torch.aten.transpose.int %175, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %177 = torch.aten.transpose.int %170, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %178 = torch.aten.broadcast_to %164, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %179 = torch.aten.view %178, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %180 = torch.aten.broadcast_to %177, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %181 = torch.aten.view %180, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %182 = torch.aten.bmm %179, %181 : !torch.vtensor<[8,15,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,15,15],f32> | |
| %183 = torch.aten.view %182, %81 : !torch.vtensor<[8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %184 = torch.aten.add.Tensor %183, %111, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %values_0, %indices_1 = torch.aten.max.dim %184, %int-1, %true : !torch.vtensor<[1,8,15,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,15,1],f32>, !torch.vtensor<[1,8,15,1],si64> | |
| %185 = torch.aten.sub.Tensor %184, %values_0, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %186 = torch.aten.exp %185 : !torch.vtensor<[1,8,15,15],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %187 = torch.aten.sum.dim_IntList %186, %43, %true, %none : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,15,1],f32> | |
| %188 = torch.aten.div.Tensor %186, %187 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %189 = torch.aten.broadcast_to %188, %81 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %190 = torch.aten.view %189, %118 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %191 = torch.aten.broadcast_to %176, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %192 = torch.aten.view %191, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %193 = torch.aten.bmm %190, %192 : !torch.vtensor<[8,15,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,15,64],f32> | |
| %194 = torch.aten.view %193, %72 : !torch.vtensor<[8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %195 = torch.aten.transpose.int %194, %int1, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %196 = torch.aten.clone %195, %int0 : !torch.vtensor<[1,15,8,64],f32>, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %197 = torch.aten.view %196, %126 : !torch.vtensor<[1,15,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %198 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %199 = torch.aten.view %197, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %200 = torch.aten.mm %199, %198 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %201 = torch.aten.view %200, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %202 = torch.aten.add.Tensor %151, %201, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %203 = torch.aten.pow.Tensor_Scalar %202, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %204 = torch.aten.sum.dim_IntList %203, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %205 = torch.aten.div.Scalar %204, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %206 = torch.aten.add.Scalar %205, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %207 = torch.aten.rsqrt %206 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %208 = torch.aten.mul.Tensor %202, %207 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %209 = torch.aten.mul.Tensor %4, %208 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %210 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %211 = torch.aten.view %209, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %212 = torch.aten.mm %211, %210 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[15,2048],f32> | |
| %213 = torch.aten.view %212, %143 : !torch.vtensor<[15,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,15,2048],f32> | |
| %214 = torch.aten.relu %213 : !torch.vtensor<[1,15,2048],f32> -> !torch.vtensor<[1,15,2048],f32> | |
| %215 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %216 = torch.aten.view %214, %147 : !torch.vtensor<[1,15,2048],f32>, !torch.list<int> -> !torch.vtensor<[15,2048],f32> | |
| %217 = torch.aten.mm %216, %215 : !torch.vtensor<[15,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %218 = torch.aten.view %217, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %219 = torch.aten.add.Tensor %202, %218, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %220 = torch.aten.pow.Tensor_Scalar %219, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %221 = torch.aten.sum.dim_IntList %220, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %222 = torch.aten.div.Scalar %221, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %223 = torch.aten.add.Scalar %222, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %224 = torch.aten.rsqrt %223 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %225 = torch.aten.mul.Tensor %219, %224 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %226 = torch.aten.mul.Tensor %4, %225 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %227 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %228 = torch.aten.view %226, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %229 = torch.aten.mm %228, %227 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %230 = torch.aten.view %229, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %231 = torch.aten.view %230, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %232 = torch.aten.transpose.int %231, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %233 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %234 = torch.aten.view %226, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %235 = torch.aten.mm %234, %233 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %236 = torch.aten.view %235, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %237 = torch.aten.view %236, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %238 = torch.aten.transpose.int %237, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %239 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %240 = torch.aten.view %226, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %241 = torch.aten.mm %240, %239 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %242 = torch.aten.view %241, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %243 = torch.aten.view %242, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %244 = torch.aten.transpose.int %243, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %245 = torch.aten.transpose.int %238, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %246 = torch.aten.broadcast_to %232, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %247 = torch.aten.view %246, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %248 = torch.aten.broadcast_to %245, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %249 = torch.aten.view %248, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %250 = torch.aten.bmm %247, %249 : !torch.vtensor<[8,15,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,15,15],f32> | |
| %251 = torch.aten.view %250, %81 : !torch.vtensor<[8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %252 = torch.aten.add.Tensor %251, %111, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %values_2, %indices_3 = torch.aten.max.dim %252, %int-1, %true : !torch.vtensor<[1,8,15,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,15,1],f32>, !torch.vtensor<[1,8,15,1],si64> | |
| %253 = torch.aten.sub.Tensor %252, %values_2, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %254 = torch.aten.exp %253 : !torch.vtensor<[1,8,15,15],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %255 = torch.aten.sum.dim_IntList %254, %43, %true, %none : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,15,1],f32> | |
| %256 = torch.aten.div.Tensor %254, %255 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %257 = torch.aten.broadcast_to %256, %81 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %258 = torch.aten.view %257, %118 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %259 = torch.aten.broadcast_to %244, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %260 = torch.aten.view %259, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %261 = torch.aten.bmm %258, %260 : !torch.vtensor<[8,15,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,15,64],f32> | |
| %262 = torch.aten.view %261, %72 : !torch.vtensor<[8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %263 = torch.aten.transpose.int %262, %int1, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %264 = torch.aten.clone %263, %int0 : !torch.vtensor<[1,15,8,64],f32>, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %265 = torch.aten.view %264, %126 : !torch.vtensor<[1,15,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %266 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %267 = torch.aten.view %265, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %268 = torch.aten.mm %267, %266 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %269 = torch.aten.view %268, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %270 = torch.aten.add.Tensor %219, %269, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %271 = torch.aten.pow.Tensor_Scalar %270, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %272 = torch.aten.sum.dim_IntList %271, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %273 = torch.aten.div.Scalar %272, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %274 = torch.aten.add.Scalar %273, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %275 = torch.aten.rsqrt %274 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %276 = torch.aten.mul.Tensor %270, %275 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %277 = torch.aten.mul.Tensor %4, %276 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %278 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %279 = torch.aten.view %277, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %280 = torch.aten.mm %279, %278 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[15,2048],f32> | |
| %281 = torch.aten.view %280, %143 : !torch.vtensor<[15,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,15,2048],f32> | |
| %282 = torch.aten.relu %281 : !torch.vtensor<[1,15,2048],f32> -> !torch.vtensor<[1,15,2048],f32> | |
| %283 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %284 = torch.aten.view %282, %147 : !torch.vtensor<[1,15,2048],f32>, !torch.list<int> -> !torch.vtensor<[15,2048],f32> | |
| %285 = torch.aten.mm %284, %283 : !torch.vtensor<[15,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %286 = torch.aten.view %285, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %287 = torch.aten.add.Tensor %270, %286, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %288 = torch.aten.pow.Tensor_Scalar %287, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %289 = torch.aten.sum.dim_IntList %288, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %290 = torch.aten.div.Scalar %289, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %291 = torch.aten.add.Scalar %290, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %292 = torch.aten.rsqrt %291 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %293 = torch.aten.mul.Tensor %287, %292 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %294 = torch.aten.mul.Tensor %4, %293 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %295 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %296 = torch.aten.view %294, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %297 = torch.aten.mm %296, %295 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %298 = torch.aten.view %297, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %299 = torch.aten.view %298, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %300 = torch.aten.transpose.int %299, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %301 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %302 = torch.aten.view %294, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %303 = torch.aten.mm %302, %301 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %304 = torch.aten.view %303, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %305 = torch.aten.view %304, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %306 = torch.aten.transpose.int %305, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %307 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %308 = torch.aten.view %294, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %309 = torch.aten.mm %308, %307 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %310 = torch.aten.view %309, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %311 = torch.aten.view %310, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %312 = torch.aten.transpose.int %311, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %313 = torch.aten.transpose.int %306, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %314 = torch.aten.broadcast_to %300, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %315 = torch.aten.view %314, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %316 = torch.aten.broadcast_to %313, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %317 = torch.aten.view %316, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %318 = torch.aten.bmm %315, %317 : !torch.vtensor<[8,15,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,15,15],f32> | |
| %319 = torch.aten.view %318, %81 : !torch.vtensor<[8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %320 = torch.aten.add.Tensor %319, %111, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %values_4, %indices_5 = torch.aten.max.dim %320, %int-1, %true : !torch.vtensor<[1,8,15,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,15,1],f32>, !torch.vtensor<[1,8,15,1],si64> | |
| %321 = torch.aten.sub.Tensor %320, %values_4, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %322 = torch.aten.exp %321 : !torch.vtensor<[1,8,15,15],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %323 = torch.aten.sum.dim_IntList %322, %43, %true, %none : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,15,1],f32> | |
| %324 = torch.aten.div.Tensor %322, %323 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %325 = torch.aten.broadcast_to %324, %81 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %326 = torch.aten.view %325, %118 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %327 = torch.aten.broadcast_to %312, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %328 = torch.aten.view %327, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %329 = torch.aten.bmm %326, %328 : !torch.vtensor<[8,15,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,15,64],f32> | |
| %330 = torch.aten.view %329, %72 : !torch.vtensor<[8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %331 = torch.aten.transpose.int %330, %int1, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %332 = torch.aten.clone %331, %int0 : !torch.vtensor<[1,15,8,64],f32>, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %333 = torch.aten.view %332, %126 : !torch.vtensor<[1,15,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %334 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %335 = torch.aten.view %333, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %336 = torch.aten.mm %335, %334 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %337 = torch.aten.view %336, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %338 = torch.aten.add.Tensor %287, %337, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %339 = torch.aten.pow.Tensor_Scalar %338, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %340 = torch.aten.sum.dim_IntList %339, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %341 = torch.aten.div.Scalar %340, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %342 = torch.aten.add.Scalar %341, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %343 = torch.aten.rsqrt %342 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %344 = torch.aten.mul.Tensor %338, %343 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %345 = torch.aten.mul.Tensor %4, %344 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %346 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %347 = torch.aten.view %345, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %348 = torch.aten.mm %347, %346 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[15,2048],f32> | |
| %349 = torch.aten.view %348, %143 : !torch.vtensor<[15,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,15,2048],f32> | |
| %350 = torch.aten.relu %349 : !torch.vtensor<[1,15,2048],f32> -> !torch.vtensor<[1,15,2048],f32> | |
| %351 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %352 = torch.aten.view %350, %147 : !torch.vtensor<[1,15,2048],f32>, !torch.list<int> -> !torch.vtensor<[15,2048],f32> | |
| %353 = torch.aten.mm %352, %351 : !torch.vtensor<[15,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %354 = torch.aten.view %353, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %355 = torch.aten.add.Tensor %338, %354, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %356 = torch.aten.pow.Tensor_Scalar %355, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %357 = torch.aten.sum.dim_IntList %356, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %358 = torch.aten.div.Scalar %357, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %359 = torch.aten.add.Scalar %358, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %360 = torch.aten.rsqrt %359 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %361 = torch.aten.mul.Tensor %355, %360 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %362 = torch.aten.mul.Tensor %4, %361 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %363 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %364 = torch.aten.view %362, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %365 = torch.aten.mm %364, %363 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %366 = torch.aten.view %365, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %367 = torch.aten.view %366, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %368 = torch.aten.transpose.int %367, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %369 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %370 = torch.aten.view %362, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %371 = torch.aten.mm %370, %369 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %372 = torch.aten.view %371, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %373 = torch.aten.view %372, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %374 = torch.aten.transpose.int %373, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %375 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %376 = torch.aten.view %362, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %377 = torch.aten.mm %376, %375 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %378 = torch.aten.view %377, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %379 = torch.aten.view %378, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %380 = torch.aten.transpose.int %379, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %381 = torch.aten.transpose.int %374, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %382 = torch.aten.broadcast_to %368, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %383 = torch.aten.view %382, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %384 = torch.aten.broadcast_to %381, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %385 = torch.aten.view %384, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %386 = torch.aten.bmm %383, %385 : !torch.vtensor<[8,15,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,15,15],f32> | |
| %387 = torch.aten.view %386, %81 : !torch.vtensor<[8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %388 = torch.aten.add.Tensor %387, %111, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %values_6, %indices_7 = torch.aten.max.dim %388, %int-1, %true : !torch.vtensor<[1,8,15,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,15,1],f32>, !torch.vtensor<[1,8,15,1],si64> | |
| %389 = torch.aten.sub.Tensor %388, %values_6, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %390 = torch.aten.exp %389 : !torch.vtensor<[1,8,15,15],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %391 = torch.aten.sum.dim_IntList %390, %43, %true, %none : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,15,1],f32> | |
| %392 = torch.aten.div.Tensor %390, %391 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %393 = torch.aten.broadcast_to %392, %81 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %394 = torch.aten.view %393, %118 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %395 = torch.aten.broadcast_to %380, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %396 = torch.aten.view %395, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %397 = torch.aten.bmm %394, %396 : !torch.vtensor<[8,15,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,15,64],f32> | |
| %398 = torch.aten.view %397, %72 : !torch.vtensor<[8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %399 = torch.aten.transpose.int %398, %int1, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %400 = torch.aten.clone %399, %int0 : !torch.vtensor<[1,15,8,64],f32>, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %401 = torch.aten.view %400, %126 : !torch.vtensor<[1,15,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %402 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %403 = torch.aten.view %401, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %404 = torch.aten.mm %403, %402 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %405 = torch.aten.view %404, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %406 = torch.aten.add.Tensor %355, %405, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %407 = torch.aten.pow.Tensor_Scalar %406, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %408 = torch.aten.sum.dim_IntList %407, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %409 = torch.aten.div.Scalar %408, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %410 = torch.aten.add.Scalar %409, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %411 = torch.aten.rsqrt %410 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %412 = torch.aten.mul.Tensor %406, %411 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %413 = torch.aten.mul.Tensor %4, %412 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %414 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %415 = torch.aten.view %413, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %416 = torch.aten.mm %415, %414 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[15,2048],f32> | |
| %417 = torch.aten.view %416, %143 : !torch.vtensor<[15,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,15,2048],f32> | |
| %418 = torch.aten.relu %417 : !torch.vtensor<[1,15,2048],f32> -> !torch.vtensor<[1,15,2048],f32> | |
| %419 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %420 = torch.aten.view %418, %147 : !torch.vtensor<[1,15,2048],f32>, !torch.list<int> -> !torch.vtensor<[15,2048],f32> | |
| %421 = torch.aten.mm %420, %419 : !torch.vtensor<[15,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %422 = torch.aten.view %421, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %423 = torch.aten.add.Tensor %406, %422, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %424 = torch.aten.pow.Tensor_Scalar %423, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %425 = torch.aten.sum.dim_IntList %424, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %426 = torch.aten.div.Scalar %425, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %427 = torch.aten.add.Scalar %426, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %428 = torch.aten.rsqrt %427 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %429 = torch.aten.mul.Tensor %423, %428 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %430 = torch.aten.mul.Tensor %4, %429 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %431 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %432 = torch.aten.view %430, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %433 = torch.aten.mm %432, %431 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %434 = torch.aten.view %433, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %435 = torch.aten.view %434, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %436 = torch.aten.transpose.int %435, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %437 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %438 = torch.aten.view %430, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %439 = torch.aten.mm %438, %437 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %440 = torch.aten.view %439, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %441 = torch.aten.view %440, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %442 = torch.aten.transpose.int %441, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %443 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %444 = torch.aten.view %430, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %445 = torch.aten.mm %444, %443 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %446 = torch.aten.view %445, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %447 = torch.aten.view %446, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %448 = torch.aten.transpose.int %447, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %449 = torch.aten.transpose.int %442, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %450 = torch.aten.broadcast_to %436, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %451 = torch.aten.view %450, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %452 = torch.aten.broadcast_to %449, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %453 = torch.aten.view %452, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %454 = torch.aten.bmm %451, %453 : !torch.vtensor<[8,15,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,15,15],f32> | |
| %455 = torch.aten.view %454, %81 : !torch.vtensor<[8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %456 = torch.aten.add.Tensor %455, %111, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,15],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %values_8, %indices_9 = torch.aten.max.dim %456, %int-1, %true : !torch.vtensor<[1,8,15,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,15,1],f32>, !torch.vtensor<[1,8,15,1],si64> | |
| %457 = torch.aten.sub.Tensor %456, %values_8, %int1 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32>, !torch.int -> !torch.vtensor<[1,8,15,15],f32> | |
| %458 = torch.aten.exp %457 : !torch.vtensor<[1,8,15,15],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %459 = torch.aten.sum.dim_IntList %458, %43, %true, %none : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,15,1],f32> | |
| %460 = torch.aten.div.Tensor %458, %459 : !torch.vtensor<[1,8,15,15],f32>, !torch.vtensor<[1,8,15,1],f32> -> !torch.vtensor<[1,8,15,15],f32> | |
| %461 = torch.aten.broadcast_to %460, %81 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,15],f32> | |
| %462 = torch.aten.view %461, %118 : !torch.vtensor<[1,8,15,15],f32>, !torch.list<int> -> !torch.vtensor<[8,15,15],f32> | |
| %463 = torch.aten.broadcast_to %448, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %464 = torch.aten.view %463, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %465 = torch.aten.bmm %462, %464 : !torch.vtensor<[8,15,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,15,64],f32> | |
| %466 = torch.aten.view %465, %72 : !torch.vtensor<[8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %467 = torch.aten.transpose.int %466, %int1, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %468 = torch.aten.clone %467, %int0 : !torch.vtensor<[1,15,8,64],f32>, !torch.int -> !torch.vtensor<[1,15,8,64],f32> | |
| %469 = torch.aten.view %468, %126 : !torch.vtensor<[1,15,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %470 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %471 = torch.aten.view %469, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %472 = torch.aten.mm %471, %470 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %473 = torch.aten.view %472, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %474 = torch.aten.add.Tensor %423, %473, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %475 = torch.aten.pow.Tensor_Scalar %474, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %476 = torch.aten.sum.dim_IntList %475, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %477 = torch.aten.div.Scalar %476, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %478 = torch.aten.add.Scalar %477, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %479 = torch.aten.rsqrt %478 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %480 = torch.aten.mul.Tensor %474, %479 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %481 = torch.aten.mul.Tensor %4, %480 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %482 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %483 = torch.aten.view %481, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %484 = torch.aten.mm %483, %482 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[15,2048],f32> | |
| %485 = torch.aten.view %484, %143 : !torch.vtensor<[15,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,15,2048],f32> | |
| %486 = torch.aten.relu %485 : !torch.vtensor<[1,15,2048],f32> -> !torch.vtensor<[1,15,2048],f32> | |
| %487 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %488 = torch.aten.view %486, %147 : !torch.vtensor<[1,15,2048],f32>, !torch.list<int> -> !torch.vtensor<[15,2048],f32> | |
| %489 = torch.aten.mm %488, %487 : !torch.vtensor<[15,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %490 = torch.aten.view %489, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %491 = torch.aten.add.Tensor %474, %490, %int1 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %492 = torch.aten.pow.Tensor_Scalar %491, %int2 : !torch.vtensor<[1,15,512],f32>, !torch.int -> !torch.vtensor<[1,15,512],f32> | |
| %493 = torch.aten.sum.dim_IntList %492, %43, %true, %none : !torch.vtensor<[1,15,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,15,1],f32> | |
| %494 = torch.aten.div.Scalar %493, %int512 : !torch.vtensor<[1,15,1],f32>, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %495 = torch.aten.add.Scalar %494, %float9.999990e-07, %int1 : !torch.vtensor<[1,15,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,15,1],f32> | |
| %496 = torch.aten.rsqrt %495 : !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,1],f32> | |
| %497 = torch.aten.mul.Tensor %491, %496 : !torch.vtensor<[1,15,512],f32>, !torch.vtensor<[1,15,1],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %498 = torch.aten.mul.Tensor %4, %497 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,15,512],f32> -> !torch.vtensor<[1,15,512],f32> | |
| %499 = torch.prim.ListConstruct %int-1, %int4 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %500 = torch.aten.view %10, %499 : !torch.tensor<*,si64>, !torch.list<int> -> !torch.tensor<[1,4],si64> | |
| %501 = torch.tensor_static_info_cast %500 : !torch.tensor<[1,4],si64> to !torch.tensor<*,si64> | |
| %502 = torch.copy.to_vtensor %501 : !torch.vtensor<*,si64> | |
| %503 = torch.aten.embedding %5, %502, %int-1, %false, %false : !torch.vtensor<[32128,512],f32>, !torch.vtensor<*,si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[1,4,512],f32> | |
| %504 = torch.aten.ones %7, %none, %none, %cpu, %false : !torch.list<int>, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[1,4],f32> | |
| %505 = torch.aten.ones %36, %int4, %none, %cpu, %false : !torch.list<int>, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[1,15],si64> | |
| %506 = torch.aten.arange.start_step %int0, %int4, %int1, %none, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[4],si64> | |
| %507 = torch.aten.unsqueeze %506, %int0 : !torch.vtensor<[4],si64>, !torch.int -> !torch.vtensor<[1,4],si64> | |
| %508 = torch.aten.unsqueeze %507, %int1 : !torch.vtensor<[1,4],si64>, !torch.int -> !torch.vtensor<[1,1,4],si64> | |
| %509 = torch.prim.ListConstruct %int1, %int1, %int1, %int1, %int1, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %510 = torch.prim.ListConstruct %int1, %int1, %int4, %int1, %int1, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %511 = torch.prim.ListConstruct %int1, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %512 = torch.aten.view %508, %509 : !torch.vtensor<[1,1,4],si64>, !torch.list<int> -> !torch.vtensor<[1,1,1,1,1,4],si64> | |
| %513 = torch.aten.broadcast_to %512, %510 : !torch.vtensor<[1,1,1,1,1,4],si64>, !torch.list<int> -> !torch.vtensor<[1,1,4,1,1,4],si64> | |
| %514 = torch.aten.view %513, %511 : !torch.vtensor<[1,1,4,1,1,4],si64>, !torch.list<int> -> !torch.vtensor<[1,4,4],si64> | |
| %515 = torch.aten.unsqueeze %506, %int0 : !torch.vtensor<[4],si64>, !torch.int -> !torch.vtensor<[1,4],si64> | |
| %516 = torch.aten.unsqueeze %515, %int2 : !torch.vtensor<[1,4],si64>, !torch.int -> !torch.vtensor<[1,4,1],si64> | |
| %517 = torch.aten.le.Tensor %514, %516 : !torch.vtensor<[1,4,4],si64>, !torch.vtensor<[1,4,1],si64> -> !torch.vtensor<[1,4,4],i1> | |
| %518 = torch.aten.to.dtype %517, %int6, %false, %false, %none : !torch.vtensor<[1,4,4],i1>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,4,4],f32> | |
| %519 = torch.aten.unsqueeze %518, %int1 : !torch.vtensor<[1,4,4],f32>, !torch.int -> !torch.vtensor<[1,1,4,4],f32> | |
| %520 = torch.aten.unsqueeze %504, %int1 : !torch.vtensor<[1,4],f32>, !torch.int -> !torch.vtensor<[1,1,4],f32> | |
| %521 = torch.aten.unsqueeze %520, %int2 : !torch.vtensor<[1,1,4],f32>, !torch.int -> !torch.vtensor<[1,1,1,4],f32> | |
| %522 = torch.aten.mul.Tensor %519, %521 : !torch.vtensor<[1,1,4,4],f32>, !torch.vtensor<[1,1,1,4],f32> -> !torch.vtensor<[1,1,4,4],f32> | |
| %523 = torch.aten.rsub.Scalar %522, %float1.000000e00, %int1 : !torch.vtensor<[1,1,4,4],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,1,4,4],f32> | |
| %524 = torch.aten.mul.Scalar %523, %float-3.402820e38 : !torch.vtensor<[1,1,4,4],f32>, !torch.float -> !torch.vtensor<[1,1,4,4],f32> | |
| %525 = torch.aten.unsqueeze %505, %int1 : !torch.vtensor<[1,15],si64>, !torch.int -> !torch.vtensor<[1,1,15],si64> | |
| %526 = torch.aten.unsqueeze %525, %int2 : !torch.vtensor<[1,1,15],si64>, !torch.int -> !torch.vtensor<[1,1,1,15],si64> | |
| %527 = torch.aten.to.dtype %526, %int6, %false, %false, %none : !torch.vtensor<[1,1,1,15],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,1,15],f32> | |
| %528 = torch.aten.rsub.Scalar %527, %float1.000000e00, %int1 : !torch.vtensor<[1,1,1,15],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,1,1,15],f32> | |
| %529 = torch.aten.mul.Scalar %528, %float-3.402820e38 : !torch.vtensor<[1,1,1,15],f32>, !torch.float -> !torch.vtensor<[1,1,1,15],f32> | |
| %530 = torch.aten.pow.Tensor_Scalar %503, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %531 = torch.aten.sum.dim_IntList %530, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %532 = torch.aten.div.Scalar %531, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %533 = torch.aten.add.Scalar %532, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %534 = torch.aten.rsqrt %533 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %535 = torch.aten.mul.Tensor %503, %534 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %536 = torch.aten.mul.Tensor %4, %535 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %537 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %538 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %539 = torch.aten.view %536, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %540 = torch.aten.mm %539, %537 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %541 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %542 = torch.aten.view %540, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %543 = torch.aten.view %542, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %544 = torch.aten.transpose.int %543, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %545 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %546 = torch.aten.view %536, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %547 = torch.aten.mm %546, %545 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %548 = torch.aten.view %547, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %549 = torch.aten.view %548, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %550 = torch.aten.transpose.int %549, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %551 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %552 = torch.aten.view %536, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %553 = torch.aten.mm %552, %551 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %554 = torch.aten.view %553, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %555 = torch.aten.view %554, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %556 = torch.aten.transpose.int %555, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %557 = torch.aten.transpose.int %550, %int3, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,4],f32> | |
| %558 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %559 = torch.aten.broadcast_to %544, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %560 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %561 = torch.aten.view %559, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %562 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %563 = torch.aten.broadcast_to %557, %562 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,4],f32> | |
| %564 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %565 = torch.aten.view %563, %564 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[8,64,4],f32> | |
| %566 = torch.aten.bmm %561, %565 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,4],f32> -> !torch.vtensor<[8,4,4],f32> | |
| %567 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %568 = torch.aten.view %566, %567 : !torch.vtensor<[8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %569 = torch.aten.arange.start_step %int0, %int4, %int1, %int4, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[4],si64> | |
| %570 = torch.aten.unsqueeze %569, %int1 : !torch.vtensor<[4],si64>, !torch.int -> !torch.vtensor<[4,1],si64> | |
| %571 = torch.aten.arange.start_step %int0, %int4, %int1, %int4, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[4],si64> | |
| %572 = torch.aten.unsqueeze %571, %int0 : !torch.vtensor<[4],si64>, !torch.int -> !torch.vtensor<[1,4],si64> | |
| %573 = torch.aten.sub.Tensor %572, %570, %int1 : !torch.vtensor<[1,4],si64>, !torch.vtensor<[4,1],si64>, !torch.int -> !torch.vtensor<[4,4],si64> | |
| %574 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> | |
| %575 = torch.prim.ListConstruct %int4, %int4 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %576 = torch.aten.broadcast_to %574, %575 : !torch.vtensor<[],si64>, !torch.list<int> -> !torch.vtensor<[4,4],si64> | |
| %577 = torch.aten.minimum %573, %576 : !torch.vtensor<[4,4],si64>, !torch.vtensor<[4,4],si64> -> !torch.vtensor<[4,4],si64> | |
| %578 = torch.aten.neg %577 : !torch.vtensor<[4,4],si64> -> !torch.vtensor<[4,4],si64> | |
| %579 = torch.aten.lt.Scalar %578, %int16 : !torch.vtensor<[4,4],si64>, !torch.int -> !torch.vtensor<[4,4],i1> | |
| %580 = torch.aten.to.dtype %578, %int6, %false, %false, %none : !torch.vtensor<[4,4],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[4,4],f32> | |
| %581 = torch.aten.div.Scalar %580, %int16 : !torch.vtensor<[4,4],f32>, !torch.int -> !torch.vtensor<[4,4],f32> | |
| %582 = torch.aten.log %581 : !torch.vtensor<[4,4],f32> -> !torch.vtensor<[4,4],f32> | |
| %583 = torch.aten.div.Scalar %582, %float2.079440e00 : !torch.vtensor<[4,4],f32>, !torch.float -> !torch.vtensor<[4,4],f32> | |
| %584 = torch.aten.mul.Scalar %583, %int16 : !torch.vtensor<[4,4],f32>, !torch.int -> !torch.vtensor<[4,4],f32> | |
| %585 = torch.aten.to.dtype %584, %int4, %false, %false, %none : !torch.vtensor<[4,4],f32>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[4,4],si64> | |
| %586 = torch.aten.add.Scalar %585, %int16, %int1 : !torch.vtensor<[4,4],si64>, !torch.int, !torch.int -> !torch.vtensor<[4,4],si64> | |
| %587 = torch.prim.NumToTensor.Scalar %int31 : !torch.int -> !torch.vtensor<[],si64> | |
| %588 = torch.aten.broadcast_to %587, %575 : !torch.vtensor<[],si64>, !torch.list<int> -> !torch.vtensor<[4,4],si64> | |
| %589 = torch.aten.minimum %586, %588 : !torch.vtensor<[4,4],si64>, !torch.vtensor<[4,4],si64> -> !torch.vtensor<[4,4],si64> | |
| %590 = torch.aten.where.self %579, %578, %589 : !torch.vtensor<[4,4],i1>, !torch.vtensor<[4,4],si64>, !torch.vtensor<[4,4],si64> -> !torch.vtensor<[4,4],si64> | |
| %591 = torch.aten.add.Scalar %590, %int0, %int1 : !torch.vtensor<[4,4],si64>, !torch.int, !torch.int -> !torch.vtensor<[4,4],si64> | |
| %592 = torch.aten.embedding %2, %591, %int-1, %false, %false : !torch.vtensor<[32,8],f32>, !torch.vtensor<[4,4],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[4,4,8],f32> | |
| %593 = torch.aten.permute %592, %108 : !torch.vtensor<[4,4,8],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %594 = torch.aten.unsqueeze %593, %int0 : !torch.vtensor<[8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %595 = torch.aten.add.Tensor %594, %524, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,1,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %596 = torch.aten.add.Tensor %568, %595, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %values_10, %indices_11 = torch.aten.max.dim %596, %int-1, %true : !torch.vtensor<[1,8,4,4],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %597 = torch.aten.sub.Tensor %596, %values_10, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %598 = torch.aten.exp %597 : !torch.vtensor<[1,8,4,4],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %599 = torch.aten.sum.dim_IntList %598, %43, %true, %none : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %600 = torch.aten.div.Tensor %598, %599 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %601 = torch.aten.broadcast_to %600, %567 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %602 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %603 = torch.aten.view %601, %602 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %604 = torch.aten.broadcast_to %556, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %605 = torch.aten.view %604, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %606 = torch.aten.bmm %603, %605 : !torch.vtensor<[8,4,4],f32>, !torch.vtensor<[8,4,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %607 = torch.aten.view %606, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %608 = torch.aten.transpose.int %607, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %609 = torch.aten.clone %608, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %610 = torch.aten.view %609, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %611 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %612 = torch.aten.view %610, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %613 = torch.aten.mm %612, %611 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %614 = torch.aten.view %613, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %615 = torch.aten.add.Tensor %503, %614, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %616 = torch.aten.pow.Tensor_Scalar %615, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %617 = torch.aten.sum.dim_IntList %616, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %618 = torch.aten.div.Scalar %617, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %619 = torch.aten.add.Scalar %618, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %620 = torch.aten.rsqrt %619 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %621 = torch.aten.mul.Tensor %615, %620 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %622 = torch.aten.mul.Tensor %4, %621 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %623 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %624 = torch.aten.view %622, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %625 = torch.aten.mm %624, %623 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %626 = torch.aten.view %625, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %627 = torch.aten.view %626, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %628 = torch.aten.transpose.int %627, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %629 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %630 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %631 = torch.aten.mm %630, %629 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %632 = torch.aten.view %631, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %633 = torch.aten.view %632, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %634 = torch.aten.transpose.int %633, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %635 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %636 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %637 = torch.aten.mm %636, %635 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %638 = torch.aten.view %637, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %639 = torch.aten.view %638, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %640 = torch.aten.transpose.int %639, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %641 = torch.aten.transpose.int %634, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %642 = torch.aten.broadcast_to %628, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %643 = torch.aten.view %642, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %644 = torch.aten.broadcast_to %641, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %645 = torch.aten.view %644, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %646 = torch.aten.bmm %643, %645 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,4,15],f32> | |
| %647 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %648 = torch.aten.view %646, %647 : !torch.vtensor<[8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %649 = torch.aten.zeros %647, %int6, %none, %cpu, %false : !torch.list<int>, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[1,8,4,15],f32> | |
| %650 = torch.aten.add.Tensor %649, %529, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,1,1,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %651 = torch.aten.add.Tensor %648, %650, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %values_12, %indices_13 = torch.aten.max.dim %651, %int-1, %true : !torch.vtensor<[1,8,4,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %652 = torch.aten.sub.Tensor %651, %values_12, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %653 = torch.aten.exp %652 : !torch.vtensor<[1,8,4,15],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %654 = torch.aten.sum.dim_IntList %653, %43, %true, %none : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %655 = torch.aten.div.Tensor %653, %654 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %656 = torch.aten.broadcast_to %655, %647 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %657 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %658 = torch.aten.view %656, %657 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[8,4,15],f32> | |
| %659 = torch.aten.broadcast_to %640, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %660 = torch.aten.view %659, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %661 = torch.aten.bmm %658, %660 : !torch.vtensor<[8,4,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %662 = torch.aten.view %661, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %663 = torch.aten.transpose.int %662, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %664 = torch.aten.clone %663, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %665 = torch.aten.view %664, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %666 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %667 = torch.aten.view %665, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %668 = torch.aten.mm %667, %666 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %669 = torch.aten.view %668, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %670 = torch.aten.add.Tensor %615, %669, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %671 = torch.aten.pow.Tensor_Scalar %670, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %672 = torch.aten.sum.dim_IntList %671, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %673 = torch.aten.div.Scalar %672, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %674 = torch.aten.add.Scalar %673, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %675 = torch.aten.rsqrt %674 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %676 = torch.aten.mul.Tensor %670, %675 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %677 = torch.aten.mul.Tensor %4, %676 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %678 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %679 = torch.aten.view %677, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %680 = torch.aten.mm %679, %678 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[4,2048],f32> | |
| %681 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %682 = torch.aten.view %680, %681 : !torch.vtensor<[4,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,4,2048],f32> | |
| %683 = torch.aten.relu %682 : !torch.vtensor<[1,4,2048],f32> -> !torch.vtensor<[1,4,2048],f32> | |
| %684 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %685 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> | |
| %686 = torch.aten.view %683, %685 : !torch.vtensor<[1,4,2048],f32>, !torch.list<int> -> !torch.vtensor<[4,2048],f32> | |
| %687 = torch.aten.mm %686, %684 : !torch.vtensor<[4,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %688 = torch.aten.view %687, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %689 = torch.aten.add.Tensor %670, %688, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %690 = torch.aten.pow.Tensor_Scalar %689, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %691 = torch.aten.sum.dim_IntList %690, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %692 = torch.aten.div.Scalar %691, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %693 = torch.aten.add.Scalar %692, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %694 = torch.aten.rsqrt %693 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %695 = torch.aten.mul.Tensor %689, %694 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %696 = torch.aten.mul.Tensor %4, %695 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %697 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %698 = torch.aten.view %696, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %699 = torch.aten.mm %698, %697 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %700 = torch.aten.view %699, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %701 = torch.aten.view %700, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %702 = torch.aten.transpose.int %701, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %703 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %704 = torch.aten.view %696, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %705 = torch.aten.mm %704, %703 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %706 = torch.aten.view %705, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %707 = torch.aten.view %706, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %708 = torch.aten.transpose.int %707, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %709 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %710 = torch.aten.view %696, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %711 = torch.aten.mm %710, %709 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %712 = torch.aten.view %711, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %713 = torch.aten.view %712, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %714 = torch.aten.transpose.int %713, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %715 = torch.aten.transpose.int %708, %int3, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,4],f32> | |
| %716 = torch.aten.broadcast_to %702, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %717 = torch.aten.view %716, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %718 = torch.aten.broadcast_to %715, %562 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,4],f32> | |
| %719 = torch.aten.view %718, %564 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[8,64,4],f32> | |
| %720 = torch.aten.bmm %717, %719 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,4],f32> -> !torch.vtensor<[8,4,4],f32> | |
| %721 = torch.aten.view %720, %567 : !torch.vtensor<[8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %722 = torch.aten.add.Tensor %721, %595, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %values_14, %indices_15 = torch.aten.max.dim %722, %int-1, %true : !torch.vtensor<[1,8,4,4],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %723 = torch.aten.sub.Tensor %722, %values_14, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %724 = torch.aten.exp %723 : !torch.vtensor<[1,8,4,4],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %725 = torch.aten.sum.dim_IntList %724, %43, %true, %none : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %726 = torch.aten.div.Tensor %724, %725 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %727 = torch.aten.broadcast_to %726, %567 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %728 = torch.aten.view %727, %602 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %729 = torch.aten.broadcast_to %714, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %730 = torch.aten.view %729, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %731 = torch.aten.bmm %728, %730 : !torch.vtensor<[8,4,4],f32>, !torch.vtensor<[8,4,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %732 = torch.aten.view %731, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %733 = torch.aten.transpose.int %732, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %734 = torch.aten.clone %733, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %735 = torch.aten.view %734, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %736 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %737 = torch.aten.view %735, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %738 = torch.aten.mm %737, %736 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %739 = torch.aten.view %738, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %740 = torch.aten.add.Tensor %689, %739, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %741 = torch.aten.pow.Tensor_Scalar %740, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %742 = torch.aten.sum.dim_IntList %741, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %743 = torch.aten.div.Scalar %742, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %744 = torch.aten.add.Scalar %743, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %745 = torch.aten.rsqrt %744 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %746 = torch.aten.mul.Tensor %740, %745 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %747 = torch.aten.mul.Tensor %4, %746 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %748 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %749 = torch.aten.view %747, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %750 = torch.aten.mm %749, %748 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %751 = torch.aten.view %750, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %752 = torch.aten.view %751, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %753 = torch.aten.transpose.int %752, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %754 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %755 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %756 = torch.aten.mm %755, %754 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %757 = torch.aten.view %756, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %758 = torch.aten.view %757, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %759 = torch.aten.transpose.int %758, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %760 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %761 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %762 = torch.aten.mm %761, %760 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %763 = torch.aten.view %762, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %764 = torch.aten.view %763, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %765 = torch.aten.transpose.int %764, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %766 = torch.aten.transpose.int %759, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %767 = torch.aten.broadcast_to %753, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %768 = torch.aten.view %767, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %769 = torch.aten.broadcast_to %766, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %770 = torch.aten.view %769, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %771 = torch.aten.bmm %768, %770 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,4,15],f32> | |
| %772 = torch.aten.view %771, %647 : !torch.vtensor<[8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %773 = torch.aten.add.Tensor %772, %650, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %values_16, %indices_17 = torch.aten.max.dim %773, %int-1, %true : !torch.vtensor<[1,8,4,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %774 = torch.aten.sub.Tensor %773, %values_16, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %775 = torch.aten.exp %774 : !torch.vtensor<[1,8,4,15],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %776 = torch.aten.sum.dim_IntList %775, %43, %true, %none : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %777 = torch.aten.div.Tensor %775, %776 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %778 = torch.aten.broadcast_to %777, %647 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %779 = torch.aten.view %778, %657 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[8,4,15],f32> | |
| %780 = torch.aten.broadcast_to %765, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %781 = torch.aten.view %780, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %782 = torch.aten.bmm %779, %781 : !torch.vtensor<[8,4,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %783 = torch.aten.view %782, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %784 = torch.aten.transpose.int %783, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %785 = torch.aten.clone %784, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %786 = torch.aten.view %785, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %787 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %788 = torch.aten.view %786, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %789 = torch.aten.mm %788, %787 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %790 = torch.aten.view %789, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %791 = torch.aten.add.Tensor %740, %790, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %792 = torch.aten.pow.Tensor_Scalar %791, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %793 = torch.aten.sum.dim_IntList %792, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %794 = torch.aten.div.Scalar %793, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %795 = torch.aten.add.Scalar %794, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %796 = torch.aten.rsqrt %795 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %797 = torch.aten.mul.Tensor %791, %796 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %798 = torch.aten.mul.Tensor %4, %797 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %799 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %800 = torch.aten.view %798, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %801 = torch.aten.mm %800, %799 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[4,2048],f32> | |
| %802 = torch.aten.view %801, %681 : !torch.vtensor<[4,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,4,2048],f32> | |
| %803 = torch.aten.relu %802 : !torch.vtensor<[1,4,2048],f32> -> !torch.vtensor<[1,4,2048],f32> | |
| %804 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %805 = torch.aten.view %803, %685 : !torch.vtensor<[1,4,2048],f32>, !torch.list<int> -> !torch.vtensor<[4,2048],f32> | |
| %806 = torch.aten.mm %805, %804 : !torch.vtensor<[4,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %807 = torch.aten.view %806, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %808 = torch.aten.add.Tensor %791, %807, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %809 = torch.aten.pow.Tensor_Scalar %808, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %810 = torch.aten.sum.dim_IntList %809, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %811 = torch.aten.div.Scalar %810, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %812 = torch.aten.add.Scalar %811, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %813 = torch.aten.rsqrt %812 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %814 = torch.aten.mul.Tensor %808, %813 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %815 = torch.aten.mul.Tensor %4, %814 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %816 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %817 = torch.aten.view %815, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %818 = torch.aten.mm %817, %816 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %819 = torch.aten.view %818, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %820 = torch.aten.view %819, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %821 = torch.aten.transpose.int %820, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %822 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %823 = torch.aten.view %815, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %824 = torch.aten.mm %823, %822 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %825 = torch.aten.view %824, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %826 = torch.aten.view %825, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %827 = torch.aten.transpose.int %826, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %828 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %829 = torch.aten.view %815, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %830 = torch.aten.mm %829, %828 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %831 = torch.aten.view %830, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %832 = torch.aten.view %831, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %833 = torch.aten.transpose.int %832, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %834 = torch.aten.transpose.int %827, %int3, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,4],f32> | |
| %835 = torch.aten.broadcast_to %821, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %836 = torch.aten.view %835, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %837 = torch.aten.broadcast_to %834, %562 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,4],f32> | |
| %838 = torch.aten.view %837, %564 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[8,64,4],f32> | |
| %839 = torch.aten.bmm %836, %838 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,4],f32> -> !torch.vtensor<[8,4,4],f32> | |
| %840 = torch.aten.view %839, %567 : !torch.vtensor<[8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %841 = torch.aten.add.Tensor %840, %595, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %values_18, %indices_19 = torch.aten.max.dim %841, %int-1, %true : !torch.vtensor<[1,8,4,4],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %842 = torch.aten.sub.Tensor %841, %values_18, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %843 = torch.aten.exp %842 : !torch.vtensor<[1,8,4,4],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %844 = torch.aten.sum.dim_IntList %843, %43, %true, %none : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %845 = torch.aten.div.Tensor %843, %844 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %846 = torch.aten.broadcast_to %845, %567 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %847 = torch.aten.view %846, %602 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %848 = torch.aten.broadcast_to %833, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %849 = torch.aten.view %848, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %850 = torch.aten.bmm %847, %849 : !torch.vtensor<[8,4,4],f32>, !torch.vtensor<[8,4,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %851 = torch.aten.view %850, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %852 = torch.aten.transpose.int %851, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %853 = torch.aten.clone %852, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %854 = torch.aten.view %853, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %855 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %856 = torch.aten.view %854, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %857 = torch.aten.mm %856, %855 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %858 = torch.aten.view %857, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %859 = torch.aten.add.Tensor %808, %858, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %860 = torch.aten.pow.Tensor_Scalar %859, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %861 = torch.aten.sum.dim_IntList %860, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %862 = torch.aten.div.Scalar %861, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %863 = torch.aten.add.Scalar %862, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %864 = torch.aten.rsqrt %863 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %865 = torch.aten.mul.Tensor %859, %864 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %866 = torch.aten.mul.Tensor %4, %865 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %867 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %868 = torch.aten.view %866, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %869 = torch.aten.mm %868, %867 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %870 = torch.aten.view %869, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %871 = torch.aten.view %870, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %872 = torch.aten.transpose.int %871, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %873 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %874 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %875 = torch.aten.mm %874, %873 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %876 = torch.aten.view %875, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %877 = torch.aten.view %876, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %878 = torch.aten.transpose.int %877, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %879 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %880 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %881 = torch.aten.mm %880, %879 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %882 = torch.aten.view %881, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %883 = torch.aten.view %882, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %884 = torch.aten.transpose.int %883, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %885 = torch.aten.transpose.int %878, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %886 = torch.aten.broadcast_to %872, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %887 = torch.aten.view %886, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %888 = torch.aten.broadcast_to %885, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %889 = torch.aten.view %888, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %890 = torch.aten.bmm %887, %889 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,4,15],f32> | |
| %891 = torch.aten.view %890, %647 : !torch.vtensor<[8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %892 = torch.aten.add.Tensor %891, %650, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %values_20, %indices_21 = torch.aten.max.dim %892, %int-1, %true : !torch.vtensor<[1,8,4,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %893 = torch.aten.sub.Tensor %892, %values_20, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %894 = torch.aten.exp %893 : !torch.vtensor<[1,8,4,15],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %895 = torch.aten.sum.dim_IntList %894, %43, %true, %none : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %896 = torch.aten.div.Tensor %894, %895 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %897 = torch.aten.broadcast_to %896, %647 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %898 = torch.aten.view %897, %657 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[8,4,15],f32> | |
| %899 = torch.aten.broadcast_to %884, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %900 = torch.aten.view %899, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %901 = torch.aten.bmm %898, %900 : !torch.vtensor<[8,4,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %902 = torch.aten.view %901, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %903 = torch.aten.transpose.int %902, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %904 = torch.aten.clone %903, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %905 = torch.aten.view %904, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %906 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %907 = torch.aten.view %905, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %908 = torch.aten.mm %907, %906 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %909 = torch.aten.view %908, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %910 = torch.aten.add.Tensor %859, %909, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %911 = torch.aten.pow.Tensor_Scalar %910, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %912 = torch.aten.sum.dim_IntList %911, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %913 = torch.aten.div.Scalar %912, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %914 = torch.aten.add.Scalar %913, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %915 = torch.aten.rsqrt %914 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %916 = torch.aten.mul.Tensor %910, %915 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %917 = torch.aten.mul.Tensor %4, %916 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %918 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %919 = torch.aten.view %917, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %920 = torch.aten.mm %919, %918 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[4,2048],f32> | |
| %921 = torch.aten.view %920, %681 : !torch.vtensor<[4,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,4,2048],f32> | |
| %922 = torch.aten.relu %921 : !torch.vtensor<[1,4,2048],f32> -> !torch.vtensor<[1,4,2048],f32> | |
| %923 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %924 = torch.aten.view %922, %685 : !torch.vtensor<[1,4,2048],f32>, !torch.list<int> -> !torch.vtensor<[4,2048],f32> | |
| %925 = torch.aten.mm %924, %923 : !torch.vtensor<[4,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %926 = torch.aten.view %925, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %927 = torch.aten.add.Tensor %910, %926, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %928 = torch.aten.pow.Tensor_Scalar %927, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %929 = torch.aten.sum.dim_IntList %928, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %930 = torch.aten.div.Scalar %929, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %931 = torch.aten.add.Scalar %930, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %932 = torch.aten.rsqrt %931 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %933 = torch.aten.mul.Tensor %927, %932 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %934 = torch.aten.mul.Tensor %4, %933 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %935 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %936 = torch.aten.view %934, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %937 = torch.aten.mm %936, %935 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %938 = torch.aten.view %937, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %939 = torch.aten.view %938, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %940 = torch.aten.transpose.int %939, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %941 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %942 = torch.aten.view %934, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %943 = torch.aten.mm %942, %941 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %944 = torch.aten.view %943, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %945 = torch.aten.view %944, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %946 = torch.aten.transpose.int %945, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %947 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %948 = torch.aten.view %934, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %949 = torch.aten.mm %948, %947 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %950 = torch.aten.view %949, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %951 = torch.aten.view %950, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %952 = torch.aten.transpose.int %951, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %953 = torch.aten.transpose.int %946, %int3, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,4],f32> | |
| %954 = torch.aten.broadcast_to %940, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %955 = torch.aten.view %954, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %956 = torch.aten.broadcast_to %953, %562 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,4],f32> | |
| %957 = torch.aten.view %956, %564 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[8,64,4],f32> | |
| %958 = torch.aten.bmm %955, %957 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,4],f32> -> !torch.vtensor<[8,4,4],f32> | |
| %959 = torch.aten.view %958, %567 : !torch.vtensor<[8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %960 = torch.aten.add.Tensor %959, %595, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %values_22, %indices_23 = torch.aten.max.dim %960, %int-1, %true : !torch.vtensor<[1,8,4,4],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %961 = torch.aten.sub.Tensor %960, %values_22, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %962 = torch.aten.exp %961 : !torch.vtensor<[1,8,4,4],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %963 = torch.aten.sum.dim_IntList %962, %43, %true, %none : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %964 = torch.aten.div.Tensor %962, %963 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %965 = torch.aten.broadcast_to %964, %567 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %966 = torch.aten.view %965, %602 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %967 = torch.aten.broadcast_to %952, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %968 = torch.aten.view %967, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %969 = torch.aten.bmm %966, %968 : !torch.vtensor<[8,4,4],f32>, !torch.vtensor<[8,4,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %970 = torch.aten.view %969, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %971 = torch.aten.transpose.int %970, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %972 = torch.aten.clone %971, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %973 = torch.aten.view %972, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %974 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %975 = torch.aten.view %973, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %976 = torch.aten.mm %975, %974 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %977 = torch.aten.view %976, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %978 = torch.aten.add.Tensor %927, %977, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %979 = torch.aten.pow.Tensor_Scalar %978, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %980 = torch.aten.sum.dim_IntList %979, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %981 = torch.aten.div.Scalar %980, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %982 = torch.aten.add.Scalar %981, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %983 = torch.aten.rsqrt %982 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %984 = torch.aten.mul.Tensor %978, %983 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %985 = torch.aten.mul.Tensor %4, %984 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %986 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %987 = torch.aten.view %985, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %988 = torch.aten.mm %987, %986 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %989 = torch.aten.view %988, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %990 = torch.aten.view %989, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %991 = torch.aten.transpose.int %990, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %992 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %993 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %994 = torch.aten.mm %993, %992 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %995 = torch.aten.view %994, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %996 = torch.aten.view %995, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %997 = torch.aten.transpose.int %996, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %998 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %999 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %1000 = torch.aten.mm %999, %998 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %1001 = torch.aten.view %1000, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %1002 = torch.aten.view %1001, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %1003 = torch.aten.transpose.int %1002, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %1004 = torch.aten.transpose.int %997, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %1005 = torch.aten.broadcast_to %991, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1006 = torch.aten.view %1005, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1007 = torch.aten.broadcast_to %1004, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %1008 = torch.aten.view %1007, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %1009 = torch.aten.bmm %1006, %1008 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,4,15],f32> | |
| %1010 = torch.aten.view %1009, %647 : !torch.vtensor<[8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1011 = torch.aten.add.Tensor %1010, %650, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %values_24, %indices_25 = torch.aten.max.dim %1011, %int-1, %true : !torch.vtensor<[1,8,4,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %1012 = torch.aten.sub.Tensor %1011, %values_24, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %1013 = torch.aten.exp %1012 : !torch.vtensor<[1,8,4,15],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1014 = torch.aten.sum.dim_IntList %1013, %43, %true, %none : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %1015 = torch.aten.div.Tensor %1013, %1014 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1016 = torch.aten.broadcast_to %1015, %647 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1017 = torch.aten.view %1016, %657 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[8,4,15],f32> | |
| %1018 = torch.aten.broadcast_to %1003, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %1019 = torch.aten.view %1018, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %1020 = torch.aten.bmm %1017, %1019 : !torch.vtensor<[8,4,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %1021 = torch.aten.view %1020, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1022 = torch.aten.transpose.int %1021, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1023 = torch.aten.clone %1022, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1024 = torch.aten.view %1023, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1025 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1026 = torch.aten.view %1024, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1027 = torch.aten.mm %1026, %1025 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1028 = torch.aten.view %1027, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1029 = torch.aten.add.Tensor %978, %1028, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1030 = torch.aten.pow.Tensor_Scalar %1029, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1031 = torch.aten.sum.dim_IntList %1030, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1032 = torch.aten.div.Scalar %1031, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1033 = torch.aten.add.Scalar %1032, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1034 = torch.aten.rsqrt %1033 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1035 = torch.aten.mul.Tensor %1029, %1034 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1036 = torch.aten.mul.Tensor %4, %1035 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1037 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %1038 = torch.aten.view %1036, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1039 = torch.aten.mm %1038, %1037 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[4,2048],f32> | |
| %1040 = torch.aten.view %1039, %681 : !torch.vtensor<[4,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,4,2048],f32> | |
| %1041 = torch.aten.relu %1040 : !torch.vtensor<[1,4,2048],f32> -> !torch.vtensor<[1,4,2048],f32> | |
| %1042 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %1043 = torch.aten.view %1041, %685 : !torch.vtensor<[1,4,2048],f32>, !torch.list<int> -> !torch.vtensor<[4,2048],f32> | |
| %1044 = torch.aten.mm %1043, %1042 : !torch.vtensor<[4,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1045 = torch.aten.view %1044, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1046 = torch.aten.add.Tensor %1029, %1045, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1047 = torch.aten.pow.Tensor_Scalar %1046, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1048 = torch.aten.sum.dim_IntList %1047, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1049 = torch.aten.div.Scalar %1048, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1050 = torch.aten.add.Scalar %1049, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1051 = torch.aten.rsqrt %1050 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1052 = torch.aten.mul.Tensor %1046, %1051 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1053 = torch.aten.mul.Tensor %4, %1052 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1054 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1055 = torch.aten.view %1053, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1056 = torch.aten.mm %1055, %1054 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1057 = torch.aten.view %1056, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1058 = torch.aten.view %1057, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1059 = torch.aten.transpose.int %1058, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1060 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1061 = torch.aten.view %1053, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1062 = torch.aten.mm %1061, %1060 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1063 = torch.aten.view %1062, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1064 = torch.aten.view %1063, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1065 = torch.aten.transpose.int %1064, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1066 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1067 = torch.aten.view %1053, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1068 = torch.aten.mm %1067, %1066 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1069 = torch.aten.view %1068, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1070 = torch.aten.view %1069, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1071 = torch.aten.transpose.int %1070, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1072 = torch.aten.transpose.int %1065, %int3, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,4],f32> | |
| %1073 = torch.aten.broadcast_to %1059, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1074 = torch.aten.view %1073, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1075 = torch.aten.broadcast_to %1072, %562 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,4],f32> | |
| %1076 = torch.aten.view %1075, %564 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[8,64,4],f32> | |
| %1077 = torch.aten.bmm %1074, %1076 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,4],f32> -> !torch.vtensor<[8,4,4],f32> | |
| %1078 = torch.aten.view %1077, %567 : !torch.vtensor<[8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1079 = torch.aten.add.Tensor %1078, %595, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %values_26, %indices_27 = torch.aten.max.dim %1079, %int-1, %true : !torch.vtensor<[1,8,4,4],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %1080 = torch.aten.sub.Tensor %1079, %values_26, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %1081 = torch.aten.exp %1080 : !torch.vtensor<[1,8,4,4],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1082 = torch.aten.sum.dim_IntList %1081, %43, %true, %none : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %1083 = torch.aten.div.Tensor %1081, %1082 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1084 = torch.aten.broadcast_to %1083, %567 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1085 = torch.aten.view %1084, %602 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %1086 = torch.aten.broadcast_to %1071, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1087 = torch.aten.view %1086, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1088 = torch.aten.bmm %1085, %1087 : !torch.vtensor<[8,4,4],f32>, !torch.vtensor<[8,4,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %1089 = torch.aten.view %1088, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1090 = torch.aten.transpose.int %1089, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1091 = torch.aten.clone %1090, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1092 = torch.aten.view %1091, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1093 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1094 = torch.aten.view %1092, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1095 = torch.aten.mm %1094, %1093 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1096 = torch.aten.view %1095, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1097 = torch.aten.add.Tensor %1046, %1096, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1098 = torch.aten.pow.Tensor_Scalar %1097, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1099 = torch.aten.sum.dim_IntList %1098, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1100 = torch.aten.div.Scalar %1099, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1101 = torch.aten.add.Scalar %1100, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1102 = torch.aten.rsqrt %1101 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1103 = torch.aten.mul.Tensor %1097, %1102 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1104 = torch.aten.mul.Tensor %4, %1103 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1105 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1106 = torch.aten.view %1104, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1107 = torch.aten.mm %1106, %1105 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1108 = torch.aten.view %1107, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1109 = torch.aten.view %1108, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1110 = torch.aten.transpose.int %1109, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1111 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1112 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %1113 = torch.aten.mm %1112, %1111 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %1114 = torch.aten.view %1113, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %1115 = torch.aten.view %1114, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %1116 = torch.aten.transpose.int %1115, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %1117 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1118 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %1119 = torch.aten.mm %1118, %1117 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %1120 = torch.aten.view %1119, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %1121 = torch.aten.view %1120, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %1122 = torch.aten.transpose.int %1121, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %1123 = torch.aten.transpose.int %1116, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %1124 = torch.aten.broadcast_to %1110, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1125 = torch.aten.view %1124, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1126 = torch.aten.broadcast_to %1123, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %1127 = torch.aten.view %1126, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %1128 = torch.aten.bmm %1125, %1127 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,4,15],f32> | |
| %1129 = torch.aten.view %1128, %647 : !torch.vtensor<[8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1130 = torch.aten.add.Tensor %1129, %650, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %values_28, %indices_29 = torch.aten.max.dim %1130, %int-1, %true : !torch.vtensor<[1,8,4,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %1131 = torch.aten.sub.Tensor %1130, %values_28, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %1132 = torch.aten.exp %1131 : !torch.vtensor<[1,8,4,15],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1133 = torch.aten.sum.dim_IntList %1132, %43, %true, %none : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %1134 = torch.aten.div.Tensor %1132, %1133 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1135 = torch.aten.broadcast_to %1134, %647 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1136 = torch.aten.view %1135, %657 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[8,4,15],f32> | |
| %1137 = torch.aten.broadcast_to %1122, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %1138 = torch.aten.view %1137, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %1139 = torch.aten.bmm %1136, %1138 : !torch.vtensor<[8,4,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %1140 = torch.aten.view %1139, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1141 = torch.aten.transpose.int %1140, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1142 = torch.aten.clone %1141, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1143 = torch.aten.view %1142, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1144 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1145 = torch.aten.view %1143, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1146 = torch.aten.mm %1145, %1144 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1147 = torch.aten.view %1146, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1148 = torch.aten.add.Tensor %1097, %1147, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1149 = torch.aten.pow.Tensor_Scalar %1148, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1150 = torch.aten.sum.dim_IntList %1149, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1151 = torch.aten.div.Scalar %1150, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1152 = torch.aten.add.Scalar %1151, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1153 = torch.aten.rsqrt %1152 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1154 = torch.aten.mul.Tensor %1148, %1153 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1155 = torch.aten.mul.Tensor %4, %1154 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1156 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %1157 = torch.aten.view %1155, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1158 = torch.aten.mm %1157, %1156 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[4,2048],f32> | |
| %1159 = torch.aten.view %1158, %681 : !torch.vtensor<[4,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,4,2048],f32> | |
| %1160 = torch.aten.relu %1159 : !torch.vtensor<[1,4,2048],f32> -> !torch.vtensor<[1,4,2048],f32> | |
| %1161 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %1162 = torch.aten.view %1160, %685 : !torch.vtensor<[1,4,2048],f32>, !torch.list<int> -> !torch.vtensor<[4,2048],f32> | |
| %1163 = torch.aten.mm %1162, %1161 : !torch.vtensor<[4,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1164 = torch.aten.view %1163, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1165 = torch.aten.add.Tensor %1148, %1164, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1166 = torch.aten.pow.Tensor_Scalar %1165, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1167 = torch.aten.sum.dim_IntList %1166, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1168 = torch.aten.div.Scalar %1167, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1169 = torch.aten.add.Scalar %1168, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1170 = torch.aten.rsqrt %1169 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1171 = torch.aten.mul.Tensor %1165, %1170 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1172 = torch.aten.mul.Tensor %4, %1171 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1173 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1174 = torch.aten.view %1172, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1175 = torch.aten.mm %1174, %1173 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1176 = torch.aten.view %1175, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1177 = torch.aten.view %1176, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1178 = torch.aten.transpose.int %1177, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1179 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1180 = torch.aten.view %1172, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1181 = torch.aten.mm %1180, %1179 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1182 = torch.aten.view %1181, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1183 = torch.aten.view %1182, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1184 = torch.aten.transpose.int %1183, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1185 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1186 = torch.aten.view %1172, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1187 = torch.aten.mm %1186, %1185 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1188 = torch.aten.view %1187, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1189 = torch.aten.view %1188, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1190 = torch.aten.transpose.int %1189, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1191 = torch.aten.transpose.int %1184, %int3, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,4],f32> | |
| %1192 = torch.aten.broadcast_to %1178, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1193 = torch.aten.view %1192, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1194 = torch.aten.broadcast_to %1191, %562 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,4],f32> | |
| %1195 = torch.aten.view %1194, %564 : !torch.vtensor<[1,8,64,4],f32>, !torch.list<int> -> !torch.vtensor<[8,64,4],f32> | |
| %1196 = torch.aten.bmm %1193, %1195 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,4],f32> -> !torch.vtensor<[8,4,4],f32> | |
| %1197 = torch.aten.view %1196, %567 : !torch.vtensor<[8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1198 = torch.aten.add.Tensor %1197, %595, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,4],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %values_30, %indices_31 = torch.aten.max.dim %1198, %int-1, %true : !torch.vtensor<[1,8,4,4],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %1199 = torch.aten.sub.Tensor %1198, %values_30, %int1 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,4],f32> | |
| %1200 = torch.aten.exp %1199 : !torch.vtensor<[1,8,4,4],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1201 = torch.aten.sum.dim_IntList %1200, %43, %true, %none : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %1202 = torch.aten.div.Tensor %1200, %1201 : !torch.vtensor<[1,8,4,4],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1203 = torch.aten.broadcast_to %1202, %567 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,4],f32> | |
| %1204 = torch.aten.view %1203, %602 : !torch.vtensor<[1,8,4,4],f32>, !torch.list<int> -> !torch.vtensor<[8,4,4],f32> | |
| %1205 = torch.aten.broadcast_to %1190, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1206 = torch.aten.view %1205, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1207 = torch.aten.bmm %1204, %1206 : !torch.vtensor<[8,4,4],f32>, !torch.vtensor<[8,4,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %1208 = torch.aten.view %1207, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1209 = torch.aten.transpose.int %1208, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1210 = torch.aten.clone %1209, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1211 = torch.aten.view %1210, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1212 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1213 = torch.aten.view %1211, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1214 = torch.aten.mm %1213, %1212 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1215 = torch.aten.view %1214, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1216 = torch.aten.add.Tensor %1165, %1215, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1217 = torch.aten.pow.Tensor_Scalar %1216, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1218 = torch.aten.sum.dim_IntList %1217, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1219 = torch.aten.div.Scalar %1218, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1220 = torch.aten.add.Scalar %1219, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1221 = torch.aten.rsqrt %1220 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1222 = torch.aten.mul.Tensor %1216, %1221 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1223 = torch.aten.mul.Tensor %4, %1222 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1224 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1225 = torch.aten.view %1223, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1226 = torch.aten.mm %1225, %1224 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1227 = torch.aten.view %1226, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1228 = torch.aten.view %1227, %56 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,8,64],f32> | |
| %1229 = torch.aten.transpose.int %1228, %int1, %int2 : !torch.vtensor<[1,4,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,4,64],f32> | |
| %1230 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1231 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %1232 = torch.aten.mm %1231, %1230 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %1233 = torch.aten.view %1232, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %1234 = torch.aten.view %1233, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %1235 = torch.aten.transpose.int %1234, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %1236 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1237 = torch.aten.view %498, %51 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[15,512],f32> | |
| %1238 = torch.aten.mm %1237, %1236 : !torch.vtensor<[15,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[15,512],f32> | |
| %1239 = torch.aten.view %1238, %54 : !torch.vtensor<[15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,512],f32> | |
| %1240 = torch.aten.view %1239, %56 : !torch.vtensor<[1,15,512],f32>, !torch.list<int> -> !torch.vtensor<[1,15,8,64],f32> | |
| %1241 = torch.aten.transpose.int %1240, %int1, %int2 : !torch.vtensor<[1,15,8,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,15,64],f32> | |
| %1242 = torch.aten.transpose.int %1235, %int3, %int2 : !torch.vtensor<[1,8,15,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,8,64,15],f32> | |
| %1243 = torch.aten.broadcast_to %1229, %558 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1244 = torch.aten.view %1243, %560 : !torch.vtensor<[1,8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[8,4,64],f32> | |
| %1245 = torch.aten.broadcast_to %1242, %76 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,64,15],f32> | |
| %1246 = torch.aten.view %1245, %78 : !torch.vtensor<[1,8,64,15],f32>, !torch.list<int> -> !torch.vtensor<[8,64,15],f32> | |
| %1247 = torch.aten.bmm %1244, %1246 : !torch.vtensor<[8,4,64],f32>, !torch.vtensor<[8,64,15],f32> -> !torch.vtensor<[8,4,15],f32> | |
| %1248 = torch.aten.view %1247, %647 : !torch.vtensor<[8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1249 = torch.aten.add.Tensor %1248, %650, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,15],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %values_32, %indices_33 = torch.aten.max.dim %1249, %int-1, %true : !torch.vtensor<[1,8,4,15],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,8,4,1],f32>, !torch.vtensor<[1,8,4,1],si64> | |
| %1250 = torch.aten.sub.Tensor %1249, %values_32, %int1 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32>, !torch.int -> !torch.vtensor<[1,8,4,15],f32> | |
| %1251 = torch.aten.exp %1250 : !torch.vtensor<[1,8,4,15],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1252 = torch.aten.sum.dim_IntList %1251, %43, %true, %none : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,8,4,1],f32> | |
| %1253 = torch.aten.div.Tensor %1251, %1252 : !torch.vtensor<[1,8,4,15],f32>, !torch.vtensor<[1,8,4,1],f32> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1254 = torch.aten.broadcast_to %1253, %647 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,15],f32> | |
| %1255 = torch.aten.view %1254, %657 : !torch.vtensor<[1,8,4,15],f32>, !torch.list<int> -> !torch.vtensor<[8,4,15],f32> | |
| %1256 = torch.aten.broadcast_to %1241, %72 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,15,64],f32> | |
| %1257 = torch.aten.view %1256, %74 : !torch.vtensor<[1,8,15,64],f32>, !torch.list<int> -> !torch.vtensor<[8,15,64],f32> | |
| %1258 = torch.aten.bmm %1255, %1257 : !torch.vtensor<[8,4,15],f32>, !torch.vtensor<[8,15,64],f32> -> !torch.vtensor<[8,4,64],f32> | |
| %1259 = torch.aten.view %1258, %558 : !torch.vtensor<[8,4,64],f32>, !torch.list<int> -> !torch.vtensor<[1,8,4,64],f32> | |
| %1260 = torch.aten.transpose.int %1259, %int1, %int2 : !torch.vtensor<[1,8,4,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1261 = torch.aten.clone %1260, %int0 : !torch.vtensor<[1,4,8,64],f32>, !torch.int -> !torch.vtensor<[1,4,8,64],f32> | |
| %1262 = torch.aten.view %1261, %126 : !torch.vtensor<[1,4,8,64],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1263 = torch.aten.transpose.int %3, %int0, %int1 : !torch.vtensor<[512,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,512],f32> | |
| %1264 = torch.aten.view %1262, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1265 = torch.aten.mm %1264, %1263 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1266 = torch.aten.view %1265, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1267 = torch.aten.add.Tensor %1216, %1266, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1268 = torch.aten.pow.Tensor_Scalar %1267, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1269 = torch.aten.sum.dim_IntList %1268, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1270 = torch.aten.div.Scalar %1269, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1271 = torch.aten.add.Scalar %1270, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1272 = torch.aten.rsqrt %1271 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1273 = torch.aten.mul.Tensor %1267, %1272 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1274 = torch.aten.mul.Tensor %4, %1273 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1275 = torch.aten.transpose.int %1, %int0, %int1 : !torch.vtensor<[2048,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,2048],f32> | |
| %1276 = torch.aten.view %1274, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1277 = torch.aten.mm %1276, %1275 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,2048],f32> -> !torch.vtensor<[4,2048],f32> | |
| %1278 = torch.aten.view %1277, %681 : !torch.vtensor<[4,2048],f32>, !torch.list<int> -> !torch.vtensor<[1,4,2048],f32> | |
| %1279 = torch.aten.relu %1278 : !torch.vtensor<[1,4,2048],f32> -> !torch.vtensor<[1,4,2048],f32> | |
| %1280 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[512,2048],f32>, !torch.int, !torch.int -> !torch.vtensor<[2048,512],f32> | |
| %1281 = torch.aten.view %1279, %685 : !torch.vtensor<[1,4,2048],f32>, !torch.list<int> -> !torch.vtensor<[4,2048],f32> | |
| %1282 = torch.aten.mm %1281, %1280 : !torch.vtensor<[4,2048],f32>, !torch.vtensor<[2048,512],f32> -> !torch.vtensor<[4,512],f32> | |
| %1283 = torch.aten.view %1282, %541 : !torch.vtensor<[4,512],f32>, !torch.list<int> -> !torch.vtensor<[1,4,512],f32> | |
| %1284 = torch.aten.add.Tensor %1267, %1283, %int1 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1285 = torch.aten.pow.Tensor_Scalar %1284, %int2 : !torch.vtensor<[1,4,512],f32>, !torch.int -> !torch.vtensor<[1,4,512],f32> | |
| %1286 = torch.aten.sum.dim_IntList %1285, %43, %true, %none : !torch.vtensor<[1,4,512],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,4,1],f32> | |
| %1287 = torch.aten.div.Scalar %1286, %int512 : !torch.vtensor<[1,4,1],f32>, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1288 = torch.aten.add.Scalar %1287, %float9.999990e-07, %int1 : !torch.vtensor<[1,4,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,4,1],f32> | |
| %1289 = torch.aten.rsqrt %1288 : !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,1],f32> | |
| %1290 = torch.aten.mul.Tensor %1284, %1289 : !torch.vtensor<[1,4,512],f32>, !torch.vtensor<[1,4,1],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1291 = torch.aten.mul.Tensor %4, %1290 : !torch.vtensor<[512],f32>, !torch.vtensor<[1,4,512],f32> -> !torch.vtensor<[1,4,512],f32> | |
| %1292 = torch.aten.mul.Scalar %1291, %float4.419420e-02 : !torch.vtensor<[1,4,512],f32>, !torch.float -> !torch.vtensor<[1,4,512],f32> | |
| %1293 = torch.aten.transpose.int %5, %int0, %int1 : !torch.vtensor<[32128,512],f32>, !torch.int, !torch.int -> !torch.vtensor<[512,32128],f32> | |
| %1294 = torch.aten.view %1292, %538 : !torch.vtensor<[1,4,512],f32>, !torch.list<int> -> !torch.vtensor<[4,512],f32> | |
| %1295 = torch.aten.mm %1294, %1293 : !torch.vtensor<[4,512],f32>, !torch.vtensor<[512,32128],f32> -> !torch.vtensor<[4,32128],f32> | |
| %1296 = torch.prim.ListConstruct %int1, %int4, %int32128 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
| %1297 = torch.aten.view %1295, %1296 : !torch.vtensor<[4,32128],f32>, !torch.list<int> -> !torch.vtensor<[1,4,32128],f32> | |
| return %1297 : !torch.vtensor<[1,4,32128],f32> | |
| } | |
| } | |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment