Skip to content

Instantly share code, notes, and snippets.

@AmosLewis
Created May 25, 2023 15:56
Show Gist options
  • Save AmosLewis/ca8a6523e88bb69c5cae3bcb1b50555d to your computer and use it in GitHub Desktop.
Save AmosLewis/ca8a6523e88bb69c5cae3bcb1b50555d to your computer and use it in GitHub Desktop.
#loc = loc(unknown)
module attributes {torch.debug_module_name = "_lambda"} {
func.func @forward(%arg0: !torch.vtensor<[1,128],si64> loc(unknown)) -> !torch.vtensor<[1,2],f32> {
%int128 = torch.constant.int 128 loc(#loc1)
%int127 = torch.constant.int 127 loc(#loc2)
%int1 = torch.constant.int 1 loc(#loc3)
%true = torch.constant.bool true loc(#loc4)
%int0 = torch.constant.int 0 loc(#loc5)
%int2 = torch.constant.int 2 loc(#loc6)
%none = torch.constant.none loc(#loc)
%int768 = torch.constant.int 768 loc(#loc7)
%float1.000000e00 = torch.constant.float 1.000000e+00 loc(#loc8)
%int6 = torch.constant.int 6 loc(#loc9)
%false = torch.constant.bool false loc(#loc10)
%int11 = torch.constant.int 11 loc(#loc11)
%0 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2x768xf32>) : !torch.vtensor<[2,768],f32> loc(#loc)
%1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072x768xf32>) : !torch.vtensor<[3072,768],f32> loc(#loc)
%4 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x3072xf32>) : !torch.vtensor<[768,3072],f32> loc(#loc)
%6 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072xf32>) : !torch.vtensor<[3072],f32> loc(#loc)
%7 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%8 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%9 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x768xf32>) : !torch.vtensor<[768,768],f32> loc(#loc)
%10 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%11 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x2304xf32>) : !torch.vtensor<[768,2304],f32> loc(#loc)
%12 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2304xf32>) : !torch.vtensor<[2304],f32> loc(#loc)
%13 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%14 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%15 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072x768xf32>) : !torch.vtensor<[3072,768],f32> loc(#loc)
%16 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%17 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x3072xf32>) : !torch.vtensor<[768,3072],f32> loc(#loc)
%18 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072xf32>) : !torch.vtensor<[3072],f32> loc(#loc)
%19 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%20 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%21 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x768xf32>) : !torch.vtensor<[768,768],f32> loc(#loc)
%22 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%23 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x2304xf32>) : !torch.vtensor<[768,2304],f32> loc(#loc)
%24 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2304xf32>) : !torch.vtensor<[2304],f32> loc(#loc)
%25 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%26 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%27 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072x768xf32>) : !torch.vtensor<[3072,768],f32> loc(#loc)
%28 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%29 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x3072xf32>) : !torch.vtensor<[768,3072],f32> loc(#loc)
%30 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072xf32>) : !torch.vtensor<[3072],f32> loc(#loc)
%31 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%32 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%33 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x768xf32>) : !torch.vtensor<[768,768],f32> loc(#loc)
%34 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%35 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x2304xf32>) : !torch.vtensor<[768,2304],f32> loc(#loc)
%36 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2304xf32>) : !torch.vtensor<[2304],f32> loc(#loc)
%37 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%38 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%39 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072x768xf32>) : !torch.vtensor<[3072,768],f32> loc(#loc)
%40 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%41 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x3072xf32>) : !torch.vtensor<[768,3072],f32> loc(#loc)
%42 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072xf32>) : !torch.vtensor<[3072],f32> loc(#loc)
%43 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%44 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%45 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x768xf32>) : !torch.vtensor<[768,768],f32> loc(#loc)
%46 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%47 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x2304xf32>) : !torch.vtensor<[768,2304],f32> loc(#loc)
%48 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2304xf32>) : !torch.vtensor<[2304],f32> loc(#loc)
%49 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%50 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%51 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072x768xf32>) : !torch.vtensor<[3072,768],f32> loc(#loc)
%52 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%53 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x3072xf32>) : !torch.vtensor<[768,3072],f32> loc(#loc)
%54 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072xf32>) : !torch.vtensor<[3072],f32> loc(#loc)
%55 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%56 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%57 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x768xf32>) : !torch.vtensor<[768,768],f32> loc(#loc)
%58 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%59 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x2304xf32>) : !torch.vtensor<[768,2304],f32> loc(#loc)
%60 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2304xf32>) : !torch.vtensor<[2304],f32> loc(#loc)
%61 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%62 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%63 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072x768xf32>) : !torch.vtensor<[3072,768],f32> loc(#loc)
%64 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%65 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x3072xf32>) : !torch.vtensor<[768,3072],f32> loc(#loc)
%66 = torch.vtensor.literal(dense_resource<__elided__> : tensor<3072xf32>) : !torch.vtensor<[3072],f32> loc(#loc)
%67 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%68 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%69 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x768xf32>) : !torch.vtensor<[768,768],f32> loc(#loc)
%70 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%71 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x1x1024x1024xui8>) : !torch.vtensor<[1,1,1024,1024],ui8> loc(#loc)
%72 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768x2304xf32>) : !torch.vtensor<[768,2304],f32> loc(#loc)
%73 = torch.vtensor.literal(dense_resource<__elided__> : tensor<2304xf32>) : !torch.vtensor<[2304],f32> loc(#loc)
%74 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%75 = torch.vtensor.literal(dense_resource<__elided__> : tensor<768xf32>) : !torch.vtensor<[768],f32> loc(#loc)
%76 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1024x768xf32>) : !torch.vtensor<[1024,768],f32> loc(#loc)
%77 = torch.vtensor.literal(dense_resource<__elided__> : tensor<50257x768xf32>) : !torch.vtensor<[50257,768],f32> loc(#loc)
%int-2 = torch.constant.int -2 loc(#loc12)
%float-3.402820e38 = torch.constant.float -3.4028234663852886E+38 loc(#loc13)
%int4 = torch.constant.int 4 loc(#loc14)
%int-1 = torch.constant.int -1 loc(#loc15)
%float1.000000e-05 = torch.constant.float 1.000000e-05 loc(#loc16)
%int2304 = torch.constant.int 2304 loc(#loc17)
%int1536 = torch.constant.int 1536 loc(#loc18)
%int12 = torch.constant.int 12 loc(#loc19)
%int64 = torch.constant.int 64 loc(#loc20)
%int3 = torch.constant.int 3 loc(#loc21)
%float8.000000e00 = torch.constant.float 8.000000e+00 loc(#loc22)
%int3072 = torch.constant.int 3072 loc(#loc23)
%float5.000000e-01 = torch.constant.float 5.000000e-01 loc(#loc24)
%float3.000000e00 = torch.constant.float 3.000000e+00 loc(#loc25)
%float4.471500e-02 = torch.constant.float 4.471500e-02 loc(#loc26)
%float7.978850e-01 = torch.constant.float 0.79788456080286541 loc(#loc27)
%cpu = torch.constant.device "cpu" loc(#loc)
%78 = torch.prim.ListConstruct %int-1, %int128 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc)
%79 = torch.aten.view %arg0, %78 : !torch.vtensor<[1,128],si64>, !torch.list<int> -> !torch.vtensor<[1,128],si64> loc(#loc28)
%80 = torch.aten.arange.start_step %int0, %int128, %int1, %int4, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[128],si64> loc(#loc29)
%81 = torch.aten.unsqueeze %80, %int0 : !torch.vtensor<[128],si64>, !torch.int -> !torch.vtensor<[1,128],si64> loc(#loc30)
%82 = torch.aten.view %81, %78 : !torch.vtensor<[1,128],si64>, !torch.list<int> -> !torch.vtensor<[1,128],si64> loc(#loc31)
%83 = torch.aten.embedding %77, %79, %int-1, %false, %false : !torch.vtensor<[50257,768],f32>, !torch.vtensor<[1,128],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[1,128,768],f32> loc(#loc32)
%84 = torch.aten.embedding %76, %82, %int-1, %false, %false : !torch.vtensor<[1024,768],f32>, !torch.vtensor<[1,128],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[1,128,768],f32> loc(#loc33)
%85 = torch.aten.add.Tensor %83, %84, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc34)
%86 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc4)
%87 = torch.aten.sum.dim_IntList %85, %86, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc4)
%88 = torch.aten.div.Scalar %87, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc4)
%89 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc4)
%90 = torch.aten.broadcast_to %88, %89 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%91 = torch.aten.sub.Tensor %85, %90, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%92 = torch.aten.mul.Tensor %91, %91 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%93 = torch.aten.sum.dim_IntList %92, %86, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc4)
%94 = torch.aten.div.Scalar %93, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc4)
%95 = torch.aten.add.Scalar %94, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc4)
%96 = torch.aten.rsqrt %95 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc4)
%97 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc4)
%98 = torch.aten.broadcast_to %96, %97 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%99 = torch.aten.mul.Tensor %91, %98 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%100 = torch.aten.mul.Tensor %99, %75 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%101 = torch.aten.add.Tensor %100, %74, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc4)
%102 = torch.prim.ListConstruct %int-1, %int768 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc)
%103 = torch.aten.view %101, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc35)
%104 = torch.aten.mm %103, %72 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2304],f32> -> !torch.vtensor<[128,2304],f32> loc(#loc36)
%105 = torch.aten.mul.Scalar %73, %int1 : !torch.vtensor<[2304],f32>, !torch.int -> !torch.vtensor<[2304],f32> loc(#loc36)
%106 = torch.aten.add.Tensor %105, %104, %int1 : !torch.vtensor<[2304],f32>, !torch.vtensor<[128,2304],f32>, !torch.int -> !torch.vtensor<[128,2304],f32> loc(#loc36)
%107 = torch.prim.ListConstruct %int1, %int128, %int2304 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc393)
%108 = torch.aten.view %106, %107 : !torch.vtensor<[128,2304],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2304],f32> loc(#loc39)
%109 = torch.aten.slice.Tensor %108, %int2, %int0, %int768, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc40)
%110 = torch.aten.slice.Tensor %108, %int2, %int768, %int1536, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc41)
%111 = torch.aten.slice.Tensor %108, %int2, %int1536, %int2304, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc42)
%112 = torch.prim.ListConstruct %int1, %int128, %int12, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc394)
%113 = torch.aten.view %109, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc43)
%114 = torch.prim.ListConstruct %int0, %int2, %int1, %int3 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc)
%115 = torch.aten.permute %113, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc44)
%116 = torch.aten.view %110, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc45)
%117 = torch.aten.permute %116, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc46)
%118 = torch.aten.view %111, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc47)
%119 = torch.aten.permute %118, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc48)
%120 = torch.aten.transpose.int %117, %int-1, %int-2 : !torch.vtensor<[1,12,128,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,12,64,128],f32> loc(#loc49)
%121 = torch.prim.ListConstruct %int1, %int12, %int128, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc395)
%122 = torch.aten.broadcast_to %115, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc52)
%123 = torch.prim.ListConstruct %int12, %int128, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc396)
%124 = torch.aten.view %122, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc53)
%125 = torch.prim.ListConstruct %int1, %int12, %int64, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc397)
%126 = torch.aten.broadcast_to %120, %125 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,64,128],f32> loc(#loc54)
%127 = torch.prim.ListConstruct %int12, %int64, %int128 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc398)
%128 = torch.aten.view %126, %127 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[12,64,128],f32> loc(#loc55)
%129 = torch.aten.bmm %124, %128 : !torch.vtensor<[12,128,64],f32>, !torch.vtensor<[12,64,128],f32> -> !torch.vtensor<[12,128,128],f32> loc(#loc56)
%130 = torch.prim.ListConstruct %int1, %int12, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc399)
%131 = torch.aten.view %129, %130 : !torch.vtensor<[12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc60)
%132 = torch.prim.ListConstruct : () -> !torch.list<int> loc(#loc)
%133 = torch.prim.NumToTensor.Scalar %float8.000000e00 : !torch.float -> !torch.vtensor<[],f64> loc(#loc61)
%134 = torch.aten.to.dtype %133, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc61)
%135 = torch.aten.broadcast_to %134, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc61)
%136 = torch.aten.div.Tensor %131, %135 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc59)
%137 = torch.aten.slice.Tensor %71, %int2, %int0, %int128, %int1 : !torch.vtensor<[1,1,1024,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,1024],ui8> loc(#loc62)
%138 = torch.aten.slice.Tensor %137, %int3, %int0, %int128, %int1 : !torch.vtensor<[1,1,128,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,128],ui8> loc(#loc63)
%139 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> loc(#loc64)
%140 = torch.aten.to.dtype %139, %int11, %false, %false, %none : !torch.vtensor<[],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],i1> loc(#loc64)
%141 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc64)
%142 = torch.aten.broadcast_to %140, %141 : !torch.vtensor<[],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc64)
%143 = torch.aten.to.dtype %138, %int11, %false, %false, %none : !torch.vtensor<[1,1,128,128],ui8>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,128,128],i1> loc(#loc64)
%144 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc64)
%145 = torch.aten.broadcast_to %143, %144 : !torch.vtensor<[1,1,128,128],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc64)
%146 = torch.prim.NumToTensor.Scalar %float-3.402820e38 : !torch.float -> !torch.vtensor<[],f64> loc(#loc65)
%147 = torch.aten.to.dtype %146, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc65)
%148 = torch.aten.broadcast_to %147, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc65)
%149 = torch.aten.where.self %145, %136, %148 : !torch.vtensor<[1,1,128,128],i1>, !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc66)
%values, %indices = torch.aten.max.dim %149, %int-1, %true : !torch.vtensor<[1,12,128,128],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,12,128,1],f32>, !torch.vtensor<[1,12,128,1],si64> loc(#loc67)
%150 = torch.aten.sub.Tensor %149, %values, %float1.000000e00 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32>, !torch.float -> !torch.vtensor<[1,12,128,128],f32> loc(#loc67)
%151 = torch.aten.exp %150 : !torch.vtensor<[1,12,128,128],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc67)
%152 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc67)
%153 = torch.aten.sum.dim_IntList %151, %152, %true, %none : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,12,128,1],f32> loc(#loc67)
%154 = torch.aten.div.Tensor %151, %153 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc67)
%155 = torch.aten.broadcast_to %154, %130 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc68)
%156 = torch.prim.ListConstruct %int12, %int128, %int128 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc400)
%157 = torch.aten.view %155, %156 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[12,128,128],f32> loc(#loc69)
%158 = torch.aten.broadcast_to %119, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc70)
%159 = torch.aten.view %158, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc71)
%160 = torch.aten.bmm %157, %159 : !torch.vtensor<[12,128,128],f32>, !torch.vtensor<[12,128,64],f32> -> !torch.vtensor<[12,128,64],f32> loc(#loc72)
%161 = torch.aten.view %160, %121 : !torch.vtensor<[12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc73)
%162 = torch.aten.permute %161, %114 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc74)
%163 = torch.aten.clone %162, %int0 : !torch.vtensor<[1,128,12,64],f32>, !torch.int -> !torch.vtensor<[1,128,12,64],f32> loc(#loc75)
%164 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc401)
%165 = torch.aten.view %163, %164 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc76)
%166 = torch.aten.view %165, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc77)
%167 = torch.aten.mm %166, %69 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc78)
%168 = torch.aten.mul.Scalar %70, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc78)
%169 = torch.aten.add.Tensor %168, %167, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc78)
%170 = torch.aten.view %169, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc79)
%171 = torch.aten.add.Tensor %170, %85, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc80)
%172 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc81)
%173 = torch.aten.sum.dim_IntList %171, %172, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc81)
%174 = torch.aten.div.Scalar %173, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc81)
%175 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc81)
%176 = torch.aten.broadcast_to %174, %175 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%177 = torch.aten.sub.Tensor %171, %176, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%178 = torch.aten.mul.Tensor %177, %177 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%179 = torch.aten.sum.dim_IntList %178, %172, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc81)
%180 = torch.aten.div.Scalar %179, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc81)
%181 = torch.aten.add.Scalar %180, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc81)
%182 = torch.aten.rsqrt %181 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc81)
%183 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc81)
%184 = torch.aten.broadcast_to %182, %183 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%185 = torch.aten.mul.Tensor %177, %184 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%186 = torch.aten.mul.Tensor %185, %68 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%187 = torch.aten.add.Tensor %186, %67, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc81)
%188 = torch.aten.view %187, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc82)
%189 = torch.aten.mm %188, %65 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,3072],f32> -> !torch.vtensor<[128,3072],f32> loc(#loc83)
%190 = torch.aten.mul.Scalar %66, %int1 : !torch.vtensor<[3072],f32>, !torch.int -> !torch.vtensor<[3072],f32> loc(#loc83)
%191 = torch.aten.add.Tensor %190, %189, %int1 : !torch.vtensor<[3072],f32>, !torch.vtensor<[128,3072],f32>, !torch.int -> !torch.vtensor<[128,3072],f32> loc(#loc83)
%192 = torch.prim.ListConstruct %int1, %int128, %int3072 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc402)
%193 = torch.aten.view %191, %192 : !torch.vtensor<[128,3072],f32>, !torch.list<int> -> !torch.vtensor<[1,128,3072],f32> loc(#loc84)
%194 = torch.aten.mul.Scalar %193, %float5.000000e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc85)
%195 = torch.aten.pow.Tensor_Scalar %193, %float3.000000e00 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc86)
%196 = torch.aten.mul.Scalar %195, %float4.471500e-02 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc87)
%197 = torch.aten.add.Tensor %193, %196, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32>, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc88)
%198 = torch.aten.mul.Scalar %197, %float7.978850e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc89)
%199 = torch.aten.tanh %198 : !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc90)
%200 = torch.aten.add.Scalar %199, %float1.000000e00, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc91)
%201 = torch.aten.mul.Tensor %194, %200 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc92)
%202 = torch.prim.ListConstruct %int-1, %int3072 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc)
%203 = torch.aten.view %201, %202 : !torch.vtensor<[1,128,3072],f32>, !torch.list<int> -> !torch.vtensor<[128,3072],f32> loc(#loc93)
%204 = torch.aten.mm %203, %63 : !torch.vtensor<[128,3072],f32>, !torch.vtensor<[3072,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc94)
%205 = torch.aten.mul.Scalar %64, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc94)
%206 = torch.aten.add.Tensor %205, %204, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc94)
%207 = torch.aten.view %206, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc95)
%208 = torch.aten.add.Tensor %171, %207, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc96)
%209 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc97)
%210 = torch.aten.sum.dim_IntList %208, %209, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc97)
%211 = torch.aten.div.Scalar %210, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc97)
%212 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc97)
%213 = torch.aten.broadcast_to %211, %212 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%214 = torch.aten.sub.Tensor %208, %213, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%215 = torch.aten.mul.Tensor %214, %214 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%216 = torch.aten.sum.dim_IntList %215, %209, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc97)
%217 = torch.aten.div.Scalar %216, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc97)
%218 = torch.aten.add.Scalar %217, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc97)
%219 = torch.aten.rsqrt %218 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc97)
%220 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc97)
%221 = torch.aten.broadcast_to %219, %220 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%222 = torch.aten.mul.Tensor %214, %221 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%223 = torch.aten.mul.Tensor %222, %62 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%224 = torch.aten.add.Tensor %223, %61, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc97)
%225 = torch.aten.view %224, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc98)
%226 = torch.aten.mm %225, %59 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2304],f32> -> !torch.vtensor<[128,2304],f32> loc(#loc99)
%227 = torch.aten.mul.Scalar %60, %int1 : !torch.vtensor<[2304],f32>, !torch.int -> !torch.vtensor<[2304],f32> loc(#loc99)
%228 = torch.aten.add.Tensor %227, %226, %int1 : !torch.vtensor<[2304],f32>, !torch.vtensor<[128,2304],f32>, !torch.int -> !torch.vtensor<[128,2304],f32> loc(#loc99)
%229 = torch.aten.view %228, %107 : !torch.vtensor<[128,2304],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2304],f32> loc(#loc100)
%230 = torch.aten.slice.Tensor %229, %int2, %int0, %int768, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc101)
%231 = torch.aten.slice.Tensor %229, %int2, %int768, %int1536, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc102)
%232 = torch.aten.slice.Tensor %229, %int2, %int1536, %int2304, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc103)
%233 = torch.aten.view %230, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc104)
%234 = torch.aten.permute %233, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc105)
%235 = torch.aten.view %231, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc106)
%236 = torch.aten.permute %235, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc107)
%237 = torch.aten.view %232, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc108)
%238 = torch.aten.permute %237, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc109)
%239 = torch.aten.transpose.int %236, %int-1, %int-2 : !torch.vtensor<[1,12,128,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,12,64,128],f32> loc(#loc110)
%240 = torch.aten.broadcast_to %234, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc111)
%241 = torch.aten.view %240, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc112)
%242 = torch.aten.broadcast_to %239, %125 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,64,128],f32> loc(#loc113)
%243 = torch.aten.view %242, %127 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[12,64,128],f32> loc(#loc114)
%244 = torch.aten.bmm %241, %243 : !torch.vtensor<[12,128,64],f32>, !torch.vtensor<[12,64,128],f32> -> !torch.vtensor<[12,128,128],f32> loc(#loc115)
%245 = torch.aten.view %244, %130 : !torch.vtensor<[12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc116)
%246 = torch.prim.NumToTensor.Scalar %float8.000000e00 : !torch.float -> !torch.vtensor<[],f64> loc(#loc117)
%247 = torch.aten.to.dtype %246, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc117)
%248 = torch.aten.broadcast_to %247, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc117)
%249 = torch.aten.div.Tensor %245, %248 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc118)
%250 = torch.aten.slice.Tensor %71, %int2, %int0, %int128, %int1 : !torch.vtensor<[1,1,1024,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,1024],ui8> loc(#loc119)
%251 = torch.aten.slice.Tensor %250, %int3, %int0, %int128, %int1 : !torch.vtensor<[1,1,128,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,128],ui8> loc(#loc120)
%252 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> loc(#loc121)
%253 = torch.aten.to.dtype %252, %int11, %false, %false, %none : !torch.vtensor<[],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],i1> loc(#loc121)
%254 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc121)
%255 = torch.aten.broadcast_to %253, %254 : !torch.vtensor<[],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc121)
%256 = torch.aten.to.dtype %251, %int11, %false, %false, %none : !torch.vtensor<[1,1,128,128],ui8>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,128,128],i1> loc(#loc121)
%257 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc121)
%258 = torch.aten.broadcast_to %256, %257 : !torch.vtensor<[1,1,128,128],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc121)
%259 = torch.prim.NumToTensor.Scalar %float-3.402820e38 : !torch.float -> !torch.vtensor<[],f64> loc(#loc122)
%260 = torch.aten.to.dtype %259, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc122)
%261 = torch.aten.broadcast_to %260, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc122)
%262 = torch.aten.where.self %258, %249, %261 : !torch.vtensor<[1,1,128,128],i1>, !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc123)
%values_0, %indices_1 = torch.aten.max.dim %262, %int-1, %true : !torch.vtensor<[1,12,128,128],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,12,128,1],f32>, !torch.vtensor<[1,12,128,1],si64> loc(#loc124)
%263 = torch.aten.sub.Tensor %262, %values_0, %float1.000000e00 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32>, !torch.float -> !torch.vtensor<[1,12,128,128],f32> loc(#loc124)
%264 = torch.aten.exp %263 : !torch.vtensor<[1,12,128,128],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc124)
%265 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc124)
%266 = torch.aten.sum.dim_IntList %264, %265, %true, %none : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,12,128,1],f32> loc(#loc124)
%267 = torch.aten.div.Tensor %264, %266 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc124)
%268 = torch.aten.broadcast_to %267, %130 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc125)
%269 = torch.aten.view %268, %156 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[12,128,128],f32> loc(#loc126)
%270 = torch.aten.broadcast_to %238, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc127)
%271 = torch.aten.view %270, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc128)
%272 = torch.aten.bmm %269, %271 : !torch.vtensor<[12,128,128],f32>, !torch.vtensor<[12,128,64],f32> -> !torch.vtensor<[12,128,64],f32> loc(#loc129)
%273 = torch.aten.view %272, %121 : !torch.vtensor<[12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc130)
%274 = torch.aten.permute %273, %114 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc131)
%275 = torch.aten.clone %274, %int0 : !torch.vtensor<[1,128,12,64],f32>, !torch.int -> !torch.vtensor<[1,128,12,64],f32> loc(#loc132)
%276 = torch.aten.view %275, %164 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc133)
%277 = torch.aten.view %276, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc134)
%278 = torch.aten.mm %277, %57 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc135)
%279 = torch.aten.mul.Scalar %58, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc135)
%280 = torch.aten.add.Tensor %279, %278, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc135)
%281 = torch.aten.view %280, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc136)
%282 = torch.aten.add.Tensor %281, %208, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc137)
%283 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc138)
%284 = torch.aten.sum.dim_IntList %282, %283, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc138)
%285 = torch.aten.div.Scalar %284, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc138)
%286 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc138)
%287 = torch.aten.broadcast_to %285, %286 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%288 = torch.aten.sub.Tensor %282, %287, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%289 = torch.aten.mul.Tensor %288, %288 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%290 = torch.aten.sum.dim_IntList %289, %283, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc138)
%291 = torch.aten.div.Scalar %290, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc138)
%292 = torch.aten.add.Scalar %291, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc138)
%293 = torch.aten.rsqrt %292 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc138)
%294 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc138)
%295 = torch.aten.broadcast_to %293, %294 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%296 = torch.aten.mul.Tensor %288, %295 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%297 = torch.aten.mul.Tensor %296, %56 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%298 = torch.aten.add.Tensor %297, %55, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc138)
%299 = torch.aten.view %298, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc139)
%300 = torch.aten.mm %299, %53 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,3072],f32> -> !torch.vtensor<[128,3072],f32> loc(#loc140)
%301 = torch.aten.mul.Scalar %54, %int1 : !torch.vtensor<[3072],f32>, !torch.int -> !torch.vtensor<[3072],f32> loc(#loc140)
%302 = torch.aten.add.Tensor %301, %300, %int1 : !torch.vtensor<[3072],f32>, !torch.vtensor<[128,3072],f32>, !torch.int -> !torch.vtensor<[128,3072],f32> loc(#loc140)
%303 = torch.aten.view %302, %192 : !torch.vtensor<[128,3072],f32>, !torch.list<int> -> !torch.vtensor<[1,128,3072],f32> loc(#loc141)
%304 = torch.aten.mul.Scalar %303, %float5.000000e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc142)
%305 = torch.aten.pow.Tensor_Scalar %303, %float3.000000e00 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc143)
%306 = torch.aten.mul.Scalar %305, %float4.471500e-02 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc144)
%307 = torch.aten.add.Tensor %303, %306, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32>, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc145)
%308 = torch.aten.mul.Scalar %307, %float7.978850e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc146)
%309 = torch.aten.tanh %308 : !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc147)
%310 = torch.aten.add.Scalar %309, %float1.000000e00, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc148)
%311 = torch.aten.mul.Tensor %304, %310 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc149)
%312 = torch.aten.view %311, %202 : !torch.vtensor<[1,128,3072],f32>, !torch.list<int> -> !torch.vtensor<[128,3072],f32> loc(#loc150)
%313 = torch.aten.mm %312, %51 : !torch.vtensor<[128,3072],f32>, !torch.vtensor<[3072,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc151)
%314 = torch.aten.mul.Scalar %52, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc151)
%315 = torch.aten.add.Tensor %314, %313, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc151)
%316 = torch.aten.view %315, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc152)
%317 = torch.aten.add.Tensor %282, %316, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc153)
%318 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc154)
%319 = torch.aten.sum.dim_IntList %317, %318, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc154)
%320 = torch.aten.div.Scalar %319, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc154)
%321 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc154)
%322 = torch.aten.broadcast_to %320, %321 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%323 = torch.aten.sub.Tensor %317, %322, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%324 = torch.aten.mul.Tensor %323, %323 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%325 = torch.aten.sum.dim_IntList %324, %318, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc154)
%326 = torch.aten.div.Scalar %325, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc154)
%327 = torch.aten.add.Scalar %326, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc154)
%328 = torch.aten.rsqrt %327 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc154)
%329 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc154)
%330 = torch.aten.broadcast_to %328, %329 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%331 = torch.aten.mul.Tensor %323, %330 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%332 = torch.aten.mul.Tensor %331, %50 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%333 = torch.aten.add.Tensor %332, %49, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc154)
%334 = torch.aten.view %333, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc155)
%335 = torch.aten.mm %334, %47 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2304],f32> -> !torch.vtensor<[128,2304],f32> loc(#loc156)
%336 = torch.aten.mul.Scalar %48, %int1 : !torch.vtensor<[2304],f32>, !torch.int -> !torch.vtensor<[2304],f32> loc(#loc156)
%337 = torch.aten.add.Tensor %336, %335, %int1 : !torch.vtensor<[2304],f32>, !torch.vtensor<[128,2304],f32>, !torch.int -> !torch.vtensor<[128,2304],f32> loc(#loc156)
%338 = torch.aten.view %337, %107 : !torch.vtensor<[128,2304],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2304],f32> loc(#loc157)
%339 = torch.aten.slice.Tensor %338, %int2, %int0, %int768, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc158)
%340 = torch.aten.slice.Tensor %338, %int2, %int768, %int1536, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc159)
%341 = torch.aten.slice.Tensor %338, %int2, %int1536, %int2304, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc160)
%342 = torch.aten.view %339, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc161)
%343 = torch.aten.permute %342, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc162)
%344 = torch.aten.view %340, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc163)
%345 = torch.aten.permute %344, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc164)
%346 = torch.aten.view %341, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc165)
%347 = torch.aten.permute %346, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc166)
%348 = torch.aten.transpose.int %345, %int-1, %int-2 : !torch.vtensor<[1,12,128,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,12,64,128],f32> loc(#loc167)
%349 = torch.aten.broadcast_to %343, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc168)
%350 = torch.aten.view %349, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc169)
%351 = torch.aten.broadcast_to %348, %125 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,64,128],f32> loc(#loc170)
%352 = torch.aten.view %351, %127 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[12,64,128],f32> loc(#loc171)
%353 = torch.aten.bmm %350, %352 : !torch.vtensor<[12,128,64],f32>, !torch.vtensor<[12,64,128],f32> -> !torch.vtensor<[12,128,128],f32> loc(#loc172)
%354 = torch.aten.view %353, %130 : !torch.vtensor<[12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc173)
%355 = torch.prim.NumToTensor.Scalar %float8.000000e00 : !torch.float -> !torch.vtensor<[],f64> loc(#loc174)
%356 = torch.aten.to.dtype %355, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc174)
%357 = torch.aten.broadcast_to %356, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc174)
%358 = torch.aten.div.Tensor %354, %357 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc175)
%359 = torch.aten.slice.Tensor %71, %int2, %int0, %int128, %int1 : !torch.vtensor<[1,1,1024,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,1024],ui8> loc(#loc176)
%360 = torch.aten.slice.Tensor %359, %int3, %int0, %int128, %int1 : !torch.vtensor<[1,1,128,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,128],ui8> loc(#loc177)
%361 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> loc(#loc178)
%362 = torch.aten.to.dtype %361, %int11, %false, %false, %none : !torch.vtensor<[],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],i1> loc(#loc178)
%363 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc178)
%364 = torch.aten.broadcast_to %362, %363 : !torch.vtensor<[],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc178)
%365 = torch.aten.to.dtype %360, %int11, %false, %false, %none : !torch.vtensor<[1,1,128,128],ui8>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,128,128],i1> loc(#loc178)
%366 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc178)
%367 = torch.aten.broadcast_to %365, %366 : !torch.vtensor<[1,1,128,128],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc178)
%368 = torch.prim.NumToTensor.Scalar %float-3.402820e38 : !torch.float -> !torch.vtensor<[],f64> loc(#loc179)
%369 = torch.aten.to.dtype %368, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc179)
%370 = torch.aten.broadcast_to %369, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc179)
%371 = torch.aten.where.self %367, %358, %370 : !torch.vtensor<[1,1,128,128],i1>, !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc180)
%values_2, %indices_3 = torch.aten.max.dim %371, %int-1, %true : !torch.vtensor<[1,12,128,128],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,12,128,1],f32>, !torch.vtensor<[1,12,128,1],si64> loc(#loc181)
%372 = torch.aten.sub.Tensor %371, %values_2, %float1.000000e00 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32>, !torch.float -> !torch.vtensor<[1,12,128,128],f32> loc(#loc181)
%373 = torch.aten.exp %372 : !torch.vtensor<[1,12,128,128],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc181)
%374 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc181)
%375 = torch.aten.sum.dim_IntList %373, %374, %true, %none : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,12,128,1],f32> loc(#loc181)
%376 = torch.aten.div.Tensor %373, %375 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc181)
%377 = torch.aten.broadcast_to %376, %130 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc182)
%378 = torch.aten.view %377, %156 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[12,128,128],f32> loc(#loc183)
%379 = torch.aten.broadcast_to %347, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc184)
%380 = torch.aten.view %379, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc185)
%381 = torch.aten.bmm %378, %380 : !torch.vtensor<[12,128,128],f32>, !torch.vtensor<[12,128,64],f32> -> !torch.vtensor<[12,128,64],f32> loc(#loc186)
%382 = torch.aten.view %381, %121 : !torch.vtensor<[12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc187)
%383 = torch.aten.permute %382, %114 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc188)
%384 = torch.aten.clone %383, %int0 : !torch.vtensor<[1,128,12,64],f32>, !torch.int -> !torch.vtensor<[1,128,12,64],f32> loc(#loc189)
%385 = torch.aten.view %384, %164 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc190)
%386 = torch.aten.view %385, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc191)
%387 = torch.aten.mm %386, %45 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc192)
%388 = torch.aten.mul.Scalar %46, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc192)
%389 = torch.aten.add.Tensor %388, %387, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc192)
%390 = torch.aten.view %389, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc193)
%391 = torch.aten.add.Tensor %390, %317, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc194)
%392 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc195)
%393 = torch.aten.sum.dim_IntList %391, %392, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc195)
%394 = torch.aten.div.Scalar %393, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc195)
%395 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc195)
%396 = torch.aten.broadcast_to %394, %395 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%397 = torch.aten.sub.Tensor %391, %396, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%398 = torch.aten.mul.Tensor %397, %397 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%399 = torch.aten.sum.dim_IntList %398, %392, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc195)
%400 = torch.aten.div.Scalar %399, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc195)
%401 = torch.aten.add.Scalar %400, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc195)
%402 = torch.aten.rsqrt %401 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc195)
%403 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc195)
%404 = torch.aten.broadcast_to %402, %403 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%405 = torch.aten.mul.Tensor %397, %404 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%406 = torch.aten.mul.Tensor %405, %44 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%407 = torch.aten.add.Tensor %406, %43, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc195)
%408 = torch.aten.view %407, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc196)
%409 = torch.aten.mm %408, %41 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,3072],f32> -> !torch.vtensor<[128,3072],f32> loc(#loc197)
%410 = torch.aten.mul.Scalar %42, %int1 : !torch.vtensor<[3072],f32>, !torch.int -> !torch.vtensor<[3072],f32> loc(#loc197)
%411 = torch.aten.add.Tensor %410, %409, %int1 : !torch.vtensor<[3072],f32>, !torch.vtensor<[128,3072],f32>, !torch.int -> !torch.vtensor<[128,3072],f32> loc(#loc197)
%412 = torch.aten.view %411, %192 : !torch.vtensor<[128,3072],f32>, !torch.list<int> -> !torch.vtensor<[1,128,3072],f32> loc(#loc198)
%413 = torch.aten.mul.Scalar %412, %float5.000000e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc199)
%414 = torch.aten.pow.Tensor_Scalar %412, %float3.000000e00 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc200)
%415 = torch.aten.mul.Scalar %414, %float4.471500e-02 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc201)
%416 = torch.aten.add.Tensor %412, %415, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32>, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc202)
%417 = torch.aten.mul.Scalar %416, %float7.978850e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc203)
%418 = torch.aten.tanh %417 : !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc204)
%419 = torch.aten.add.Scalar %418, %float1.000000e00, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc205)
%420 = torch.aten.mul.Tensor %413, %419 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc206)
%421 = torch.aten.view %420, %202 : !torch.vtensor<[1,128,3072],f32>, !torch.list<int> -> !torch.vtensor<[128,3072],f32> loc(#loc207)
%422 = torch.aten.mm %421, %39 : !torch.vtensor<[128,3072],f32>, !torch.vtensor<[3072,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc208)
%423 = torch.aten.mul.Scalar %40, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc208)
%424 = torch.aten.add.Tensor %423, %422, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc208)
%425 = torch.aten.view %424, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc209)
%426 = torch.aten.add.Tensor %391, %425, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc210)
%427 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc211)
%428 = torch.aten.sum.dim_IntList %426, %427, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc211)
%429 = torch.aten.div.Scalar %428, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc211)
%430 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc211)
%431 = torch.aten.broadcast_to %429, %430 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%432 = torch.aten.sub.Tensor %426, %431, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%433 = torch.aten.mul.Tensor %432, %432 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%434 = torch.aten.sum.dim_IntList %433, %427, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc211)
%435 = torch.aten.div.Scalar %434, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc211)
%436 = torch.aten.add.Scalar %435, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc211)
%437 = torch.aten.rsqrt %436 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc211)
%438 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc211)
%439 = torch.aten.broadcast_to %437, %438 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%440 = torch.aten.mul.Tensor %432, %439 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%441 = torch.aten.mul.Tensor %440, %38 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%442 = torch.aten.add.Tensor %441, %37, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc211)
%443 = torch.aten.view %442, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc212)
%444 = torch.aten.mm %443, %35 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2304],f32> -> !torch.vtensor<[128,2304],f32> loc(#loc213)
%445 = torch.aten.mul.Scalar %36, %int1 : !torch.vtensor<[2304],f32>, !torch.int -> !torch.vtensor<[2304],f32> loc(#loc213)
%446 = torch.aten.add.Tensor %445, %444, %int1 : !torch.vtensor<[2304],f32>, !torch.vtensor<[128,2304],f32>, !torch.int -> !torch.vtensor<[128,2304],f32> loc(#loc213)
%447 = torch.aten.view %446, %107 : !torch.vtensor<[128,2304],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2304],f32> loc(#loc214)
%448 = torch.aten.slice.Tensor %447, %int2, %int0, %int768, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc215)
%449 = torch.aten.slice.Tensor %447, %int2, %int768, %int1536, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc216)
%450 = torch.aten.slice.Tensor %447, %int2, %int1536, %int2304, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc217)
%451 = torch.aten.view %448, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc218)
%452 = torch.aten.permute %451, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc219)
%453 = torch.aten.view %449, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc220)
%454 = torch.aten.permute %453, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc221)
%455 = torch.aten.view %450, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc222)
%456 = torch.aten.permute %455, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc223)
%457 = torch.aten.transpose.int %454, %int-1, %int-2 : !torch.vtensor<[1,12,128,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,12,64,128],f32> loc(#loc224)
%458 = torch.aten.broadcast_to %452, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc225)
%459 = torch.aten.view %458, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc226)
%460 = torch.aten.broadcast_to %457, %125 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,64,128],f32> loc(#loc227)
%461 = torch.aten.view %460, %127 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[12,64,128],f32> loc(#loc228)
%462 = torch.aten.bmm %459, %461 : !torch.vtensor<[12,128,64],f32>, !torch.vtensor<[12,64,128],f32> -> !torch.vtensor<[12,128,128],f32> loc(#loc229)
%463 = torch.aten.view %462, %130 : !torch.vtensor<[12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc230)
%464 = torch.prim.NumToTensor.Scalar %float8.000000e00 : !torch.float -> !torch.vtensor<[],f64> loc(#loc231)
%465 = torch.aten.to.dtype %464, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc231)
%466 = torch.aten.broadcast_to %465, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc231)
%467 = torch.aten.div.Tensor %463, %466 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc232)
%468 = torch.aten.slice.Tensor %71, %int2, %int0, %int128, %int1 : !torch.vtensor<[1,1,1024,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,1024],ui8> loc(#loc233)
%469 = torch.aten.slice.Tensor %468, %int3, %int0, %int128, %int1 : !torch.vtensor<[1,1,128,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,128],ui8> loc(#loc234)
%470 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> loc(#loc235)
%471 = torch.aten.to.dtype %470, %int11, %false, %false, %none : !torch.vtensor<[],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],i1> loc(#loc235)
%472 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc235)
%473 = torch.aten.broadcast_to %471, %472 : !torch.vtensor<[],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc235)
%474 = torch.aten.to.dtype %469, %int11, %false, %false, %none : !torch.vtensor<[1,1,128,128],ui8>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,128,128],i1> loc(#loc235)
%475 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc235)
%476 = torch.aten.broadcast_to %474, %475 : !torch.vtensor<[1,1,128,128],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc235)
%477 = torch.prim.NumToTensor.Scalar %float-3.402820e38 : !torch.float -> !torch.vtensor<[],f64> loc(#loc236)
%478 = torch.aten.to.dtype %477, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc236)
%479 = torch.aten.broadcast_to %478, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc236)
%480 = torch.aten.where.self %476, %467, %479 : !torch.vtensor<[1,1,128,128],i1>, !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc237)
%values_4, %indices_5 = torch.aten.max.dim %480, %int-1, %true : !torch.vtensor<[1,12,128,128],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,12,128,1],f32>, !torch.vtensor<[1,12,128,1],si64> loc(#loc238)
%481 = torch.aten.sub.Tensor %480, %values_4, %float1.000000e00 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32>, !torch.float -> !torch.vtensor<[1,12,128,128],f32> loc(#loc238)
%482 = torch.aten.exp %481 : !torch.vtensor<[1,12,128,128],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc238)
%483 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc238)
%484 = torch.aten.sum.dim_IntList %482, %483, %true, %none : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,12,128,1],f32> loc(#loc238)
%485 = torch.aten.div.Tensor %482, %484 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc238)
%486 = torch.aten.broadcast_to %485, %130 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc239)
%487 = torch.aten.view %486, %156 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[12,128,128],f32> loc(#loc240)
%488 = torch.aten.broadcast_to %456, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc241)
%489 = torch.aten.view %488, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc242)
%490 = torch.aten.bmm %487, %489 : !torch.vtensor<[12,128,128],f32>, !torch.vtensor<[12,128,64],f32> -> !torch.vtensor<[12,128,64],f32> loc(#loc243)
%491 = torch.aten.view %490, %121 : !torch.vtensor<[12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc244)
%492 = torch.aten.permute %491, %114 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc245)
%493 = torch.aten.clone %492, %int0 : !torch.vtensor<[1,128,12,64],f32>, !torch.int -> !torch.vtensor<[1,128,12,64],f32> loc(#loc246)
%494 = torch.aten.view %493, %164 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc247)
%495 = torch.aten.view %494, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc248)
%496 = torch.aten.mm %495, %33 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc249)
%497 = torch.aten.mul.Scalar %34, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc249)
%498 = torch.aten.add.Tensor %497, %496, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc249)
%499 = torch.aten.view %498, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc250)
%500 = torch.aten.add.Tensor %499, %426, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc251)
%501 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc252)
%502 = torch.aten.sum.dim_IntList %500, %501, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc252)
%503 = torch.aten.div.Scalar %502, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc252)
%504 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc252)
%505 = torch.aten.broadcast_to %503, %504 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%506 = torch.aten.sub.Tensor %500, %505, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%507 = torch.aten.mul.Tensor %506, %506 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%508 = torch.aten.sum.dim_IntList %507, %501, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc252)
%509 = torch.aten.div.Scalar %508, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc252)
%510 = torch.aten.add.Scalar %509, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc252)
%511 = torch.aten.rsqrt %510 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc252)
%512 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc252)
%513 = torch.aten.broadcast_to %511, %512 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%514 = torch.aten.mul.Tensor %506, %513 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%515 = torch.aten.mul.Tensor %514, %32 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%516 = torch.aten.add.Tensor %515, %31, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc252)
%517 = torch.aten.view %516, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc253)
%518 = torch.aten.mm %517, %29 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,3072],f32> -> !torch.vtensor<[128,3072],f32> loc(#loc254)
%519 = torch.aten.mul.Scalar %30, %int1 : !torch.vtensor<[3072],f32>, !torch.int -> !torch.vtensor<[3072],f32> loc(#loc254)
%520 = torch.aten.add.Tensor %519, %518, %int1 : !torch.vtensor<[3072],f32>, !torch.vtensor<[128,3072],f32>, !torch.int -> !torch.vtensor<[128,3072],f32> loc(#loc254)
%521 = torch.aten.view %520, %192 : !torch.vtensor<[128,3072],f32>, !torch.list<int> -> !torch.vtensor<[1,128,3072],f32> loc(#loc255)
%522 = torch.aten.mul.Scalar %521, %float5.000000e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc256)
%523 = torch.aten.pow.Tensor_Scalar %521, %float3.000000e00 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc257)
%524 = torch.aten.mul.Scalar %523, %float4.471500e-02 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc258)
%525 = torch.aten.add.Tensor %521, %524, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32>, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc259)
%526 = torch.aten.mul.Scalar %525, %float7.978850e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc260)
%527 = torch.aten.tanh %526 : !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc261)
%528 = torch.aten.add.Scalar %527, %float1.000000e00, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc262)
%529 = torch.aten.mul.Tensor %522, %528 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc263)
%530 = torch.aten.view %529, %202 : !torch.vtensor<[1,128,3072],f32>, !torch.list<int> -> !torch.vtensor<[128,3072],f32> loc(#loc264)
%531 = torch.aten.mm %530, %27 : !torch.vtensor<[128,3072],f32>, !torch.vtensor<[3072,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc265)
%532 = torch.aten.mul.Scalar %28, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc265)
%533 = torch.aten.add.Tensor %532, %531, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc265)
%534 = torch.aten.view %533, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc266)
%535 = torch.aten.add.Tensor %500, %534, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc267)
%536 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc268)
%537 = torch.aten.sum.dim_IntList %535, %536, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc268)
%538 = torch.aten.div.Scalar %537, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc268)
%539 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc268)
%540 = torch.aten.broadcast_to %538, %539 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%541 = torch.aten.sub.Tensor %535, %540, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%542 = torch.aten.mul.Tensor %541, %541 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%543 = torch.aten.sum.dim_IntList %542, %536, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc268)
%544 = torch.aten.div.Scalar %543, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc268)
%545 = torch.aten.add.Scalar %544, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc268)
%546 = torch.aten.rsqrt %545 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc268)
%547 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc268)
%548 = torch.aten.broadcast_to %546, %547 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%549 = torch.aten.mul.Tensor %541, %548 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%550 = torch.aten.mul.Tensor %549, %26 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%551 = torch.aten.add.Tensor %550, %25, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc268)
%552 = torch.aten.view %551, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc269)
%553 = torch.aten.mm %552, %23 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2304],f32> -> !torch.vtensor<[128,2304],f32> loc(#loc270)
%554 = torch.aten.mul.Scalar %24, %int1 : !torch.vtensor<[2304],f32>, !torch.int -> !torch.vtensor<[2304],f32> loc(#loc270)
%555 = torch.aten.add.Tensor %554, %553, %int1 : !torch.vtensor<[2304],f32>, !torch.vtensor<[128,2304],f32>, !torch.int -> !torch.vtensor<[128,2304],f32> loc(#loc270)
%556 = torch.aten.view %555, %107 : !torch.vtensor<[128,2304],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2304],f32> loc(#loc271)
%557 = torch.aten.slice.Tensor %556, %int2, %int0, %int768, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc272)
%558 = torch.aten.slice.Tensor %556, %int2, %int768, %int1536, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc273)
%559 = torch.aten.slice.Tensor %556, %int2, %int1536, %int2304, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc274)
%560 = torch.aten.view %557, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc275)
%561 = torch.aten.permute %560, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc276)
%562 = torch.aten.view %558, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc277)
%563 = torch.aten.permute %562, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc278)
%564 = torch.aten.view %559, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc279)
%565 = torch.aten.permute %564, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc280)
%566 = torch.aten.transpose.int %563, %int-1, %int-2 : !torch.vtensor<[1,12,128,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,12,64,128],f32> loc(#loc281)
%567 = torch.aten.broadcast_to %561, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc282)
%568 = torch.aten.view %567, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc283)
%569 = torch.aten.broadcast_to %566, %125 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,64,128],f32> loc(#loc284)
%570 = torch.aten.view %569, %127 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[12,64,128],f32> loc(#loc285)
%571 = torch.aten.bmm %568, %570 : !torch.vtensor<[12,128,64],f32>, !torch.vtensor<[12,64,128],f32> -> !torch.vtensor<[12,128,128],f32> loc(#loc286)
%572 = torch.aten.view %571, %130 : !torch.vtensor<[12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc287)
%573 = torch.prim.NumToTensor.Scalar %float8.000000e00 : !torch.float -> !torch.vtensor<[],f64> loc(#loc288)
%574 = torch.aten.to.dtype %573, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc288)
%575 = torch.aten.broadcast_to %574, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc288)
%576 = torch.aten.div.Tensor %572, %575 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc289)
%577 = torch.aten.slice.Tensor %71, %int2, %int0, %int128, %int1 : !torch.vtensor<[1,1,1024,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,1024],ui8> loc(#loc290)
%578 = torch.aten.slice.Tensor %577, %int3, %int0, %int128, %int1 : !torch.vtensor<[1,1,128,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,128],ui8> loc(#loc291)
%579 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> loc(#loc292)
%580 = torch.aten.to.dtype %579, %int11, %false, %false, %none : !torch.vtensor<[],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],i1> loc(#loc292)
%581 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc292)
%582 = torch.aten.broadcast_to %580, %581 : !torch.vtensor<[],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc292)
%583 = torch.aten.to.dtype %578, %int11, %false, %false, %none : !torch.vtensor<[1,1,128,128],ui8>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,128,128],i1> loc(#loc292)
%584 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc292)
%585 = torch.aten.broadcast_to %583, %584 : !torch.vtensor<[1,1,128,128],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc292)
%586 = torch.prim.NumToTensor.Scalar %float-3.402820e38 : !torch.float -> !torch.vtensor<[],f64> loc(#loc293)
%587 = torch.aten.to.dtype %586, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc293)
%588 = torch.aten.broadcast_to %587, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc293)
%589 = torch.aten.where.self %585, %576, %588 : !torch.vtensor<[1,1,128,128],i1>, !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc294)
%values_6, %indices_7 = torch.aten.max.dim %589, %int-1, %true : !torch.vtensor<[1,12,128,128],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,12,128,1],f32>, !torch.vtensor<[1,12,128,1],si64> loc(#loc295)
%590 = torch.aten.sub.Tensor %589, %values_6, %float1.000000e00 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32>, !torch.float -> !torch.vtensor<[1,12,128,128],f32> loc(#loc295)
%591 = torch.aten.exp %590 : !torch.vtensor<[1,12,128,128],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc295)
%592 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc295)
%593 = torch.aten.sum.dim_IntList %591, %592, %true, %none : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,12,128,1],f32> loc(#loc295)
%594 = torch.aten.div.Tensor %591, %593 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc295)
%595 = torch.aten.broadcast_to %594, %130 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc296)
%596 = torch.aten.view %595, %156 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[12,128,128],f32> loc(#loc297)
%597 = torch.aten.broadcast_to %565, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc298)
%598 = torch.aten.view %597, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc299)
%599 = torch.aten.bmm %596, %598 : !torch.vtensor<[12,128,128],f32>, !torch.vtensor<[12,128,64],f32> -> !torch.vtensor<[12,128,64],f32> loc(#loc300)
%600 = torch.aten.view %599, %121 : !torch.vtensor<[12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc301)
%601 = torch.aten.permute %600, %114 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc302)
%602 = torch.aten.clone %601, %int0 : !torch.vtensor<[1,128,12,64],f32>, !torch.int -> !torch.vtensor<[1,128,12,64],f32> loc(#loc303)
%603 = torch.aten.view %602, %164 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc304)
%604 = torch.aten.view %603, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc305)
%605 = torch.aten.mm %604, %21 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc306)
%606 = torch.aten.mul.Scalar %22, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc306)
%607 = torch.aten.add.Tensor %606, %605, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc306)
%608 = torch.aten.view %607, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc307)
%609 = torch.aten.add.Tensor %608, %535, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc308)
%610 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc309)
%611 = torch.aten.sum.dim_IntList %609, %610, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc309)
%612 = torch.aten.div.Scalar %611, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc309)
%613 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc309)
%614 = torch.aten.broadcast_to %612, %613 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%615 = torch.aten.sub.Tensor %609, %614, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%616 = torch.aten.mul.Tensor %615, %615 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%617 = torch.aten.sum.dim_IntList %616, %610, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc309)
%618 = torch.aten.div.Scalar %617, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc309)
%619 = torch.aten.add.Scalar %618, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc309)
%620 = torch.aten.rsqrt %619 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc309)
%621 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc309)
%622 = torch.aten.broadcast_to %620, %621 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%623 = torch.aten.mul.Tensor %615, %622 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%624 = torch.aten.mul.Tensor %623, %20 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%625 = torch.aten.add.Tensor %624, %19, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc309)
%626 = torch.aten.view %625, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc310)
%627 = torch.aten.mm %626, %17 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,3072],f32> -> !torch.vtensor<[128,3072],f32> loc(#loc311)
%628 = torch.aten.mul.Scalar %18, %int1 : !torch.vtensor<[3072],f32>, !torch.int -> !torch.vtensor<[3072],f32> loc(#loc311)
%629 = torch.aten.add.Tensor %628, %627, %int1 : !torch.vtensor<[3072],f32>, !torch.vtensor<[128,3072],f32>, !torch.int -> !torch.vtensor<[128,3072],f32> loc(#loc311)
%630 = torch.aten.view %629, %192 : !torch.vtensor<[128,3072],f32>, !torch.list<int> -> !torch.vtensor<[1,128,3072],f32> loc(#loc312)
%631 = torch.aten.mul.Scalar %630, %float5.000000e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc313)
%632 = torch.aten.pow.Tensor_Scalar %630, %float3.000000e00 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc314)
%633 = torch.aten.mul.Scalar %632, %float4.471500e-02 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc315)
%634 = torch.aten.add.Tensor %630, %633, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32>, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc316)
%635 = torch.aten.mul.Scalar %634, %float7.978850e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc317)
%636 = torch.aten.tanh %635 : !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc318)
%637 = torch.aten.add.Scalar %636, %float1.000000e00, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc319)
%638 = torch.aten.mul.Tensor %631, %637 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc320)
%639 = torch.aten.view %638, %202 : !torch.vtensor<[1,128,3072],f32>, !torch.list<int> -> !torch.vtensor<[128,3072],f32> loc(#loc321)
%640 = torch.aten.mm %639, %15 : !torch.vtensor<[128,3072],f32>, !torch.vtensor<[3072,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc322)
%641 = torch.aten.mul.Scalar %16, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc322)
%642 = torch.aten.add.Tensor %641, %640, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc322)
%643 = torch.aten.view %642, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc323)
%644 = torch.aten.add.Tensor %609, %643, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc324)
%645 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc325)
%646 = torch.aten.sum.dim_IntList %644, %645, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc325)
%647 = torch.aten.div.Scalar %646, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc325)
%648 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc325)
%649 = torch.aten.broadcast_to %647, %648 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%650 = torch.aten.sub.Tensor %644, %649, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%651 = torch.aten.mul.Tensor %650, %650 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%652 = torch.aten.sum.dim_IntList %651, %645, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc325)
%653 = torch.aten.div.Scalar %652, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc325)
%654 = torch.aten.add.Scalar %653, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc325)
%655 = torch.aten.rsqrt %654 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc325)
%656 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc325)
%657 = torch.aten.broadcast_to %655, %656 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%658 = torch.aten.mul.Tensor %650, %657 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%659 = torch.aten.mul.Tensor %658, %14 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%660 = torch.aten.add.Tensor %659, %13, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc325)
%661 = torch.aten.view %660, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc326)
%662 = torch.aten.mm %661, %11 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2304],f32> -> !torch.vtensor<[128,2304],f32> loc(#loc327)
%663 = torch.aten.mul.Scalar %12, %int1 : !torch.vtensor<[2304],f32>, !torch.int -> !torch.vtensor<[2304],f32> loc(#loc327)
%664 = torch.aten.add.Tensor %663, %662, %int1 : !torch.vtensor<[2304],f32>, !torch.vtensor<[128,2304],f32>, !torch.int -> !torch.vtensor<[128,2304],f32> loc(#loc327)
%665 = torch.aten.view %664, %107 : !torch.vtensor<[128,2304],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2304],f32> loc(#loc328)
%666 = torch.aten.slice.Tensor %665, %int2, %int0, %int768, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc329)
%667 = torch.aten.slice.Tensor %665, %int2, %int768, %int1536, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc330)
%668 = torch.aten.slice.Tensor %665, %int2, %int1536, %int2304, %int1 : !torch.vtensor<[1,128,2304],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc331)
%669 = torch.aten.view %666, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc332)
%670 = torch.aten.permute %669, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc333)
%671 = torch.aten.view %667, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc334)
%672 = torch.aten.permute %671, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc335)
%673 = torch.aten.view %668, %112 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc336)
%674 = torch.aten.permute %673, %114 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc337)
%675 = torch.aten.transpose.int %672, %int-1, %int-2 : !torch.vtensor<[1,12,128,64],f32>, !torch.int, !torch.int -> !torch.vtensor<[1,12,64,128],f32> loc(#loc338)
%676 = torch.aten.broadcast_to %670, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc339)
%677 = torch.aten.view %676, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc340)
%678 = torch.aten.broadcast_to %675, %125 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,64,128],f32> loc(#loc341)
%679 = torch.aten.view %678, %127 : !torch.vtensor<[1,12,64,128],f32>, !torch.list<int> -> !torch.vtensor<[12,64,128],f32> loc(#loc342)
%680 = torch.aten.bmm %677, %679 : !torch.vtensor<[12,128,64],f32>, !torch.vtensor<[12,64,128],f32> -> !torch.vtensor<[12,128,128],f32> loc(#loc343)
%681 = torch.aten.view %680, %130 : !torch.vtensor<[12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc344)
%682 = torch.prim.NumToTensor.Scalar %float8.000000e00 : !torch.float -> !torch.vtensor<[],f64> loc(#loc345)
%683 = torch.aten.to.dtype %682, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc345)
%684 = torch.aten.broadcast_to %683, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc345)
%685 = torch.aten.div.Tensor %681, %684 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc346)
%686 = torch.aten.slice.Tensor %71, %int2, %int0, %int128, %int1 : !torch.vtensor<[1,1,1024,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,1024],ui8> loc(#loc347)
%687 = torch.aten.slice.Tensor %686, %int3, %int0, %int128, %int1 : !torch.vtensor<[1,1,128,1024],ui8>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,128,128],ui8> loc(#loc348)
%688 = torch.prim.NumToTensor.Scalar %int0 : !torch.int -> !torch.vtensor<[],si64> loc(#loc349)
%689 = torch.aten.to.dtype %688, %int11, %false, %false, %none : !torch.vtensor<[],si64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],i1> loc(#loc349)
%690 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc349)
%691 = torch.aten.broadcast_to %689, %690 : !torch.vtensor<[],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc349)
%692 = torch.aten.to.dtype %687, %int11, %false, %false, %none : !torch.vtensor<[1,1,128,128],ui8>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[1,1,128,128],i1> loc(#loc349)
%693 = torch.prim.ListConstruct %int1, %int1, %int128, %int128 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc349)
%694 = torch.aten.broadcast_to %692, %693 : !torch.vtensor<[1,1,128,128],i1>, !torch.list<int> -> !torch.vtensor<[1,1,128,128],i1> loc(#loc349)
%695 = torch.prim.NumToTensor.Scalar %float-3.402820e38 : !torch.float -> !torch.vtensor<[],f64> loc(#loc350)
%696 = torch.aten.to.dtype %695, %int6, %false, %false, %none : !torch.vtensor<[],f64>, !torch.int, !torch.bool, !torch.bool, !torch.none -> !torch.vtensor<[],f32> loc(#loc350)
%697 = torch.aten.broadcast_to %696, %132 : !torch.vtensor<[],f32>, !torch.list<int> -> !torch.vtensor<[],f32> loc(#loc350)
%698 = torch.aten.where.self %694, %685, %697 : !torch.vtensor<[1,1,128,128],i1>, !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc351)
%values_8, %indices_9 = torch.aten.max.dim %698, %int-1, %true : !torch.vtensor<[1,12,128,128],f32>, !torch.int, !torch.bool -> !torch.vtensor<[1,12,128,1],f32>, !torch.vtensor<[1,12,128,1],si64> loc(#loc352)
%699 = torch.aten.sub.Tensor %698, %values_8, %float1.000000e00 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32>, !torch.float -> !torch.vtensor<[1,12,128,128],f32> loc(#loc352)
%700 = torch.aten.exp %699 : !torch.vtensor<[1,12,128,128],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc352)
%701 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc352)
%702 = torch.aten.sum.dim_IntList %700, %701, %true, %none : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,12,128,1],f32> loc(#loc352)
%703 = torch.aten.div.Tensor %700, %702 : !torch.vtensor<[1,12,128,128],f32>, !torch.vtensor<[1,12,128,1],f32> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc352)
%704 = torch.aten.broadcast_to %703, %130 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,128],f32> loc(#loc353)
%705 = torch.aten.view %704, %156 : !torch.vtensor<[1,12,128,128],f32>, !torch.list<int> -> !torch.vtensor<[12,128,128],f32> loc(#loc354)
%706 = torch.aten.broadcast_to %674, %121 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc355)
%707 = torch.aten.view %706, %123 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[12,128,64],f32> loc(#loc356)
%708 = torch.aten.bmm %705, %707 : !torch.vtensor<[12,128,128],f32>, !torch.vtensor<[12,128,64],f32> -> !torch.vtensor<[12,128,64],f32> loc(#loc357)
%709 = torch.aten.view %708, %121 : !torch.vtensor<[12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,12,128,64],f32> loc(#loc358)
%710 = torch.aten.permute %709, %114 : !torch.vtensor<[1,12,128,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,12,64],f32> loc(#loc359)
%711 = torch.aten.clone %710, %int0 : !torch.vtensor<[1,128,12,64],f32>, !torch.int -> !torch.vtensor<[1,128,12,64],f32> loc(#loc360)
%712 = torch.aten.view %711, %164 : !torch.vtensor<[1,128,12,64],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc361)
%713 = torch.aten.view %712, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc362)
%714 = torch.aten.mm %713, %9 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc363)
%715 = torch.aten.mul.Scalar %10, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc363)
%716 = torch.aten.add.Tensor %715, %714, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc363)
%717 = torch.aten.view %716, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc364)
%718 = torch.aten.add.Tensor %717, %644, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc365)
%719 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc366)
%720 = torch.aten.sum.dim_IntList %718, %719, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc366)
%721 = torch.aten.div.Scalar %720, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc366)
%722 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc366)
%723 = torch.aten.broadcast_to %721, %722 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%724 = torch.aten.sub.Tensor %718, %723, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%725 = torch.aten.mul.Tensor %724, %724 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%726 = torch.aten.sum.dim_IntList %725, %719, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc366)
%727 = torch.aten.div.Scalar %726, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc366)
%728 = torch.aten.add.Scalar %727, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc366)
%729 = torch.aten.rsqrt %728 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc366)
%730 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc366)
%731 = torch.aten.broadcast_to %729, %730 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%732 = torch.aten.mul.Tensor %724, %731 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%733 = torch.aten.mul.Tensor %732, %8 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%734 = torch.aten.add.Tensor %733, %7, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc366)
%735 = torch.aten.view %734, %102 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc367)
%736 = torch.aten.mm %735, %5 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,3072],f32> -> !torch.vtensor<[128,3072],f32> loc(#loc368)
%737 = torch.aten.mul.Scalar %6, %int1 : !torch.vtensor<[3072],f32>, !torch.int -> !torch.vtensor<[3072],f32> loc(#loc368)
%738 = torch.aten.add.Tensor %737, %736, %int1 : !torch.vtensor<[3072],f32>, !torch.vtensor<[128,3072],f32>, !torch.int -> !torch.vtensor<[128,3072],f32> loc(#loc368)
%739 = torch.aten.view %738, %192 : !torch.vtensor<[128,3072],f32>, !torch.list<int> -> !torch.vtensor<[1,128,3072],f32> loc(#loc369)
%740 = torch.aten.mul.Scalar %739, %float5.000000e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc370)
%741 = torch.aten.pow.Tensor_Scalar %739, %float3.000000e00 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc371)
%742 = torch.aten.mul.Scalar %741, %float4.471500e-02 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc372)
%743 = torch.aten.add.Tensor %739, %742, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32>, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc373)
%744 = torch.aten.mul.Scalar %743, %float7.978850e-01 : !torch.vtensor<[1,128,3072],f32>, !torch.float -> !torch.vtensor<[1,128,3072],f32> loc(#loc374)
%745 = torch.aten.tanh %744 : !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc375)
%746 = torch.aten.add.Scalar %745, %float1.000000e00, %int1 : !torch.vtensor<[1,128,3072],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,3072],f32> loc(#loc376)
%747 = torch.aten.mul.Tensor %740, %746 : !torch.vtensor<[1,128,3072],f32>, !torch.vtensor<[1,128,3072],f32> -> !torch.vtensor<[1,128,3072],f32> loc(#loc377)
%748 = torch.aten.view %747, %202 : !torch.vtensor<[1,128,3072],f32>, !torch.list<int> -> !torch.vtensor<[128,3072],f32> loc(#loc378)
%749 = torch.aten.mm %748, %3 : !torch.vtensor<[128,3072],f32>, !torch.vtensor<[3072,768],f32> -> !torch.vtensor<[128,768],f32> loc(#loc379)
%750 = torch.aten.mul.Scalar %4, %int1 : !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[768],f32> loc(#loc379)
%751 = torch.aten.add.Tensor %750, %749, %int1 : !torch.vtensor<[768],f32>, !torch.vtensor<[128,768],f32>, !torch.int -> !torch.vtensor<[128,768],f32> loc(#loc379)
%752 = torch.aten.view %751, %164 : !torch.vtensor<[128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc380)
%753 = torch.aten.add.Tensor %718, %752, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc381)
%754 = torch.prim.ListConstruct %int2 : (!torch.int) -> !torch.list<int> loc(#loc382)
%755 = torch.aten.sum.dim_IntList %753, %754, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc382)
%756 = torch.aten.div.Scalar %755, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc382)
%757 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc382)
%758 = torch.aten.broadcast_to %756, %757 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%759 = torch.aten.sub.Tensor %753, %758, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%760 = torch.aten.mul.Tensor %759, %759 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%761 = torch.aten.sum.dim_IntList %760, %754, %true, %none : !torch.vtensor<[1,128,768],f32>, !torch.list<int>, !torch.bool, !torch.none -> !torch.vtensor<[1,128,1],f32> loc(#loc382)
%762 = torch.aten.div.Scalar %761, %int768 : !torch.vtensor<[1,128,1],f32>, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc382)
%763 = torch.aten.add.Scalar %762, %float1.000000e-05, %int1 : !torch.vtensor<[1,128,1],f32>, !torch.float, !torch.int -> !torch.vtensor<[1,128,1],f32> loc(#loc382)
%764 = torch.aten.rsqrt %763 : !torch.vtensor<[1,128,1],f32> -> !torch.vtensor<[1,128,1],f32> loc(#loc382)
%765 = torch.prim.ListConstruct %int1, %int128, %int768 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc382)
%766 = torch.aten.broadcast_to %764, %765 : !torch.vtensor<[1,128,1],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%767 = torch.aten.mul.Tensor %759, %766 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[1,128,768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%768 = torch.aten.mul.Tensor %767, %2 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32> -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%769 = torch.aten.add.Tensor %768, %1, %int1 : !torch.vtensor<[1,128,768],f32>, !torch.vtensor<[768],f32>, !torch.int -> !torch.vtensor<[1,128,768],f32> loc(#loc382)
%770 = torch.aten.view %769, %164 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[1,128,768],f32> loc(#loc383)
%771 = torch.aten.transpose.int %0, %int0, %int1 : !torch.vtensor<[2,768],f32>, !torch.int, !torch.int -> !torch.vtensor<[768,2],f32> loc(#loc384)
%772 = torch.prim.ListConstruct %int128, %int768 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc403)
%773 = torch.aten.view %770, %772 : !torch.vtensor<[1,128,768],f32>, !torch.list<int> -> !torch.vtensor<[128,768],f32> loc(#loc385)
%774 = torch.aten.mm %773, %771 : !torch.vtensor<[128,768],f32>, !torch.vtensor<[768,2],f32> -> !torch.vtensor<[128,2],f32> loc(#loc386)
%775 = torch.prim.ListConstruct %int1, %int128, %int2 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc)
%776 = torch.aten.view %774, %775 : !torch.vtensor<[128,2],f32>, !torch.list<int> -> !torch.vtensor<[1,128,2],f32> loc(#loc387)
%777 = torch.aten.arange.start_step %int0, %int1, %int1, %none, %none, %cpu, %false : !torch.int, !torch.int, !torch.int, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.vtensor<[1],si64> loc(#loc388)
%778 = torch.aten.slice.Tensor %776, %int1, %int127, %int128, %int1 : !torch.vtensor<[1,128,2],f32>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,1,2],f32> loc(#loc2)
%779 = torch.aten.squeeze.dim %778, %int1 : !torch.vtensor<[1,1,2],f32>, !torch.int -> !torch.vtensor<[1,2],f32> loc(#loc2)
%780 = torch.prim.ListConstruct %777 : (!torch.vtensor<[1],si64>) -> !torch.list<vtensor> loc(#loc389)
%781 = torch.aten.index.Tensor %779, %780 : !torch.vtensor<[1,2],f32>, !torch.list<vtensor> -> !torch.vtensor<[1,2],f32> loc(#loc389)
return %781 : !torch.vtensor<[1,2],f32> loc(#loc)
} loc(#loc)
} loc(#loc)
#loc1 = loc("<eval_with_key>.2":5:44)
#loc2 = loc("<eval_with_key>.2":507:13)
#loc3 = loc("<eval_with_key>.2":5:41)
#loc4 = loc("<eval_with_key>.2":16:24)
#loc5 = loc("<eval_with_key>.2":6:35)
#loc6 = loc("<eval_with_key>.2":19:34)
#loc7 = loc("<eval_with_key>.2":16:63)
#loc8 = loc("<eval_with_key>.2":86:37)
#loc9 = loc("<eval_with_key>.2":41:48)
#loc10 = loc("<eval_with_key>.2":6:106)
#loc11 = loc("<eval_with_key>.2":48:56)
#loc12 = loc("<eval_with_key>.2":34:56)
#loc13 = loc("<eval_with_key>.2":49:37)
#loc14 = loc("<eval_with_key>.2":6:51)
#loc15 = loc("<eval_with_key>.2":5:40)
#loc16 = loc("<eval_with_key>.2":16:105)
#loc17 = loc("<eval_with_key>.2":24:49)
#loc18 = loc("<eval_with_key>.2":26:51)
#loc19 = loc("<eval_with_key>.2":28:51)
#loc20 = loc("<eval_with_key>.2":28:55)
#loc21 = loc("<eval_with_key>.2":29:55)
#loc22 = loc("<eval_with_key>.2":41:35)
#loc23 = loc("<eval_with_key>.2":78:52)
#loc24 = loc("<eval_with_key>.2":79:38)
#loc25 = loc("<eval_with_key>.2":80:40)
#loc26 = loc("<eval_with_key>.2":81:38)
#loc27 = loc("<eval_with_key>.2":83:38)
#loc28 = loc("<eval_with_key>.2":5:11)
#loc29 = loc("<eval_with_key>.2":6:13)
#loc30 = loc("<eval_with_key>.2":7:16)
#loc31 = loc("<eval_with_key>.2":8:13)
#loc32 = loc("<eval_with_key>.2":10:16)
#loc33 = loc("<eval_with_key>.2":12:18)
#loc34 = loc("<eval_with_key>.2":13:10)
#loc35 = loc("<eval_with_key>.2":20:13)
#loc36 = loc("<eval_with_key>.2":23:12)
#loc37 = loc("-":4783:13)
#loc38 = loc("-":6691:10)
#loc39 = loc("<eval_with_key>.2":24:13)
#loc40 = loc("<eval_with_key>.2":25:14)
#loc41 = loc("<eval_with_key>.2":26:14)
#loc42 = loc("<eval_with_key>.2":27:14)
#loc43 = loc("<eval_with_key>.2":28:13)
#loc44 = loc("<eval_with_key>.2":29:14)
#loc45 = loc("<eval_with_key>.2":30:13)
#loc46 = loc("<eval_with_key>.2":31:16)
#loc47 = loc("<eval_with_key>.2":32:13)
#loc48 = loc("<eval_with_key>.2":33:16)
#loc49 = loc("<eval_with_key>.2":34:16)
#loc50 = loc("-":4951:15)
#loc51 = loc("-":6679:10)
#loc52 = loc("<eval_with_key>.2":35:13)
#loc53 = loc("<eval_with_key>.2":36:13)
#loc54 = loc("<eval_with_key>.2":37:15)
#loc55 = loc("<eval_with_key>.2":38:13)
#loc56 = loc("<eval_with_key>.2":39:10)
#loc57 = loc("-":1322:13)
#loc58 = loc("-":7035:10)
#loc59 = loc("<eval_with_key>.2":42:10)
#loc60 = loc("<eval_with_key>.2":40:19)
#loc61 = loc("<eval_with_key>.2":41:11)
#loc62 = loc("<eval_with_key>.2":46:14)
#loc63 = loc("<eval_with_key>.2":47:14)
#loc64 = loc("<eval_with_key>.2":48:15)
#loc65 = loc("<eval_with_key>.2":49:13)
#loc66 = loc("<eval_with_key>.2":50:12)
#loc67 = loc("<eval_with_key>.2":51:15)
#loc68 = loc("<eval_with_key>.2":53:15)
#loc69 = loc("<eval_with_key>.2":54:13)
#loc70 = loc("<eval_with_key>.2":55:15)
#loc71 = loc("<eval_with_key>.2":56:14)
#loc72 = loc("<eval_with_key>.2":57:12)
#loc73 = loc("<eval_with_key>.2":58:21)
#loc74 = loc("<eval_with_key>.2":59:16)
#loc75 = loc("<eval_with_key>.2":60:12)
#loc76 = loc("<eval_with_key>.2":61:14)
#loc77 = loc("<eval_with_key>.2":62:14)
#loc78 = loc("<eval_with_key>.2":65:14)
#loc79 = loc("<eval_with_key>.2":66:14)
#loc80 = loc("<eval_with_key>.2":67:12)
#loc81 = loc("<eval_with_key>.2":70:26)
#loc82 = loc("<eval_with_key>.2":74:14)
#loc83 = loc("<eval_with_key>.2":77:14)
#loc84 = loc("<eval_with_key>.2":78:14)
#loc85 = loc("<eval_with_key>.2":79:10)
#loc86 = loc("<eval_with_key>.2":80:12)
#loc87 = loc("<eval_with_key>.2":81:12)
#loc88 = loc("<eval_with_key>.2":82:12)
#loc89 = loc("<eval_with_key>.2":83:12)
#loc90 = loc("<eval_with_key>.2":84:11)
#loc91 = loc("<eval_with_key>.2":86:12)
#loc92 = loc("<eval_with_key>.2":87:12)
#loc93 = loc("<eval_with_key>.2":88:14)
#loc94 = loc("<eval_with_key>.2":91:14)
#loc95 = loc("<eval_with_key>.2":92:14)
#loc96 = loc("<eval_with_key>.2":93:12)
#loc97 = loc("<eval_with_key>.2":96:26)
#loc98 = loc("<eval_with_key>.2":100:14)
#loc99 = loc("<eval_with_key>.2":103:14)
#loc100 = loc("<eval_with_key>.2":104:14)
#loc101 = loc("<eval_with_key>.2":105:14)
#loc102 = loc("<eval_with_key>.2":106:14)
#loc103 = loc("<eval_with_key>.2":107:15)
#loc104 = loc("<eval_with_key>.2":108:14)
#loc105 = loc("<eval_with_key>.2":109:16)
#loc106 = loc("<eval_with_key>.2":110:14)
#loc107 = loc("<eval_with_key>.2":111:16)
#loc108 = loc("<eval_with_key>.2":112:14)
#loc109 = loc("<eval_with_key>.2":113:16)
#loc110 = loc("<eval_with_key>.2":114:18)
#loc111 = loc("<eval_with_key>.2":115:15)
#loc112 = loc("<eval_with_key>.2":116:14)
#loc113 = loc("<eval_with_key>.2":117:15)
#loc114 = loc("<eval_with_key>.2":118:14)
#loc115 = loc("<eval_with_key>.2":119:12)
#loc116 = loc("<eval_with_key>.2":120:21)
#loc117 = loc("<eval_with_key>.2":121:13)
#loc118 = loc("<eval_with_key>.2":122:12)
#loc119 = loc("<eval_with_key>.2":126:15)
#loc120 = loc("<eval_with_key>.2":127:15)
#loc121 = loc("<eval_with_key>.2":128:17)
#loc122 = loc("<eval_with_key>.2":129:13)
#loc123 = loc("<eval_with_key>.2":130:14)
#loc124 = loc("<eval_with_key>.2":131:17)
#loc125 = loc("<eval_with_key>.2":133:15)
#loc126 = loc("<eval_with_key>.2":134:14)
#loc127 = loc("<eval_with_key>.2":135:15)
#loc128 = loc("<eval_with_key>.2":136:14)
#loc129 = loc("<eval_with_key>.2":137:12)
#loc130 = loc("<eval_with_key>.2":138:21)
#loc131 = loc("<eval_with_key>.2":139:16)
#loc132 = loc("<eval_with_key>.2":140:14)
#loc133 = loc("<eval_with_key>.2":141:14)
#loc134 = loc("<eval_with_key>.2":142:14)
#loc135 = loc("<eval_with_key>.2":145:14)
#loc136 = loc("<eval_with_key>.2":146:14)
#loc137 = loc("<eval_with_key>.2":147:12)
#loc138 = loc("<eval_with_key>.2":150:26)
#loc139 = loc("<eval_with_key>.2":154:14)
#loc140 = loc("<eval_with_key>.2":157:14)
#loc141 = loc("<eval_with_key>.2":158:14)
#loc142 = loc("<eval_with_key>.2":159:12)
#loc143 = loc("<eval_with_key>.2":160:12)
#loc144 = loc("<eval_with_key>.2":161:12)
#loc145 = loc("<eval_with_key>.2":162:12)
#loc146 = loc("<eval_with_key>.2":163:12)
#loc147 = loc("<eval_with_key>.2":164:13)
#loc148 = loc("<eval_with_key>.2":166:12)
#loc149 = loc("<eval_with_key>.2":167:12)
#loc150 = loc("<eval_with_key>.2":168:14)
#loc151 = loc("<eval_with_key>.2":171:14)
#loc152 = loc("<eval_with_key>.2":172:14)
#loc153 = loc("<eval_with_key>.2":173:12)
#loc154 = loc("<eval_with_key>.2":176:26)
#loc155 = loc("<eval_with_key>.2":180:14)
#loc156 = loc("<eval_with_key>.2":183:14)
#loc157 = loc("<eval_with_key>.2":184:14)
#loc158 = loc("<eval_with_key>.2":185:15)
#loc159 = loc("<eval_with_key>.2":186:15)
#loc160 = loc("<eval_with_key>.2":187:15)
#loc161 = loc("<eval_with_key>.2":188:14)
#loc162 = loc("<eval_with_key>.2":189:16)
#loc163 = loc("<eval_with_key>.2":190:14)
#loc164 = loc("<eval_with_key>.2":191:16)
#loc165 = loc("<eval_with_key>.2":192:14)
#loc166 = loc("<eval_with_key>.2":193:17)
#loc167 = loc("<eval_with_key>.2":194:18)
#loc168 = loc("<eval_with_key>.2":195:15)
#loc169 = loc("<eval_with_key>.2":196:14)
#loc170 = loc("<eval_with_key>.2":197:15)
#loc171 = loc("<eval_with_key>.2":198:14)
#loc172 = loc("<eval_with_key>.2":199:12)
#loc173 = loc("<eval_with_key>.2":200:21)
#loc174 = loc("<eval_with_key>.2":201:13)
#loc175 = loc("<eval_with_key>.2":202:12)
#loc176 = loc("<eval_with_key>.2":206:15)
#loc177 = loc("<eval_with_key>.2":207:15)
#loc178 = loc("<eval_with_key>.2":208:17)
#loc179 = loc("<eval_with_key>.2":209:13)
#loc180 = loc("<eval_with_key>.2":210:14)
#loc181 = loc("<eval_with_key>.2":211:17)
#loc182 = loc("<eval_with_key>.2":213:16)
#loc183 = loc("<eval_with_key>.2":214:14)
#loc184 = loc("<eval_with_key>.2":215:16)
#loc185 = loc("<eval_with_key>.2":216:14)
#loc186 = loc("<eval_with_key>.2":217:12)
#loc187 = loc("<eval_with_key>.2":218:21)
#loc188 = loc("<eval_with_key>.2":219:17)
#loc189 = loc("<eval_with_key>.2":220:14)
#loc190 = loc("<eval_with_key>.2":221:14)
#loc191 = loc("<eval_with_key>.2":222:14)
#loc192 = loc("<eval_with_key>.2":225:14)
#loc193 = loc("<eval_with_key>.2":226:14)
#loc194 = loc("<eval_with_key>.2":227:12)
#loc195 = loc("<eval_with_key>.2":230:26)
#loc196 = loc("<eval_with_key>.2":234:14)
#loc197 = loc("<eval_with_key>.2":237:15)
#loc198 = loc("<eval_with_key>.2":238:14)
#loc199 = loc("<eval_with_key>.2":239:12)
#loc200 = loc("<eval_with_key>.2":240:12)
#loc201 = loc("<eval_with_key>.2":241:12)
#loc202 = loc("<eval_with_key>.2":242:13)
#loc203 = loc("<eval_with_key>.2":243:13)
#loc204 = loc("<eval_with_key>.2":244:13)
#loc205 = loc("<eval_with_key>.2":246:13)
#loc206 = loc("<eval_with_key>.2":247:13)
#loc207 = loc("<eval_with_key>.2":248:14)
#loc208 = loc("<eval_with_key>.2":251:15)
#loc209 = loc("<eval_with_key>.2":252:14)
#loc210 = loc("<eval_with_key>.2":253:13)
#loc211 = loc("<eval_with_key>.2":256:26)
#loc212 = loc("<eval_with_key>.2":260:14)
#loc213 = loc("<eval_with_key>.2":263:15)
#loc214 = loc("<eval_with_key>.2":264:14)
#loc215 = loc("<eval_with_key>.2":265:15)
#loc216 = loc("<eval_with_key>.2":266:15)
#loc217 = loc("<eval_with_key>.2":267:15)
#loc218 = loc("<eval_with_key>.2":268:14)
#loc219 = loc("<eval_with_key>.2":269:17)
#loc220 = loc("<eval_with_key>.2":270:14)
#loc221 = loc("<eval_with_key>.2":271:17)
#loc222 = loc("<eval_with_key>.2":272:14)
#loc223 = loc("<eval_with_key>.2":273:17)
#loc224 = loc("<eval_with_key>.2":274:18)
#loc225 = loc("<eval_with_key>.2":275:16)
#loc226 = loc("<eval_with_key>.2":276:14)
#loc227 = loc("<eval_with_key>.2":277:16)
#loc228 = loc("<eval_with_key>.2":278:14)
#loc229 = loc("<eval_with_key>.2":279:12)
#loc230 = loc("<eval_with_key>.2":280:21)
#loc231 = loc("<eval_with_key>.2":281:13)
#loc232 = loc("<eval_with_key>.2":282:12)
#loc233 = loc("<eval_with_key>.2":286:15)
#loc234 = loc("<eval_with_key>.2":287:15)
#loc235 = loc("<eval_with_key>.2":288:17)
#loc236 = loc("<eval_with_key>.2":289:13)
#loc237 = loc("<eval_with_key>.2":290:14)
#loc238 = loc("<eval_with_key>.2":291:17)
#loc239 = loc("<eval_with_key>.2":293:16)
#loc240 = loc("<eval_with_key>.2":294:14)
#loc241 = loc("<eval_with_key>.2":295:16)
#loc242 = loc("<eval_with_key>.2":296:14)
#loc243 = loc("<eval_with_key>.2":297:12)
#loc244 = loc("<eval_with_key>.2":298:21)
#loc245 = loc("<eval_with_key>.2":299:17)
#loc246 = loc("<eval_with_key>.2":300:14)
#loc247 = loc("<eval_with_key>.2":301:14)
#loc248 = loc("<eval_with_key>.2":302:14)
#loc249 = loc("<eval_with_key>.2":305:15)
#loc250 = loc("<eval_with_key>.2":306:14)
#loc251 = loc("<eval_with_key>.2":307:13)
#loc252 = loc("<eval_with_key>.2":310:26)
#loc253 = loc("<eval_with_key>.2":314:14)
#loc254 = loc("<eval_with_key>.2":317:15)
#loc255 = loc("<eval_with_key>.2":318:14)
#loc256 = loc("<eval_with_key>.2":319:13)
#loc257 = loc("<eval_with_key>.2":320:12)
#loc258 = loc("<eval_with_key>.2":321:13)
#loc259 = loc("<eval_with_key>.2":322:13)
#loc260 = loc("<eval_with_key>.2":323:13)
#loc261 = loc("<eval_with_key>.2":324:13)
#loc262 = loc("<eval_with_key>.2":326:13)
#loc263 = loc("<eval_with_key>.2":327:13)
#loc264 = loc("<eval_with_key>.2":328:14)
#loc265 = loc("<eval_with_key>.2":331:15)
#loc266 = loc("<eval_with_key>.2":332:14)
#loc267 = loc("<eval_with_key>.2":333:13)
#loc268 = loc("<eval_with_key>.2":336:26)
#loc269 = loc("<eval_with_key>.2":340:14)
#loc270 = loc("<eval_with_key>.2":343:15)
#loc271 = loc("<eval_with_key>.2":344:14)
#loc272 = loc("<eval_with_key>.2":345:15)
#loc273 = loc("<eval_with_key>.2":346:15)
#loc274 = loc("<eval_with_key>.2":347:15)
#loc275 = loc("<eval_with_key>.2":348:14)
#loc276 = loc("<eval_with_key>.2":349:17)
#loc277 = loc("<eval_with_key>.2":350:14)
#loc278 = loc("<eval_with_key>.2":351:17)
#loc279 = loc("<eval_with_key>.2":352:14)
#loc280 = loc("<eval_with_key>.2":353:17)
#loc281 = loc("<eval_with_key>.2":354:18)
#loc282 = loc("<eval_with_key>.2":355:16)
#loc283 = loc("<eval_with_key>.2":356:14)
#loc284 = loc("<eval_with_key>.2":357:16)
#loc285 = loc("<eval_with_key>.2":358:14)
#loc286 = loc("<eval_with_key>.2":359:12)
#loc287 = loc("<eval_with_key>.2":360:21)
#loc288 = loc("<eval_with_key>.2":361:13)
#loc289 = loc("<eval_with_key>.2":362:12)
#loc290 = loc("<eval_with_key>.2":366:15)
#loc291 = loc("<eval_with_key>.2":367:15)
#loc292 = loc("<eval_with_key>.2":368:17)
#loc293 = loc("<eval_with_key>.2":369:13)
#loc294 = loc("<eval_with_key>.2":370:14)
#loc295 = loc("<eval_with_key>.2":371:17)
#loc296 = loc("<eval_with_key>.2":373:16)
#loc297 = loc("<eval_with_key>.2":374:14)
#loc298 = loc("<eval_with_key>.2":375:16)
#loc299 = loc("<eval_with_key>.2":376:14)
#loc300 = loc("<eval_with_key>.2":377:12)
#loc301 = loc("<eval_with_key>.2":378:21)
#loc302 = loc("<eval_with_key>.2":379:17)
#loc303 = loc("<eval_with_key>.2":380:14)
#loc304 = loc("<eval_with_key>.2":381:14)
#loc305 = loc("<eval_with_key>.2":382:14)
#loc306 = loc("<eval_with_key>.2":385:15)
#loc307 = loc("<eval_with_key>.2":386:14)
#loc308 = loc("<eval_with_key>.2":387:13)
#loc309 = loc("<eval_with_key>.2":390:26)
#loc310 = loc("<eval_with_key>.2":394:14)
#loc311 = loc("<eval_with_key>.2":397:15)
#loc312 = loc("<eval_with_key>.2":398:14)
#loc313 = loc("<eval_with_key>.2":399:13)
#loc314 = loc("<eval_with_key>.2":400:12)
#loc315 = loc("<eval_with_key>.2":401:13)
#loc316 = loc("<eval_with_key>.2":402:13)
#loc317 = loc("<eval_with_key>.2":403:13)
#loc318 = loc("<eval_with_key>.2":404:13)
#loc319 = loc("<eval_with_key>.2":406:13)
#loc320 = loc("<eval_with_key>.2":407:13)
#loc321 = loc("<eval_with_key>.2":408:14)
#loc322 = loc("<eval_with_key>.2":411:15)
#loc323 = loc("<eval_with_key>.2":412:14)
#loc324 = loc("<eval_with_key>.2":413:13)
#loc325 = loc("<eval_with_key>.2":416:27)
#loc326 = loc("<eval_with_key>.2":420:14)
#loc327 = loc("<eval_with_key>.2":423:15)
#loc328 = loc("<eval_with_key>.2":424:14)
#loc329 = loc("<eval_with_key>.2":425:15)
#loc330 = loc("<eval_with_key>.2":426:15)
#loc331 = loc("<eval_with_key>.2":427:15)
#loc332 = loc("<eval_with_key>.2":428:14)
#loc333 = loc("<eval_with_key>.2":429:17)
#loc334 = loc("<eval_with_key>.2":430:14)
#loc335 = loc("<eval_with_key>.2":431:17)
#loc336 = loc("<eval_with_key>.2":432:14)
#loc337 = loc("<eval_with_key>.2":433:17)
#loc338 = loc("<eval_with_key>.2":434:18)
#loc339 = loc("<eval_with_key>.2":435:16)
#loc340 = loc("<eval_with_key>.2":436:14)
#loc341 = loc("<eval_with_key>.2":437:16)
#loc342 = loc("<eval_with_key>.2":438:14)
#loc343 = loc("<eval_with_key>.2":439:13)
#loc344 = loc("<eval_with_key>.2":440:22)
#loc345 = loc("<eval_with_key>.2":441:14)
#loc346 = loc("<eval_with_key>.2":442:12)
#loc347 = loc("<eval_with_key>.2":446:15)
#loc348 = loc("<eval_with_key>.2":447:15)
#loc349 = loc("<eval_with_key>.2":448:17)
#loc350 = loc("<eval_with_key>.2":449:14)
#loc351 = loc("<eval_with_key>.2":450:14)
#loc352 = loc("<eval_with_key>.2":451:17)
#loc353 = loc("<eval_with_key>.2":453:16)
#loc354 = loc("<eval_with_key>.2":454:14)
#loc355 = loc("<eval_with_key>.2":455:16)
#loc356 = loc("<eval_with_key>.2":456:14)
#loc357 = loc("<eval_with_key>.2":457:13)
#loc358 = loc("<eval_with_key>.2":458:22)
#loc359 = loc("<eval_with_key>.2":459:17)
#loc360 = loc("<eval_with_key>.2":460:14)
#loc361 = loc("<eval_with_key>.2":461:14)
#loc362 = loc("<eval_with_key>.2":462:14)
#loc363 = loc("<eval_with_key>.2":465:15)
#loc364 = loc("<eval_with_key>.2":466:14)
#loc365 = loc("<eval_with_key>.2":467:13)
#loc366 = loc("<eval_with_key>.2":470:27)
#loc367 = loc("<eval_with_key>.2":474:14)
#loc368 = loc("<eval_with_key>.2":477:15)
#loc369 = loc("<eval_with_key>.2":478:14)
#loc370 = loc("<eval_with_key>.2":479:13)
#loc371 = loc("<eval_with_key>.2":480:12)
#loc372 = loc("<eval_with_key>.2":481:13)
#loc373 = loc("<eval_with_key>.2":482:13)
#loc374 = loc("<eval_with_key>.2":483:13)
#loc375 = loc("<eval_with_key>.2":484:13)
#loc376 = loc("<eval_with_key>.2":486:13)
#loc377 = loc("<eval_with_key>.2":487:13)
#loc378 = loc("<eval_with_key>.2":488:14)
#loc379 = loc("<eval_with_key>.2":491:15)
#loc380 = loc("<eval_with_key>.2":492:14)
#loc381 = loc("<eval_with_key>.2":493:13)
#loc382 = loc("<eval_with_key>.2":496:27)
#loc383 = loc("<eval_with_key>.2":500:14)
#loc384 = loc("<eval_with_key>.2":502:8)
#loc385 = loc("<eval_with_key>.2":503:14)
#loc386 = loc("<eval_with_key>.2":504:9)
#loc387 = loc("<eval_with_key>.2":505:22)
#loc388 = loc("<eval_with_key>.2":506:15)
#loc389 = loc("<eval_with_key>.2":508:12)
#loc390 = loc(callsite(#loc37 at #loc38))
#loc391 = loc(callsite(#loc50 at #loc51))
#loc392 = loc(callsite(#loc57 at #loc58))
#loc393 = loc(callsite(#loc390 at #loc39))
#loc394 = loc(callsite(#loc390 at #loc43))
#loc395 = loc(callsite(#loc391 at #loc52))
#loc396 = loc(callsite(#loc390 at #loc53))
#loc397 = loc(callsite(#loc391 at #loc54))
#loc398 = loc(callsite(#loc390 at #loc55))
#loc399 = loc(callsite(#loc392 at #loc59))
#loc400 = loc(callsite(#loc390 at #loc69))
#loc401 = loc(callsite(#loc390 at #loc76))
#loc402 = loc(callsite(#loc390 at #loc84))
#loc403 = loc(callsite(#loc390 at #loc385))
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment