Last active
February 14, 2023 19:17
-
-
Save AmosLewis/28dce2a1a36b5904de89b30feb34d643 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| #loc = loc(unknown) | |
| module attributes {torch.debug_module_name = "_lambda"} { | |
| func.func private @__torch__.torch.fx.graph_module._lambda.__code_getter(%arg0: !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> loc(unknown)) -> !torch.str { | |
| %133 = torch.prim.GetAttr %arg0["_code"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.str loc(#loc) | |
| return %133 : !torch.str loc(#loc) | |
| } loc(#loc) | |
| func.func private @__torch__.torch.fx.graph_module._lambda.forward(%arg0: !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> loc(unknown), %arg1: !torch.tensor {torch.type_bound = !torch.vtensor<[1,15],si64>} loc(unknown), %arg2: !torch.tensor {torch.type_bound = !torch.vtensor<[1,4],si64>} loc(unknown)) -> !torch.tensor { | |
| %int6 = torch.constant.int 6 loc(#loc1) | |
| %true_0 = torch.constant.bool true loc(#loc2) | |
| %float-3.402820e38 = torch.constant.float -3.4028234663852886E+38 loc(#loc3) | |
| %int-100 = torch.constant.int -100 loc(#loc4) | |
| %none_1 = torch.constant.none loc(#loc) | |
| %int-1 = torch.constant.int -1 loc(#loc5) | |
| %false = torch.constant.bool false loc(#loc6) | |
| %cpu = torch.constant.device "cpu" loc(#loc) | |
| %int1 = torch.constant.int 1 loc(#loc7) | |
| %int4 = torch.constant.int 4 loc(#loc8) | |
| %int0 = torch.constant.int 0 loc(#loc9) | |
| %int9223372036854775807 = torch.constant.int 9223372036854775807 loc(#loc10) | |
| %int15 = torch.constant.int 15 loc(#loc11) | |
| %int2 = torch.constant.int 2 loc(#loc12) | |
| %int3 = torch.constant.int 3 loc(#loc13) | |
| %float1.000000e00 = torch.constant.float 1.000000e+00 loc(#loc14) | |
| %float9.999990e-07 = torch.constant.float 9.9999999999999995E-7 loc(#loc15) | |
| %int512 = torch.constant.int 512 loc(#loc16) | |
| %int8 = torch.constant.int 8 loc(#loc17) | |
| %int64 = torch.constant.int 64 loc(#loc18) | |
| %int16 = torch.constant.int 16 loc(#loc19) | |
| %float2.772590e00 = torch.constant.float 2.7725887222397811 loc(#loc20) | |
| %int2048 = torch.constant.int 2048 loc(#loc21) | |
| %float2.079440e00 = torch.constant.float 2.0794415416798357 loc(#loc22) | |
| %int31 = torch.constant.int 31 loc(#loc23) | |
| %float4.419420e-02 = torch.constant.float 0.044194173824159223 loc(#loc24) | |
| %int32128 = torch.constant.int 32128 loc(#loc25) | |
| %133 = torch.prim.ListConstruct %int1, %int4 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %134 = torch.aten.new_zeros %arg2, %133, %int4, %int0, %cpu, %false : !torch.tensor, !torch.list<int>, !torch.int, !torch.int, !torch.Device, !torch.bool -> !torch.tensor loc(#loc26) | |
| %135 = torch.aten.slice.Tensor %arg2, %int1, %int0, %int-1, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc27) | |
| %136 = torch.aten.clone %135, %none_1 : !torch.tensor, !torch.none -> !torch.tensor loc(#loc28) | |
| %137 = torch.aten.slice.Tensor %134, %int1, %int1, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc29) | |
| %138 = torch.aten.copy_ %137, %136, %false : !torch.tensor, !torch.tensor, !torch.bool -> !torch.tensor loc(#loc30) | |
| %139 = torch.prim.GetAttr %arg0["_tensor_constant0"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %140 = torch.aten.lift_fresh_copy %139 : !torch.tensor -> !torch.tensor loc(#loc31) | |
| %141 = torch.aten.select.int %134, %int1, %int0 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc32) | |
| %142 = torch.aten.fill_.Tensor %141, %140 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc33) | |
| %143 = torch.aten.eq.Scalar %134, %int-100 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc34) | |
| %144 = torch.aten.masked_fill_.Scalar %134, %143, %int0 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc35) | |
| %145 = torch.prim.ListConstruct %int-1, %int15 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %146 = torch.aten.view %arg1, %145 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc36) | |
| %147 = torch.prim.GetAttr %arg0["_param_constant0"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %148 = torch.aten.embedding %147, %146, %int-1, %false, %false : !torch.tensor, !torch.tensor, !torch.int, !torch.bool, !torch.bool -> !torch.tensor loc(#loc37) | |
| %149 = torch.prim.ListConstruct %int1, %int15 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %150 = torch.aten.ones %149, %none_1, %none_1, %cpu, %false : !torch.list<int>, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc38) | |
| %151 = torch.aten.slice.Tensor %150, %int0, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc39) | |
| %152 = torch.aten.unsqueeze %151, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc40) | |
| %153 = torch.aten.unsqueeze %152, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc41) | |
| %154 = torch.aten.slice.Tensor %153, %int3, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc42) | |
| %155 = torch.aten.rsub.Scalar %154, %float1.000000e00, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc43) | |
| %156 = torch.aten.mul.Scalar %155, %float-3.402820e38 : !torch.tensor, !torch.float -> !torch.tensor loc(#loc44) | |
| %157 = torch.aten.pow.Tensor_Scalar %148, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc45) | |
| %158 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %159 = torch.aten.mean.dim %157, %158, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc46) | |
| %160 = torch.aten.add.Scalar %159, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc47) | |
| %161 = torch.aten.rsqrt %160 : !torch.tensor -> !torch.tensor loc(#loc48) | |
| %162 = torch.aten.mul.Tensor %148, %161 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc49) | |
| %163 = torch.prim.GetAttr %arg0["_param_constant1"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %164 = torch.aten.mul.Tensor %163, %162 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc50) | |
| %165 = torch.prim.GetAttr %arg0["_param_constant2"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %166 = torch.aten.t %165 : !torch.tensor -> !torch.tensor loc(#loc51) | |
| %167 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %168 = torch.aten.view %164, %167 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc52) | |
| %169 = torch.aten.mm %168, %166 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc53) | |
| %170 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %171 = torch.aten._unsafe_view %169, %170 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc54) | |
| %172 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %173 = torch.aten.view %171, %172 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc55) | |
| %174 = torch.aten.transpose.int %173, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc56) | |
| %175 = torch.prim.GetAttr %arg0["_param_constant3"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %176 = torch.aten.t %175 : !torch.tensor -> !torch.tensor loc(#loc57) | |
| %177 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %178 = torch.aten.view %164, %177 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc58) | |
| %179 = torch.aten.mm %178, %176 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc59) | |
| %180 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %181 = torch.aten._unsafe_view %179, %180 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc60) | |
| %182 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %183 = torch.aten.view %181, %182 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc61) | |
| %184 = torch.aten.transpose.int %183, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc62) | |
| %185 = torch.prim.GetAttr %arg0["_param_constant4"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %186 = torch.aten.t %185 : !torch.tensor -> !torch.tensor loc(#loc63) | |
| %187 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %188 = torch.aten.view %164, %187 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc64) | |
| %189 = torch.aten.mm %188, %186 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc65) | |
| %190 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %191 = torch.aten._unsafe_view %189, %190 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc66) | |
| %192 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %193 = torch.aten.view %191, %192 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc67) | |
| %194 = torch.aten.transpose.int %193, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc68) | |
| %195 = torch.aten.transpose.int %184, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc69) | |
| %196 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %197 = torch.aten.expand %174, %196, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc70) | |
| %198 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %199 = torch.aten.view %197, %198 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc71) | |
| %200 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %201 = torch.aten.expand %195, %200, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc72) | |
| %202 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %203 = torch.aten.view %201, %202 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc73) | |
| %204 = torch.aten.bmm %199, %203 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc74) | |
| %205 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %206 = torch.aten._unsafe_view %204, %205 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc75) | |
| %207 = torch.aten.arange %int15, %int4, %none_1, %cpu, %false : !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc76) | |
| %208 = torch.aten.slice.Tensor %207, %int0, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc77) | |
| %209 = torch.aten.unsqueeze %208, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc78) | |
| %210 = torch.aten.arange %int15, %int4, %none_1, %cpu, %false : !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc79) | |
| %211 = torch.aten.unsqueeze %210, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc80) | |
| %212 = torch.aten.slice.Tensor %211, %int1, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc81) | |
| %213 = torch.aten.sub.Tensor %212, %209, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc82) | |
| %214 = torch.aten.gt.Scalar %213, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc83) | |
| %215 = torch.prims.convert_element_type %214, %int4 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc84) | |
| %216 = torch.aten.mul.Scalar %215, %int16 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc85) | |
| %217 = torch.aten.add.Scalar %216, %int0, %int1 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc86) | |
| %218 = torch.aten.abs %213 : !torch.tensor -> !torch.tensor loc(#loc87) | |
| %219 = torch.aten.lt.Scalar %218, %int8 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc88) | |
| %220 = torch.prims.convert_element_type %218, %int6 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc89) | |
| %221 = torch.aten.div.Scalar %220, %int8 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc90) | |
| %222 = torch.aten.log %221 : !torch.tensor -> !torch.tensor loc(#loc91) | |
| %223 = torch.aten.div.Scalar %222, %float2.772590e00 : !torch.tensor, !torch.float -> !torch.tensor loc(#loc92) | |
| %224 = torch.aten.mul.Scalar %223, %int8 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc93) | |
| %225 = torch.prims.convert_element_type %224, %int4 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc94) | |
| %226 = torch.aten.add.Scalar %225, %int8, %int1 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc95) | |
| %227 = torch.aten.full_like %226, %int15, %int4, %int0, %cpu, %false, %none_1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.Device, !torch.bool, !torch.none -> !torch.tensor loc(#loc96) | |
| %228 = torch.aten.minimum %226, %227 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc97) | |
| %229 = torch.aten.where.self %219, %218, %228 : !torch.tensor, !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc98) | |
| %230 = torch.aten.add_.Tensor %217, %229, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc99) | |
| %231 = torch.prim.GetAttr %arg0["_param_constant5"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %232 = torch.aten.embedding %231, %230, %int-1, %false, %false : !torch.tensor, !torch.tensor, !torch.int, !torch.bool, !torch.bool -> !torch.tensor loc(#loc100) | |
| %233 = torch.prim.ListConstruct %int2, %int0, %int1 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %234 = torch.aten.permute %232, %233 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc101) | |
| %235 = torch.aten.unsqueeze %234, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc102) | |
| %236 = torch.aten.add.Tensor %235, %156, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc103) | |
| %237 = torch.aten.add_.Tensor %206, %236, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc104) | |
| %238 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %239 = torch.aten.amax %237, %238, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc105) | |
| %240 = torch.aten.sub.Tensor %237, %239, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc106) | |
| %241 = torch.aten.exp %240 : !torch.tensor -> !torch.tensor loc(#loc107) | |
| %242 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %243 = torch.aten.sum.dim_IntList %241, %242, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc108) | |
| %244 = torch.aten.div.Tensor %241, %243 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc109) | |
| %245 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %246 = torch.aten.expand %244, %245, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc110) | |
| %247 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %248 = torch.aten.view %246, %247 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc111) | |
| %249 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %250 = torch.aten.expand %194, %249, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc112) | |
| %251 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %252 = torch.aten.view %250, %251 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc113) | |
| %253 = torch.aten.bmm %248, %252 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc114) | |
| %254 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %255 = torch.aten._unsafe_view %253, %254 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc115) | |
| %256 = torch.aten.transpose.int %255, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc116) | |
| %257 = torch.aten.clone %256, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc117) | |
| %258 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %259 = torch.aten.view %257, %258 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc118) | |
| %260 = torch.prim.GetAttr %arg0["_param_constant6"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %261 = torch.aten.t %260 : !torch.tensor -> !torch.tensor loc(#loc119) | |
| %262 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %263 = torch.aten.view %259, %262 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc120) | |
| %264 = torch.aten.mm %263, %261 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc121) | |
| %265 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %266 = torch.aten._unsafe_view %264, %265 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc122) | |
| %267 = torch.aten.add.Tensor %148, %266, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc123) | |
| %268 = torch.aten.pow.Tensor_Scalar %267, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc124) | |
| %269 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %270 = torch.aten.mean.dim %268, %269, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc125) | |
| %271 = torch.aten.add.Scalar %270, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc126) | |
| %272 = torch.aten.rsqrt %271 : !torch.tensor -> !torch.tensor loc(#loc127) | |
| %273 = torch.aten.mul.Tensor %267, %272 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc128) | |
| %274 = torch.prim.GetAttr %arg0["_param_constant7"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %275 = torch.aten.mul.Tensor %274, %273 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc129) | |
| %276 = torch.prim.GetAttr %arg0["_param_constant8"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %277 = torch.aten.t %276 : !torch.tensor -> !torch.tensor loc(#loc130) | |
| %278 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %279 = torch.aten.view %275, %278 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc131) | |
| %280 = torch.aten.mm %279, %277 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc132) | |
| %281 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %282 = torch.aten._unsafe_view %280, %281 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc133) | |
| %283 = torch.aten.relu %282 : !torch.tensor -> !torch.tensor loc(#loc134) | |
| %284 = torch.prim.GetAttr %arg0["_param_constant9"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %285 = torch.aten.t %284 : !torch.tensor -> !torch.tensor loc(#loc135) | |
| %286 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %287 = torch.aten.view %283, %286 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc136) | |
| %288 = torch.aten.mm %287, %285 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc137) | |
| %289 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %290 = torch.aten._unsafe_view %288, %289 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc138) | |
| %291 = torch.aten.add.Tensor %267, %290, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc139) | |
| %292 = torch.aten.pow.Tensor_Scalar %291, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc140) | |
| %293 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %294 = torch.aten.mean.dim %292, %293, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc141) | |
| %295 = torch.aten.add.Scalar %294, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc142) | |
| %296 = torch.aten.rsqrt %295 : !torch.tensor -> !torch.tensor loc(#loc143) | |
| %297 = torch.aten.mul.Tensor %291, %296 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc144) | |
| %298 = torch.prim.GetAttr %arg0["_param_constant10"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %299 = torch.aten.mul.Tensor %298, %297 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc145) | |
| %300 = torch.prim.GetAttr %arg0["_param_constant11"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %301 = torch.aten.t %300 : !torch.tensor -> !torch.tensor loc(#loc146) | |
| %302 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %303 = torch.aten.view %299, %302 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc147) | |
| %304 = torch.aten.mm %303, %301 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc148) | |
| %305 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %306 = torch.aten._unsafe_view %304, %305 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc149) | |
| %307 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %308 = torch.aten.view %306, %307 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc150) | |
| %309 = torch.aten.transpose.int %308, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc151) | |
| %310 = torch.prim.GetAttr %arg0["_param_constant12"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %311 = torch.aten.t %310 : !torch.tensor -> !torch.tensor loc(#loc152) | |
| %312 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %313 = torch.aten.view %299, %312 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc153) | |
| %314 = torch.aten.mm %313, %311 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc154) | |
| %315 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %316 = torch.aten._unsafe_view %314, %315 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc155) | |
| %317 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %318 = torch.aten.view %316, %317 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc156) | |
| %319 = torch.aten.transpose.int %318, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc157) | |
| %320 = torch.prim.GetAttr %arg0["_param_constant13"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %321 = torch.aten.t %320 : !torch.tensor -> !torch.tensor loc(#loc158) | |
| %322 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %323 = torch.aten.view %299, %322 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc159) | |
| %324 = torch.aten.mm %323, %321 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc160) | |
| %325 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %326 = torch.aten._unsafe_view %324, %325 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc161) | |
| %327 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %328 = torch.aten.view %326, %327 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc162) | |
| %329 = torch.aten.transpose.int %328, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc163) | |
| %330 = torch.aten.transpose.int %319, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc164) | |
| %331 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %332 = torch.aten.expand %309, %331, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc165) | |
| %333 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %334 = torch.aten.view %332, %333 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc166) | |
| %335 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %336 = torch.aten.expand %330, %335, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc167) | |
| %337 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %338 = torch.aten.view %336, %337 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc168) | |
| %339 = torch.aten.bmm %334, %338 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc169) | |
| %340 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %341 = torch.aten._unsafe_view %339, %340 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc170) | |
| %342 = torch.aten.add_.Tensor %341, %236, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc171) | |
| %343 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %344 = torch.aten.amax %342, %343, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc172) | |
| %345 = torch.aten.sub.Tensor %342, %344, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc173) | |
| %346 = torch.aten.exp %345 : !torch.tensor -> !torch.tensor loc(#loc174) | |
| %347 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %348 = torch.aten.sum.dim_IntList %346, %347, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc175) | |
| %349 = torch.aten.div.Tensor %346, %348 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc176) | |
| %350 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %351 = torch.aten.expand %349, %350, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc177) | |
| %352 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %353 = torch.aten.view %351, %352 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc178) | |
| %354 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %355 = torch.aten.expand %329, %354, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc179) | |
| %356 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %357 = torch.aten.view %355, %356 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc180) | |
| %358 = torch.aten.bmm %353, %357 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc181) | |
| %359 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %360 = torch.aten._unsafe_view %358, %359 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc182) | |
| %361 = torch.aten.transpose.int %360, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc183) | |
| %362 = torch.aten.clone %361, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc184) | |
| %363 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %364 = torch.aten.view %362, %363 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc185) | |
| %365 = torch.prim.GetAttr %arg0["_param_constant14"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %366 = torch.aten.t %365 : !torch.tensor -> !torch.tensor loc(#loc186) | |
| %367 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %368 = torch.aten.view %364, %367 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc187) | |
| %369 = torch.aten.mm %368, %366 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc188) | |
| %370 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %371 = torch.aten._unsafe_view %369, %370 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc189) | |
| %372 = torch.aten.add.Tensor %291, %371, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc190) | |
| %373 = torch.aten.pow.Tensor_Scalar %372, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc191) | |
| %374 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %375 = torch.aten.mean.dim %373, %374, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc192) | |
| %376 = torch.aten.add.Scalar %375, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc193) | |
| %377 = torch.aten.rsqrt %376 : !torch.tensor -> !torch.tensor loc(#loc194) | |
| %378 = torch.aten.mul.Tensor %372, %377 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc195) | |
| %379 = torch.prim.GetAttr %arg0["_param_constant15"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %380 = torch.aten.mul.Tensor %379, %378 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc196) | |
| %381 = torch.prim.GetAttr %arg0["_param_constant16"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %382 = torch.aten.t %381 : !torch.tensor -> !torch.tensor loc(#loc197) | |
| %383 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %384 = torch.aten.view %380, %383 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc198) | |
| %385 = torch.aten.mm %384, %382 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc199) | |
| %386 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %387 = torch.aten._unsafe_view %385, %386 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc200) | |
| %388 = torch.aten.relu %387 : !torch.tensor -> !torch.tensor loc(#loc201) | |
| %389 = torch.prim.GetAttr %arg0["_param_constant17"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %390 = torch.aten.t %389 : !torch.tensor -> !torch.tensor loc(#loc202) | |
| %391 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %392 = torch.aten.view %388, %391 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc203) | |
| %393 = torch.aten.mm %392, %390 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc204) | |
| %394 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %395 = torch.aten._unsafe_view %393, %394 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc205) | |
| %396 = torch.aten.add.Tensor %372, %395, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc206) | |
| %397 = torch.aten.pow.Tensor_Scalar %396, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc207) | |
| %398 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %399 = torch.aten.mean.dim %397, %398, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc208) | |
| %400 = torch.aten.add.Scalar %399, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc209) | |
| %401 = torch.aten.rsqrt %400 : !torch.tensor -> !torch.tensor loc(#loc210) | |
| %402 = torch.aten.mul.Tensor %396, %401 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc211) | |
| %403 = torch.prim.GetAttr %arg0["_param_constant18"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %404 = torch.aten.mul.Tensor %403, %402 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc212) | |
| %405 = torch.prim.GetAttr %arg0["_param_constant19"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %406 = torch.aten.t %405 : !torch.tensor -> !torch.tensor loc(#loc213) | |
| %407 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %408 = torch.aten.view %404, %407 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc214) | |
| %409 = torch.aten.mm %408, %406 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc215) | |
| %410 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %411 = torch.aten._unsafe_view %409, %410 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc216) | |
| %412 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %413 = torch.aten.view %411, %412 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc217) | |
| %414 = torch.aten.transpose.int %413, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc218) | |
| %415 = torch.prim.GetAttr %arg0["_param_constant20"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %416 = torch.aten.t %415 : !torch.tensor -> !torch.tensor loc(#loc219) | |
| %417 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %418 = torch.aten.view %404, %417 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc220) | |
| %419 = torch.aten.mm %418, %416 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc221) | |
| %420 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %421 = torch.aten._unsafe_view %419, %420 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc222) | |
| %422 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %423 = torch.aten.view %421, %422 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc223) | |
| %424 = torch.aten.transpose.int %423, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc224) | |
| %425 = torch.prim.GetAttr %arg0["_param_constant21"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %426 = torch.aten.t %425 : !torch.tensor -> !torch.tensor loc(#loc225) | |
| %427 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %428 = torch.aten.view %404, %427 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc226) | |
| %429 = torch.aten.mm %428, %426 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc227) | |
| %430 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %431 = torch.aten._unsafe_view %429, %430 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc228) | |
| %432 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %433 = torch.aten.view %431, %432 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc229) | |
| %434 = torch.aten.transpose.int %433, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc230) | |
| %435 = torch.aten.transpose.int %424, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc231) | |
| %436 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %437 = torch.aten.expand %414, %436, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc232) | |
| %438 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %439 = torch.aten.view %437, %438 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc233) | |
| %440 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %441 = torch.aten.expand %435, %440, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc234) | |
| %442 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %443 = torch.aten.view %441, %442 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc235) | |
| %444 = torch.aten.bmm %439, %443 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc236) | |
| %445 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %446 = torch.aten._unsafe_view %444, %445 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc237) | |
| %447 = torch.aten.add_.Tensor %446, %236, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc238) | |
| %448 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %449 = torch.aten.amax %447, %448, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc239) | |
| %450 = torch.aten.sub.Tensor %447, %449, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc240) | |
| %451 = torch.aten.exp %450 : !torch.tensor -> !torch.tensor loc(#loc241) | |
| %452 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %453 = torch.aten.sum.dim_IntList %451, %452, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc242) | |
| %454 = torch.aten.div.Tensor %451, %453 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc243) | |
| %455 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %456 = torch.aten.expand %454, %455, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc244) | |
| %457 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %458 = torch.aten.view %456, %457 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc245) | |
| %459 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %460 = torch.aten.expand %434, %459, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc246) | |
| %461 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %462 = torch.aten.view %460, %461 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc247) | |
| %463 = torch.aten.bmm %458, %462 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc248) | |
| %464 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %465 = torch.aten._unsafe_view %463, %464 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc249) | |
| %466 = torch.aten.transpose.int %465, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc250) | |
| %467 = torch.aten.clone %466, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc251) | |
| %468 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %469 = torch.aten.view %467, %468 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc252) | |
| %470 = torch.prim.GetAttr %arg0["_param_constant22"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %471 = torch.aten.t %470 : !torch.tensor -> !torch.tensor loc(#loc253) | |
| %472 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %473 = torch.aten.view %469, %472 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc254) | |
| %474 = torch.aten.mm %473, %471 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc255) | |
| %475 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %476 = torch.aten._unsafe_view %474, %475 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc256) | |
| %477 = torch.aten.add.Tensor %396, %476, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc257) | |
| %478 = torch.aten.pow.Tensor_Scalar %477, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc258) | |
| %479 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %480 = torch.aten.mean.dim %478, %479, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc259) | |
| %481 = torch.aten.add.Scalar %480, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc260) | |
| %482 = torch.aten.rsqrt %481 : !torch.tensor -> !torch.tensor loc(#loc261) | |
| %483 = torch.aten.mul.Tensor %477, %482 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc262) | |
| %484 = torch.prim.GetAttr %arg0["_param_constant23"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %485 = torch.aten.mul.Tensor %484, %483 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc263) | |
| %486 = torch.prim.GetAttr %arg0["_param_constant24"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %487 = torch.aten.t %486 : !torch.tensor -> !torch.tensor loc(#loc264) | |
| %488 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %489 = torch.aten.view %485, %488 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc265) | |
| %490 = torch.aten.mm %489, %487 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc266) | |
| %491 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %492 = torch.aten._unsafe_view %490, %491 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc267) | |
| %493 = torch.aten.relu %492 : !torch.tensor -> !torch.tensor loc(#loc268) | |
| %494 = torch.prim.GetAttr %arg0["_param_constant25"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %495 = torch.aten.t %494 : !torch.tensor -> !torch.tensor loc(#loc269) | |
| %496 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %497 = torch.aten.view %493, %496 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc270) | |
| %498 = torch.aten.mm %497, %495 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc271) | |
| %499 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %500 = torch.aten._unsafe_view %498, %499 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc272) | |
| %501 = torch.aten.add.Tensor %477, %500, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc273) | |
| %502 = torch.aten.pow.Tensor_Scalar %501, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc274) | |
| %503 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %504 = torch.aten.mean.dim %502, %503, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc275) | |
| %505 = torch.aten.add.Scalar %504, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc276) | |
| %506 = torch.aten.rsqrt %505 : !torch.tensor -> !torch.tensor loc(#loc277) | |
| %507 = torch.aten.mul.Tensor %501, %506 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc278) | |
| %508 = torch.prim.GetAttr %arg0["_param_constant26"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %509 = torch.aten.mul.Tensor %508, %507 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc279) | |
| %510 = torch.prim.GetAttr %arg0["_param_constant27"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %511 = torch.aten.t %510 : !torch.tensor -> !torch.tensor loc(#loc280) | |
| %512 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %513 = torch.aten.view %509, %512 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc281) | |
| %514 = torch.aten.mm %513, %511 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc282) | |
| %515 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %516 = torch.aten._unsafe_view %514, %515 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc283) | |
| %517 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %518 = torch.aten.view %516, %517 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc284) | |
| %519 = torch.aten.transpose.int %518, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc285) | |
| %520 = torch.prim.GetAttr %arg0["_param_constant28"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %521 = torch.aten.t %520 : !torch.tensor -> !torch.tensor loc(#loc286) | |
| %522 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %523 = torch.aten.view %509, %522 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc287) | |
| %524 = torch.aten.mm %523, %521 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc288) | |
| %525 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %526 = torch.aten._unsafe_view %524, %525 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc289) | |
| %527 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %528 = torch.aten.view %526, %527 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc290) | |
| %529 = torch.aten.transpose.int %528, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc291) | |
| %530 = torch.prim.GetAttr %arg0["_param_constant29"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %531 = torch.aten.t %530 : !torch.tensor -> !torch.tensor loc(#loc292) | |
| %532 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %533 = torch.aten.view %509, %532 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc293) | |
| %534 = torch.aten.mm %533, %531 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc294) | |
| %535 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %536 = torch.aten._unsafe_view %534, %535 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc295) | |
| %537 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %538 = torch.aten.view %536, %537 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc296) | |
| %539 = torch.aten.transpose.int %538, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc297) | |
| %540 = torch.aten.transpose.int %529, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc298) | |
| %541 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %542 = torch.aten.expand %519, %541, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc299) | |
| %543 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %544 = torch.aten.view %542, %543 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc300) | |
| %545 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %546 = torch.aten.expand %540, %545, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc301) | |
| %547 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %548 = torch.aten.view %546, %547 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc302) | |
| %549 = torch.aten.bmm %544, %548 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc303) | |
| %550 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %551 = torch.aten._unsafe_view %549, %550 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc304) | |
| %552 = torch.aten.add_.Tensor %551, %236, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc305) | |
| %553 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %554 = torch.aten.amax %552, %553, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc306) | |
| %555 = torch.aten.sub.Tensor %552, %554, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc307) | |
| %556 = torch.aten.exp %555 : !torch.tensor -> !torch.tensor loc(#loc308) | |
| %557 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %558 = torch.aten.sum.dim_IntList %556, %557, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc309) | |
| %559 = torch.aten.div.Tensor %556, %558 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc310) | |
| %560 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %561 = torch.aten.expand %559, %560, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc311) | |
| %562 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %563 = torch.aten.view %561, %562 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc312) | |
| %564 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %565 = torch.aten.expand %539, %564, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc313) | |
| %566 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %567 = torch.aten.view %565, %566 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc314) | |
| %568 = torch.aten.bmm %563, %567 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc315) | |
| %569 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %570 = torch.aten._unsafe_view %568, %569 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc316) | |
| %571 = torch.aten.transpose.int %570, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc317) | |
| %572 = torch.aten.clone %571, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc318) | |
| %573 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %574 = torch.aten.view %572, %573 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc319) | |
| %575 = torch.prim.GetAttr %arg0["_param_constant30"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %576 = torch.aten.t %575 : !torch.tensor -> !torch.tensor loc(#loc320) | |
| %577 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %578 = torch.aten.view %574, %577 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc321) | |
| %579 = torch.aten.mm %578, %576 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc322) | |
| %580 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %581 = torch.aten._unsafe_view %579, %580 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc323) | |
| %582 = torch.aten.add.Tensor %501, %581, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc324) | |
| %583 = torch.aten.pow.Tensor_Scalar %582, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc325) | |
| %584 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %585 = torch.aten.mean.dim %583, %584, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc326) | |
| %586 = torch.aten.add.Scalar %585, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc327) | |
| %587 = torch.aten.rsqrt %586 : !torch.tensor -> !torch.tensor loc(#loc328) | |
| %588 = torch.aten.mul.Tensor %582, %587 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc329) | |
| %589 = torch.prim.GetAttr %arg0["_param_constant31"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %590 = torch.aten.mul.Tensor %589, %588 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc330) | |
| %591 = torch.prim.GetAttr %arg0["_param_constant32"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %592 = torch.aten.t %591 : !torch.tensor -> !torch.tensor loc(#loc331) | |
| %593 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %594 = torch.aten.view %590, %593 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc332) | |
| %595 = torch.aten.mm %594, %592 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc333) | |
| %596 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %597 = torch.aten._unsafe_view %595, %596 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc334) | |
| %598 = torch.aten.relu %597 : !torch.tensor -> !torch.tensor loc(#loc335) | |
| %599 = torch.prim.GetAttr %arg0["_param_constant33"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %600 = torch.aten.t %599 : !torch.tensor -> !torch.tensor loc(#loc336) | |
| %601 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %602 = torch.aten.view %598, %601 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc337) | |
| %603 = torch.aten.mm %602, %600 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc338) | |
| %604 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %605 = torch.aten._unsafe_view %603, %604 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc339) | |
| %606 = torch.aten.add.Tensor %582, %605, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc340) | |
| %607 = torch.aten.pow.Tensor_Scalar %606, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc341) | |
| %608 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %609 = torch.aten.mean.dim %607, %608, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc342) | |
| %610 = torch.aten.add.Scalar %609, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc343) | |
| %611 = torch.aten.rsqrt %610 : !torch.tensor -> !torch.tensor loc(#loc344) | |
| %612 = torch.aten.mul.Tensor %606, %611 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc345) | |
| %613 = torch.prim.GetAttr %arg0["_param_constant34"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %614 = torch.aten.mul.Tensor %613, %612 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc346) | |
| %615 = torch.prim.GetAttr %arg0["_param_constant35"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %616 = torch.aten.t %615 : !torch.tensor -> !torch.tensor loc(#loc347) | |
| %617 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %618 = torch.aten.view %614, %617 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc348) | |
| %619 = torch.aten.mm %618, %616 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc349) | |
| %620 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %621 = torch.aten._unsafe_view %619, %620 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc350) | |
| %622 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %623 = torch.aten.view %621, %622 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc351) | |
| %624 = torch.aten.transpose.int %623, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc352) | |
| %625 = torch.prim.GetAttr %arg0["_param_constant36"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %626 = torch.aten.t %625 : !torch.tensor -> !torch.tensor loc(#loc353) | |
| %627 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %628 = torch.aten.view %614, %627 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc354) | |
| %629 = torch.aten.mm %628, %626 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc355) | |
| %630 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %631 = torch.aten._unsafe_view %629, %630 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc356) | |
| %632 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %633 = torch.aten.view %631, %632 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc357) | |
| %634 = torch.aten.transpose.int %633, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc358) | |
| %635 = torch.prim.GetAttr %arg0["_param_constant37"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %636 = torch.aten.t %635 : !torch.tensor -> !torch.tensor loc(#loc359) | |
| %637 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %638 = torch.aten.view %614, %637 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc360) | |
| %639 = torch.aten.mm %638, %636 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc361) | |
| %640 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %641 = torch.aten._unsafe_view %639, %640 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc362) | |
| %642 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %643 = torch.aten.view %641, %642 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc363) | |
| %644 = torch.aten.transpose.int %643, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc364) | |
| %645 = torch.aten.transpose.int %634, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc365) | |
| %646 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %647 = torch.aten.expand %624, %646, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc366) | |
| %648 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %649 = torch.aten.view %647, %648 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc367) | |
| %650 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %651 = torch.aten.expand %645, %650, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc368) | |
| %652 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %653 = torch.aten.view %651, %652 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc369) | |
| %654 = torch.aten.bmm %649, %653 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc370) | |
| %655 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %656 = torch.aten._unsafe_view %654, %655 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc371) | |
| %657 = torch.aten.add_.Tensor %656, %236, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc372) | |
| %658 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %659 = torch.aten.amax %657, %658, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc373) | |
| %660 = torch.aten.sub.Tensor %657, %659, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc374) | |
| %661 = torch.aten.exp %660 : !torch.tensor -> !torch.tensor loc(#loc375) | |
| %662 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %663 = torch.aten.sum.dim_IntList %661, %662, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc376) | |
| %664 = torch.aten.div.Tensor %661, %663 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc377) | |
| %665 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %666 = torch.aten.expand %664, %665, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc378) | |
| %667 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %668 = torch.aten.view %666, %667 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc379) | |
| %669 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %670 = torch.aten.expand %644, %669, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc380) | |
| %671 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %672 = torch.aten.view %670, %671 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc381) | |
| %673 = torch.aten.bmm %668, %672 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc382) | |
| %674 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %675 = torch.aten._unsafe_view %673, %674 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc383) | |
| %676 = torch.aten.transpose.int %675, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc384) | |
| %677 = torch.aten.clone %676, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc385) | |
| %678 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %679 = torch.aten.view %677, %678 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc386) | |
| %680 = torch.prim.GetAttr %arg0["_param_constant38"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %681 = torch.aten.t %680 : !torch.tensor -> !torch.tensor loc(#loc387) | |
| %682 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %683 = torch.aten.view %679, %682 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc388) | |
| %684 = torch.aten.mm %683, %681 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc389) | |
| %685 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %686 = torch.aten._unsafe_view %684, %685 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc390) | |
| %687 = torch.aten.add.Tensor %606, %686, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc391) | |
| %688 = torch.aten.pow.Tensor_Scalar %687, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc392) | |
| %689 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %690 = torch.aten.mean.dim %688, %689, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc393) | |
| %691 = torch.aten.add.Scalar %690, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc394) | |
| %692 = torch.aten.rsqrt %691 : !torch.tensor -> !torch.tensor loc(#loc395) | |
| %693 = torch.aten.mul.Tensor %687, %692 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc396) | |
| %694 = torch.prim.GetAttr %arg0["_param_constant39"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %695 = torch.aten.mul.Tensor %694, %693 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc397) | |
| %696 = torch.prim.GetAttr %arg0["_param_constant40"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %697 = torch.aten.t %696 : !torch.tensor -> !torch.tensor loc(#loc398) | |
| %698 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %699 = torch.aten.view %695, %698 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc399) | |
| %700 = torch.aten.mm %699, %697 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc400) | |
| %701 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %702 = torch.aten._unsafe_view %700, %701 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc401) | |
| %703 = torch.aten.relu %702 : !torch.tensor -> !torch.tensor loc(#loc402) | |
| %704 = torch.prim.GetAttr %arg0["_param_constant41"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %705 = torch.aten.t %704 : !torch.tensor -> !torch.tensor loc(#loc403) | |
| %706 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %707 = torch.aten.view %703, %706 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc404) | |
| %708 = torch.aten.mm %707, %705 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc405) | |
| %709 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %710 = torch.aten._unsafe_view %708, %709 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc406) | |
| %711 = torch.aten.add.Tensor %687, %710, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc407) | |
| %712 = torch.aten.pow.Tensor_Scalar %711, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc408) | |
| %713 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %714 = torch.aten.mean.dim %712, %713, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc409) | |
| %715 = torch.aten.add.Scalar %714, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc410) | |
| %716 = torch.aten.rsqrt %715 : !torch.tensor -> !torch.tensor loc(#loc411) | |
| %717 = torch.aten.mul.Tensor %711, %716 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc412) | |
| %718 = torch.prim.GetAttr %arg0["_param_constant42"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %719 = torch.aten.mul.Tensor %718, %717 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc413) | |
| %720 = torch.prim.GetAttr %arg0["_param_constant43"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %721 = torch.aten.t %720 : !torch.tensor -> !torch.tensor loc(#loc414) | |
| %722 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %723 = torch.aten.view %719, %722 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc415) | |
| %724 = torch.aten.mm %723, %721 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc416) | |
| %725 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %726 = torch.aten._unsafe_view %724, %725 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc417) | |
| %727 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %728 = torch.aten.view %726, %727 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc418) | |
| %729 = torch.aten.transpose.int %728, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc419) | |
| %730 = torch.prim.GetAttr %arg0["_param_constant44"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %731 = torch.aten.t %730 : !torch.tensor -> !torch.tensor loc(#loc420) | |
| %732 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %733 = torch.aten.view %719, %732 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc421) | |
| %734 = torch.aten.mm %733, %731 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc422) | |
| %735 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %736 = torch.aten._unsafe_view %734, %735 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc423) | |
| %737 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %738 = torch.aten.view %736, %737 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc424) | |
| %739 = torch.aten.transpose.int %738, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc425) | |
| %740 = torch.prim.GetAttr %arg0["_param_constant45"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %741 = torch.aten.t %740 : !torch.tensor -> !torch.tensor loc(#loc426) | |
| %742 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %743 = torch.aten.view %719, %742 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc427) | |
| %744 = torch.aten.mm %743, %741 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc428) | |
| %745 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %746 = torch.aten._unsafe_view %744, %745 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc429) | |
| %747 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %748 = torch.aten.view %746, %747 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc430) | |
| %749 = torch.aten.transpose.int %748, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc431) | |
| %750 = torch.aten.transpose.int %739, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc432) | |
| %751 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %752 = torch.aten.expand %729, %751, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc433) | |
| %753 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %754 = torch.aten.view %752, %753 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc434) | |
| %755 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %756 = torch.aten.expand %750, %755, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc435) | |
| %757 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %758 = torch.aten.view %756, %757 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc436) | |
| %759 = torch.aten.bmm %754, %758 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc437) | |
| %760 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %761 = torch.aten._unsafe_view %759, %760 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc438) | |
| %762 = torch.aten.add_.Tensor %761, %236, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc439) | |
| %763 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %764 = torch.aten.amax %762, %763, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc440) | |
| %765 = torch.aten.sub.Tensor %762, %764, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc441) | |
| %766 = torch.aten.exp %765 : !torch.tensor -> !torch.tensor loc(#loc442) | |
| %767 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %768 = torch.aten.sum.dim_IntList %766, %767, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc443) | |
| %769 = torch.aten.div.Tensor %766, %768 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc444) | |
| %770 = torch.prim.ListConstruct %int1, %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %771 = torch.aten.expand %769, %770, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc445) | |
| %772 = torch.prim.ListConstruct %int8, %int15, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %773 = torch.aten.view %771, %772 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc446) | |
| %774 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %775 = torch.aten.expand %749, %774, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc447) | |
| %776 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %777 = torch.aten.view %775, %776 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc448) | |
| %778 = torch.aten.bmm %773, %777 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc449) | |
| %779 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %780 = torch.aten._unsafe_view %778, %779 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc450) | |
| %781 = torch.aten.transpose.int %780, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc451) | |
| %782 = torch.aten.clone %781, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc452) | |
| %783 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %784 = torch.aten.view %782, %783 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc453) | |
| %785 = torch.prim.GetAttr %arg0["_param_constant46"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %786 = torch.aten.t %785 : !torch.tensor -> !torch.tensor loc(#loc454) | |
| %787 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %788 = torch.aten.view %784, %787 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc455) | |
| %789 = torch.aten.mm %788, %786 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc456) | |
| %790 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %791 = torch.aten._unsafe_view %789, %790 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc457) | |
| %792 = torch.aten.add.Tensor %711, %791, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc458) | |
| %793 = torch.aten.pow.Tensor_Scalar %792, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc459) | |
| %794 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %795 = torch.aten.mean.dim %793, %794, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc460) | |
| %796 = torch.aten.add.Scalar %795, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc461) | |
| %797 = torch.aten.rsqrt %796 : !torch.tensor -> !torch.tensor loc(#loc462) | |
| %798 = torch.aten.mul.Tensor %792, %797 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc463) | |
| %799 = torch.prim.GetAttr %arg0["_param_constant47"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %800 = torch.aten.mul.Tensor %799, %798 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc464) | |
| %801 = torch.prim.GetAttr %arg0["_param_constant48"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %802 = torch.aten.t %801 : !torch.tensor -> !torch.tensor loc(#loc465) | |
| %803 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %804 = torch.aten.view %800, %803 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc466) | |
| %805 = torch.aten.mm %804, %802 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc467) | |
| %806 = torch.prim.ListConstruct %int1, %int15, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %807 = torch.aten._unsafe_view %805, %806 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc468) | |
| %808 = torch.aten.relu %807 : !torch.tensor -> !torch.tensor loc(#loc469) | |
| %809 = torch.prim.GetAttr %arg0["_param_constant49"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %810 = torch.aten.t %809 : !torch.tensor -> !torch.tensor loc(#loc470) | |
| %811 = torch.prim.ListConstruct %int15, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %812 = torch.aten.view %808, %811 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc471) | |
| %813 = torch.aten.mm %812, %810 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc472) | |
| %814 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %815 = torch.aten._unsafe_view %813, %814 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc473) | |
| %816 = torch.aten.add.Tensor %792, %815, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc474) | |
| %817 = torch.aten.pow.Tensor_Scalar %816, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc475) | |
| %818 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %819 = torch.aten.mean.dim %817, %818, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc476) | |
| %820 = torch.aten.add.Scalar %819, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc477) | |
| %821 = torch.aten.rsqrt %820 : !torch.tensor -> !torch.tensor loc(#loc478) | |
| %822 = torch.aten.mul.Tensor %816, %821 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc479) | |
| %823 = torch.prim.GetAttr %arg0["_param_constant50"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %824 = torch.aten.mul.Tensor %823, %822 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc480) | |
| %825 = torch.prim.ListConstruct %int-1, %int4 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %826 = torch.aten.view %144, %825 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc481) | |
| %827 = torch.prim.GetAttr %arg0["_param_constant0"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %828 = torch.aten.embedding %827, %826, %int-1, %false, %false : !torch.tensor, !torch.tensor, !torch.int, !torch.bool, !torch.bool -> !torch.tensor loc(#loc482) | |
| %829 = torch.prim.ListConstruct %int1, %int4 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %830 = torch.aten.ones %829, %none_1, %none_1, %cpu, %false : !torch.list<int>, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc483) | |
| %831 = torch.prim.ListConstruct %int1, %int15 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %832 = torch.aten.ones %831, %int4, %none_1, %cpu, %false : !torch.list<int>, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc484) | |
| %833 = torch.aten.arange %int4, %none_1, %none_1, %cpu, %false : !torch.int, !torch.none, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc485) | |
| %834 = torch.aten.unsqueeze %833, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc486) | |
| %835 = torch.aten.unsqueeze %834, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc487) | |
| %836 = torch.aten.slice.Tensor %835, %int2, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc488) | |
| %837 = torch.prim.ListConstruct %int1, %int4, %int1 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %838 = torch.aten.repeat %836, %837 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc489) | |
| %839 = torch.aten.unsqueeze %833, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc490) | |
| %840 = torch.aten.slice.Tensor %839, %int1, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc491) | |
| %841 = torch.aten.unsqueeze %840, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc492) | |
| %842 = torch.aten.le.Tensor %838, %841 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc493) | |
| %843 = torch.prims.convert_element_type %842, %int6 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc494) | |
| %844 = torch.aten.slice.Tensor %843, %int0, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc495) | |
| %845 = torch.aten.unsqueeze %844, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc496) | |
| %846 = torch.aten.slice.Tensor %845, %int2, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc497) | |
| %847 = torch.aten.slice.Tensor %846, %int3, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc498) | |
| %848 = torch.aten.slice.Tensor %830, %int0, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc499) | |
| %849 = torch.aten.unsqueeze %848, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc500) | |
| %850 = torch.aten.unsqueeze %849, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc501) | |
| %851 = torch.aten.slice.Tensor %850, %int3, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc502) | |
| %852 = torch.aten.mul.Tensor %847, %851 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc503) | |
| %853 = torch.aten.rsub.Scalar %852, %float1.000000e00, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc504) | |
| %854 = torch.aten.mul.Scalar %853, %float-3.402820e38 : !torch.tensor, !torch.float -> !torch.tensor loc(#loc505) | |
| %855 = torch.aten.slice.Tensor %832, %int0, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc506) | |
| %856 = torch.aten.unsqueeze %855, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc507) | |
| %857 = torch.aten.unsqueeze %856, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc508) | |
| %858 = torch.aten.slice.Tensor %857, %int3, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc509) | |
| %859 = torch.prims.convert_element_type %858, %int6 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc510) | |
| %860 = torch.aten.rsub.Scalar %859, %float1.000000e00, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc511) | |
| %861 = torch.aten.mul.Scalar %860, %float-3.402820e38 : !torch.tensor, !torch.float -> !torch.tensor loc(#loc512) | |
| %862 = torch.aten.pow.Tensor_Scalar %828, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc513) | |
| %863 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %864 = torch.aten.mean.dim %862, %863, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc514) | |
| %865 = torch.aten.add.Scalar %864, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc515) | |
| %866 = torch.aten.rsqrt %865 : !torch.tensor -> !torch.tensor loc(#loc516) | |
| %867 = torch.aten.mul.Tensor %828, %866 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc517) | |
| %868 = torch.prim.GetAttr %arg0["_param_constant51"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %869 = torch.aten.mul.Tensor %868, %867 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc518) | |
| %870 = torch.prim.GetAttr %arg0["_param_constant52"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %871 = torch.aten.t %870 : !torch.tensor -> !torch.tensor loc(#loc519) | |
| %872 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %873 = torch.aten.view %869, %872 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc520) | |
| %874 = torch.aten.mm %873, %871 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc521) | |
| %875 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %876 = torch.aten._unsafe_view %874, %875 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc522) | |
| %877 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %878 = torch.aten.view %876, %877 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc523) | |
| %879 = torch.aten.transpose.int %878, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc524) | |
| %880 = torch.prim.GetAttr %arg0["_param_constant53"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %881 = torch.aten.t %880 : !torch.tensor -> !torch.tensor loc(#loc525) | |
| %882 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %883 = torch.aten.view %869, %882 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc526) | |
| %884 = torch.aten.mm %883, %881 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc527) | |
| %885 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %886 = torch.aten._unsafe_view %884, %885 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc528) | |
| %887 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %888 = torch.aten.view %886, %887 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc529) | |
| %889 = torch.aten.transpose.int %888, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc530) | |
| %890 = torch.prim.GetAttr %arg0["_param_constant54"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %891 = torch.aten.t %890 : !torch.tensor -> !torch.tensor loc(#loc531) | |
| %892 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %893 = torch.aten.view %869, %892 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc532) | |
| %894 = torch.aten.mm %893, %891 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc533) | |
| %895 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %896 = torch.aten._unsafe_view %894, %895 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc534) | |
| %897 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %898 = torch.aten.view %896, %897 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc535) | |
| %899 = torch.aten.transpose.int %898, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc536) | |
| %900 = torch.aten.transpose.int %889, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc537) | |
| %901 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %902 = torch.aten.expand %879, %901, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc538) | |
| %903 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %904 = torch.aten.view %902, %903 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc539) | |
| %905 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %906 = torch.aten.expand %900, %905, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc540) | |
| %907 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %908 = torch.aten.view %906, %907 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc541) | |
| %909 = torch.aten.bmm %904, %908 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc542) | |
| %910 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %911 = torch.aten._unsafe_view %909, %910 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc543) | |
| %912 = torch.aten.arange %int4, %int4, %none_1, %cpu, %false : !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc544) | |
| %913 = torch.aten.slice.Tensor %912, %int0, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc545) | |
| %914 = torch.aten.unsqueeze %913, %int1 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc546) | |
| %915 = torch.aten.arange %int4, %int4, %none_1, %cpu, %false : !torch.int, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc547) | |
| %916 = torch.aten.unsqueeze %915, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc548) | |
| %917 = torch.aten.slice.Tensor %916, %int1, %int0, %int9223372036854775807, %int1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.tensor loc(#loc549) | |
| %918 = torch.aten.sub.Tensor %917, %914, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc550) | |
| %919 = torch.aten.zeros_like %918, %int4, %int0, %cpu, %false, %none_1 : !torch.tensor, !torch.int, !torch.int, !torch.Device, !torch.bool, !torch.none -> !torch.tensor loc(#loc551) | |
| %920 = torch.aten.minimum %918, %919 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc552) | |
| %921 = torch.aten.neg %920 : !torch.tensor -> !torch.tensor loc(#loc553) | |
| %922 = torch.aten.lt.Scalar %921, %int16 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc554) | |
| %923 = torch.prims.convert_element_type %921, %int6 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc555) | |
| %924 = torch.aten.div.Scalar %923, %int16 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc556) | |
| %925 = torch.aten.log %924 : !torch.tensor -> !torch.tensor loc(#loc557) | |
| %926 = torch.aten.div.Scalar %925, %float2.079440e00 : !torch.tensor, !torch.float -> !torch.tensor loc(#loc558) | |
| %927 = torch.aten.mul.Scalar %926, %int16 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc559) | |
| %928 = torch.prims.convert_element_type %927, %int4 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc560) | |
| %929 = torch.aten.add.Scalar %928, %int16, %int1 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc561) | |
| %930 = torch.aten.full_like %929, %int31, %int4, %int0, %cpu, %false, %none_1 : !torch.tensor, !torch.int, !torch.int, !torch.int, !torch.Device, !torch.bool, !torch.none -> !torch.tensor loc(#loc562) | |
| %931 = torch.aten.minimum %929, %930 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc563) | |
| %932 = torch.aten.where.self %922, %921, %931 : !torch.tensor, !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc564) | |
| %933 = torch.aten.add.Scalar %932, %int0, %int1 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc565) | |
| %934 = torch.prim.GetAttr %arg0["_param_constant55"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %935 = torch.aten.embedding %934, %933, %int-1, %false, %false : !torch.tensor, !torch.tensor, !torch.int, !torch.bool, !torch.bool -> !torch.tensor loc(#loc566) | |
| %936 = torch.prim.ListConstruct %int2, %int0, %int1 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %937 = torch.aten.permute %935, %936 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc567) | |
| %938 = torch.aten.unsqueeze %937, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc568) | |
| %939 = torch.aten.add.Tensor %938, %854, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc569) | |
| %940 = torch.aten.add_.Tensor %911, %939, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc570) | |
| %941 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %942 = torch.aten.amax %940, %941, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc571) | |
| %943 = torch.aten.sub.Tensor %940, %942, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc572) | |
| %944 = torch.aten.exp %943 : !torch.tensor -> !torch.tensor loc(#loc573) | |
| %945 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %946 = torch.aten.sum.dim_IntList %944, %945, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc574) | |
| %947 = torch.aten.div.Tensor %944, %946 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc575) | |
| %948 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %949 = torch.aten.expand %947, %948, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc576) | |
| %950 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %951 = torch.aten.view %949, %950 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc577) | |
| %952 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %953 = torch.aten.expand %899, %952, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc578) | |
| %954 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %955 = torch.aten.view %953, %954 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc579) | |
| %956 = torch.aten.bmm %951, %955 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc580) | |
| %957 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %958 = torch.aten._unsafe_view %956, %957 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc581) | |
| %959 = torch.aten.transpose.int %958, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc582) | |
| %960 = torch.aten.clone %959, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc583) | |
| %961 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %962 = torch.aten.view %960, %961 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc584) | |
| %963 = torch.prim.GetAttr %arg0["_param_constant56"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %964 = torch.aten.t %963 : !torch.tensor -> !torch.tensor loc(#loc585) | |
| %965 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %966 = torch.aten.view %962, %965 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc586) | |
| %967 = torch.aten.mm %966, %964 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc587) | |
| %968 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %969 = torch.aten._unsafe_view %967, %968 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc588) | |
| %970 = torch.aten.add.Tensor %828, %969, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc589) | |
| %971 = torch.aten.pow.Tensor_Scalar %970, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc590) | |
| %972 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %973 = torch.aten.mean.dim %971, %972, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc591) | |
| %974 = torch.aten.add.Scalar %973, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc592) | |
| %975 = torch.aten.rsqrt %974 : !torch.tensor -> !torch.tensor loc(#loc593) | |
| %976 = torch.aten.mul.Tensor %970, %975 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc594) | |
| %977 = torch.prim.GetAttr %arg0["_param_constant57"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %978 = torch.aten.mul.Tensor %977, %976 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc595) | |
| %979 = torch.prim.GetAttr %arg0["_param_constant58"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %980 = torch.aten.t %979 : !torch.tensor -> !torch.tensor loc(#loc596) | |
| %981 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %982 = torch.aten.view %978, %981 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc597) | |
| %983 = torch.aten.mm %982, %980 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc598) | |
| %984 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %985 = torch.aten._unsafe_view %983, %984 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc599) | |
| %986 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %987 = torch.aten.view %985, %986 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc600) | |
| %988 = torch.aten.transpose.int %987, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc601) | |
| %989 = torch.prim.GetAttr %arg0["_param_constant59"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %990 = torch.aten.t %989 : !torch.tensor -> !torch.tensor loc(#loc602) | |
| %991 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %992 = torch.aten.view %824, %991 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc603) | |
| %993 = torch.aten.mm %992, %990 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc604) | |
| %994 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %995 = torch.aten._unsafe_view %993, %994 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc605) | |
| %996 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %997 = torch.aten.view %995, %996 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc606) | |
| %998 = torch.aten.transpose.int %997, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc607) | |
| %999 = torch.prim.GetAttr %arg0["_param_constant60"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1000 = torch.aten.t %999 : !torch.tensor -> !torch.tensor loc(#loc608) | |
| %1001 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1002 = torch.aten.view %824, %1001 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc609) | |
| %1003 = torch.aten.mm %1002, %1000 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc610) | |
| %1004 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1005 = torch.aten._unsafe_view %1003, %1004 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc611) | |
| %1006 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1007 = torch.aten.view %1005, %1006 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc612) | |
| %1008 = torch.aten.transpose.int %1007, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc613) | |
| %1009 = torch.aten.transpose.int %998, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc614) | |
| %1010 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1011 = torch.aten.expand %988, %1010, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc615) | |
| %1012 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1013 = torch.aten.view %1011, %1012 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc616) | |
| %1014 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1015 = torch.aten.expand %1009, %1014, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc617) | |
| %1016 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1017 = torch.aten.view %1015, %1016 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc618) | |
| %1018 = torch.aten.bmm %1013, %1017 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc619) | |
| %1019 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1020 = torch.aten._unsafe_view %1018, %1019 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc620) | |
| %1021 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1022 = torch.aten.zeros %1021, %int6, %none_1, %cpu, %false : !torch.list<int>, !torch.int, !torch.none, !torch.Device, !torch.bool -> !torch.tensor loc(#loc621) | |
| %1023 = torch.aten.add.Tensor %1022, %861, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc622) | |
| %1024 = torch.aten.add_.Tensor %1020, %1023, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc623) | |
| %1025 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1026 = torch.aten.amax %1024, %1025, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc624) | |
| %1027 = torch.aten.sub.Tensor %1024, %1026, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc625) | |
| %1028 = torch.aten.exp %1027 : !torch.tensor -> !torch.tensor loc(#loc626) | |
| %1029 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1030 = torch.aten.sum.dim_IntList %1028, %1029, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc627) | |
| %1031 = torch.aten.div.Tensor %1028, %1030 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc628) | |
| %1032 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1033 = torch.aten.expand %1031, %1032, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc629) | |
| %1034 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1035 = torch.aten.view %1033, %1034 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc630) | |
| %1036 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1037 = torch.aten.expand %1008, %1036, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc631) | |
| %1038 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1039 = torch.aten.view %1037, %1038 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc632) | |
| %1040 = torch.aten.bmm %1035, %1039 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc633) | |
| %1041 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1042 = torch.aten._unsafe_view %1040, %1041 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc634) | |
| %1043 = torch.aten.transpose.int %1042, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc635) | |
| %1044 = torch.aten.clone %1043, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc636) | |
| %1045 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1046 = torch.aten.view %1044, %1045 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc637) | |
| %1047 = torch.prim.GetAttr %arg0["_param_constant61"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1048 = torch.aten.t %1047 : !torch.tensor -> !torch.tensor loc(#loc638) | |
| %1049 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1050 = torch.aten.view %1046, %1049 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc639) | |
| %1051 = torch.aten.mm %1050, %1048 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc640) | |
| %1052 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1053 = torch.aten._unsafe_view %1051, %1052 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc641) | |
| %1054 = torch.aten.add.Tensor %970, %1053, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc642) | |
| %1055 = torch.aten.pow.Tensor_Scalar %1054, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc643) | |
| %1056 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1057 = torch.aten.mean.dim %1055, %1056, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc644) | |
| %1058 = torch.aten.add.Scalar %1057, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc645) | |
| %1059 = torch.aten.rsqrt %1058 : !torch.tensor -> !torch.tensor loc(#loc646) | |
| %1060 = torch.aten.mul.Tensor %1054, %1059 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc647) | |
| %1061 = torch.prim.GetAttr %arg0["_param_constant62"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1062 = torch.aten.mul.Tensor %1061, %1060 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc648) | |
| %1063 = torch.prim.GetAttr %arg0["_param_constant63"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1064 = torch.aten.t %1063 : !torch.tensor -> !torch.tensor loc(#loc649) | |
| %1065 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1066 = torch.aten.view %1062, %1065 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc650) | |
| %1067 = torch.aten.mm %1066, %1064 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc651) | |
| %1068 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1069 = torch.aten._unsafe_view %1067, %1068 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc652) | |
| %1070 = torch.aten.relu %1069 : !torch.tensor -> !torch.tensor loc(#loc653) | |
| %1071 = torch.prim.GetAttr %arg0["_param_constant64"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1072 = torch.aten.t %1071 : !torch.tensor -> !torch.tensor loc(#loc654) | |
| %1073 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1074 = torch.aten.view %1070, %1073 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc655) | |
| %1075 = torch.aten.mm %1074, %1072 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc656) | |
| %1076 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1077 = torch.aten._unsafe_view %1075, %1076 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc657) | |
| %1078 = torch.aten.add.Tensor %1054, %1077, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc658) | |
| %1079 = torch.aten.pow.Tensor_Scalar %1078, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc659) | |
| %1080 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1081 = torch.aten.mean.dim %1079, %1080, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc660) | |
| %1082 = torch.aten.add.Scalar %1081, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc661) | |
| %1083 = torch.aten.rsqrt %1082 : !torch.tensor -> !torch.tensor loc(#loc662) | |
| %1084 = torch.aten.mul.Tensor %1078, %1083 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc663) | |
| %1085 = torch.prim.GetAttr %arg0["_param_constant65"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1086 = torch.aten.mul.Tensor %1085, %1084 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc664) | |
| %1087 = torch.prim.GetAttr %arg0["_param_constant66"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1088 = torch.aten.t %1087 : !torch.tensor -> !torch.tensor loc(#loc665) | |
| %1089 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1090 = torch.aten.view %1086, %1089 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc666) | |
| %1091 = torch.aten.mm %1090, %1088 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc667) | |
| %1092 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1093 = torch.aten._unsafe_view %1091, %1092 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc668) | |
| %1094 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1095 = torch.aten.view %1093, %1094 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc669) | |
| %1096 = torch.aten.transpose.int %1095, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc670) | |
| %1097 = torch.prim.GetAttr %arg0["_param_constant67"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1098 = torch.aten.t %1097 : !torch.tensor -> !torch.tensor loc(#loc671) | |
| %1099 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1100 = torch.aten.view %1086, %1099 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc672) | |
| %1101 = torch.aten.mm %1100, %1098 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc673) | |
| %1102 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1103 = torch.aten._unsafe_view %1101, %1102 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc674) | |
| %1104 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1105 = torch.aten.view %1103, %1104 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc675) | |
| %1106 = torch.aten.transpose.int %1105, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc676) | |
| %1107 = torch.prim.GetAttr %arg0["_param_constant68"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1108 = torch.aten.t %1107 : !torch.tensor -> !torch.tensor loc(#loc677) | |
| %1109 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1110 = torch.aten.view %1086, %1109 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc678) | |
| %1111 = torch.aten.mm %1110, %1108 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc679) | |
| %1112 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1113 = torch.aten._unsafe_view %1111, %1112 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc680) | |
| %1114 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1115 = torch.aten.view %1113, %1114 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc681) | |
| %1116 = torch.aten.transpose.int %1115, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc682) | |
| %1117 = torch.aten.transpose.int %1106, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc683) | |
| %1118 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1119 = torch.aten.expand %1096, %1118, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc684) | |
| %1120 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1121 = torch.aten.view %1119, %1120 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc685) | |
| %1122 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1123 = torch.aten.expand %1117, %1122, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc686) | |
| %1124 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1125 = torch.aten.view %1123, %1124 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc687) | |
| %1126 = torch.aten.bmm %1121, %1125 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc688) | |
| %1127 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1128 = torch.aten._unsafe_view %1126, %1127 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc689) | |
| %1129 = torch.aten.add_.Tensor %1128, %939, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc690) | |
| %1130 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1131 = torch.aten.amax %1129, %1130, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc691) | |
| %1132 = torch.aten.sub.Tensor %1129, %1131, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc692) | |
| %1133 = torch.aten.exp %1132 : !torch.tensor -> !torch.tensor loc(#loc693) | |
| %1134 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1135 = torch.aten.sum.dim_IntList %1133, %1134, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc694) | |
| %1136 = torch.aten.div.Tensor %1133, %1135 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc695) | |
| %1137 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1138 = torch.aten.expand %1136, %1137, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc696) | |
| %1139 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1140 = torch.aten.view %1138, %1139 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc697) | |
| %1141 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1142 = torch.aten.expand %1116, %1141, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc698) | |
| %1143 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1144 = torch.aten.view %1142, %1143 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc699) | |
| %1145 = torch.aten.bmm %1140, %1144 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc700) | |
| %1146 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1147 = torch.aten._unsafe_view %1145, %1146 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc701) | |
| %1148 = torch.aten.transpose.int %1147, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc702) | |
| %1149 = torch.aten.clone %1148, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc703) | |
| %1150 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1151 = torch.aten.view %1149, %1150 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc704) | |
| %1152 = torch.prim.GetAttr %arg0["_param_constant69"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1153 = torch.aten.t %1152 : !torch.tensor -> !torch.tensor loc(#loc705) | |
| %1154 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1155 = torch.aten.view %1151, %1154 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc706) | |
| %1156 = torch.aten.mm %1155, %1153 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc707) | |
| %1157 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1158 = torch.aten._unsafe_view %1156, %1157 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc708) | |
| %1159 = torch.aten.add.Tensor %1078, %1158, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc709) | |
| %1160 = torch.aten.pow.Tensor_Scalar %1159, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc710) | |
| %1161 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1162 = torch.aten.mean.dim %1160, %1161, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc711) | |
| %1163 = torch.aten.add.Scalar %1162, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc712) | |
| %1164 = torch.aten.rsqrt %1163 : !torch.tensor -> !torch.tensor loc(#loc713) | |
| %1165 = torch.aten.mul.Tensor %1159, %1164 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc714) | |
| %1166 = torch.prim.GetAttr %arg0["_param_constant70"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1167 = torch.aten.mul.Tensor %1166, %1165 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc715) | |
| %1168 = torch.prim.GetAttr %arg0["_param_constant71"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1169 = torch.aten.t %1168 : !torch.tensor -> !torch.tensor loc(#loc716) | |
| %1170 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1171 = torch.aten.view %1167, %1170 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc717) | |
| %1172 = torch.aten.mm %1171, %1169 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc718) | |
| %1173 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1174 = torch.aten._unsafe_view %1172, %1173 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc719) | |
| %1175 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1176 = torch.aten.view %1174, %1175 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc720) | |
| %1177 = torch.aten.transpose.int %1176, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc721) | |
| %1178 = torch.prim.GetAttr %arg0["_param_constant72"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1179 = torch.aten.t %1178 : !torch.tensor -> !torch.tensor loc(#loc722) | |
| %1180 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1181 = torch.aten.view %824, %1180 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc723) | |
| %1182 = torch.aten.mm %1181, %1179 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc724) | |
| %1183 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1184 = torch.aten._unsafe_view %1182, %1183 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc725) | |
| %1185 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1186 = torch.aten.view %1184, %1185 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc726) | |
| %1187 = torch.aten.transpose.int %1186, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc727) | |
| %1188 = torch.prim.GetAttr %arg0["_param_constant73"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1189 = torch.aten.t %1188 : !torch.tensor -> !torch.tensor loc(#loc728) | |
| %1190 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1191 = torch.aten.view %824, %1190 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc729) | |
| %1192 = torch.aten.mm %1191, %1189 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc730) | |
| %1193 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1194 = torch.aten._unsafe_view %1192, %1193 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc731) | |
| %1195 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1196 = torch.aten.view %1194, %1195 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc732) | |
| %1197 = torch.aten.transpose.int %1196, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc733) | |
| %1198 = torch.aten.transpose.int %1187, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc734) | |
| %1199 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1200 = torch.aten.expand %1177, %1199, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc735) | |
| %1201 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1202 = torch.aten.view %1200, %1201 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc736) | |
| %1203 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1204 = torch.aten.expand %1198, %1203, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc737) | |
| %1205 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1206 = torch.aten.view %1204, %1205 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc738) | |
| %1207 = torch.aten.bmm %1202, %1206 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc739) | |
| %1208 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1209 = torch.aten._unsafe_view %1207, %1208 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc740) | |
| %1210 = torch.aten.add_.Tensor %1209, %1023, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc741) | |
| %1211 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1212 = torch.aten.amax %1210, %1211, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc742) | |
| %1213 = torch.aten.sub.Tensor %1210, %1212, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc743) | |
| %1214 = torch.aten.exp %1213 : !torch.tensor -> !torch.tensor loc(#loc744) | |
| %1215 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1216 = torch.aten.sum.dim_IntList %1214, %1215, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc745) | |
| %1217 = torch.aten.div.Tensor %1214, %1216 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc746) | |
| %1218 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1219 = torch.aten.expand %1217, %1218, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc747) | |
| %1220 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1221 = torch.aten.view %1219, %1220 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc748) | |
| %1222 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1223 = torch.aten.expand %1197, %1222, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc749) | |
| %1224 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1225 = torch.aten.view %1223, %1224 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc750) | |
| %1226 = torch.aten.bmm %1221, %1225 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc751) | |
| %1227 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1228 = torch.aten._unsafe_view %1226, %1227 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc752) | |
| %1229 = torch.aten.transpose.int %1228, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc753) | |
| %1230 = torch.aten.clone %1229, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc754) | |
| %1231 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1232 = torch.aten.view %1230, %1231 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc755) | |
| %1233 = torch.prim.GetAttr %arg0["_param_constant74"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1234 = torch.aten.t %1233 : !torch.tensor -> !torch.tensor loc(#loc756) | |
| %1235 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1236 = torch.aten.view %1232, %1235 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc757) | |
| %1237 = torch.aten.mm %1236, %1234 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc758) | |
| %1238 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1239 = torch.aten._unsafe_view %1237, %1238 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc759) | |
| %1240 = torch.aten.add.Tensor %1159, %1239, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc760) | |
| %1241 = torch.aten.pow.Tensor_Scalar %1240, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc761) | |
| %1242 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1243 = torch.aten.mean.dim %1241, %1242, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc762) | |
| %1244 = torch.aten.add.Scalar %1243, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc763) | |
| %1245 = torch.aten.rsqrt %1244 : !torch.tensor -> !torch.tensor loc(#loc764) | |
| %1246 = torch.aten.mul.Tensor %1240, %1245 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc765) | |
| %1247 = torch.prim.GetAttr %arg0["_param_constant75"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1248 = torch.aten.mul.Tensor %1247, %1246 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc766) | |
| %1249 = torch.prim.GetAttr %arg0["_param_constant76"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1250 = torch.aten.t %1249 : !torch.tensor -> !torch.tensor loc(#loc767) | |
| %1251 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1252 = torch.aten.view %1248, %1251 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc768) | |
| %1253 = torch.aten.mm %1252, %1250 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc769) | |
| %1254 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1255 = torch.aten._unsafe_view %1253, %1254 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc770) | |
| %1256 = torch.aten.relu %1255 : !torch.tensor -> !torch.tensor loc(#loc771) | |
| %1257 = torch.prim.GetAttr %arg0["_param_constant77"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1258 = torch.aten.t %1257 : !torch.tensor -> !torch.tensor loc(#loc772) | |
| %1259 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1260 = torch.aten.view %1256, %1259 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc773) | |
| %1261 = torch.aten.mm %1260, %1258 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc774) | |
| %1262 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1263 = torch.aten._unsafe_view %1261, %1262 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc775) | |
| %1264 = torch.aten.add.Tensor %1240, %1263, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc776) | |
| %1265 = torch.aten.pow.Tensor_Scalar %1264, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc777) | |
| %1266 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1267 = torch.aten.mean.dim %1265, %1266, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc778) | |
| %1268 = torch.aten.add.Scalar %1267, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc779) | |
| %1269 = torch.aten.rsqrt %1268 : !torch.tensor -> !torch.tensor loc(#loc780) | |
| %1270 = torch.aten.mul.Tensor %1264, %1269 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc781) | |
| %1271 = torch.prim.GetAttr %arg0["_param_constant78"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1272 = torch.aten.mul.Tensor %1271, %1270 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc782) | |
| %1273 = torch.prim.GetAttr %arg0["_param_constant79"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1274 = torch.aten.t %1273 : !torch.tensor -> !torch.tensor loc(#loc783) | |
| %1275 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1276 = torch.aten.view %1272, %1275 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc784) | |
| %1277 = torch.aten.mm %1276, %1274 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc785) | |
| %1278 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1279 = torch.aten._unsafe_view %1277, %1278 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc786) | |
| %1280 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1281 = torch.aten.view %1279, %1280 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc787) | |
| %1282 = torch.aten.transpose.int %1281, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc788) | |
| %1283 = torch.prim.GetAttr %arg0["_param_constant80"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1284 = torch.aten.t %1283 : !torch.tensor -> !torch.tensor loc(#loc789) | |
| %1285 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1286 = torch.aten.view %1272, %1285 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc790) | |
| %1287 = torch.aten.mm %1286, %1284 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc791) | |
| %1288 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1289 = torch.aten._unsafe_view %1287, %1288 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc792) | |
| %1290 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1291 = torch.aten.view %1289, %1290 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc793) | |
| %1292 = torch.aten.transpose.int %1291, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc794) | |
| %1293 = torch.prim.GetAttr %arg0["_param_constant81"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1294 = torch.aten.t %1293 : !torch.tensor -> !torch.tensor loc(#loc795) | |
| %1295 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1296 = torch.aten.view %1272, %1295 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc796) | |
| %1297 = torch.aten.mm %1296, %1294 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc797) | |
| %1298 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1299 = torch.aten._unsafe_view %1297, %1298 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc798) | |
| %1300 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1301 = torch.aten.view %1299, %1300 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc799) | |
| %1302 = torch.aten.transpose.int %1301, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc800) | |
| %1303 = torch.aten.transpose.int %1292, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc801) | |
| %1304 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1305 = torch.aten.expand %1282, %1304, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc802) | |
| %1306 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1307 = torch.aten.view %1305, %1306 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc803) | |
| %1308 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1309 = torch.aten.expand %1303, %1308, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc804) | |
| %1310 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1311 = torch.aten.view %1309, %1310 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc805) | |
| %1312 = torch.aten.bmm %1307, %1311 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc806) | |
| %1313 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1314 = torch.aten._unsafe_view %1312, %1313 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc807) | |
| %1315 = torch.aten.add_.Tensor %1314, %939, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc808) | |
| %1316 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1317 = torch.aten.amax %1315, %1316, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc809) | |
| %1318 = torch.aten.sub.Tensor %1315, %1317, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc810) | |
| %1319 = torch.aten.exp %1318 : !torch.tensor -> !torch.tensor loc(#loc811) | |
| %1320 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1321 = torch.aten.sum.dim_IntList %1319, %1320, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc812) | |
| %1322 = torch.aten.div.Tensor %1319, %1321 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc813) | |
| %1323 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1324 = torch.aten.expand %1322, %1323, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc814) | |
| %1325 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1326 = torch.aten.view %1324, %1325 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc815) | |
| %1327 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1328 = torch.aten.expand %1302, %1327, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc816) | |
| %1329 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1330 = torch.aten.view %1328, %1329 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc817) | |
| %1331 = torch.aten.bmm %1326, %1330 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc818) | |
| %1332 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1333 = torch.aten._unsafe_view %1331, %1332 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc819) | |
| %1334 = torch.aten.transpose.int %1333, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc820) | |
| %1335 = torch.aten.clone %1334, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc821) | |
| %1336 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1337 = torch.aten.view %1335, %1336 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc822) | |
| %1338 = torch.prim.GetAttr %arg0["_param_constant82"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1339 = torch.aten.t %1338 : !torch.tensor -> !torch.tensor loc(#loc823) | |
| %1340 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1341 = torch.aten.view %1337, %1340 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc824) | |
| %1342 = torch.aten.mm %1341, %1339 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc825) | |
| %1343 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1344 = torch.aten._unsafe_view %1342, %1343 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc826) | |
| %1345 = torch.aten.add.Tensor %1264, %1344, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc827) | |
| %1346 = torch.aten.pow.Tensor_Scalar %1345, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc828) | |
| %1347 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1348 = torch.aten.mean.dim %1346, %1347, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc829) | |
| %1349 = torch.aten.add.Scalar %1348, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc830) | |
| %1350 = torch.aten.rsqrt %1349 : !torch.tensor -> !torch.tensor loc(#loc831) | |
| %1351 = torch.aten.mul.Tensor %1345, %1350 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc832) | |
| %1352 = torch.prim.GetAttr %arg0["_param_constant83"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1353 = torch.aten.mul.Tensor %1352, %1351 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc833) | |
| %1354 = torch.prim.GetAttr %arg0["_param_constant84"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1355 = torch.aten.t %1354 : !torch.tensor -> !torch.tensor loc(#loc834) | |
| %1356 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1357 = torch.aten.view %1353, %1356 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc835) | |
| %1358 = torch.aten.mm %1357, %1355 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc836) | |
| %1359 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1360 = torch.aten._unsafe_view %1358, %1359 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc837) | |
| %1361 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1362 = torch.aten.view %1360, %1361 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc838) | |
| %1363 = torch.aten.transpose.int %1362, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc839) | |
| %1364 = torch.prim.GetAttr %arg0["_param_constant85"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1365 = torch.aten.t %1364 : !torch.tensor -> !torch.tensor loc(#loc840) | |
| %1366 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1367 = torch.aten.view %824, %1366 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc841) | |
| %1368 = torch.aten.mm %1367, %1365 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc842) | |
| %1369 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1370 = torch.aten._unsafe_view %1368, %1369 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc843) | |
| %1371 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1372 = torch.aten.view %1370, %1371 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc844) | |
| %1373 = torch.aten.transpose.int %1372, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc845) | |
| %1374 = torch.prim.GetAttr %arg0["_param_constant86"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1375 = torch.aten.t %1374 : !torch.tensor -> !torch.tensor loc(#loc846) | |
| %1376 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1377 = torch.aten.view %824, %1376 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc847) | |
| %1378 = torch.aten.mm %1377, %1375 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc848) | |
| %1379 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1380 = torch.aten._unsafe_view %1378, %1379 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc849) | |
| %1381 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1382 = torch.aten.view %1380, %1381 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc850) | |
| %1383 = torch.aten.transpose.int %1382, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc851) | |
| %1384 = torch.aten.transpose.int %1373, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc852) | |
| %1385 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1386 = torch.aten.expand %1363, %1385, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc853) | |
| %1387 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1388 = torch.aten.view %1386, %1387 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc854) | |
| %1389 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1390 = torch.aten.expand %1384, %1389, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc855) | |
| %1391 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1392 = torch.aten.view %1390, %1391 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc856) | |
| %1393 = torch.aten.bmm %1388, %1392 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc857) | |
| %1394 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1395 = torch.aten._unsafe_view %1393, %1394 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc858) | |
| %1396 = torch.aten.add_.Tensor %1395, %1023, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc859) | |
| %1397 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1398 = torch.aten.amax %1396, %1397, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc860) | |
| %1399 = torch.aten.sub.Tensor %1396, %1398, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc861) | |
| %1400 = torch.aten.exp %1399 : !torch.tensor -> !torch.tensor loc(#loc862) | |
| %1401 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1402 = torch.aten.sum.dim_IntList %1400, %1401, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc863) | |
| %1403 = torch.aten.div.Tensor %1400, %1402 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc864) | |
| %1404 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1405 = torch.aten.expand %1403, %1404, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc865) | |
| %1406 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1407 = torch.aten.view %1405, %1406 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc866) | |
| %1408 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1409 = torch.aten.expand %1383, %1408, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc867) | |
| %1410 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1411 = torch.aten.view %1409, %1410 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc868) | |
| %1412 = torch.aten.bmm %1407, %1411 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc869) | |
| %1413 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1414 = torch.aten._unsafe_view %1412, %1413 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc870) | |
| %1415 = torch.aten.transpose.int %1414, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc871) | |
| %1416 = torch.aten.clone %1415, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc872) | |
| %1417 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1418 = torch.aten.view %1416, %1417 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc873) | |
| %1419 = torch.prim.GetAttr %arg0["_param_constant87"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1420 = torch.aten.t %1419 : !torch.tensor -> !torch.tensor loc(#loc874) | |
| %1421 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1422 = torch.aten.view %1418, %1421 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc875) | |
| %1423 = torch.aten.mm %1422, %1420 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc876) | |
| %1424 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1425 = torch.aten._unsafe_view %1423, %1424 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc877) | |
| %1426 = torch.aten.add.Tensor %1345, %1425, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc878) | |
| %1427 = torch.aten.pow.Tensor_Scalar %1426, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc879) | |
| %1428 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1429 = torch.aten.mean.dim %1427, %1428, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc880) | |
| %1430 = torch.aten.add.Scalar %1429, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc881) | |
| %1431 = torch.aten.rsqrt %1430 : !torch.tensor -> !torch.tensor loc(#loc882) | |
| %1432 = torch.aten.mul.Tensor %1426, %1431 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc883) | |
| %1433 = torch.prim.GetAttr %arg0["_param_constant88"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1434 = torch.aten.mul.Tensor %1433, %1432 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc884) | |
| %1435 = torch.prim.GetAttr %arg0["_param_constant89"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1436 = torch.aten.t %1435 : !torch.tensor -> !torch.tensor loc(#loc885) | |
| %1437 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1438 = torch.aten.view %1434, %1437 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc886) | |
| %1439 = torch.aten.mm %1438, %1436 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc887) | |
| %1440 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1441 = torch.aten._unsafe_view %1439, %1440 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc888) | |
| %1442 = torch.aten.relu %1441 : !torch.tensor -> !torch.tensor loc(#loc889) | |
| %1443 = torch.prim.GetAttr %arg0["_param_constant90"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1444 = torch.aten.t %1443 : !torch.tensor -> !torch.tensor loc(#loc890) | |
| %1445 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1446 = torch.aten.view %1442, %1445 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc891) | |
| %1447 = torch.aten.mm %1446, %1444 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc892) | |
| %1448 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1449 = torch.aten._unsafe_view %1447, %1448 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc893) | |
| %1450 = torch.aten.add.Tensor %1426, %1449, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc894) | |
| %1451 = torch.aten.pow.Tensor_Scalar %1450, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc895) | |
| %1452 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1453 = torch.aten.mean.dim %1451, %1452, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc896) | |
| %1454 = torch.aten.add.Scalar %1453, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc897) | |
| %1455 = torch.aten.rsqrt %1454 : !torch.tensor -> !torch.tensor loc(#loc898) | |
| %1456 = torch.aten.mul.Tensor %1450, %1455 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc899) | |
| %1457 = torch.prim.GetAttr %arg0["_param_constant91"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1458 = torch.aten.mul.Tensor %1457, %1456 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc900) | |
| %1459 = torch.prim.GetAttr %arg0["_param_constant92"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1460 = torch.aten.t %1459 : !torch.tensor -> !torch.tensor loc(#loc901) | |
| %1461 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1462 = torch.aten.view %1458, %1461 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc902) | |
| %1463 = torch.aten.mm %1462, %1460 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc903) | |
| %1464 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1465 = torch.aten._unsafe_view %1463, %1464 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc904) | |
| %1466 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1467 = torch.aten.view %1465, %1466 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc905) | |
| %1468 = torch.aten.transpose.int %1467, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc906) | |
| %1469 = torch.prim.GetAttr %arg0["_param_constant93"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1470 = torch.aten.t %1469 : !torch.tensor -> !torch.tensor loc(#loc907) | |
| %1471 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1472 = torch.aten.view %1458, %1471 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc908) | |
| %1473 = torch.aten.mm %1472, %1470 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc909) | |
| %1474 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1475 = torch.aten._unsafe_view %1473, %1474 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc910) | |
| %1476 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1477 = torch.aten.view %1475, %1476 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc911) | |
| %1478 = torch.aten.transpose.int %1477, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc912) | |
| %1479 = torch.prim.GetAttr %arg0["_param_constant94"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1480 = torch.aten.t %1479 : !torch.tensor -> !torch.tensor loc(#loc913) | |
| %1481 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1482 = torch.aten.view %1458, %1481 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc914) | |
| %1483 = torch.aten.mm %1482, %1480 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc915) | |
| %1484 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1485 = torch.aten._unsafe_view %1483, %1484 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc916) | |
| %1486 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1487 = torch.aten.view %1485, %1486 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc917) | |
| %1488 = torch.aten.transpose.int %1487, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc918) | |
| %1489 = torch.aten.transpose.int %1478, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc919) | |
| %1490 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1491 = torch.aten.expand %1468, %1490, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc920) | |
| %1492 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1493 = torch.aten.view %1491, %1492 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc921) | |
| %1494 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1495 = torch.aten.expand %1489, %1494, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc922) | |
| %1496 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1497 = torch.aten.view %1495, %1496 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc923) | |
| %1498 = torch.aten.bmm %1493, %1497 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc924) | |
| %1499 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1500 = torch.aten._unsafe_view %1498, %1499 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc925) | |
| %1501 = torch.aten.add_.Tensor %1500, %939, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc926) | |
| %1502 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1503 = torch.aten.amax %1501, %1502, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc927) | |
| %1504 = torch.aten.sub.Tensor %1501, %1503, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc928) | |
| %1505 = torch.aten.exp %1504 : !torch.tensor -> !torch.tensor loc(#loc929) | |
| %1506 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1507 = torch.aten.sum.dim_IntList %1505, %1506, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc930) | |
| %1508 = torch.aten.div.Tensor %1505, %1507 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc931) | |
| %1509 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1510 = torch.aten.expand %1508, %1509, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc932) | |
| %1511 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1512 = torch.aten.view %1510, %1511 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc933) | |
| %1513 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1514 = torch.aten.expand %1488, %1513, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc934) | |
| %1515 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1516 = torch.aten.view %1514, %1515 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc935) | |
| %1517 = torch.aten.bmm %1512, %1516 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc936) | |
| %1518 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1519 = torch.aten._unsafe_view %1517, %1518 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc937) | |
| %1520 = torch.aten.transpose.int %1519, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc938) | |
| %1521 = torch.aten.clone %1520, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc939) | |
| %1522 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1523 = torch.aten.view %1521, %1522 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc940) | |
| %1524 = torch.prim.GetAttr %arg0["_param_constant95"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1525 = torch.aten.t %1524 : !torch.tensor -> !torch.tensor loc(#loc941) | |
| %1526 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1527 = torch.aten.view %1523, %1526 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc942) | |
| %1528 = torch.aten.mm %1527, %1525 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc943) | |
| %1529 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1530 = torch.aten._unsafe_view %1528, %1529 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc944) | |
| %1531 = torch.aten.add.Tensor %1450, %1530, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc945) | |
| %1532 = torch.aten.pow.Tensor_Scalar %1531, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc946) | |
| %1533 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1534 = torch.aten.mean.dim %1532, %1533, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc947) | |
| %1535 = torch.aten.add.Scalar %1534, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc948) | |
| %1536 = torch.aten.rsqrt %1535 : !torch.tensor -> !torch.tensor loc(#loc949) | |
| %1537 = torch.aten.mul.Tensor %1531, %1536 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc950) | |
| %1538 = torch.prim.GetAttr %arg0["_param_constant96"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1539 = torch.aten.mul.Tensor %1538, %1537 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc951) | |
| %1540 = torch.prim.GetAttr %arg0["_param_constant97"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1541 = torch.aten.t %1540 : !torch.tensor -> !torch.tensor loc(#loc952) | |
| %1542 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1543 = torch.aten.view %1539, %1542 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc953) | |
| %1544 = torch.aten.mm %1543, %1541 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc954) | |
| %1545 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1546 = torch.aten._unsafe_view %1544, %1545 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc955) | |
| %1547 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1548 = torch.aten.view %1546, %1547 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc956) | |
| %1549 = torch.aten.transpose.int %1548, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc957) | |
| %1550 = torch.prim.GetAttr %arg0["_param_constant98"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1551 = torch.aten.t %1550 : !torch.tensor -> !torch.tensor loc(#loc958) | |
| %1552 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1553 = torch.aten.view %824, %1552 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc959) | |
| %1554 = torch.aten.mm %1553, %1551 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc960) | |
| %1555 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1556 = torch.aten._unsafe_view %1554, %1555 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc961) | |
| %1557 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1558 = torch.aten.view %1556, %1557 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc962) | |
| %1559 = torch.aten.transpose.int %1558, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc963) | |
| %1560 = torch.prim.GetAttr %arg0["_param_constant99"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1561 = torch.aten.t %1560 : !torch.tensor -> !torch.tensor loc(#loc964) | |
| %1562 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1563 = torch.aten.view %824, %1562 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc965) | |
| %1564 = torch.aten.mm %1563, %1561 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc966) | |
| %1565 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1566 = torch.aten._unsafe_view %1564, %1565 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc967) | |
| %1567 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1568 = torch.aten.view %1566, %1567 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc968) | |
| %1569 = torch.aten.transpose.int %1568, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc969) | |
| %1570 = torch.aten.transpose.int %1559, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc970) | |
| %1571 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1572 = torch.aten.expand %1549, %1571, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc971) | |
| %1573 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1574 = torch.aten.view %1572, %1573 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc972) | |
| %1575 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1576 = torch.aten.expand %1570, %1575, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc973) | |
| %1577 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1578 = torch.aten.view %1576, %1577 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc974) | |
| %1579 = torch.aten.bmm %1574, %1578 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc975) | |
| %1580 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1581 = torch.aten._unsafe_view %1579, %1580 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc976) | |
| %1582 = torch.aten.add_.Tensor %1581, %1023, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc977) | |
| %1583 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1584 = torch.aten.amax %1582, %1583, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc978) | |
| %1585 = torch.aten.sub.Tensor %1582, %1584, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc979) | |
| %1586 = torch.aten.exp %1585 : !torch.tensor -> !torch.tensor loc(#loc980) | |
| %1587 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1588 = torch.aten.sum.dim_IntList %1586, %1587, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc981) | |
| %1589 = torch.aten.div.Tensor %1586, %1588 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc982) | |
| %1590 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1591 = torch.aten.expand %1589, %1590, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc983) | |
| %1592 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1593 = torch.aten.view %1591, %1592 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc984) | |
| %1594 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1595 = torch.aten.expand %1569, %1594, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc985) | |
| %1596 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1597 = torch.aten.view %1595, %1596 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc986) | |
| %1598 = torch.aten.bmm %1593, %1597 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc987) | |
| %1599 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1600 = torch.aten._unsafe_view %1598, %1599 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc988) | |
| %1601 = torch.aten.transpose.int %1600, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc989) | |
| %1602 = torch.aten.clone %1601, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc990) | |
| %1603 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1604 = torch.aten.view %1602, %1603 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc991) | |
| %1605 = torch.prim.GetAttr %arg0["_param_constant100"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1606 = torch.aten.t %1605 : !torch.tensor -> !torch.tensor loc(#loc992) | |
| %1607 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1608 = torch.aten.view %1604, %1607 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc993) | |
| %1609 = torch.aten.mm %1608, %1606 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc994) | |
| %1610 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1611 = torch.aten._unsafe_view %1609, %1610 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc995) | |
| %1612 = torch.aten.add.Tensor %1531, %1611, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc996) | |
| %1613 = torch.aten.pow.Tensor_Scalar %1612, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc997) | |
| %1614 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1615 = torch.aten.mean.dim %1613, %1614, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc998) | |
| %1616 = torch.aten.add.Scalar %1615, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc999) | |
| %1617 = torch.aten.rsqrt %1616 : !torch.tensor -> !torch.tensor loc(#loc1000) | |
| %1618 = torch.aten.mul.Tensor %1612, %1617 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1001) | |
| %1619 = torch.prim.GetAttr %arg0["_param_constant101"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1620 = torch.aten.mul.Tensor %1619, %1618 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1002) | |
| %1621 = torch.prim.GetAttr %arg0["_param_constant102"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1622 = torch.aten.t %1621 : !torch.tensor -> !torch.tensor loc(#loc1003) | |
| %1623 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1624 = torch.aten.view %1620, %1623 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1004) | |
| %1625 = torch.aten.mm %1624, %1622 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1005) | |
| %1626 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1627 = torch.aten._unsafe_view %1625, %1626 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1006) | |
| %1628 = torch.aten.relu %1627 : !torch.tensor -> !torch.tensor loc(#loc1007) | |
| %1629 = torch.prim.GetAttr %arg0["_param_constant103"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1630 = torch.aten.t %1629 : !torch.tensor -> !torch.tensor loc(#loc1008) | |
| %1631 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1632 = torch.aten.view %1628, %1631 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1009) | |
| %1633 = torch.aten.mm %1632, %1630 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1010) | |
| %1634 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1635 = torch.aten._unsafe_view %1633, %1634 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1011) | |
| %1636 = torch.aten.add.Tensor %1612, %1635, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1012) | |
| %1637 = torch.aten.pow.Tensor_Scalar %1636, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1013) | |
| %1638 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1639 = torch.aten.mean.dim %1637, %1638, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1014) | |
| %1640 = torch.aten.add.Scalar %1639, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1015) | |
| %1641 = torch.aten.rsqrt %1640 : !torch.tensor -> !torch.tensor loc(#loc1016) | |
| %1642 = torch.aten.mul.Tensor %1636, %1641 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1017) | |
| %1643 = torch.prim.GetAttr %arg0["_param_constant104"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1644 = torch.aten.mul.Tensor %1643, %1642 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1018) | |
| %1645 = torch.prim.GetAttr %arg0["_param_constant105"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1646 = torch.aten.t %1645 : !torch.tensor -> !torch.tensor loc(#loc1019) | |
| %1647 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1648 = torch.aten.view %1644, %1647 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1020) | |
| %1649 = torch.aten.mm %1648, %1646 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1021) | |
| %1650 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1651 = torch.aten._unsafe_view %1649, %1650 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1022) | |
| %1652 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1653 = torch.aten.view %1651, %1652 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1023) | |
| %1654 = torch.aten.transpose.int %1653, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1024) | |
| %1655 = torch.prim.GetAttr %arg0["_param_constant106"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1656 = torch.aten.t %1655 : !torch.tensor -> !torch.tensor loc(#loc1025) | |
| %1657 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1658 = torch.aten.view %1644, %1657 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1026) | |
| %1659 = torch.aten.mm %1658, %1656 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1027) | |
| %1660 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1661 = torch.aten._unsafe_view %1659, %1660 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1028) | |
| %1662 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1663 = torch.aten.view %1661, %1662 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1029) | |
| %1664 = torch.aten.transpose.int %1663, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1030) | |
| %1665 = torch.prim.GetAttr %arg0["_param_constant107"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1666 = torch.aten.t %1665 : !torch.tensor -> !torch.tensor loc(#loc1031) | |
| %1667 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1668 = torch.aten.view %1644, %1667 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1032) | |
| %1669 = torch.aten.mm %1668, %1666 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1033) | |
| %1670 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1671 = torch.aten._unsafe_view %1669, %1670 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1034) | |
| %1672 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1673 = torch.aten.view %1671, %1672 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1035) | |
| %1674 = torch.aten.transpose.int %1673, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1036) | |
| %1675 = torch.aten.transpose.int %1664, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1037) | |
| %1676 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1677 = torch.aten.expand %1654, %1676, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1038) | |
| %1678 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1679 = torch.aten.view %1677, %1678 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1039) | |
| %1680 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1681 = torch.aten.expand %1675, %1680, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1040) | |
| %1682 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1683 = torch.aten.view %1681, %1682 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1041) | |
| %1684 = torch.aten.bmm %1679, %1683 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1042) | |
| %1685 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1686 = torch.aten._unsafe_view %1684, %1685 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1043) | |
| %1687 = torch.aten.add_.Tensor %1686, %939, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1044) | |
| %1688 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1689 = torch.aten.amax %1687, %1688, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1045) | |
| %1690 = torch.aten.sub.Tensor %1687, %1689, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1046) | |
| %1691 = torch.aten.exp %1690 : !torch.tensor -> !torch.tensor loc(#loc1047) | |
| %1692 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1693 = torch.aten.sum.dim_IntList %1691, %1692, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1048) | |
| %1694 = torch.aten.div.Tensor %1691, %1693 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1049) | |
| %1695 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1696 = torch.aten.expand %1694, %1695, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1050) | |
| %1697 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1698 = torch.aten.view %1696, %1697 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1051) | |
| %1699 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1700 = torch.aten.expand %1674, %1699, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1052) | |
| %1701 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1702 = torch.aten.view %1700, %1701 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1053) | |
| %1703 = torch.aten.bmm %1698, %1702 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1054) | |
| %1704 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1705 = torch.aten._unsafe_view %1703, %1704 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1055) | |
| %1706 = torch.aten.transpose.int %1705, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1056) | |
| %1707 = torch.aten.clone %1706, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1057) | |
| %1708 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1709 = torch.aten.view %1707, %1708 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1058) | |
| %1710 = torch.prim.GetAttr %arg0["_param_constant108"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1711 = torch.aten.t %1710 : !torch.tensor -> !torch.tensor loc(#loc1059) | |
| %1712 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1713 = torch.aten.view %1709, %1712 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1060) | |
| %1714 = torch.aten.mm %1713, %1711 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1061) | |
| %1715 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1716 = torch.aten._unsafe_view %1714, %1715 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1062) | |
| %1717 = torch.aten.add.Tensor %1636, %1716, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1063) | |
| %1718 = torch.aten.pow.Tensor_Scalar %1717, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1064) | |
| %1719 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1720 = torch.aten.mean.dim %1718, %1719, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1065) | |
| %1721 = torch.aten.add.Scalar %1720, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1066) | |
| %1722 = torch.aten.rsqrt %1721 : !torch.tensor -> !torch.tensor loc(#loc1067) | |
| %1723 = torch.aten.mul.Tensor %1717, %1722 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1068) | |
| %1724 = torch.prim.GetAttr %arg0["_param_constant109"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1725 = torch.aten.mul.Tensor %1724, %1723 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1069) | |
| %1726 = torch.prim.GetAttr %arg0["_param_constant110"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1727 = torch.aten.t %1726 : !torch.tensor -> !torch.tensor loc(#loc1070) | |
| %1728 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1729 = torch.aten.view %1725, %1728 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1071) | |
| %1730 = torch.aten.mm %1729, %1727 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1072) | |
| %1731 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1732 = torch.aten._unsafe_view %1730, %1731 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1073) | |
| %1733 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1734 = torch.aten.view %1732, %1733 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1074) | |
| %1735 = torch.aten.transpose.int %1734, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1075) | |
| %1736 = torch.prim.GetAttr %arg0["_param_constant111"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1737 = torch.aten.t %1736 : !torch.tensor -> !torch.tensor loc(#loc1076) | |
| %1738 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1739 = torch.aten.view %824, %1738 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1077) | |
| %1740 = torch.aten.mm %1739, %1737 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1078) | |
| %1741 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1742 = torch.aten._unsafe_view %1740, %1741 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1079) | |
| %1743 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1744 = torch.aten.view %1742, %1743 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1080) | |
| %1745 = torch.aten.transpose.int %1744, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1081) | |
| %1746 = torch.prim.GetAttr %arg0["_param_constant112"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1747 = torch.aten.t %1746 : !torch.tensor -> !torch.tensor loc(#loc1082) | |
| %1748 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1749 = torch.aten.view %824, %1748 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1083) | |
| %1750 = torch.aten.mm %1749, %1747 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1084) | |
| %1751 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1752 = torch.aten._unsafe_view %1750, %1751 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1085) | |
| %1753 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1754 = torch.aten.view %1752, %1753 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1086) | |
| %1755 = torch.aten.transpose.int %1754, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1087) | |
| %1756 = torch.aten.transpose.int %1745, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1088) | |
| %1757 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1758 = torch.aten.expand %1735, %1757, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1089) | |
| %1759 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1760 = torch.aten.view %1758, %1759 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1090) | |
| %1761 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1762 = torch.aten.expand %1756, %1761, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1091) | |
| %1763 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1764 = torch.aten.view %1762, %1763 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1092) | |
| %1765 = torch.aten.bmm %1760, %1764 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1093) | |
| %1766 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1767 = torch.aten._unsafe_view %1765, %1766 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1094) | |
| %1768 = torch.aten.add_.Tensor %1767, %1023, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1095) | |
| %1769 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1770 = torch.aten.amax %1768, %1769, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1096) | |
| %1771 = torch.aten.sub.Tensor %1768, %1770, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1097) | |
| %1772 = torch.aten.exp %1771 : !torch.tensor -> !torch.tensor loc(#loc1098) | |
| %1773 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1774 = torch.aten.sum.dim_IntList %1772, %1773, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1099) | |
| %1775 = torch.aten.div.Tensor %1772, %1774 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1100) | |
| %1776 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1777 = torch.aten.expand %1775, %1776, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1101) | |
| %1778 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1779 = torch.aten.view %1777, %1778 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1102) | |
| %1780 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1781 = torch.aten.expand %1755, %1780, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1103) | |
| %1782 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1783 = torch.aten.view %1781, %1782 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1104) | |
| %1784 = torch.aten.bmm %1779, %1783 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1105) | |
| %1785 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1786 = torch.aten._unsafe_view %1784, %1785 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1106) | |
| %1787 = torch.aten.transpose.int %1786, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1107) | |
| %1788 = torch.aten.clone %1787, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1108) | |
| %1789 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1790 = torch.aten.view %1788, %1789 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1109) | |
| %1791 = torch.prim.GetAttr %arg0["_param_constant113"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1792 = torch.aten.t %1791 : !torch.tensor -> !torch.tensor loc(#loc1110) | |
| %1793 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1794 = torch.aten.view %1790, %1793 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1111) | |
| %1795 = torch.aten.mm %1794, %1792 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1112) | |
| %1796 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1797 = torch.aten._unsafe_view %1795, %1796 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1113) | |
| %1798 = torch.aten.add.Tensor %1717, %1797, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1114) | |
| %1799 = torch.aten.pow.Tensor_Scalar %1798, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1115) | |
| %1800 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1801 = torch.aten.mean.dim %1799, %1800, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1116) | |
| %1802 = torch.aten.add.Scalar %1801, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1117) | |
| %1803 = torch.aten.rsqrt %1802 : !torch.tensor -> !torch.tensor loc(#loc1118) | |
| %1804 = torch.aten.mul.Tensor %1798, %1803 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1119) | |
| %1805 = torch.prim.GetAttr %arg0["_param_constant114"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1806 = torch.aten.mul.Tensor %1805, %1804 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1120) | |
| %1807 = torch.prim.GetAttr %arg0["_param_constant115"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1808 = torch.aten.t %1807 : !torch.tensor -> !torch.tensor loc(#loc1121) | |
| %1809 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1810 = torch.aten.view %1806, %1809 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1122) | |
| %1811 = torch.aten.mm %1810, %1808 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1123) | |
| %1812 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1813 = torch.aten._unsafe_view %1811, %1812 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1124) | |
| %1814 = torch.aten.relu %1813 : !torch.tensor -> !torch.tensor loc(#loc1125) | |
| %1815 = torch.prim.GetAttr %arg0["_param_constant116"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1816 = torch.aten.t %1815 : !torch.tensor -> !torch.tensor loc(#loc1126) | |
| %1817 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1818 = torch.aten.view %1814, %1817 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1127) | |
| %1819 = torch.aten.mm %1818, %1816 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1128) | |
| %1820 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1821 = torch.aten._unsafe_view %1819, %1820 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1129) | |
| %1822 = torch.aten.add.Tensor %1798, %1821, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1130) | |
| %1823 = torch.aten.pow.Tensor_Scalar %1822, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1131) | |
| %1824 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1825 = torch.aten.mean.dim %1823, %1824, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1132) | |
| %1826 = torch.aten.add.Scalar %1825, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1133) | |
| %1827 = torch.aten.rsqrt %1826 : !torch.tensor -> !torch.tensor loc(#loc1134) | |
| %1828 = torch.aten.mul.Tensor %1822, %1827 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1135) | |
| %1829 = torch.prim.GetAttr %arg0["_param_constant117"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1830 = torch.aten.mul.Tensor %1829, %1828 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1136) | |
| %1831 = torch.prim.GetAttr %arg0["_param_constant118"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1832 = torch.aten.t %1831 : !torch.tensor -> !torch.tensor loc(#loc1137) | |
| %1833 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1834 = torch.aten.view %1830, %1833 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1138) | |
| %1835 = torch.aten.mm %1834, %1832 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1139) | |
| %1836 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1837 = torch.aten._unsafe_view %1835, %1836 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1140) | |
| %1838 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1839 = torch.aten.view %1837, %1838 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1141) | |
| %1840 = torch.aten.transpose.int %1839, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1142) | |
| %1841 = torch.prim.GetAttr %arg0["_param_constant119"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1842 = torch.aten.t %1841 : !torch.tensor -> !torch.tensor loc(#loc1143) | |
| %1843 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1844 = torch.aten.view %1830, %1843 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1144) | |
| %1845 = torch.aten.mm %1844, %1842 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1145) | |
| %1846 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1847 = torch.aten._unsafe_view %1845, %1846 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1146) | |
| %1848 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1849 = torch.aten.view %1847, %1848 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1147) | |
| %1850 = torch.aten.transpose.int %1849, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1148) | |
| %1851 = torch.prim.GetAttr %arg0["_param_constant120"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1852 = torch.aten.t %1851 : !torch.tensor -> !torch.tensor loc(#loc1149) | |
| %1853 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1854 = torch.aten.view %1830, %1853 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1150) | |
| %1855 = torch.aten.mm %1854, %1852 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1151) | |
| %1856 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1857 = torch.aten._unsafe_view %1855, %1856 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1152) | |
| %1858 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1859 = torch.aten.view %1857, %1858 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1153) | |
| %1860 = torch.aten.transpose.int %1859, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1154) | |
| %1861 = torch.aten.transpose.int %1850, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1155) | |
| %1862 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1863 = torch.aten.expand %1840, %1862, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1156) | |
| %1864 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1865 = torch.aten.view %1863, %1864 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1157) | |
| %1866 = torch.prim.ListConstruct %int1, %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1867 = torch.aten.expand %1861, %1866, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1158) | |
| %1868 = torch.prim.ListConstruct %int8, %int64, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1869 = torch.aten.view %1867, %1868 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1159) | |
| %1870 = torch.aten.bmm %1865, %1869 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1160) | |
| %1871 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1872 = torch.aten._unsafe_view %1870, %1871 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1161) | |
| %1873 = torch.aten.add_.Tensor %1872, %939, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1162) | |
| %1874 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1875 = torch.aten.amax %1873, %1874, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1163) | |
| %1876 = torch.aten.sub.Tensor %1873, %1875, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1164) | |
| %1877 = torch.aten.exp %1876 : !torch.tensor -> !torch.tensor loc(#loc1165) | |
| %1878 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1879 = torch.aten.sum.dim_IntList %1877, %1878, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1166) | |
| %1880 = torch.aten.div.Tensor %1877, %1879 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1167) | |
| %1881 = torch.prim.ListConstruct %int1, %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1882 = torch.aten.expand %1880, %1881, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1168) | |
| %1883 = torch.prim.ListConstruct %int8, %int4, %int4 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1884 = torch.aten.view %1882, %1883 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1169) | |
| %1885 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1886 = torch.aten.expand %1860, %1885, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1170) | |
| %1887 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1888 = torch.aten.view %1886, %1887 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1171) | |
| %1889 = torch.aten.bmm %1884, %1888 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1172) | |
| %1890 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1891 = torch.aten._unsafe_view %1889, %1890 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1173) | |
| %1892 = torch.aten.transpose.int %1891, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1174) | |
| %1893 = torch.aten.clone %1892, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1175) | |
| %1894 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1895 = torch.aten.view %1893, %1894 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1176) | |
| %1896 = torch.prim.GetAttr %arg0["_param_constant121"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1897 = torch.aten.t %1896 : !torch.tensor -> !torch.tensor loc(#loc1177) | |
| %1898 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1899 = torch.aten.view %1895, %1898 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1178) | |
| %1900 = torch.aten.mm %1899, %1897 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1179) | |
| %1901 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1902 = torch.aten._unsafe_view %1900, %1901 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1180) | |
| %1903 = torch.aten.add.Tensor %1822, %1902, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1181) | |
| %1904 = torch.aten.pow.Tensor_Scalar %1903, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1182) | |
| %1905 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1906 = torch.aten.mean.dim %1904, %1905, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1183) | |
| %1907 = torch.aten.add.Scalar %1906, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1184) | |
| %1908 = torch.aten.rsqrt %1907 : !torch.tensor -> !torch.tensor loc(#loc1185) | |
| %1909 = torch.aten.mul.Tensor %1903, %1908 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1186) | |
| %1910 = torch.prim.GetAttr %arg0["_param_constant122"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1911 = torch.aten.mul.Tensor %1910, %1909 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1187) | |
| %1912 = torch.prim.GetAttr %arg0["_param_constant123"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1913 = torch.aten.t %1912 : !torch.tensor -> !torch.tensor loc(#loc1188) | |
| %1914 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1915 = torch.aten.view %1911, %1914 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1189) | |
| %1916 = torch.aten.mm %1915, %1913 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1190) | |
| %1917 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1918 = torch.aten._unsafe_view %1916, %1917 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1191) | |
| %1919 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1920 = torch.aten.view %1918, %1919 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1192) | |
| %1921 = torch.aten.transpose.int %1920, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1193) | |
| %1922 = torch.prim.GetAttr %arg0["_param_constant124"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1923 = torch.aten.t %1922 : !torch.tensor -> !torch.tensor loc(#loc1194) | |
| %1924 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1925 = torch.aten.view %824, %1924 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1195) | |
| %1926 = torch.aten.mm %1925, %1923 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1196) | |
| %1927 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1928 = torch.aten._unsafe_view %1926, %1927 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1197) | |
| %1929 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1930 = torch.aten.view %1928, %1929 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1198) | |
| %1931 = torch.aten.transpose.int %1930, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1199) | |
| %1932 = torch.prim.GetAttr %arg0["_param_constant125"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1933 = torch.aten.t %1932 : !torch.tensor -> !torch.tensor loc(#loc1200) | |
| %1934 = torch.prim.ListConstruct %int15, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1935 = torch.aten.view %824, %1934 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1201) | |
| %1936 = torch.aten.mm %1935, %1933 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1202) | |
| %1937 = torch.prim.ListConstruct %int1, %int15, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1938 = torch.aten._unsafe_view %1936, %1937 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1203) | |
| %1939 = torch.prim.ListConstruct %int1, %int-1, %int8, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1940 = torch.aten.view %1938, %1939 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1204) | |
| %1941 = torch.aten.transpose.int %1940, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1205) | |
| %1942 = torch.aten.transpose.int %1931, %int3, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1206) | |
| %1943 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1944 = torch.aten.expand %1921, %1943, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1207) | |
| %1945 = torch.prim.ListConstruct %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1946 = torch.aten.view %1944, %1945 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1208) | |
| %1947 = torch.prim.ListConstruct %int1, %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1948 = torch.aten.expand %1942, %1947, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1209) | |
| %1949 = torch.prim.ListConstruct %int8, %int64, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1950 = torch.aten.view %1948, %1949 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1210) | |
| %1951 = torch.aten.bmm %1946, %1950 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1211) | |
| %1952 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1953 = torch.aten._unsafe_view %1951, %1952 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1212) | |
| %1954 = torch.aten.add_.Tensor %1953, %1023, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1213) | |
| %1955 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1956 = torch.aten.amax %1954, %1955, %true_0 : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1214) | |
| %1957 = torch.aten.sub.Tensor %1954, %1956, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1215) | |
| %1958 = torch.aten.exp %1957 : !torch.tensor -> !torch.tensor loc(#loc1216) | |
| %1959 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1960 = torch.aten.sum.dim_IntList %1958, %1959, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1217) | |
| %1961 = torch.aten.div.Tensor %1958, %1960 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1218) | |
| %1962 = torch.prim.ListConstruct %int1, %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1963 = torch.aten.expand %1961, %1962, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1219) | |
| %1964 = torch.prim.ListConstruct %int8, %int4, %int15 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1965 = torch.aten.view %1963, %1964 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1220) | |
| %1966 = torch.prim.ListConstruct %int1, %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1967 = torch.aten.expand %1941, %1966, %false : !torch.tensor, !torch.list<int>, !torch.bool -> !torch.tensor loc(#loc1221) | |
| %1968 = torch.prim.ListConstruct %int8, %int15, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1969 = torch.aten.view %1967, %1968 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1222) | |
| %1970 = torch.aten.bmm %1965, %1969 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1223) | |
| %1971 = torch.prim.ListConstruct %int1, %int8, %int4, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1972 = torch.aten._unsafe_view %1970, %1971 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1224) | |
| %1973 = torch.aten.transpose.int %1972, %int1, %int2 : !torch.tensor, !torch.int, !torch.int -> !torch.tensor loc(#loc1225) | |
| %1974 = torch.aten.clone %1973, %int0 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1226) | |
| %1975 = torch.prim.ListConstruct %int1, %int-1, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1976 = torch.aten.view %1974, %1975 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1227) | |
| %1977 = torch.prim.GetAttr %arg0["_param_constant126"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1978 = torch.aten.t %1977 : !torch.tensor -> !torch.tensor loc(#loc1228) | |
| %1979 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1980 = torch.aten.view %1976, %1979 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1229) | |
| %1981 = torch.aten.mm %1980, %1978 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1230) | |
| %1982 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1983 = torch.aten._unsafe_view %1981, %1982 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1231) | |
| %1984 = torch.aten.add.Tensor %1903, %1983, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1232) | |
| %1985 = torch.aten.pow.Tensor_Scalar %1984, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1233) | |
| %1986 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %1987 = torch.aten.mean.dim %1985, %1986, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1234) | |
| %1988 = torch.aten.add.Scalar %1987, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1235) | |
| %1989 = torch.aten.rsqrt %1988 : !torch.tensor -> !torch.tensor loc(#loc1236) | |
| %1990 = torch.aten.mul.Tensor %1984, %1989 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1237) | |
| %1991 = torch.prim.GetAttr %arg0["_param_constant127"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1992 = torch.aten.mul.Tensor %1991, %1990 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1238) | |
| %1993 = torch.prim.GetAttr %arg0["_param_constant128"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %1994 = torch.aten.t %1993 : !torch.tensor -> !torch.tensor loc(#loc1239) | |
| %1995 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1996 = torch.aten.view %1992, %1995 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1240) | |
| %1997 = torch.aten.mm %1996, %1994 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1241) | |
| %1998 = torch.prim.ListConstruct %int1, %int4, %int2048 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %1999 = torch.aten._unsafe_view %1997, %1998 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1242) | |
| %2000 = torch.aten.relu %1999 : !torch.tensor -> !torch.tensor loc(#loc1243) | |
| %2001 = torch.prim.GetAttr %arg0["_param_constant129"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %2002 = torch.aten.t %2001 : !torch.tensor -> !torch.tensor loc(#loc1244) | |
| %2003 = torch.prim.ListConstruct %int4, %int2048 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %2004 = torch.aten.view %2000, %2003 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1245) | |
| %2005 = torch.aten.mm %2004, %2002 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1246) | |
| %2006 = torch.prim.ListConstruct %int1, %int4, %int512 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %2007 = torch.aten._unsafe_view %2005, %2006 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1247) | |
| %2008 = torch.aten.add.Tensor %1984, %2007, %int1 : !torch.tensor, !torch.tensor, !torch.int -> !torch.tensor loc(#loc1248) | |
| %2009 = torch.aten.pow.Tensor_Scalar %2008, %int2 : !torch.tensor, !torch.int -> !torch.tensor loc(#loc1249) | |
| %2010 = torch.prim.ListConstruct %int-1 : (!torch.int) -> !torch.list<int> loc(#loc) | |
| %2011 = torch.aten.mean.dim %2009, %2010, %true_0, %none_1 : !torch.tensor, !torch.list<int>, !torch.bool, !torch.none -> !torch.tensor loc(#loc1250) | |
| %2012 = torch.aten.add.Scalar %2011, %float9.999990e-07, %int1 : !torch.tensor, !torch.float, !torch.int -> !torch.tensor loc(#loc1251) | |
| %2013 = torch.aten.rsqrt %2012 : !torch.tensor -> !torch.tensor loc(#loc1252) | |
| %2014 = torch.aten.mul.Tensor %2008, %2013 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1253) | |
| %2015 = torch.prim.GetAttr %arg0["_param_constant130"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %2016 = torch.aten.mul.Tensor %2015, %2014 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1254) | |
| %2017 = torch.aten.mul.Scalar %2016, %float4.419420e-02 : !torch.tensor, !torch.float -> !torch.tensor loc(#loc1255) | |
| %2018 = torch.prim.GetAttr %arg0["_param_constant0"] : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> -> !torch.tensor loc(#loc) | |
| %2019 = torch.aten.t %2018 : !torch.tensor -> !torch.tensor loc(#loc1256) | |
| %2020 = torch.prim.ListConstruct %int4, %int512 : (!torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %2021 = torch.aten.view %2017, %2020 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1257) | |
| %2022 = torch.aten.mm %2021, %2019 : !torch.tensor, !torch.tensor -> !torch.tensor loc(#loc1258) | |
| %2023 = torch.prim.ListConstruct %int1, %int4, %int32128 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> loc(#loc) | |
| %2024 = torch.aten._unsafe_view %2022, %2023 : !torch.tensor, !torch.list<int> -> !torch.tensor loc(#loc1259) | |
| return %2024 : !torch.tensor loc(#loc) | |
| } loc(#loc) | |
| torch.class_type @__torch__.torch.fx.graph_module._lambda { | |
| torch.attr private "_param_constant0" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant1" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant2" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant3" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant4" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant5" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant6" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant7" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant8" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant9" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant10" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant11" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant12" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant13" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant14" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant15" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant16" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant17" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant18" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant19" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant20" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant21" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant22" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant23" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant24" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant25" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant26" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant27" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant28" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant29" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant30" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant31" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant32" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant33" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant34" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant35" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant36" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant37" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant38" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant39" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant40" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant41" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant42" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant43" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant44" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant45" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant46" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant47" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant48" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant49" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant50" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant51" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant52" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant53" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant54" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant55" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant56" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant57" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant58" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant59" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant60" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant61" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant62" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant63" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant64" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant65" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant66" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant67" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant68" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant69" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant70" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant71" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant72" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant73" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant74" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant75" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant76" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant77" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant78" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant79" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant80" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant81" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant82" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant83" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant84" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant85" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant86" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant87" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant88" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant89" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant90" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant91" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant92" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant93" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant94" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant95" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant96" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant97" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant98" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant99" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant100" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant101" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant102" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant103" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant104" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant105" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant106" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant107" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant108" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant109" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant110" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant111" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant112" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant113" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant114" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant115" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant116" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant117" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant118" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant119" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant120" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant121" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant122" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant123" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant124" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant125" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant126" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant127" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant128" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant129" : !torch.tensor loc(#loc) | |
| torch.attr private "_param_constant130" : !torch.tensor loc(#loc) | |
| torch.attr private "_tensor_constant0" : !torch.tensor loc(#loc) | |
| torch.attr private "training" : !torch.bool loc(#loc) | |
| torch.attr private "_is_full_backward_hook" : !torch.optional<bool> loc(#loc) | |
| torch.attr private "_code" : !torch.str loc(#loc) | |
| torch.method private "__code_getter", @__torch__.torch.fx.graph_module._lambda.__code_getter loc(#loc) | |
| torch.method "forward", @__torch__.torch.fx.graph_module._lambda.forward loc(#loc) | |
| } loc(#loc) | |
| %0 = torch.tensor.literal(dense_resource<__elided__> : tensor<32128x512xf32>) : !torch.tensor<[32128,512],f32> loc(#loc) | |
| %1 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %2 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %3 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %4 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %5 = torch.tensor.literal(dense_resource<__elided__> : tensor<32x8xf32>) : !torch.tensor<[32,8],f32> loc(#loc) | |
| %6 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %7 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %8 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %9 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %10 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %11 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %12 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %13 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %14 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %15 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %16 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %17 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %18 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %19 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %20 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %21 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %22 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %23 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %24 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %25 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %26 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %27 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %28 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %29 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %30 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %31 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %32 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %33 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %34 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %35 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %36 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %37 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %38 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %39 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %40 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %41 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %42 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %43 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %44 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %45 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %46 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %47 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %48 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %49 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %50 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %51 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %52 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %53 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %54 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %55 = torch.tensor.literal(dense_resource<__elided__> : tensor<32x8xf32>) : !torch.tensor<[32,8],f32> loc(#loc) | |
| %56 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %57 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %58 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %59 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %60 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %61 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %62 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %63 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %64 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %65 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %66 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %67 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %68 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %69 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %70 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %71 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %72 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %73 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %74 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %75 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %76 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %77 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %78 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %79 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %80 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %81 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %82 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %83 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %84 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %85 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %86 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %87 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %88 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %89 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %90 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %91 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %92 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %93 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %94 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %95 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %96 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %97 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %98 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %99 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %100 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %101 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %102 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %103 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %104 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %105 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %106 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %107 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %108 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %109 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %110 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %111 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %112 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %113 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %114 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %115 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %116 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %117 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %118 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %119 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %120 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %121 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %122 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %123 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %124 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %125 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %126 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x512xf32>) : !torch.tensor<[512,512],f32> loc(#loc) | |
| %127 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %128 = torch.tensor.literal(dense_resource<__elided__> : tensor<2048x512xf32>) : !torch.tensor<[2048,512],f32> loc(#loc) | |
| %129 = torch.tensor.literal(dense_resource<__elided__> : tensor<512x2048xf32>) : !torch.tensor<[512,2048],f32> loc(#loc) | |
| %130 = torch.tensor.literal(dense_resource<__elided__> : tensor<512xf32>) : !torch.tensor<[512],f32> loc(#loc) | |
| %131 = torch.tensor.literal(dense<0> : tensor<si64>) : !torch.tensor<[],si64> loc(#loc) | |
| %true = torch.constant.bool true loc(#loc) | |
| %none = torch.constant.none loc(#loc) | |
| %str = torch.constant.str "\0A\0A\0Adef forward(self, arg0_1, arg1_1):\0A new_zeros = torch.ops.aten.new_zeros(arg1_1, [1, 4], dtype = torch.int64, layout = torch.strided, device = device(type='cpu'), pin_memory = False)\0A slice_1 = torch.ops.aten.slice(arg1_1, 1, 0, -1); arg1_1 = None\0A clone = torch.ops.aten.clone(slice_1); slice_1 = None\0A slice_2 = torch.ops.aten.slice(new_zeros, 1, 1, 9223372036854775807)\0A copy_ = torch.ops.aten.copy_(slice_2, clone); slice_2 = clone = None\0A _tensor_constant0 = self._tensor_constant0\0A lift_fresh_copy = torch.ops.aten.lift_fresh_copy(_tensor_constant0); _tensor_constant0 = None\0A select = torch.ops.aten.select(new_zeros, 1, 0)\0A fill_ = torch.ops.aten.fill_(select, lift_fresh_copy); select = lift_fresh_copy = None\0A eq = torch.ops.aten.eq(new_zeros, -100)\0A masked_fill_ = torch.ops.aten.masked_fill_(new_zeros, eq, 0); new_zeros = eq = None\0A view = torch.ops.aten.view(arg0_1, [-1, 15]); arg0_1 = None\0A _param_constant0 = self._param_constant0\0A embedding = torch.ops.aten.embedding(_param_constant0, view); _param_constant0 = view = None\0A ones = torch.ops.aten.ones([1, 15], device = device(type='cpu'), pin_memory = False)\0A slice_3 = torch.ops.aten.slice(ones, 0, 0, 9223372036854775807); ones = None\0A unsqueeze = torch.ops.aten.unsqueeze(slice_3, 1); slice_3 = None\0A unsqueeze_1 = torch.ops.aten.unsqueeze(unsqueeze, 2); unsqueeze = None\0A slice_4 = torch.ops.aten.slice(unsqueeze_1, 3, 0, 9223372036854775807); unsqueeze_1 = None\0A rsub = torch.ops.aten.rsub(slice_4, 1.0); slice_4 = None\0A mul = torch.ops.aten.mul(rsub, -3.4028234663852886e+38); rsub = None\0A pow_1 = torch.ops.aten.pow(embedding, 2)\0A mean = torch.ops.aten.mean(pow_1, [-1], True); pow_1 = None\0A add = torch.ops.aten.add(mean, 1e-06); mean = None\0A rsqrt = torch.ops.aten.rsqrt(add); add = None\0A detach = torch.ops.aten.detach(rsqrt)\0A mul_1 = torch.ops.aten.mul(embedding, rsqrt); rsqrt = None\0A _param_constant1 = self._param_constant1\0A mul_2 = torch.ops.aten.mul(_param_constant1, mul_1); _param_constant1 = mul_1 = None\0A _param_constant2 = self._param_constant2\0A t = torch.ops.aten.t(_param_constant2); _param_constant2 = None\0A view_1 = torch.ops.aten.view(mul_2, [15, 512])\0A mm = torch.ops.aten.mm(view_1, t); view_1 = t = None\0A _unsafe_view = torch.ops.aten._unsafe_view(mm, [1, 15, 512]); mm = None\0A view_2 = torch.ops.aten.view(_unsafe_view, [1, -1, 8, 64]); _unsafe_view = None\0A transpose = torch.ops.aten.transpose(view_2, 1, 2); view_2 = None\0A _param_constant3 = self._param_constant3\0A t_1 = torch.ops.aten.t(_param_constant3); _param_constant3 = None\0A view_3 = torch.ops.aten.view(mul_2, [15, 512])\0A mm_1 = torch.ops.aten.mm(view_3, t_1); view_3 = t_1 = None\0A _unsafe_view_1 = torch.ops.aten._unsafe_view(mm_1, [1, 15, 512]); mm_1 = None\0A view_4 = torch.ops.aten.view(_unsafe_view_1, [1, -1, 8, 64]); _unsafe_view_1 = None\0A transpose_1 = torch.ops.aten.transpose(view_4, 1, 2); view_4 = None\0A _param_constant4 = self._param_constant4\0A t_2 = torch.ops.aten.t(_param_constant4); _param_constant4 = None\0A view_5 = torch.ops.aten.view(mul_2, [15, 512]); mul_2 = None\0A mm_2 = torch.ops.aten.mm(view_5, t_2); view_5 = t_2 = None\0A _unsafe_view_2 = torch.ops.aten._unsafe_view(mm_2, [1, 15, 512]); mm_2 = None\0A view_6 = torch.ops.aten.view(_unsafe_view_2, [1, -1, 8, 64]); _unsafe_view_2 = None\0A transpose_2 = torch.ops.aten.transpose(view_6, 1, 2); view_6 = None\0A transpose_3 = torch.ops.aten.transpose(transpose_1, 3, 2); transpose_1 = None\0A expand = torch.ops.aten.expand(transpose, [1, 8, 15, 64]); transpose = None\0A view_7 = torch.ops.aten.view(expand, [8, 15, 64]); expand = None\0A expand_1 = torch.ops.aten.expand(transpose_3, [1, 8, 64, 15]); transpose_3 = None\0A view_8 = torch.ops.aten.view(expand_1, [8, 64, 15]); expand_1 = None\0A bmm = torch.ops.aten.bmm(view_7, view_8); view_7 = view_8 = None\0A _unsafe_view_3 = torch.ops.aten._unsafe_view(bmm, [1, 8, 15, 15]); bmm = None\0A arange = torch.ops.aten.arange(15, dtype = torch.int64, device = device(type='cpu'), pin_memory = False)\0A slice_5 = torch.ops.aten.slice(arange, 0, 0, 9223372036854775807); arange = None\0A unsqueeze_2 = torch.ops.aten.unsqueeze(slice_5, 1); slice_5 = None\0A arange_1 = torch.ops.aten.arange(15, dtype = torch.int64, device = device(type='cpu'), pin_memory = False)\0A unsqueeze_3 = torch.ops.aten.unsqueeze(arange_1, 0); arange_1 = None\0A slice_6 = torch.ops.aten.slice(unsqueeze_3, 1, 0, 9223372036854775807); unsqueeze_3 = None\0A sub = torch.ops.aten.sub(slice_6, unsqueeze_2); slice_6 = unsqueeze_2 = None\0A gt = torch.ops.aten.gt(sub, 0)\0A convert_element_type = torch.ops.prims.convert_element_type(gt, torch.int64); gt = None\0A mul_3 = torch.ops.aten.mul(convert_element_type, 16); convert_element_type = None\0A add_1 = torch.ops.aten.add(mul_3, 0); mul_3 = None\0A abs_1 = torch.ops.aten.abs(sub); sub = None\0A lt = torch.ops.aten.lt(abs_1, 8)\0A convert_element_type_1 = torch.ops.prims.convert_element_type(abs_1, torch.float32)\0A div = torch.ops.aten.div(convert_element_type_1, 8); convert_element_type_1 = None\0A log = torch.ops.aten.log(div); div = None\0A div_1 = torch.ops.aten.div(log, 2.772588722239781); log = None\0A mul_4 = torch.ops.aten.mul(div_1, 8); div_1 = None\0A convert_element_type_2 = torch.ops.prims.convert_element_type(mul_4, torch.int64); mul_4 = None\0A add_2 = torch.ops.aten.add(convert_element_type_2, 8); convert_element_type_2 = None\0A full_like = torch.ops.aten.full_like(add_2, 15, dtype = torch.int64, layout = torch.strided, device = device(type='cpu'), pin_memory = False)\0A minimum = torch.ops.aten.minimum(add_2, full_like); add_2 = full_like = None\0A where = torch.ops.aten.where(lt, abs_1, minimum); lt = abs_1 = minimum = None\0A add_ = torch.ops.aten.add_(add_1, where); add_1 = where = None\0A _param_constant5 = self._param_constant5\0A embedding_1 = torch.ops.aten.embedding(_param_constant5, add_); _param_constant5 = add_ = None\0A permute = torch.ops.aten.permute(embedding_1, [2, 0, 1]); embedding_1 = None\0A unsqueeze_4 = torch.ops.aten.unsqueeze(permute, 0); permute = None\0A add_3 = torch.ops.aten.add(unsqueeze_4, mul); unsqueeze_4 = mul = None\0A add__1 = torch.ops.aten.add_(_unsafe_view_3, add_3); _unsafe_view_3 = None\0A amax = torch.ops.aten.amax(add__1, [-1], True)\0A sub_1 = torch.ops.aten.sub(add__1, amax); add__1 = amax = None\0A exp = torch.ops.aten.exp(sub_1); sub_1 = None\0A sum_1 = torch.ops.aten.sum(exp, [-1], True)\0A div_2 = torch.ops.aten.div(exp, sum_1); exp = sum_1 = None\0A detach_1 = torch.ops.aten.detach(div_2)\0A expand_2 = torch.ops.aten.expand(div_2, [1, 8, 15, 15]); div_2 = None\0A view_9 = torch.ops.aten.view(expand_2, [8, 15, 15]); expand_2 = None\0A expand_3 = torch.ops.aten.expand(transpose_2, [1, 8, 15, 64]); transpose_2 = None\0A view_10 = torch.ops.aten.view(expand_3, [8, 15, 64]); expand_3 = None\0A bmm_1 = torch.ops.aten.bmm(view_9, view_10); view_9 = view_10 = None\0A _unsafe_view_4 = torch.ops.aten._unsafe_view(bmm_1, [1, 8, 15, 64]); bmm_1 = None\0A transpose_4 = torch.ops.aten.transpose(_unsafe_view_4, 1, 2); _unsafe_view_4 = None\0A clone_1 = torch.ops.aten.clone(transpose_4, memory_format = torch.contiguous_format); transpose_4 = None\0A view_11 = torch.ops.aten.view(clone_1, [1, -1, 512]); clone_1 = None\0A _param_constant6 = self._param_constant6\0A t_3 = torch.ops.aten.t(_param_constant6); _param_constant6 = None\0A view_12 = torch.ops.aten.view(view_11, [15, 512]); view_11 = None\0A mm_3 = torch.ops.aten.mm(view_12, t_3); view_12 = t_3 = None\0A _unsafe_view_5 = torch.ops.aten._unsafe_view(mm_3, [1, 15, 512]); mm_3 = None\0A add_4 = torch.ops.aten.add(embedding, _unsafe_view_5); embedding = _unsafe_view_5 = None\0A pow_2 = torch.ops.aten.pow(add_4, 2)\0A mean_1 = torch.ops.aten.mean(pow_2, [-1], True); pow_2 = None\0A add_5 = torch.ops.aten.add(mean_1, 1e-06); mean_1 = None\0A rsqrt_1 = torch.ops.aten.rsqrt(add_5); add_5 = None\0A detach_2 = torch.ops.aten.detach(rsqrt_1)\0A mul_5 = torch.ops.aten.mul(add_4, rsqrt_1); rsqrt_1 = None\0A _param_constant7 = self._param_constant7\0A mul_6 = torch.ops.aten.mul(_param_constant7, mul_5); _param_constant7 = mul_5 = None\0A _param_constant8 = self._param_constant8\0A t_4 = torch.ops.aten.t(_param_constant8); _param_constant8 = None\0A view_13 = torch.ops.aten.view(mul_6, [15, 512]); mul_6 = None\0A mm_4 = torch.ops.aten.mm(view_13, t_4); view_13 = t_4 = None\0A _unsafe_view_6 = torch.ops.aten._unsafe_view(mm_4, [1, 15, 2048]); mm_4 = None\0A relu = torch.ops.aten.relu(_unsafe_view_6); _unsafe_view_6 = None\0A detach_3 = torch.ops.aten.detach(relu)\0A _param_constant9 = self._param_constant9\0A t_5 = torch.ops.aten.t(_param_constant9); _param_constant9 = None\0A view_14 = torch.ops.aten.view(relu, [15, 2048]); relu = None\0A mm_5 = torch.ops.aten.mm(view_14, t_5); view_14 = t_5 = None\0A _unsafe_view_7 = torch.ops.aten._unsafe_view(mm_5, [1, 15, 512]); mm_5 = None\0A add_6 = torch.ops.aten.add(add_4, _unsafe_view_7); add_4 = _unsafe_view_7 = None\0A pow_3 = torch.ops.aten.pow(add_6, 2)\0A mean_2 = torch.ops.aten.mean(pow_3, [-1], True); pow_3 = None\0A add_7 = torch.ops.aten.add(mean_2, 1e-06); mean_2 = None\0A rsqrt_2 = torch.ops.aten.rsqrt(add_7); add_7 = None\0A detach_4 = torch.ops.aten.detach(rsqrt_2)\0A mul_7 = torch.ops.aten.mul(add_6, rsqrt_2); rsqrt_2 = None\0A _param_constant10 = self._param_constant10\0A mul_8 = torch.ops.aten.mul(_param_constant10, mul_7); _param_constant10 = mul_7 = None\0A _param_constant11 = self._param_constant11\0A t_6 = torch.ops.aten.t(_param_constant11); _param_constant11 = None\0A view_15 = torch.ops.aten.view(mul_8, [15, 512])\0A mm_6 = torch.ops.aten.mm(view_15, t_6); view_15 = t_6 = None\0A _unsafe_view_8 = torch.ops.aten._unsafe_view(mm_6, [1, 15, 512]); mm_6 = None\0A view_16 = torch.ops.aten.view(_unsafe_view_8, [1, -1, 8, 64]); _unsafe_view_8 = None\0A transpose_5 = torch.ops.aten.transpose(view_16, 1, 2); view_16 = None\0A _param_constant12 = self._param_constant12\0A t_7 = torch.ops.aten.t(_param_constant12); _param_constant12 = None\0A view_17 = torch.ops.aten.view(mul_8, [15, 512])\0A mm_7 = torch.ops.aten.mm(view_17, t_7); view_17 = t_7 = None\0A _unsafe_view_9 = torch.ops.aten._unsafe_view(mm_7, [1, 15, 512]); mm_7 = None\0A view_18 = torch.ops.aten.view(_unsafe_view_9, [1, -1, 8, 64]); _unsafe_view_9 = None\0A transpose_6 = torch.ops.aten.transpose(view_18, 1, 2); view_18 = None\0A _param_constant13 = self._param_constant13\0A t_8 = torch.ops.aten.t(_param_constant13); _param_constant13 = None\0A view_19 = torch.ops.aten.view(mul_8, [15, 512]); mul_8 = None\0A mm_8 = torch.ops.aten.mm(view_19, t_8); view_19 = t_8 = None\0A _unsafe_view_10 = torch.ops.aten._unsafe_view(mm_8, [1, 15, 512]); mm_8 = None\0A view_20 = torch.ops.aten.view(_unsafe_view_10, [1, -1, 8, 64]); _unsafe_view_10 = None\0A transpose_7 = torch.ops.aten.transpose(view_20, 1, 2); view_20 = None\0A transpose_8 = torch.ops.aten.transpose(transpose_6, 3, 2); transpose_6 = None\0A expand_4 = torch.ops.aten.expand(transpose_5, [1, 8, 15, 64]); transpose_5 = None\0A view_21 = torch.ops.aten.view(expand_4, [8, 15, 64]); expand_4 = None\0A expand_5 = torch.ops.aten.expand(transpose_8, [1, 8, 64, 15]); transpose_8 = None\0A view_22 = torch.ops.aten.view(expand_5, [8, 64, 15]); expand_5 = None\0A bmm_2 = torch.ops.aten.bmm(view_21, view_22); view_21 = view_22 = None\0A _unsafe_view_11 = torch.ops.aten._unsafe_view(bmm_2, [1, 8, 15, 15]); bmm_2 = None\0A add__2 = torch.ops.aten.add_(_unsafe_view_11, add_3); _unsafe_view_11 = None\0A amax_1 = torch.ops.aten.amax(add__2, [-1], True)\0A sub_2 = torch.ops.aten.sub(add__2, amax_1); add__2 = amax_1 = None\0A exp_1 = torch.ops.aten.exp(sub_2); sub_2 = None\0A sum_2 = torch.ops.aten.sum(exp_1, [-1], True)\0A div_3 = torch.ops.aten.div(exp_1, sum_2); exp_1 = sum_2 = None\0A detach_5 = torch.ops.aten.detach(div_3)\0A expand_6 = torch.ops.aten.expand(div_3, [1, 8, 15, 15]); div_3 = None\0A view_23 = torch.ops.aten.view(expand_6, [8, 15, 15]); expand_6 = None\0A expand_7 = torch.ops.aten.expand(transpose_7, [1, 8, 15, 64]); transpose_7 = None\0A view_24 = torch.ops.aten.view(expand_7, [8, 15, 64]); expand_7 = None\0A bmm_3 = torch.ops.aten.bmm(view_23, view_24); view_23 = view_24 = None\0A _unsafe_view_12 = torch.ops.aten._unsafe_view(bmm_3, [1, 8, 15, 64]); bmm_3 = None\0A transpose_9 = torch.ops.aten.transpose(_unsafe_view_12, 1, 2); _unsafe_view_12 = None\0A clone_2 = torch.ops.aten.clone(transpose_9, memory_format = torch.contiguous_format); transpose_9 = None\0A view_25 = torch.ops.aten.view(clone_2, [1, -1, 512]); clone_2 = None\0A _param_constant14 = self._param_constant14\0A t_9 = torch.ops.aten.t(_param_constant14); _param_constant14 = None\0A view_26 = torch.ops.aten.view(view_25, [15, 512]); view_25 = None\0A mm_9 = torch.ops.aten.mm(view_26, t_9); view_26 = t_9 = None\0A _unsafe_view_13 = torch.ops.aten._unsafe_view(mm_9, [1, 15, 512]); mm_9 = None\0A add_8 = torch.ops.aten.add(add_6, _unsafe_view_13); add_6 = _unsafe_view_13 = None\0A pow_4 = torch.ops.aten.pow(add_8, 2)\0A mean_3 = torch.ops.aten.mean(pow_4, [-1], True); pow_4 = None\0A add_9 = torch.ops.aten.add(mean_3, 1e-06); mean_3 = None\0A rsqrt_3 = torch.ops.aten.rsqrt(add_9); add_9 = None\0A detach_6 = torch.ops.aten.detach(rsqrt_3)\0A mul_9 = torch.ops.aten.mul(add_8, rsqrt_3); rsqrt_3 = None\0A _param_constant15 = self._param_constant15\0A mul_10 = torch.ops.aten.mul(_param_constant15, mul_9); _param_constant15 = mul_9 = None\0A _param_constant16 = self._param_constant16\0A t_10 = torch.ops.aten.t(_param_constant16); _param_constant16 = None\0A view_27 = torch.ops.aten.view(mul_10, [15, 512]); mul_10 = None\0A mm_10 = torch.ops.aten.mm(view_27, t_10); view_27 = t_10 = None\0A _unsafe_view_14 = torch.ops.aten._unsafe_view(mm_10, [1, 15, 2048]); mm_10 = None\0A relu_1 = torch.ops.aten.relu(_unsafe_view_14); _unsafe_view_14 = None\0A detach_7 = torch.ops.aten.detach(relu_1)\0A _param_constant17 = self._param_constant17\0A t_11 = torch.ops.aten.t(_param_constant17); _param_constant17 = None\0A view_28 = torch.ops.aten.view(relu_1, [15, 2048]); relu_1 = None\0A mm_11 = torch.ops.aten.mm(view_28, t_11); view_28 = t_11 = None\0A _unsafe_view_15 = torch.ops.aten._unsafe_view(mm_11, [1, 15, 512]); mm_11 = None\0A add_10 = torch.ops.aten.add(add_8, _unsafe_view_15); add_8 = _unsafe_view_15 = None\0A pow_5 = torch.ops.aten.pow(add_10, 2)\0A mean_4 = torch.ops.aten.mean(pow_5, [-1], True); pow_5 = None\0A add_11 = torch.ops.aten.add(mean_4, 1e-06); mean_4 = None\0A rsqrt_4 = torch.ops.aten.rsqrt(add_11); add_11 = None\0A detach_8 = torch.ops.aten.detach(rsqrt_4)\0A mul_11 = torch.ops.aten.mul(add_10, rsqrt_4); rsqrt_4 = None\0A _param_constant18 = self._param_constant18\0A mul_12 = torch.ops.aten.mul(_param_constant18, mul_11); _param_constant18 = mul_11 = None\0A _param_constant19 = self._param_constant19\0A t_12 = torch.ops.aten.t(_param_constant19); _param_constant19 = None\0A view_29 = torch.ops.aten.view(mul_12, [15, 512])\0A mm_12 = torch.ops.aten.mm(view_29, t_12); view_29 = t_12 = None\0A _unsafe_view_16 = torch.ops.aten._unsafe_view(mm_12, [1, 15, 512]); mm_12 = None\0A view_30 = torch.ops.aten.view(_unsafe_view_16, [1, -1, 8, 64]); _unsafe_view_16 = None\0A transpose_10 = torch.ops.aten.transpose(view_30, 1, 2); view_30 = None\0A _param_constant20 = self._param_constant20\0A t_13 = torch.ops.aten.t(_param_constant20); _param_constant20 = None\0A view_31 = torch.ops.aten.view(mul_12, [15, 512])\0A mm_13 = torch.ops.aten.mm(view_31, t_13); view_31 = t_13 = None\0A _unsafe_view_17 = torch.ops.aten._unsafe_view(mm_13, [1, 15, 512]); mm_13 = None\0A view_32 = torch.ops.aten.view(_unsafe_view_17, [1, -1, 8, 64]); _unsafe_view_17 = None\0A transpose_11 = torch.ops.aten.transpose(view_32, 1, 2); view_32 = None\0A _param_constant21 = self._param_constant21\0A t_14 = torch.ops.aten.t(_param_constant21); _param_constant21 = None\0A view_33 = torch.ops.aten.view(mul_12, [15, 512]); mul_12 = None\0A mm_14 = torch.ops.aten.mm(view_33, t_14); view_33 = t_14 = None\0A _unsafe_view_18 = torch.ops.aten._unsafe_view(mm_14, [1, 15, 512]); mm_14 = None\0A view_34 = torch.ops.aten.view(_unsafe_view_18, [1, -1, 8, 64]); _unsafe_view_18 = None\0A transpose_12 = torch.ops.aten.transpose(view_34, 1, 2); view_34 = None\0A transpose_13 = torch.ops.aten.transpose(transpose_11, 3, 2); transpose_11 = None\0A expand_8 = torch.ops.aten.expand(transpose_10, [1, 8, 15, 64]); transpose_10 = None\0A view_35 = torch.ops.aten.view(expand_8, [8, 15, 64]); expand_8 = None\0A expand_9 = torch.ops.aten.expand(transpose_13, [1, 8, 64, 15]); transpose_13 = None\0A view_36 = torch.ops.aten.view(expand_9, [8, 64, 15]); expand_9 = None\0A bmm_4 = torch.ops.aten.bmm(view_35, view_36); view_35 = view_36 = None\0A _unsafe_view_19 = torch.ops.aten._unsafe_view(bmm_4, [1, 8, 15, 15]); bmm_4 = None\0A add__3 = torch.ops.aten.add_(_unsafe_view_19, add_3); _unsafe_view_19 = None\0A amax_2 = torch.ops.aten.amax(add__3, [-1], True)\0A sub_3 = torch.ops.aten.sub(add__3, amax_2); add__3 = amax_2 = None\0A exp_2 = torch.ops.aten.exp(sub_3); sub_3 = None\0A sum_3 = torch.ops.aten.sum(exp_2, [-1], True)\0A div_4 = torch.ops.aten.div(exp_2, sum_3); exp_2 = sum_3 = None\0A detach_9 = torch.ops.aten.detach(div_4)\0A expand_10 = torch.ops.aten.expand(div_4, [1, 8, 15, 15]); div_4 = None\0A view_37 = torch.ops.aten.view(expand_10, [8, 15, 15]); expand_10 = None\0A expand_11 = torch.ops.aten.expand(transpose_12, [1, 8, 15, 64]); transpose_12 = None\0A view_38 = torch.ops.aten.view(expand_11, [8, 15, 64]); expand_11 = None\0A bmm_5 = torch.ops.aten.bmm(view_37, view_38); view_37 = view_38 = None\0A _unsafe_view_20 = torch.ops.aten._unsafe_view(bmm_5, [1, 8, 15, 64]); bmm_5 = None\0A transpose_14 = torch.ops.aten.transpose(_unsafe_view_20, 1, 2); _unsafe_view_20 = None\0A clone_3 = torch.ops.aten.clone(transpose_14, memory_format = torch.contiguous_format); transpose_14 = None\0A view_39 = torch.ops.aten.view(clone_3, [1, -1, 512]); clone_3 = None\0A _param_constant22 = self._param_constant22\0A t_15 = torch.ops.aten.t(_param_constant22); _param_constant22 = None\0A view_40 = torch.ops.aten.view(view_39, [15, 512]); view_39 = None\0A mm_15 = torch.ops.aten.mm(view_40, t_15); view_40 = t_15 = None\0A _unsafe_view_21 = torch.ops.aten._unsafe_view(mm_15, [1, 15, 512]); mm_15 = None\0A add_12 = torch.ops.aten.add(add_10, _unsafe_view_21); add_10 = _unsafe_view_21 = None\0A pow_6 = torch.ops.aten.pow(add_12, 2)\0A mean_5 = torch.ops.aten.mean(pow_6, [-1], True); pow_6 = None\0A add_13 = torch.ops.aten.add(mean_5, 1e-06); mean_5 = None\0A rsqrt_5 = torch.ops.aten.rsqrt(add_13); add_13 = None\0A detach_10 = torch.ops.aten.detach(rsqrt_5)\0A mul_13 = torch.ops.aten.mul(add_12, rsqrt_5); rsqrt_5 = None\0A _param_constant23 = self._param_constant23\0A mul_14 = torch.ops.aten.mul(_param_constant23, mul_13); _param_constant23 = mul_13 = None\0A _param_constant24 = self._param_constant24\0A t_16 = torch.ops.aten.t(_param_constant24); _param_constant24 = None\0A view_41 = torch.ops.aten.view(mul_14, [15, 512]); mul_14 = None\0A mm_16 = torch.ops.aten.mm(view_41, t_16); view_41 = t_16 = None\0A _unsafe_view_22 = torch.ops.aten._unsafe_view(mm_16, [1, 15, 2048]); mm_16 = None\0A relu_2 = torch.ops.aten.relu(_unsafe_view_22); _unsafe_view_22 = None\0A detach_11 = torch.ops.aten.detach(relu_2)\0A _param_constant25 = self._param_constant25\0A t_17 = torch.ops.aten.t(_param_constant25); _param_constant25 = None\0A view_42 = torch.ops.aten.view(relu_2, [15, 2048]); relu_2 = None\0A mm_17 = torch.ops.aten.mm(view_42, t_17); view_42 = t_17 = None\0A _unsafe_view_23 = torch.ops.aten._unsafe_view(mm_17, [1, 15, 512]); mm_17 = None\0A add_14 = torch.ops.aten.add(add_12, _unsafe_view_23); add_12 = _unsafe_view_23 = None\0A pow_7 = torch.ops.aten.pow(add_14, 2)\0A mean_6 = torch.ops.aten.mean(pow_7, [-1], True); pow_7 = None\0A add_15 = torch.ops.aten.add(mean_6, 1e-06); mean_6 = None\0A rsqrt_6 = torch.ops.aten.rsqrt(add_15); add_15 = None\0A detach_12 = torch.ops.aten.detach(rsqrt_6)\0A mul_15 = torch.ops.aten.mul(add_14, rsqrt_6); rsqrt_6 = None\0A _param_constant26 = self._param_constant26\0A mul_16 = torch.ops.aten.mul(_param_constant26, mul_15); _param_constant26 = mul_15 = None\0A _param_constant27 = self._param_constant27\0A t_18 = torch.ops.aten.t(_param_constant27); _param_constant27 = None\0A view_43 = torch.ops.aten.view(mul_16, [15, 512])\0A mm_18 = torch.ops.aten.mm(view_43, t_18); view_43 = t_18 = None\0A _unsafe_view_24 = torch.ops.aten._unsafe_view(mm_18, [1, 15, 512]); mm_18 = None\0A view_44 = torch.ops.aten.view(_unsafe_view_24, [1, -1, 8, 64]); _unsafe_view_24 = None\0A transpose_15 = torch.ops.aten.transpose(view_44, 1, 2); view_44 = None\0A _param_constant28 = self._param_constant28\0A t_19 = torch.ops.aten.t(_param_constant28); _param_constant28 = None\0A view_45 = torch.ops.aten.view(mul_16, [15, 512])\0A mm_19 = torch.ops.aten.mm(view_45, t_19); view_45 = t_19 = None\0A _unsafe_view_25 = torch.ops.aten._unsafe_view(mm_19, [1, 15, 512]); mm_19 = None\0A view_46 = torch.ops.aten.view(_unsafe_view_25, [1, -1, 8, 64]); _unsafe_view_25 = None\0A transpose_16 = torch.ops.aten.transpose(view_46, 1, 2); view_46 = None\0A _param_constant29 = self._param_constant29\0A t_20 = torch.ops.aten.t(_param_constant29); _param_constant29 = None\0A view_47 = torch.ops.aten.view(mul_16, [15, 512]); mul_16 = None\0A mm_20 = torch.ops.aten.mm(view_47, t_20); view_47 = t_20 = None\0A _unsafe_view_26 = torch.ops.aten._unsafe_view(mm_20, [1, 15, 512]); mm_20 = None\0A view_48 = torch.ops.aten.view(_unsafe_view_26, [1, -1, 8, 64]); _unsafe_view_26 = None\0A transpose_17 = torch.ops.aten.transpose(view_48, 1, 2); view_48 = None\0A transpose_18 = torch.ops.aten.transpose(transpose_16, 3, 2); transpose_16 = None\0A expand_12 = torch.ops.aten.expand(transpose_15, [1, 8, 15, 64]); transpose_15 = None\0A view_49 = torch.ops.aten.view(expand_12, [8, 15, 64]); expand_12 = None\0A expand_13 = torch.ops.aten.expand(transpose_18, [1, 8, 64, 15]); transpose_18 = None\0A view_50 = torch.ops.aten.view(expand_13, [8, 64, 15]); expand_13 = None\0A bmm_6 = torch.ops.aten.bmm(view_49, view_50); view_49 = view_50 = None\0A _unsafe_view_27 = torch.ops.aten._unsafe_view(bmm_6, [1, 8, 15, 15]); bmm_6 = None\0A add__4 = torch.ops.aten.add_(_unsafe_view_27, add_3); _unsafe_view_27 = None\0A amax_3 = torch.ops.aten.amax(add__4, [-1], True)\0A sub_4 = torch.ops.aten.sub(add__4, amax_3); add__4 = amax_3 = None\0A exp_3 = torch.ops.aten.exp(sub_4); sub_4 = None\0A sum_4 = torch.ops.aten.sum(exp_3, [-1], True)\0A div_5 = torch.ops.aten.div(exp_3, sum_4); exp_3 = sum_4 = None\0A detach_13 = torch.ops.aten.detach(div_5)\0A expand_14 = torch.ops.aten.expand(div_5, [1, 8, 15, 15]); div_5 = None\0A view_51 = torch.ops.aten.view(expand_14, [8, 15, 15]); expand_14 = None\0A expand_15 = torch.ops.aten.expand(transpose_17, [1, 8, 15, 64]); transpose_17 = None\0A view_52 = torch.ops.aten.view(expand_15, [8, 15, 64]); expand_15 = None\0A bmm_7 = torch.ops.aten.bmm(view_51, view_52); view_51 = view_52 = None\0A _unsafe_view_28 = torch.ops.aten._unsafe_view(bmm_7, [1, 8, 15, 64]); bmm_7 = None\0A transpose_19 = torch.ops.aten.transpose(_unsafe_view_28, 1, 2); _unsafe_view_28 = None\0A clone_4 = torch.ops.aten.clone(transpose_19, memory_format = torch.contiguous_format); transpose_19 = None\0A view_53 = torch.ops.aten.view(clone_4, [1, -1, 512]); clone_4 = None\0A _param_constant30 = self._param_constant30\0A t_21 = torch.ops.aten.t(_param_constant30); _param_constant30 = None\0A view_54 = torch.ops.aten.view(view_53, [15, 512]); view_53 = None\0A mm_21 = torch.ops.aten.mm(view_54, t_21); view_54 = t_21 = None\0A _unsafe_view_29 = torch.ops.aten._unsafe_view(mm_21, [1, 15, 512]); mm_21 = None\0A add_16 = torch.ops.aten.add(add_14, _unsafe_view_29); add_14 = _unsafe_view_29 = None\0A pow_8 = torch.ops.aten.pow(add_16, 2)\0A mean_7 = torch.ops.aten.mean(pow_8, [-1], True); pow_8 = None\0A add_17 = torch.ops.aten.add(mean_7, 1e-06); mean_7 = None\0A rsqrt_7 = torch.ops.aten.rsqrt(add_17); add_17 = None\0A detach_14 = torch.ops.aten.detach(rsqrt_7)\0A mul_17 = torch.ops.aten.mul(add_16, rsqrt_7); rsqrt_7 = None\0A _param_constant31 = self._param_constant31\0A mul_18 = torch.ops.aten.mul(_param_constant31, mul_17); _param_constant31 = mul_17 = None\0A _param_constant32 = self._param_constant32\0A t_22 = torch.ops.aten.t(_param_constant32); _param_constant32 = None\0A view_55 = torch.ops.aten.view(mul_18, [15, 512]); mul_18 = None\0A mm_22 = torch.ops.aten.mm(view_55, t_22); view_55 = t_22 = None\0A _unsafe_view_30 = torch.ops.aten._unsafe_view(mm_22, [1, 15, 2048]); mm_22 = None\0A relu_3 = torch.ops.aten.relu(_unsafe_view_30); _unsafe_view_30 = None\0A detach_15 = torch.ops.aten.detach(relu_3)\0A _param_constant33 = self._param_constant33\0A t_23 = torch.ops.aten.t(_param_constant33); _param_constant33 = None\0A view_56 = torch.ops.aten.view(relu_3, [15, 2048]); relu_3 = None\0A mm_23 = torch.ops.aten.mm(view_56, t_23); view_56 = t_23 = None\0A _unsafe_view_31 = torch.ops.aten._unsafe_view(mm_23, [1, 15, 512]); mm_23 = None\0A add_18 = torch.ops.aten.add(add_16, _unsafe_view_31); add_16 = _unsafe_view_31 = None\0A pow_9 = torch.ops.aten.pow(add_18, 2)\0A mean_8 = torch.ops.aten.mean(pow_9, [-1], True); pow_9 = None\0A add_19 = torch.ops.aten.add(mean_8, 1e-06); mean_8 = None\0A rsqrt_8 = torch.ops.aten.rsqrt(add_19); add_19 = None\0A detach_16 = torch.ops.aten.detach(rsqrt_8)\0A mul_19 = torch.ops.aten.mul(add_18, rsqrt_8); rsqrt_8 = None\0A _param_constant34 = self._param_constant34\0A mul_20 = torch.ops.aten.mul(_param_constant34, mul_19); _param_constant34 = mul_19 = None\0A _param_constant35 = self._param_constant35\0A t_24 = torch.ops.aten.t(_param_constant35); _param_constant35 = None\0A view_57 = torch.ops.aten.view(mul_20, [15, 512])\0A mm_24 = torch.ops.aten.mm(view_57, t_24); view_57 = t_24 = None\0A _unsafe_view_32 = torch.ops.aten._unsafe_view(mm_24, [1, 15, 512]); mm_24 = None\0A view_58 = torch.ops.aten.view(_unsafe_view_32, [1, -1, 8, 64]); _unsafe_view_32 = None\0A transpose_20 = torch.ops.aten.transpose(view_58, 1, 2); view_58 = None\0A _param_constant36 = self._param_constant36\0A t_25 = torch.ops.aten.t(_param_constant36); _param_constant36 = None\0A view_59 = torch.ops.aten.view(mul_20, [15, 512])\0A mm_25 = torch.ops.aten.mm(view_59, t_25); view_59 = t_25 = None\0A _unsafe_view_33 = torch.ops.aten._unsafe_view(mm_25, [1, 15, 512]); mm_25 = None\0A view_60 = torch.ops.aten.view(_unsafe_view_33, [1, -1, 8, 64]); _unsafe_view_33 = None\0A transpose_21 = torch.ops.aten.transpose(view_60, 1, 2); view_60 = None\0A _param_constant37 = self._param_constant37\0A t_26 = torch.ops.aten.t(_param_constant37); _param_constant37 = None\0A view_61 = torch.ops.aten.view(mul_20, [15, 512]); mul_20 = None\0A mm_26 = torch.ops.aten.mm(view_61, t_26); view_61 = t_26 = None\0A _unsafe_view_34 = torch.ops.aten._unsafe_view(mm_26, [1, 15, 512]); mm_26 = None\0A view_62 = torch.ops.aten.view(_unsafe_view_34, [1, -1, 8, 64]); _unsafe_view_34 = None\0A transpose_22 = torch.ops.aten.transpose(view_62, 1, 2); view_62 = None\0A transpose_23 = torch.ops.aten.transpose(transpose_21, 3, 2); transpose_21 = None\0A expand_16 = torch.ops.aten.expand(transpose_20, [1, 8, 15, 64]); transpose_20 = None\0A view_63 = torch.ops.aten.view(expand_16, [8, 15, 64]); expand_16 = None\0A expand_17 = torch.ops.aten.expand(transpose_23, [1, 8, 64, 15]); transpose_23 = None\0A view_64 = torch.ops.aten.view(expand_17, [8, 64, 15]); expand_17 = None\0A bmm_8 = torch.ops.aten.bmm(view_63, view_64); view_63 = view_64 = None\0A _unsafe_view_35 = torch.ops.aten._unsafe_view(bmm_8, [1, 8, 15, 15]); bmm_8 = None\0A add__5 = torch.ops.aten.add_(_unsafe_view_35, add_3); _unsafe_view_35 = None\0A amax_4 = torch.ops.aten.amax(add__5, [-1], True)\0A sub_5 = torch.ops.aten.sub(add__5, amax_4); add__5 = amax_4 = None\0A exp_4 = torch.ops.aten.exp(sub_5); sub_5 = None\0A sum_5 = torch.ops.aten.sum(exp_4, [-1], True)\0A div_6 = torch.ops.aten.div(exp_4, sum_5); exp_4 = sum_5 = None\0A detach_17 = torch.ops.aten.detach(div_6)\0A expand_18 = torch.ops.aten.expand(div_6, [1, 8, 15, 15]); div_6 = None\0A view_65 = torch.ops.aten.view(expand_18, [8, 15, 15]); expand_18 = None\0A expand_19 = torch.ops.aten.expand(transpose_22, [1, 8, 15, 64]); transpose_22 = None\0A view_66 = torch.ops.aten.view(expand_19, [8, 15, 64]); expand_19 = None\0A bmm_9 = torch.ops.aten.bmm(view_65, view_66); view_65 = view_66 = None\0A _unsafe_view_36 = torch.ops.aten._unsafe_view(bmm_9, [1, 8, 15, 64]); bmm_9 = None\0A transpose_24 = torch.ops.aten.transpose(_unsafe_view_36, 1, 2); _unsafe_view_36 = None\0A clone_5 = torch.ops.aten.clone(transpose_24, memory_format = torch.contiguous_format); transpose_24 = None\0A view_67 = torch.ops.aten.view(clone_5, [1, -1, 512]); clone_5 = None\0A _param_constant38 = self._param_constant38\0A t_27 = torch.ops.aten.t(_param_constant38); _param_constant38 = None\0A view_68 = torch.ops.aten.view(view_67, [15, 512]); view_67 = None\0A mm_27 = torch.ops.aten.mm(view_68, t_27); view_68 = t_27 = None\0A _unsafe_view_37 = torch.ops.aten._unsafe_view(mm_27, [1, 15, 512]); mm_27 = None\0A add_20 = torch.ops.aten.add(add_18, _unsafe_view_37); add_18 = _unsafe_view_37 = None\0A pow_10 = torch.ops.aten.pow(add_20, 2)\0A mean_9 = torch.ops.aten.mean(pow_10, [-1], True); pow_10 = None\0A add_21 = torch.ops.aten.add(mean_9, 1e-06); mean_9 = None\0A rsqrt_9 = torch.ops.aten.rsqrt(add_21); add_21 = None\0A detach_18 = torch.ops.aten.detach(rsqrt_9)\0A mul_21 = torch.ops.aten.mul(add_20, rsqrt_9); rsqrt_9 = None\0A _param_constant39 = self._param_constant39\0A mul_22 = torch.ops.aten.mul(_param_constant39, mul_21); _param_constant39 = mul_21 = None\0A _param_constant40 = self._param_constant40\0A t_28 = torch.ops.aten.t(_param_constant40); _param_constant40 = None\0A view_69 = torch.ops.aten.view(mul_22, [15, 512]); mul_22 = None\0A mm_28 = torch.ops.aten.mm(view_69, t_28); view_69 = t_28 = None\0A _unsafe_view_38 = torch.ops.aten._unsafe_view(mm_28, [1, 15, 2048]); mm_28 = None\0A relu_4 = torch.ops.aten.relu(_unsafe_view_38); _unsafe_view_38 = None\0A detach_19 = torch.ops.aten.detach(relu_4)\0A _param_constant41 = self._param_constant41\0A t_29 = torch.ops.aten.t(_param_constant41); _param_constant41 = None\0A view_70 = torch.ops.aten.view(relu_4, [15, 2048]); relu_4 = None\0A mm_29 = torch.ops.aten.mm(view_70, t_29); view_70 = t_29 = None\0A _unsafe_view_39 = torch.ops.aten._unsafe_view(mm_29, [1, 15, 512]); mm_29 = None\0A add_22 = torch.ops.aten.add(add_20, _unsafe_view_39); add_20 = _unsafe_view_39 = None\0A pow_11 = torch.ops.aten.pow(add_22, 2)\0A mean_10 = torch.ops.aten.mean(pow_11, [-1], True); pow_11 = None\0A add_23 = torch.ops.aten.add(mean_10, 1e-06); mean_10 = None\0A rsqrt_10 = torch.ops.aten.rsqrt(add_23); add_23 = None\0A detach_20 = torch.ops.aten.detach(rsqrt_10)\0A mul_23 = torch.ops.aten.mul(add_22, rsqrt_10); rsqrt_10 = None\0A _param_constant42 = self._param_constant42\0A mul_24 = torch.ops.aten.mul(_param_constant42, mul_23); _param_constant42 = mul_23 = None\0A _param_constant43 = self._param_constant43\0A t_30 = torch.ops.aten.t(_param_constant43); _param_constant43 = None\0A view_71 = torch.ops.aten.view(mul_24, [15, 512])\0A mm_30 = torch.ops.aten.mm(view_71, t_30); view_71 = t_30 = None\0A _unsafe_view_40 = torch.ops.aten._unsafe_view(mm_30, [1, 15, 512]); mm_30 = None\0A view_72 = torch.ops.aten.view(_unsafe_view_40, [1, -1, 8, 64]); _unsafe_view_40 = None\0A transpose_25 = torch.ops.aten.transpose(view_72, 1, 2); view_72 = None\0A _param_constant44 = self._param_constant44\0A t_31 = torch.ops.aten.t(_param_constant44); _param_constant44 = None\0A view_73 = torch.ops.aten.view(mul_24, [15, 512])\0A mm_31 = torch.ops.aten.mm(view_73, t_31); view_73 = t_31 = None\0A _unsafe_view_41 = torch.ops.aten._unsafe_view(mm_31, [1, 15, 512]); mm_31 = None\0A view_74 = torch.ops.aten.view(_unsafe_view_41, [1, -1, 8, 64]); _unsafe_view_41 = None\0A transpose_26 = torch.ops.aten.transpose(view_74, 1, 2); view_74 = None\0A _param_constant45 = self._param_constant45\0A t_32 = torch.ops.aten.t(_param_constant45); _param_constant45 = None\0A view_75 = torch.ops.aten.view(mul_24, [15, 512]); mul_24 = None\0A mm_32 = torch.ops.aten.mm(view_75, t_32); view_75 = t_32 = None\0A _unsafe_view_42 = torch.ops.aten._unsafe_view(mm_32, [1, 15, 512]); mm_32 = None\0A view_76 = torch.ops.aten.view(_unsafe_view_42, [1, -1, 8, 64]); _unsafe_view_42 = None\0A transpose_27 = torch.ops.aten.transpose(view_76, 1, 2); view_76 = None\0A transpose_28 = torch.ops.aten.transpose(transpose_26, 3, 2); transpose_26 = None\0A expand_20 = torch.ops.aten.expand(transpose_25, [1, 8, 15, 64]); transpose_25 = None\0A view_77 = torch.ops.aten.view(expand_20, [8, 15, 64]); expand_20 = None\0A expand_21 = torch.ops.aten.expand(transpose_28, [1, 8, 64, 15]); transpose_28 = None\0A view_78 = torch.ops.aten.view(expand_21, [8, 64, 15]); expand_21 = None\0A bmm_10 = torch.ops.aten.bmm(view_77, view_78); view_77 = view_78 = None\0A _unsafe_view_43 = torch.ops.aten._unsafe_view(bmm_10, [1, 8, 15, 15]); bmm_10 = None\0A add__6 = torch.ops.aten.add_(_unsafe_view_43, add_3); _unsafe_view_43 = add_3 = None\0A amax_5 = torch.ops.aten.amax(add__6, [-1], True)\0A sub_6 = torch.ops.aten.sub(add__6, amax_5); add__6 = amax_5 = None\0A exp_5 = torch.ops.aten.exp(sub_6); sub_6 = None\0A sum_6 = torch.ops.aten.sum(exp_5, [-1], True)\0A div_7 = torch.ops.aten.div(exp_5, sum_6); exp_5 = sum_6 = None\0A detach_21 = torch.ops.aten.detach(div_7)\0A expand_22 = torch.ops.aten.expand(div_7, [1, 8, 15, 15]); div_7 = None\0A view_79 = torch.ops.aten.view(expand_22, [8, 15, 15]); expand_22 = None\0A expand_23 = torch.ops.aten.expand(transpose_27, [1, 8, 15, 64]); transpose_27 = None\0A view_80 = torch.ops.aten.view(expand_23, [8, 15, 64]); expand_23 = None\0A bmm_11 = torch.ops.aten.bmm(view_79, view_80); view_79 = view_80 = None\0A _unsafe_view_44 = torch.ops.aten._unsafe_view(bmm_11, [1, 8, 15, 64]); bmm_11 = None\0A transpose_29 = torch.ops.aten.transpose(_unsafe_view_44, 1, 2); _unsafe_view_44 = None\0A clone_6 = torch.ops.aten.clone(transpose_29, memory_format = torch.contiguous_format); transpose_29 = None\0A view_81 = torch.ops.aten.view(clone_6, [1, -1, 512]); clone_6 = None\0A _param_constant46 = self._param_constant46\0A t_33 = torch.ops.aten.t(_param_constant46); _param_constant46 = None\0A view_82 = torch.ops.aten.view(view_81, [15, 512]); view_81 = None\0A mm_33 = torch.ops.aten.mm(view_82, t_33); view_82 = t_33 = None\0A _unsafe_view_45 = torch.ops.aten._unsafe_view(mm_33, [1, 15, 512]); mm_33 = None\0A add_24 = torch.ops.aten.add(add_22, _unsafe_view_45); add_22 = _unsafe_view_45 = None\0A pow_12 = torch.ops.aten.pow(add_24, 2)\0A mean_11 = torch.ops.aten.mean(pow_12, [-1], True); pow_12 = None\0A add_25 = torch.ops.aten.add(mean_11, 1e-06); mean_11 = None\0A rsqrt_11 = torch.ops.aten.rsqrt(add_25); add_25 = None\0A detach_22 = torch.ops.aten.detach(rsqrt_11)\0A mul_25 = torch.ops.aten.mul(add_24, rsqrt_11); rsqrt_11 = None\0A _param_constant47 = self._param_constant47\0A mul_26 = torch.ops.aten.mul(_param_constant47, mul_25); _param_constant47 = mul_25 = None\0A _param_constant48 = self._param_constant48\0A t_34 = torch.ops.aten.t(_param_constant48); _param_constant48 = None\0A view_83 = torch.ops.aten.view(mul_26, [15, 512]); mul_26 = None\0A mm_34 = torch.ops.aten.mm(view_83, t_34); view_83 = t_34 = None\0A _unsafe_view_46 = torch.ops.aten._unsafe_view(mm_34, [1, 15, 2048]); mm_34 = None\0A relu_5 = torch.ops.aten.relu(_unsafe_view_46); _unsafe_view_46 = None\0A detach_23 = torch.ops.aten.detach(relu_5)\0A _param_constant49 = self._param_constant49\0A t_35 = torch.ops.aten.t(_param_constant49); _param_constant49 = None\0A view_84 = torch.ops.aten.view(relu_5, [15, 2048]); relu_5 = None\0A mm_35 = torch.ops.aten.mm(view_84, t_35); view_84 = t_35 = None\0A _unsafe_view_47 = torch.ops.aten._unsafe_view(mm_35, [1, 15, 512]); mm_35 = None\0A add_26 = torch.ops.aten.add(add_24, _unsafe_view_47); add_24 = _unsafe_view_47 = None\0A pow_13 = torch.ops.aten.pow(add_26, 2)\0A mean_12 = torch.ops.aten.mean(pow_13, [-1], True); pow_13 = None\0A add_27 = torch.ops.aten.add(mean_12, 1e-06); mean_12 = None\0A rsqrt_12 = torch.ops.aten.rsqrt(add_27); add_27 = None\0A detach_24 = torch.ops.aten.detach(rsqrt_12)\0A mul_27 = torch.ops.aten.mul(add_26, rsqrt_12); add_26 = rsqrt_12 = None\0A _param_constant50 = self._param_constant50\0A mul_28 = torch.ops.aten.mul(_param_constant50, mul_27); _param_constant50 = mul_27 = None\0A view_85 = torch.ops.aten.view(masked_fill_, [-1, 4]); masked_fill_ = None\0A _param_constant0_1 = self._param_constant0\0A embedding_2 = torch.ops.aten.embedding(_param_constant0_1, view_85); _param_constant0_1 = view_85 = None\0A ones_1 = torch.ops.aten.ones([1, 4], device = device(type='cpu'), pin_memory = False)\0A ones_2 = torch.ops.aten.ones([1, 15], dtype = torch.int64, device = device(type='cpu'), pin_memory = False)\0A arange_2 = torch.ops.aten.arange(4, device = device(type='cpu'), pin_memory = False)\0A unsqueeze_5 = torch.ops.aten.unsqueeze(arange_2, 0)\0A unsqueeze_6 = torch.ops.aten.unsqueeze(unsqueeze_5, 1); unsqueeze_5 = None\0A slice_7 = torch.ops.aten.slice(unsqueeze_6, 2, 0, 9223372036854775807); unsqueeze_6 = None\0A repeat = torch.ops.aten.repeat(slice_7, [1, 4, 1]); slice_7 = None\0A unsqueeze_7 = torch.ops.aten.unsqueeze(arange_2, 0); arange_2 = None\0A slice_8 = torch.ops.aten.slice(unsqueeze_7, 1, 0, 9223372036854775807); unsqueeze_7 = None\0A unsqueeze_8 = torch.ops.aten.unsqueeze(slice_8, 2); slice_8 = None\0A le = torch.ops.aten.le(repeat, unsqueeze_8); repeat = unsqueeze_8 = None\0A convert_element_type_3 = torch.ops.prims.convert_element_type(le, torch.float32); le = None\0A slice_9 = torch.ops.aten.slice(convert_element_type_3, 0, 0, 9223372036854775807); convert_element_type_3 = None\0A unsqueeze_9 = torch.ops.aten.unsqueeze(slice_9, 1); slice_9 = None\0A slice_10 = torch.ops.aten.slice(unsqueeze_9, 2, 0, 9223372036854775807); unsqueeze_9 = None\0A slice_11 = torch.ops.aten.slice(slice_10, 3, 0, 9223372036854775807); slice_10 = None\0A slice_12 = torch.ops.aten.slice(ones_1, 0, 0, 9223372036854775807); ones_1 = None\0A unsqueeze_10 = torch.ops.aten.unsqueeze(slice_12, 1); slice_12 = None\0A unsqueeze_11 = torch.ops.aten.unsqueeze(unsqueeze_10, 2); unsqueeze_10 = None\0A slice_13 = torch.ops.aten.slice(unsqueeze_11, 3, 0, 9223372036854775807); unsqueeze_11 = None\0A mul_29 = torch.ops.aten.mul(slice_11, slice_13); slice_11 = slice_13 = None\0A rsub_1 = torch.ops.aten.rsub(mul_29, 1.0); mul_29 = None\0A mul_30 = torch.ops.aten.mul(rsub_1, -3.4028234663852886e+38); rsub_1 = None\0A slice_14 = torch.ops.aten.slice(ones_2, 0, 0, 9223372036854775807); ones_2 = None\0A unsqueeze_12 = torch.ops.aten.unsqueeze(slice_14, 1); slice_14 = None\0A unsqueeze_13 = torch.ops.aten.unsqueeze(unsqueeze_12, 2); unsqueeze_12 = None\0A slice_15 = torch.ops.aten.slice(unsqueeze_13, 3, 0, 9223372036854775807); unsqueeze_13 = None\0A convert_element_type_4 = torch.ops.prims.convert_element_type(slice_15, torch.float32); slice_15 = None\0A rsub_2 = torch.ops.aten.rsub(convert_element_type_4, 1.0); convert_element_type_4 = None\0A mul_31 = torch.ops.aten.mul(rsub_2, -3.4028234663852886e+38); rsub_2 = None\0A pow_14 = torch.ops.aten.pow(embedding_2, 2)\0A mean_13 = torch.ops.aten.mean(pow_14, [-1], True); pow_14 = None\0A add_28 = torch.ops.aten.add(mean_13, 1e-06); mean_13 = None\0A rsqrt_13 = torch.ops.aten.rsqrt(add_28); add_28 = None\0A detach_25 = torch.ops.aten.detach(rsqrt_13)\0A mul_32 = torch.ops.aten.mul(embedding_2, rsqrt_13); rsqrt_13 = None\0A _param_constant51 = self._param_constant51\0A mul_33 = torch.ops.aten.mul(_param_constant51, mul_32); _param_constant51 = mul_32 = None\0A _param_constant52 = self._param_constant52\0A t_36 = torch.ops.aten.t(_param_constant52); _param_constant52 = None\0A view_86 = torch.ops.aten.view(mul_33, [4, 512])\0A mm_36 = torch.ops.aten.mm(view_86, t_36); view_86 = t_36 = None\0A _unsafe_view_48 = torch.ops.aten._unsafe_view(mm_36, [1, 4, 512]); mm_36 = None\0A view_87 = torch.ops.aten.view(_unsafe_view_48, [1, -1, 8, 64]); _unsafe_view_48 = None\0A transpose_30 = torch.ops.aten.transpose(view_87, 1, 2); view_87 = None\0A _param_constant53 = self._param_constant53\0A t_37 = torch.ops.aten.t(_param_constant53); _param_constant53 = None\0A view_88 = torch.ops.aten.view(mul_33, [4, 512])\0A mm_37 = torch.ops.aten.mm(view_88, t_37); view_88 = t_37 = None\0A _unsafe_view_49 = torch.ops.aten._unsafe_view(mm_37, [1, 4, 512]); mm_37 = None\0A view_89 = torch.ops.aten.view(_unsafe_view_49, [1, -1, 8, 64]); _unsafe_view_49 = None\0A transpose_31 = torch.ops.aten.transpose(view_89, 1, 2); view_89 = None\0A _param_constant54 = self._param_constant54\0A t_38 = torch.ops.aten.t(_param_constant54); _param_constant54 = None\0A view_90 = torch.ops.aten.view(mul_33, [4, 512]); mul_33 = None\0A mm_38 = torch.ops.aten.mm(view_90, t_38); view_90 = t_38 = None\0A _unsafe_view_50 = torch.ops.aten._unsafe_view(mm_38, [1, 4, 512]); mm_38 = None\0A view_91 = torch.ops.aten.view(_unsafe_view_50, [1, -1, 8, 64]); _unsafe_view_50 = None\0A transpose_32 = torch.ops.aten.transpose(view_91, 1, 2); view_91 = None\0A transpose_33 = torch.ops.aten.transpose(transpose_31, 3, 2); transpose_31 = None\0A expand_24 = torch.ops.aten.expand(transpose_30, [1, 8, 4, 64]); transpose_30 = None\0A view_92 = torch.ops.aten.view(expand_24, [8, 4, 64]); expand_24 = None\0A expand_25 = torch.ops.aten.expand(transpose_33, [1, 8, 64, 4]); transpose_33 = None\0A view_93 = torch.ops.aten.view(expand_25, [8, 64, 4]); expand_25 = None\0A bmm_12 = torch.ops.aten.bmm(view_92, view_93); view_92 = view_93 = None\0A _unsafe_view_51 = torch.ops.aten._unsafe_view(bmm_12, [1, 8, 4, 4]); bmm_12 = None\0A arange_3 = torch.ops.aten.arange(4, dtype = torch.int64, device = device(type='cpu'), pin_memory = False)\0A slice_16 = torch.ops.aten.slice(arange_3, 0, 0, 9223372036854775807); arange_3 = None\0A unsqueeze_14 = torch.ops.aten.unsqueeze(slice_16, 1); slice_16 = None\0A arange_4 = torch.ops.aten.arange(4, dtype = torch.int64, device = device(type='cpu'), pin_memory = False)\0A unsqueeze_15 = torch.ops.aten.unsqueeze(arange_4, 0); arange_4 = None\0A slice_17 = torch.ops.aten.slice(unsqueeze_15, 1, 0, 9223372036854775807); unsqueeze_15 = None\0A sub_7 = torch.ops.aten.sub(slice_17, unsqueeze_14); slice_17 = unsqueeze_14 = None\0A zeros_like = torch.ops.aten.zeros_like(sub_7, dtype = torch.int64, layout = torch.strided, device = device(type='cpu'), pin_memory = False)\0A minimum_1 = torch.ops.aten.minimum(sub_7, zeros_like); sub_7 = zeros_like = None\0A neg = torch.ops.aten.neg(minimum_1); minimum_1 = None\0A lt_1 = torch.ops.aten.lt(neg, 16)\0A convert_element_type_5 = torch.ops.prims.convert_element_type(neg, torch.float32)\0A div_8 = torch.ops.aten.div(convert_element_type_5, 16); convert_element_type_5 = None\0A log_1 = torch.ops.aten.log(div_8); div_8 = None\0A div_9 = torch.ops.aten.div(log_1, 2.0794415416798357); log_1 = None\0A mul_34 = torch.ops.aten.mul(div_9, 16); div_9 = None\0A convert_element_type_6 = torch.ops.prims.convert_element_type(mul_34, torch.int64); mul_34 = None\0A add_29 = torch.ops.aten.add(convert_element_type_6, 16); convert_element_type_6 = None\0A full_like_1 = torch.ops.aten.full_like(add_29, 31, dtype = torch.int64, layout = torch.strided, device = device(type='cpu'), pin_memory = False)\0A minimum_2 = torch.ops.aten.minimum(add_29, full_like_1); add_29 = full_like_1 = None\0A where_1 = torch.ops.aten.where(lt_1, neg, minimum_2); lt_1 = neg = minimum_2 = None\0A add_30 = torch.ops.aten.add(where_1, 0); where_1 = None\0A _param_constant55 = self._param_constant55\0A embedding_3 = torch.ops.aten.embedding(_param_constant55, add_30); _param_constant55 = add_30 = None\0A permute_1 = torch.ops.aten.permute(embedding_3, [2, 0, 1]); embedding_3 = None\0A unsqueeze_16 = torch.ops.aten.unsqueeze(permute_1, 0); permute_1 = None\0A add_31 = torch.ops.aten.add(unsqueeze_16, mul_30); unsqueeze_16 = mul_30 = None\0A add__7 = torch.ops.aten.add_(_unsafe_view_51, add_31); _unsafe_view_51 = None\0A amax_6 = torch.ops.aten.amax(add__7, [-1], True)\0A sub_8 = torch.ops.aten.sub(add__7, amax_6); add__7 = amax_6 = None\0A exp_6 = torch.ops.aten.exp(sub_8); sub_8 = None\0A sum_7 = torch.ops.aten.sum(exp_6, [-1], True)\0A div_10 = torch.ops.aten.div(exp_6, sum_7); exp_6 = sum_7 = None\0A detach_26 = torch.ops.aten.detach(div_10)\0A expand_26 = torch.ops.aten.expand(div_10, [1, 8, 4, 4]); div_10 = None\0A view_94 = torch.ops.aten.view(expand_26, [8, 4, 4]); expand_26 = None\0A expand_27 = torch.ops.aten.expand(transpose_32, [1, 8, 4, 64]); transpose_32 = None\0A view_95 = torch.ops.aten.view(expand_27, [8, 4, 64]); expand_27 = None\0A bmm_13 = torch.ops.aten.bmm(view_94, view_95); view_94 = view_95 = None\0A _unsafe_view_52 = torch.ops.aten._unsafe_view(bmm_13, [1, 8, 4, 64]); bmm_13 = None\0A transpose_34 = torch.ops.aten.transpose(_unsafe_view_52, 1, 2); _unsafe_view_52 = None\0A clone_7 = torch.ops.aten.clone(transpose_34, memory_format = torch.contiguous_format); transpose_34 = None\0A view_96 = torch.ops.aten.view(clone_7, [1, -1, 512]); clone_7 = None\0A _param_constant56 = self._param_constant56\0A t_39 = torch.ops.aten.t(_param_constant56); _param_constant56 = None\0A view_97 = torch.ops.aten.view(view_96, [4, 512]); view_96 = None\0A mm_39 = torch.ops.aten.mm(view_97, t_39); view_97 = t_39 = None\0A _unsafe_view_53 = torch.ops.aten._unsafe_view(mm_39, [1, 4, 512]); mm_39 = None\0A add_32 = torch.ops.aten.add(embedding_2, _unsafe_view_53); embedding_2 = _unsafe_view_53 = None\0A pow_15 = torch.ops.aten.pow(add_32, 2)\0A mean_14 = torch.ops.aten.mean(pow_15, [-1], True); pow_15 = None\0A add_33 = torch.ops.aten.add(mean_14, 1e-06); mean_14 = None\0A rsqrt_14 = torch.ops.aten.rsqrt(add_33); add_33 = None\0A detach_27 = torch.ops.aten.detach(rsqrt_14)\0A mul_35 = torch.ops.aten.mul(add_32, rsqrt_14); rsqrt_14 = None\0A _param_constant57 = self._param_constant57\0A mul_36 = torch.ops.aten.mul(_param_constant57, mul_35); _param_constant57 = mul_35 = None\0A _param_constant58 = self._param_constant58\0A t_40 = torch.ops.aten.t(_param_constant58); _param_constant58 = None\0A view_98 = torch.ops.aten.view(mul_36, [4, 512]); mul_36 = None\0A mm_40 = torch.ops.aten.mm(view_98, t_40); view_98 = t_40 = None\0A _unsafe_view_54 = torch.ops.aten._unsafe_view(mm_40, [1, 4, 512]); mm_40 = None\0A view_99 = torch.ops.aten.view(_unsafe_view_54, [1, -1, 8, 64]); _unsafe_view_54 = None\0A transpose_35 = torch.ops.aten.transpose(view_99, 1, 2); view_99 = None\0A _param_constant59 = self._param_constant59\0A t_41 = torch.ops.aten.t(_param_constant59); _param_constant59 = None\0A view_100 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_41 = torch.ops.aten.mm(view_100, t_41); view_100 = t_41 = None\0A _unsafe_view_55 = torch.ops.aten._unsafe_view(mm_41, [1, 15, 512]); mm_41 = None\0A view_101 = torch.ops.aten.view(_unsafe_view_55, [1, -1, 8, 64]); _unsafe_view_55 = None\0A transpose_36 = torch.ops.aten.transpose(view_101, 1, 2); view_101 = None\0A _param_constant60 = self._param_constant60\0A t_42 = torch.ops.aten.t(_param_constant60); _param_constant60 = None\0A view_102 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_42 = torch.ops.aten.mm(view_102, t_42); view_102 = t_42 = None\0A _unsafe_view_56 = torch.ops.aten._unsafe_view(mm_42, [1, 15, 512]); mm_42 = None\0A view_103 = torch.ops.aten.view(_unsafe_view_56, [1, -1, 8, 64]); _unsafe_view_56 = None\0A transpose_37 = torch.ops.aten.transpose(view_103, 1, 2); view_103 = None\0A transpose_38 = torch.ops.aten.transpose(transpose_36, 3, 2); transpose_36 = None\0A expand_28 = torch.ops.aten.expand(transpose_35, [1, 8, 4, 64]); transpose_35 = None\0A view_104 = torch.ops.aten.view(expand_28, [8, 4, 64]); expand_28 = None\0A expand_29 = torch.ops.aten.expand(transpose_38, [1, 8, 64, 15]); transpose_38 = None\0A view_105 = torch.ops.aten.view(expand_29, [8, 64, 15]); expand_29 = None\0A bmm_14 = torch.ops.aten.bmm(view_104, view_105); view_104 = view_105 = None\0A _unsafe_view_57 = torch.ops.aten._unsafe_view(bmm_14, [1, 8, 4, 15]); bmm_14 = None\0A zeros = torch.ops.aten.zeros([1, 8, 4, 15], dtype = torch.float32, device = device(type='cpu'), pin_memory = False)\0A add_34 = torch.ops.aten.add(zeros, mul_31); zeros = mul_31 = None\0A add__8 = torch.ops.aten.add_(_unsafe_view_57, add_34); _unsafe_view_57 = None\0A amax_7 = torch.ops.aten.amax(add__8, [-1], True)\0A sub_9 = torch.ops.aten.sub(add__8, amax_7); add__8 = amax_7 = None\0A exp_7 = torch.ops.aten.exp(sub_9); sub_9 = None\0A sum_8 = torch.ops.aten.sum(exp_7, [-1], True)\0A div_11 = torch.ops.aten.div(exp_7, sum_8); exp_7 = sum_8 = None\0A detach_28 = torch.ops.aten.detach(div_11)\0A expand_30 = torch.ops.aten.expand(div_11, [1, 8, 4, 15]); div_11 = None\0A view_106 = torch.ops.aten.view(expand_30, [8, 4, 15]); expand_30 = None\0A expand_31 = torch.ops.aten.expand(transpose_37, [1, 8, 15, 64]); transpose_37 = None\0A view_107 = torch.ops.aten.view(expand_31, [8, 15, 64]); expand_31 = None\0A bmm_15 = torch.ops.aten.bmm(view_106, view_107); view_106 = view_107 = None\0A _unsafe_view_58 = torch.ops.aten._unsafe_view(bmm_15, [1, 8, 4, 64]); bmm_15 = None\0A transpose_39 = torch.ops.aten.transpose(_unsafe_view_58, 1, 2); _unsafe_view_58 = None\0A clone_8 = torch.ops.aten.clone(transpose_39, memory_format = torch.contiguous_format); transpose_39 = None\0A view_108 = torch.ops.aten.view(clone_8, [1, -1, 512]); clone_8 = None\0A _param_constant61 = self._param_constant61\0A t_43 = torch.ops.aten.t(_param_constant61); _param_constant61 = None\0A view_109 = torch.ops.aten.view(view_108, [4, 512]); view_108 = None\0A mm_43 = torch.ops.aten.mm(view_109, t_43); view_109 = t_43 = None\0A _unsafe_view_59 = torch.ops.aten._unsafe_view(mm_43, [1, 4, 512]); mm_43 = None\0A add_35 = torch.ops.aten.add(add_32, _unsafe_view_59); add_32 = _unsafe_view_59 = None\0A pow_16 = torch.ops.aten.pow(add_35, 2)\0A mean_15 = torch.ops.aten.mean(pow_16, [-1], True); pow_16 = None\0A add_36 = torch.ops.aten.add(mean_15, 1e-06); mean_15 = None\0A rsqrt_15 = torch.ops.aten.rsqrt(add_36); add_36 = None\0A detach_29 = torch.ops.aten.detach(rsqrt_15)\0A mul_37 = torch.ops.aten.mul(add_35, rsqrt_15); rsqrt_15 = None\0A _param_constant62 = self._param_constant62\0A mul_38 = torch.ops.aten.mul(_param_constant62, mul_37); _param_constant62 = mul_37 = None\0A _param_constant63 = self._param_constant63\0A t_44 = torch.ops.aten.t(_param_constant63); _param_constant63 = None\0A view_110 = torch.ops.aten.view(mul_38, [4, 512]); mul_38 = None\0A mm_44 = torch.ops.aten.mm(view_110, t_44); view_110 = t_44 = None\0A _unsafe_view_60 = torch.ops.aten._unsafe_view(mm_44, [1, 4, 2048]); mm_44 = None\0A relu_6 = torch.ops.aten.relu(_unsafe_view_60); _unsafe_view_60 = None\0A detach_30 = torch.ops.aten.detach(relu_6)\0A _param_constant64 = self._param_constant64\0A t_45 = torch.ops.aten.t(_param_constant64); _param_constant64 = None\0A view_111 = torch.ops.aten.view(relu_6, [4, 2048]); relu_6 = None\0A mm_45 = torch.ops.aten.mm(view_111, t_45); view_111 = t_45 = None\0A _unsafe_view_61 = torch.ops.aten._unsafe_view(mm_45, [1, 4, 512]); mm_45 = None\0A add_37 = torch.ops.aten.add(add_35, _unsafe_view_61); add_35 = _unsafe_view_61 = None\0A pow_17 = torch.ops.aten.pow(add_37, 2)\0A mean_16 = torch.ops.aten.mean(pow_17, [-1], True); pow_17 = None\0A add_38 = torch.ops.aten.add(mean_16, 1e-06); mean_16 = None\0A rsqrt_16 = torch.ops.aten.rsqrt(add_38); add_38 = None\0A detach_31 = torch.ops.aten.detach(rsqrt_16)\0A mul_39 = torch.ops.aten.mul(add_37, rsqrt_16); rsqrt_16 = None\0A _param_constant65 = self._param_constant65\0A mul_40 = torch.ops.aten.mul(_param_constant65, mul_39); _param_constant65 = mul_39 = None\0A _param_constant66 = self._param_constant66\0A t_46 = torch.ops.aten.t(_param_constant66); _param_constant66 = None\0A view_112 = torch.ops.aten.view(mul_40, [4, 512])\0A mm_46 = torch.ops.aten.mm(view_112, t_46); view_112 = t_46 = None\0A _unsafe_view_62 = torch.ops.aten._unsafe_view(mm_46, [1, 4, 512]); mm_46 = None\0A view_113 = torch.ops.aten.view(_unsafe_view_62, [1, -1, 8, 64]); _unsafe_view_62 = None\0A transpose_40 = torch.ops.aten.transpose(view_113, 1, 2); view_113 = None\0A _param_constant67 = self._param_constant67\0A t_47 = torch.ops.aten.t(_param_constant67); _param_constant67 = None\0A view_114 = torch.ops.aten.view(mul_40, [4, 512])\0A mm_47 = torch.ops.aten.mm(view_114, t_47); view_114 = t_47 = None\0A _unsafe_view_63 = torch.ops.aten._unsafe_view(mm_47, [1, 4, 512]); mm_47 = None\0A view_115 = torch.ops.aten.view(_unsafe_view_63, [1, -1, 8, 64]); _unsafe_view_63 = None\0A transpose_41 = torch.ops.aten.transpose(view_115, 1, 2); view_115 = None\0A _param_constant68 = self._param_constant68\0A t_48 = torch.ops.aten.t(_param_constant68); _param_constant68 = None\0A view_116 = torch.ops.aten.view(mul_40, [4, 512]); mul_40 = None\0A mm_48 = torch.ops.aten.mm(view_116, t_48); view_116 = t_48 = None\0A _unsafe_view_64 = torch.ops.aten._unsafe_view(mm_48, [1, 4, 512]); mm_48 = None\0A view_117 = torch.ops.aten.view(_unsafe_view_64, [1, -1, 8, 64]); _unsafe_view_64 = None\0A transpose_42 = torch.ops.aten.transpose(view_117, 1, 2); view_117 = None\0A transpose_43 = torch.ops.aten.transpose(transpose_41, 3, 2); transpose_41 = None\0A expand_32 = torch.ops.aten.expand(transpose_40, [1, 8, 4, 64]); transpose_40 = None\0A view_118 = torch.ops.aten.view(expand_32, [8, 4, 64]); expand_32 = None\0A expand_33 = torch.ops.aten.expand(transpose_43, [1, 8, 64, 4]); transpose_43 = None\0A view_119 = torch.ops.aten.view(expand_33, [8, 64, 4]); expand_33 = None\0A bmm_16 = torch.ops.aten.bmm(view_118, view_119); view_118 = view_119 = None\0A _unsafe_view_65 = torch.ops.aten._unsafe_view(bmm_16, [1, 8, 4, 4]); bmm_16 = None\0A add__9 = torch.ops.aten.add_(_unsafe_view_65, add_31); _unsafe_view_65 = None\0A amax_8 = torch.ops.aten.amax(add__9, [-1], True)\0A sub_10 = torch.ops.aten.sub(add__9, amax_8); add__9 = amax_8 = None\0A exp_8 = torch.ops.aten.exp(sub_10); sub_10 = None\0A sum_9 = torch.ops.aten.sum(exp_8, [-1], True)\0A div_12 = torch.ops.aten.div(exp_8, sum_9); exp_8 = sum_9 = None\0A detach_32 = torch.ops.aten.detach(div_12)\0A expand_34 = torch.ops.aten.expand(div_12, [1, 8, 4, 4]); div_12 = None\0A view_120 = torch.ops.aten.view(expand_34, [8, 4, 4]); expand_34 = None\0A expand_35 = torch.ops.aten.expand(transpose_42, [1, 8, 4, 64]); transpose_42 = None\0A view_121 = torch.ops.aten.view(expand_35, [8, 4, 64]); expand_35 = None\0A bmm_17 = torch.ops.aten.bmm(view_120, view_121); view_120 = view_121 = None\0A _unsafe_view_66 = torch.ops.aten._unsafe_view(bmm_17, [1, 8, 4, 64]); bmm_17 = None\0A transpose_44 = torch.ops.aten.transpose(_unsafe_view_66, 1, 2); _unsafe_view_66 = None\0A clone_9 = torch.ops.aten.clone(transpose_44, memory_format = torch.contiguous_format); transpose_44 = None\0A view_122 = torch.ops.aten.view(clone_9, [1, -1, 512]); clone_9 = None\0A _param_constant69 = self._param_constant69\0A t_49 = torch.ops.aten.t(_param_constant69); _param_constant69 = None\0A view_123 = torch.ops.aten.view(view_122, [4, 512]); view_122 = None\0A mm_49 = torch.ops.aten.mm(view_123, t_49); view_123 = t_49 = None\0A _unsafe_view_67 = torch.ops.aten._unsafe_view(mm_49, [1, 4, 512]); mm_49 = None\0A add_39 = torch.ops.aten.add(add_37, _unsafe_view_67); add_37 = _unsafe_view_67 = None\0A pow_18 = torch.ops.aten.pow(add_39, 2)\0A mean_17 = torch.ops.aten.mean(pow_18, [-1], True); pow_18 = None\0A add_40 = torch.ops.aten.add(mean_17, 1e-06); mean_17 = None\0A rsqrt_17 = torch.ops.aten.rsqrt(add_40); add_40 = None\0A detach_33 = torch.ops.aten.detach(rsqrt_17)\0A mul_41 = torch.ops.aten.mul(add_39, rsqrt_17); rsqrt_17 = None\0A _param_constant70 = self._param_constant70\0A mul_42 = torch.ops.aten.mul(_param_constant70, mul_41); _param_constant70 = mul_41 = None\0A _param_constant71 = self._param_constant71\0A t_50 = torch.ops.aten.t(_param_constant71); _param_constant71 = None\0A view_124 = torch.ops.aten.view(mul_42, [4, 512]); mul_42 = None\0A mm_50 = torch.ops.aten.mm(view_124, t_50); view_124 = t_50 = None\0A _unsafe_view_68 = torch.ops.aten._unsafe_view(mm_50, [1, 4, 512]); mm_50 = None\0A view_125 = torch.ops.aten.view(_unsafe_view_68, [1, -1, 8, 64]); _unsafe_view_68 = None\0A transpose_45 = torch.ops.aten.transpose(view_125, 1, 2); view_125 = None\0A _param_constant72 = self._param_constant72\0A t_51 = torch.ops.aten.t(_param_constant72); _param_constant72 = None\0A view_126 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_51 = torch.ops.aten.mm(view_126, t_51); view_126 = t_51 = None\0A _unsafe_view_69 = torch.ops.aten._unsafe_view(mm_51, [1, 15, 512]); mm_51 = None\0A view_127 = torch.ops.aten.view(_unsafe_view_69, [1, -1, 8, 64]); _unsafe_view_69 = None\0A transpose_46 = torch.ops.aten.transpose(view_127, 1, 2); view_127 = None\0A _param_constant73 = self._param_constant73\0A t_52 = torch.ops.aten.t(_param_constant73); _param_constant73 = None\0A view_128 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_52 = torch.ops.aten.mm(view_128, t_52); view_128 = t_52 = None\0A _unsafe_view_70 = torch.ops.aten._unsafe_view(mm_52, [1, 15, 512]); mm_52 = None\0A view_129 = torch.ops.aten.view(_unsafe_view_70, [1, -1, 8, 64]); _unsafe_view_70 = None\0A transpose_47 = torch.ops.aten.transpose(view_129, 1, 2); view_129 = None\0A transpose_48 = torch.ops.aten.transpose(transpose_46, 3, 2); transpose_46 = None\0A expand_36 = torch.ops.aten.expand(transpose_45, [1, 8, 4, 64]); transpose_45 = None\0A view_130 = torch.ops.aten.view(expand_36, [8, 4, 64]); expand_36 = None\0A expand_37 = torch.ops.aten.expand(transpose_48, [1, 8, 64, 15]); transpose_48 = None\0A view_131 = torch.ops.aten.view(expand_37, [8, 64, 15]); expand_37 = None\0A bmm_18 = torch.ops.aten.bmm(view_130, view_131); view_130 = view_131 = None\0A _unsafe_view_71 = torch.ops.aten._unsafe_view(bmm_18, [1, 8, 4, 15]); bmm_18 = None\0A add__10 = torch.ops.aten.add_(_unsafe_view_71, add_34); _unsafe_view_71 = None\0A amax_9 = torch.ops.aten.amax(add__10, [-1], True)\0A sub_11 = torch.ops.aten.sub(add__10, amax_9); add__10 = amax_9 = None\0A exp_9 = torch.ops.aten.exp(sub_11); sub_11 = None\0A sum_10 = torch.ops.aten.sum(exp_9, [-1], True)\0A div_13 = torch.ops.aten.div(exp_9, sum_10); exp_9 = sum_10 = None\0A detach_34 = torch.ops.aten.detach(div_13)\0A expand_38 = torch.ops.aten.expand(div_13, [1, 8, 4, 15]); div_13 = None\0A view_132 = torch.ops.aten.view(expand_38, [8, 4, 15]); expand_38 = None\0A expand_39 = torch.ops.aten.expand(transpose_47, [1, 8, 15, 64]); transpose_47 = None\0A view_133 = torch.ops.aten.view(expand_39, [8, 15, 64]); expand_39 = None\0A bmm_19 = torch.ops.aten.bmm(view_132, view_133); view_132 = view_133 = None\0A _unsafe_view_72 = torch.ops.aten._unsafe_view(bmm_19, [1, 8, 4, 64]); bmm_19 = None\0A transpose_49 = torch.ops.aten.transpose(_unsafe_view_72, 1, 2); _unsafe_view_72 = None\0A clone_10 = torch.ops.aten.clone(transpose_49, memory_format = torch.contiguous_format); transpose_49 = None\0A view_134 = torch.ops.aten.view(clone_10, [1, -1, 512]); clone_10 = None\0A _param_constant74 = self._param_constant74\0A t_53 = torch.ops.aten.t(_param_constant74); _param_constant74 = None\0A view_135 = torch.ops.aten.view(view_134, [4, 512]); view_134 = None\0A mm_53 = torch.ops.aten.mm(view_135, t_53); view_135 = t_53 = None\0A _unsafe_view_73 = torch.ops.aten._unsafe_view(mm_53, [1, 4, 512]); mm_53 = None\0A add_41 = torch.ops.aten.add(add_39, _unsafe_view_73); add_39 = _unsafe_view_73 = None\0A pow_19 = torch.ops.aten.pow(add_41, 2)\0A mean_18 = torch.ops.aten.mean(pow_19, [-1], True); pow_19 = None\0A add_42 = torch.ops.aten.add(mean_18, 1e-06); mean_18 = None\0A rsqrt_18 = torch.ops.aten.rsqrt(add_42); add_42 = None\0A detach_35 = torch.ops.aten.detach(rsqrt_18)\0A mul_43 = torch.ops.aten.mul(add_41, rsqrt_18); rsqrt_18 = None\0A _param_constant75 = self._param_constant75\0A mul_44 = torch.ops.aten.mul(_param_constant75, mul_43); _param_constant75 = mul_43 = None\0A _param_constant76 = self._param_constant76\0A t_54 = torch.ops.aten.t(_param_constant76); _param_constant76 = None\0A view_136 = torch.ops.aten.view(mul_44, [4, 512]); mul_44 = None\0A mm_54 = torch.ops.aten.mm(view_136, t_54); view_136 = t_54 = None\0A _unsafe_view_74 = torch.ops.aten._unsafe_view(mm_54, [1, 4, 2048]); mm_54 = None\0A relu_7 = torch.ops.aten.relu(_unsafe_view_74); _unsafe_view_74 = None\0A detach_36 = torch.ops.aten.detach(relu_7)\0A _param_constant77 = self._param_constant77\0A t_55 = torch.ops.aten.t(_param_constant77); _param_constant77 = None\0A view_137 = torch.ops.aten.view(relu_7, [4, 2048]); relu_7 = None\0A mm_55 = torch.ops.aten.mm(view_137, t_55); view_137 = t_55 = None\0A _unsafe_view_75 = torch.ops.aten._unsafe_view(mm_55, [1, 4, 512]); mm_55 = None\0A add_43 = torch.ops.aten.add(add_41, _unsafe_view_75); add_41 = _unsafe_view_75 = None\0A pow_20 = torch.ops.aten.pow(add_43, 2)\0A mean_19 = torch.ops.aten.mean(pow_20, [-1], True); pow_20 = None\0A add_44 = torch.ops.aten.add(mean_19, 1e-06); mean_19 = None\0A rsqrt_19 = torch.ops.aten.rsqrt(add_44); add_44 = None\0A detach_37 = torch.ops.aten.detach(rsqrt_19)\0A mul_45 = torch.ops.aten.mul(add_43, rsqrt_19); rsqrt_19 = None\0A _param_constant78 = self._param_constant78\0A mul_46 = torch.ops.aten.mul(_param_constant78, mul_45); _param_constant78 = mul_45 = None\0A _param_constant79 = self._param_constant79\0A t_56 = torch.ops.aten.t(_param_constant79); _param_constant79 = None\0A view_138 = torch.ops.aten.view(mul_46, [4, 512])\0A mm_56 = torch.ops.aten.mm(view_138, t_56); view_138 = t_56 = None\0A _unsafe_view_76 = torch.ops.aten._unsafe_view(mm_56, [1, 4, 512]); mm_56 = None\0A view_139 = torch.ops.aten.view(_unsafe_view_76, [1, -1, 8, 64]); _unsafe_view_76 = None\0A transpose_50 = torch.ops.aten.transpose(view_139, 1, 2); view_139 = None\0A _param_constant80 = self._param_constant80\0A t_57 = torch.ops.aten.t(_param_constant80); _param_constant80 = None\0A view_140 = torch.ops.aten.view(mul_46, [4, 512])\0A mm_57 = torch.ops.aten.mm(view_140, t_57); view_140 = t_57 = None\0A _unsafe_view_77 = torch.ops.aten._unsafe_view(mm_57, [1, 4, 512]); mm_57 = None\0A view_141 = torch.ops.aten.view(_unsafe_view_77, [1, -1, 8, 64]); _unsafe_view_77 = None\0A transpose_51 = torch.ops.aten.transpose(view_141, 1, 2); view_141 = None\0A _param_constant81 = self._param_constant81\0A t_58 = torch.ops.aten.t(_param_constant81); _param_constant81 = None\0A view_142 = torch.ops.aten.view(mul_46, [4, 512]); mul_46 = None\0A mm_58 = torch.ops.aten.mm(view_142, t_58); view_142 = t_58 = None\0A _unsafe_view_78 = torch.ops.aten._unsafe_view(mm_58, [1, 4, 512]); mm_58 = None\0A view_143 = torch.ops.aten.view(_unsafe_view_78, [1, -1, 8, 64]); _unsafe_view_78 = None\0A transpose_52 = torch.ops.aten.transpose(view_143, 1, 2); view_143 = None\0A transpose_53 = torch.ops.aten.transpose(transpose_51, 3, 2); transpose_51 = None\0A expand_40 = torch.ops.aten.expand(transpose_50, [1, 8, 4, 64]); transpose_50 = None\0A view_144 = torch.ops.aten.view(expand_40, [8, 4, 64]); expand_40 = None\0A expand_41 = torch.ops.aten.expand(transpose_53, [1, 8, 64, 4]); transpose_53 = None\0A view_145 = torch.ops.aten.view(expand_41, [8, 64, 4]); expand_41 = None\0A bmm_20 = torch.ops.aten.bmm(view_144, view_145); view_144 = view_145 = None\0A _unsafe_view_79 = torch.ops.aten._unsafe_view(bmm_20, [1, 8, 4, 4]); bmm_20 = None\0A add__11 = torch.ops.aten.add_(_unsafe_view_79, add_31); _unsafe_view_79 = None\0A amax_10 = torch.ops.aten.amax(add__11, [-1], True)\0A sub_12 = torch.ops.aten.sub(add__11, amax_10); add__11 = amax_10 = None\0A exp_10 = torch.ops.aten.exp(sub_12); sub_12 = None\0A sum_11 = torch.ops.aten.sum(exp_10, [-1], True)\0A div_14 = torch.ops.aten.div(exp_10, sum_11); exp_10 = sum_11 = None\0A detach_38 = torch.ops.aten.detach(div_14)\0A expand_42 = torch.ops.aten.expand(div_14, [1, 8, 4, 4]); div_14 = None\0A view_146 = torch.ops.aten.view(expand_42, [8, 4, 4]); expand_42 = None\0A expand_43 = torch.ops.aten.expand(transpose_52, [1, 8, 4, 64]); transpose_52 = None\0A view_147 = torch.ops.aten.view(expand_43, [8, 4, 64]); expand_43 = None\0A bmm_21 = torch.ops.aten.bmm(view_146, view_147); view_146 = view_147 = None\0A _unsafe_view_80 = torch.ops.aten._unsafe_view(bmm_21, [1, 8, 4, 64]); bmm_21 = None\0A transpose_54 = torch.ops.aten.transpose(_unsafe_view_80, 1, 2); _unsafe_view_80 = None\0A clone_11 = torch.ops.aten.clone(transpose_54, memory_format = torch.contiguous_format); transpose_54 = None\0A view_148 = torch.ops.aten.view(clone_11, [1, -1, 512]); clone_11 = None\0A _param_constant82 = self._param_constant82\0A t_59 = torch.ops.aten.t(_param_constant82); _param_constant82 = None\0A view_149 = torch.ops.aten.view(view_148, [4, 512]); view_148 = None\0A mm_59 = torch.ops.aten.mm(view_149, t_59); view_149 = t_59 = None\0A _unsafe_view_81 = torch.ops.aten._unsafe_view(mm_59, [1, 4, 512]); mm_59 = None\0A add_45 = torch.ops.aten.add(add_43, _unsafe_view_81); add_43 = _unsafe_view_81 = None\0A pow_21 = torch.ops.aten.pow(add_45, 2)\0A mean_20 = torch.ops.aten.mean(pow_21, [-1], True); pow_21 = None\0A add_46 = torch.ops.aten.add(mean_20, 1e-06); mean_20 = None\0A rsqrt_20 = torch.ops.aten.rsqrt(add_46); add_46 = None\0A detach_39 = torch.ops.aten.detach(rsqrt_20)\0A mul_47 = torch.ops.aten.mul(add_45, rsqrt_20); rsqrt_20 = None\0A _param_constant83 = self._param_constant83\0A mul_48 = torch.ops.aten.mul(_param_constant83, mul_47); _param_constant83 = mul_47 = None\0A _param_constant84 = self._param_constant84\0A t_60 = torch.ops.aten.t(_param_constant84); _param_constant84 = None\0A view_150 = torch.ops.aten.view(mul_48, [4, 512]); mul_48 = None\0A mm_60 = torch.ops.aten.mm(view_150, t_60); view_150 = t_60 = None\0A _unsafe_view_82 = torch.ops.aten._unsafe_view(mm_60, [1, 4, 512]); mm_60 = None\0A view_151 = torch.ops.aten.view(_unsafe_view_82, [1, -1, 8, 64]); _unsafe_view_82 = None\0A transpose_55 = torch.ops.aten.transpose(view_151, 1, 2); view_151 = None\0A _param_constant85 = self._param_constant85\0A t_61 = torch.ops.aten.t(_param_constant85); _param_constant85 = None\0A view_152 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_61 = torch.ops.aten.mm(view_152, t_61); view_152 = t_61 = None\0A _unsafe_view_83 = torch.ops.aten._unsafe_view(mm_61, [1, 15, 512]); mm_61 = None\0A view_153 = torch.ops.aten.view(_unsafe_view_83, [1, -1, 8, 64]); _unsafe_view_83 = None\0A transpose_56 = torch.ops.aten.transpose(view_153, 1, 2); view_153 = None\0A _param_constant86 = self._param_constant86\0A t_62 = torch.ops.aten.t(_param_constant86); _param_constant86 = None\0A view_154 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_62 = torch.ops.aten.mm(view_154, t_62); view_154 = t_62 = None\0A _unsafe_view_84 = torch.ops.aten._unsafe_view(mm_62, [1, 15, 512]); mm_62 = None\0A view_155 = torch.ops.aten.view(_unsafe_view_84, [1, -1, 8, 64]); _unsafe_view_84 = None\0A transpose_57 = torch.ops.aten.transpose(view_155, 1, 2); view_155 = None\0A transpose_58 = torch.ops.aten.transpose(transpose_56, 3, 2); transpose_56 = None\0A expand_44 = torch.ops.aten.expand(transpose_55, [1, 8, 4, 64]); transpose_55 = None\0A view_156 = torch.ops.aten.view(expand_44, [8, 4, 64]); expand_44 = None\0A expand_45 = torch.ops.aten.expand(transpose_58, [1, 8, 64, 15]); transpose_58 = None\0A view_157 = torch.ops.aten.view(expand_45, [8, 64, 15]); expand_45 = None\0A bmm_22 = torch.ops.aten.bmm(view_156, view_157); view_156 = view_157 = None\0A _unsafe_view_85 = torch.ops.aten._unsafe_view(bmm_22, [1, 8, 4, 15]); bmm_22 = None\0A add__12 = torch.ops.aten.add_(_unsafe_view_85, add_34); _unsafe_view_85 = None\0A amax_11 = torch.ops.aten.amax(add__12, [-1], True)\0A sub_13 = torch.ops.aten.sub(add__12, amax_11); add__12 = amax_11 = None\0A exp_11 = torch.ops.aten.exp(sub_13); sub_13 = None\0A sum_12 = torch.ops.aten.sum(exp_11, [-1], True)\0A div_15 = torch.ops.aten.div(exp_11, sum_12); exp_11 = sum_12 = None\0A detach_40 = torch.ops.aten.detach(div_15)\0A expand_46 = torch.ops.aten.expand(div_15, [1, 8, 4, 15]); div_15 = None\0A view_158 = torch.ops.aten.view(expand_46, [8, 4, 15]); expand_46 = None\0A expand_47 = torch.ops.aten.expand(transpose_57, [1, 8, 15, 64]); transpose_57 = None\0A view_159 = torch.ops.aten.view(expand_47, [8, 15, 64]); expand_47 = None\0A bmm_23 = torch.ops.aten.bmm(view_158, view_159); view_158 = view_159 = None\0A _unsafe_view_86 = torch.ops.aten._unsafe_view(bmm_23, [1, 8, 4, 64]); bmm_23 = None\0A transpose_59 = torch.ops.aten.transpose(_unsafe_view_86, 1, 2); _unsafe_view_86 = None\0A clone_12 = torch.ops.aten.clone(transpose_59, memory_format = torch.contiguous_format); transpose_59 = None\0A view_160 = torch.ops.aten.view(clone_12, [1, -1, 512]); clone_12 = None\0A _param_constant87 = self._param_constant87\0A t_63 = torch.ops.aten.t(_param_constant87); _param_constant87 = None\0A view_161 = torch.ops.aten.view(view_160, [4, 512]); view_160 = None\0A mm_63 = torch.ops.aten.mm(view_161, t_63); view_161 = t_63 = None\0A _unsafe_view_87 = torch.ops.aten._unsafe_view(mm_63, [1, 4, 512]); mm_63 = None\0A add_47 = torch.ops.aten.add(add_45, _unsafe_view_87); add_45 = _unsafe_view_87 = None\0A pow_22 = torch.ops.aten.pow(add_47, 2)\0A mean_21 = torch.ops.aten.mean(pow_22, [-1], True); pow_22 = None\0A add_48 = torch.ops.aten.add(mean_21, 1e-06); mean_21 = None\0A rsqrt_21 = torch.ops.aten.rsqrt(add_48); add_48 = None\0A detach_41 = torch.ops.aten.detach(rsqrt_21)\0A mul_49 = torch.ops.aten.mul(add_47, rsqrt_21); rsqrt_21 = None\0A _param_constant88 = self._param_constant88\0A mul_50 = torch.ops.aten.mul(_param_constant88, mul_49); _param_constant88 = mul_49 = None\0A _param_constant89 = self._param_constant89\0A t_64 = torch.ops.aten.t(_param_constant89); _param_constant89 = None\0A view_162 = torch.ops.aten.view(mul_50, [4, 512]); mul_50 = None\0A mm_64 = torch.ops.aten.mm(view_162, t_64); view_162 = t_64 = None\0A _unsafe_view_88 = torch.ops.aten._unsafe_view(mm_64, [1, 4, 2048]); mm_64 = None\0A relu_8 = torch.ops.aten.relu(_unsafe_view_88); _unsafe_view_88 = None\0A detach_42 = torch.ops.aten.detach(relu_8)\0A _param_constant90 = self._param_constant90\0A t_65 = torch.ops.aten.t(_param_constant90); _param_constant90 = None\0A view_163 = torch.ops.aten.view(relu_8, [4, 2048]); relu_8 = None\0A mm_65 = torch.ops.aten.mm(view_163, t_65); view_163 = t_65 = None\0A _unsafe_view_89 = torch.ops.aten._unsafe_view(mm_65, [1, 4, 512]); mm_65 = None\0A add_49 = torch.ops.aten.add(add_47, _unsafe_view_89); add_47 = _unsafe_view_89 = None\0A pow_23 = torch.ops.aten.pow(add_49, 2)\0A mean_22 = torch.ops.aten.mean(pow_23, [-1], True); pow_23 = None\0A add_50 = torch.ops.aten.add(mean_22, 1e-06); mean_22 = None\0A rsqrt_22 = torch.ops.aten.rsqrt(add_50); add_50 = None\0A detach_43 = torch.ops.aten.detach(rsqrt_22)\0A mul_51 = torch.ops.aten.mul(add_49, rsqrt_22); rsqrt_22 = None\0A _param_constant91 = self._param_constant91\0A mul_52 = torch.ops.aten.mul(_param_constant91, mul_51); _param_constant91 = mul_51 = None\0A _param_constant92 = self._param_constant92\0A t_66 = torch.ops.aten.t(_param_constant92); _param_constant92 = None\0A view_164 = torch.ops.aten.view(mul_52, [4, 512])\0A mm_66 = torch.ops.aten.mm(view_164, t_66); view_164 = t_66 = None\0A _unsafe_view_90 = torch.ops.aten._unsafe_view(mm_66, [1, 4, 512]); mm_66 = None\0A view_165 = torch.ops.aten.view(_unsafe_view_90, [1, -1, 8, 64]); _unsafe_view_90 = None\0A transpose_60 = torch.ops.aten.transpose(view_165, 1, 2); view_165 = None\0A _param_constant93 = self._param_constant93\0A t_67 = torch.ops.aten.t(_param_constant93); _param_constant93 = None\0A view_166 = torch.ops.aten.view(mul_52, [4, 512])\0A mm_67 = torch.ops.aten.mm(view_166, t_67); view_166 = t_67 = None\0A _unsafe_view_91 = torch.ops.aten._unsafe_view(mm_67, [1, 4, 512]); mm_67 = None\0A view_167 = torch.ops.aten.view(_unsafe_view_91, [1, -1, 8, 64]); _unsafe_view_91 = None\0A transpose_61 = torch.ops.aten.transpose(view_167, 1, 2); view_167 = None\0A _param_constant94 = self._param_constant94\0A t_68 = torch.ops.aten.t(_param_constant94); _param_constant94 = None\0A view_168 = torch.ops.aten.view(mul_52, [4, 512]); mul_52 = None\0A mm_68 = torch.ops.aten.mm(view_168, t_68); view_168 = t_68 = None\0A _unsafe_view_92 = torch.ops.aten._unsafe_view(mm_68, [1, 4, 512]); mm_68 = None\0A view_169 = torch.ops.aten.view(_unsafe_view_92, [1, -1, 8, 64]); _unsafe_view_92 = None\0A transpose_62 = torch.ops.aten.transpose(view_169, 1, 2); view_169 = None\0A transpose_63 = torch.ops.aten.transpose(transpose_61, 3, 2); transpose_61 = None\0A expand_48 = torch.ops.aten.expand(transpose_60, [1, 8, 4, 64]); transpose_60 = None\0A view_170 = torch.ops.aten.view(expand_48, [8, 4, 64]); expand_48 = None\0A expand_49 = torch.ops.aten.expand(transpose_63, [1, 8, 64, 4]); transpose_63 = None\0A view_171 = torch.ops.aten.view(expand_49, [8, 64, 4]); expand_49 = None\0A bmm_24 = torch.ops.aten.bmm(view_170, view_171); view_170 = view_171 = None\0A _unsafe_view_93 = torch.ops.aten._unsafe_view(bmm_24, [1, 8, 4, 4]); bmm_24 = None\0A add__13 = torch.ops.aten.add_(_unsafe_view_93, add_31); _unsafe_view_93 = None\0A amax_12 = torch.ops.aten.amax(add__13, [-1], True)\0A sub_14 = torch.ops.aten.sub(add__13, amax_12); add__13 = amax_12 = None\0A exp_12 = torch.ops.aten.exp(sub_14); sub_14 = None\0A sum_13 = torch.ops.aten.sum(exp_12, [-1], True)\0A div_16 = torch.ops.aten.div(exp_12, sum_13); exp_12 = sum_13 = None\0A detach_44 = torch.ops.aten.detach(div_16)\0A expand_50 = torch.ops.aten.expand(div_16, [1, 8, 4, 4]); div_16 = None\0A view_172 = torch.ops.aten.view(expand_50, [8, 4, 4]); expand_50 = None\0A expand_51 = torch.ops.aten.expand(transpose_62, [1, 8, 4, 64]); transpose_62 = None\0A view_173 = torch.ops.aten.view(expand_51, [8, 4, 64]); expand_51 = None\0A bmm_25 = torch.ops.aten.bmm(view_172, view_173); view_172 = view_173 = None\0A _unsafe_view_94 = torch.ops.aten._unsafe_view(bmm_25, [1, 8, 4, 64]); bmm_25 = None\0A transpose_64 = torch.ops.aten.transpose(_unsafe_view_94, 1, 2); _unsafe_view_94 = None\0A clone_13 = torch.ops.aten.clone(transpose_64, memory_format = torch.contiguous_format); transpose_64 = None\0A view_174 = torch.ops.aten.view(clone_13, [1, -1, 512]); clone_13 = None\0A _param_constant95 = self._param_constant95\0A t_69 = torch.ops.aten.t(_param_constant95); _param_constant95 = None\0A view_175 = torch.ops.aten.view(view_174, [4, 512]); view_174 = None\0A mm_69 = torch.ops.aten.mm(view_175, t_69); view_175 = t_69 = None\0A _unsafe_view_95 = torch.ops.aten._unsafe_view(mm_69, [1, 4, 512]); mm_69 = None\0A add_51 = torch.ops.aten.add(add_49, _unsafe_view_95); add_49 = _unsafe_view_95 = None\0A pow_24 = torch.ops.aten.pow(add_51, 2)\0A mean_23 = torch.ops.aten.mean(pow_24, [-1], True); pow_24 = None\0A add_52 = torch.ops.aten.add(mean_23, 1e-06); mean_23 = None\0A rsqrt_23 = torch.ops.aten.rsqrt(add_52); add_52 = None\0A detach_45 = torch.ops.aten.detach(rsqrt_23)\0A mul_53 = torch.ops.aten.mul(add_51, rsqrt_23); rsqrt_23 = None\0A _param_constant96 = self._param_constant96\0A mul_54 = torch.ops.aten.mul(_param_constant96, mul_53); _param_constant96 = mul_53 = None\0A _param_constant97 = self._param_constant97\0A t_70 = torch.ops.aten.t(_param_constant97); _param_constant97 = None\0A view_176 = torch.ops.aten.view(mul_54, [4, 512]); mul_54 = None\0A mm_70 = torch.ops.aten.mm(view_176, t_70); view_176 = t_70 = None\0A _unsafe_view_96 = torch.ops.aten._unsafe_view(mm_70, [1, 4, 512]); mm_70 = None\0A view_177 = torch.ops.aten.view(_unsafe_view_96, [1, -1, 8, 64]); _unsafe_view_96 = None\0A transpose_65 = torch.ops.aten.transpose(view_177, 1, 2); view_177 = None\0A _param_constant98 = self._param_constant98\0A t_71 = torch.ops.aten.t(_param_constant98); _param_constant98 = None\0A view_178 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_71 = torch.ops.aten.mm(view_178, t_71); view_178 = t_71 = None\0A _unsafe_view_97 = torch.ops.aten._unsafe_view(mm_71, [1, 15, 512]); mm_71 = None\0A view_179 = torch.ops.aten.view(_unsafe_view_97, [1, -1, 8, 64]); _unsafe_view_97 = None\0A transpose_66 = torch.ops.aten.transpose(view_179, 1, 2); view_179 = None\0A _param_constant99 = self._param_constant99\0A t_72 = torch.ops.aten.t(_param_constant99); _param_constant99 = None\0A view_180 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_72 = torch.ops.aten.mm(view_180, t_72); view_180 = t_72 = None\0A _unsafe_view_98 = torch.ops.aten._unsafe_view(mm_72, [1, 15, 512]); mm_72 = None\0A view_181 = torch.ops.aten.view(_unsafe_view_98, [1, -1, 8, 64]); _unsafe_view_98 = None\0A transpose_67 = torch.ops.aten.transpose(view_181, 1, 2); view_181 = None\0A transpose_68 = torch.ops.aten.transpose(transpose_66, 3, 2); transpose_66 = None\0A expand_52 = torch.ops.aten.expand(transpose_65, [1, 8, 4, 64]); transpose_65 = None\0A view_182 = torch.ops.aten.view(expand_52, [8, 4, 64]); expand_52 = None\0A expand_53 = torch.ops.aten.expand(transpose_68, [1, 8, 64, 15]); transpose_68 = None\0A view_183 = torch.ops.aten.view(expand_53, [8, 64, 15]); expand_53 = None\0A bmm_26 = torch.ops.aten.bmm(view_182, view_183); view_182 = view_183 = None\0A _unsafe_view_99 = torch.ops.aten._unsafe_view(bmm_26, [1, 8, 4, 15]); bmm_26 = None\0A add__14 = torch.ops.aten.add_(_unsafe_view_99, add_34); _unsafe_view_99 = None\0A amax_13 = torch.ops.aten.amax(add__14, [-1], True)\0A sub_15 = torch.ops.aten.sub(add__14, amax_13); add__14 = amax_13 = None\0A exp_13 = torch.ops.aten.exp(sub_15); sub_15 = None\0A sum_14 = torch.ops.aten.sum(exp_13, [-1], True)\0A div_17 = torch.ops.aten.div(exp_13, sum_14); exp_13 = sum_14 = None\0A detach_46 = torch.ops.aten.detach(div_17)\0A expand_54 = torch.ops.aten.expand(div_17, [1, 8, 4, 15]); div_17 = None\0A view_184 = torch.ops.aten.view(expand_54, [8, 4, 15]); expand_54 = None\0A expand_55 = torch.ops.aten.expand(transpose_67, [1, 8, 15, 64]); transpose_67 = None\0A view_185 = torch.ops.aten.view(expand_55, [8, 15, 64]); expand_55 = None\0A bmm_27 = torch.ops.aten.bmm(view_184, view_185); view_184 = view_185 = None\0A _unsafe_view_100 = torch.ops.aten._unsafe_view(bmm_27, [1, 8, 4, 64]); bmm_27 = None\0A transpose_69 = torch.ops.aten.transpose(_unsafe_view_100, 1, 2); _unsafe_view_100 = None\0A clone_14 = torch.ops.aten.clone(transpose_69, memory_format = torch.contiguous_format); transpose_69 = None\0A view_186 = torch.ops.aten.view(clone_14, [1, -1, 512]); clone_14 = None\0A _param_constant100 = self._param_constant100\0A t_73 = torch.ops.aten.t(_param_constant100); _param_constant100 = None\0A view_187 = torch.ops.aten.view(view_186, [4, 512]); view_186 = None\0A mm_73 = torch.ops.aten.mm(view_187, t_73); view_187 = t_73 = None\0A _unsafe_view_101 = torch.ops.aten._unsafe_view(mm_73, [1, 4, 512]); mm_73 = None\0A add_53 = torch.ops.aten.add(add_51, _unsafe_view_101); add_51 = _unsafe_view_101 = None\0A pow_25 = torch.ops.aten.pow(add_53, 2)\0A mean_24 = torch.ops.aten.mean(pow_25, [-1], True); pow_25 = None\0A add_54 = torch.ops.aten.add(mean_24, 1e-06); mean_24 = None\0A rsqrt_24 = torch.ops.aten.rsqrt(add_54); add_54 = None\0A detach_47 = torch.ops.aten.detach(rsqrt_24)\0A mul_55 = torch.ops.aten.mul(add_53, rsqrt_24); rsqrt_24 = None\0A _param_constant101 = self._param_constant101\0A mul_56 = torch.ops.aten.mul(_param_constant101, mul_55); _param_constant101 = mul_55 = None\0A _param_constant102 = self._param_constant102\0A t_74 = torch.ops.aten.t(_param_constant102); _param_constant102 = None\0A view_188 = torch.ops.aten.view(mul_56, [4, 512]); mul_56 = None\0A mm_74 = torch.ops.aten.mm(view_188, t_74); view_188 = t_74 = None\0A _unsafe_view_102 = torch.ops.aten._unsafe_view(mm_74, [1, 4, 2048]); mm_74 = None\0A relu_9 = torch.ops.aten.relu(_unsafe_view_102); _unsafe_view_102 = None\0A detach_48 = torch.ops.aten.detach(relu_9)\0A _param_constant103 = self._param_constant103\0A t_75 = torch.ops.aten.t(_param_constant103); _param_constant103 = None\0A view_189 = torch.ops.aten.view(relu_9, [4, 2048]); relu_9 = None\0A mm_75 = torch.ops.aten.mm(view_189, t_75); view_189 = t_75 = None\0A _unsafe_view_103 = torch.ops.aten._unsafe_view(mm_75, [1, 4, 512]); mm_75 = None\0A add_55 = torch.ops.aten.add(add_53, _unsafe_view_103); add_53 = _unsafe_view_103 = None\0A pow_26 = torch.ops.aten.pow(add_55, 2)\0A mean_25 = torch.ops.aten.mean(pow_26, [-1], True); pow_26 = None\0A add_56 = torch.ops.aten.add(mean_25, 1e-06); mean_25 = None\0A rsqrt_25 = torch.ops.aten.rsqrt(add_56); add_56 = None\0A detach_49 = torch.ops.aten.detach(rsqrt_25)\0A mul_57 = torch.ops.aten.mul(add_55, rsqrt_25); rsqrt_25 = None\0A _param_constant104 = self._param_constant104\0A mul_58 = torch.ops.aten.mul(_param_constant104, mul_57); _param_constant104 = mul_57 = None\0A _param_constant105 = self._param_constant105\0A t_76 = torch.ops.aten.t(_param_constant105); _param_constant105 = None\0A view_190 = torch.ops.aten.view(mul_58, [4, 512])\0A mm_76 = torch.ops.aten.mm(view_190, t_76); view_190 = t_76 = None\0A _unsafe_view_104 = torch.ops.aten._unsafe_view(mm_76, [1, 4, 512]); mm_76 = None\0A view_191 = torch.ops.aten.view(_unsafe_view_104, [1, -1, 8, 64]); _unsafe_view_104 = None\0A transpose_70 = torch.ops.aten.transpose(view_191, 1, 2); view_191 = None\0A _param_constant106 = self._param_constant106\0A t_77 = torch.ops.aten.t(_param_constant106); _param_constant106 = None\0A view_192 = torch.ops.aten.view(mul_58, [4, 512])\0A mm_77 = torch.ops.aten.mm(view_192, t_77); view_192 = t_77 = None\0A _unsafe_view_105 = torch.ops.aten._unsafe_view(mm_77, [1, 4, 512]); mm_77 = None\0A view_193 = torch.ops.aten.view(_unsafe_view_105, [1, -1, 8, 64]); _unsafe_view_105 = None\0A transpose_71 = torch.ops.aten.transpose(view_193, 1, 2); view_193 = None\0A _param_constant107 = self._param_constant107\0A t_78 = torch.ops.aten.t(_param_constant107); _param_constant107 = None\0A view_194 = torch.ops.aten.view(mul_58, [4, 512]); mul_58 = None\0A mm_78 = torch.ops.aten.mm(view_194, t_78); view_194 = t_78 = None\0A _unsafe_view_106 = torch.ops.aten._unsafe_view(mm_78, [1, 4, 512]); mm_78 = None\0A view_195 = torch.ops.aten.view(_unsafe_view_106, [1, -1, 8, 64]); _unsafe_view_106 = None\0A transpose_72 = torch.ops.aten.transpose(view_195, 1, 2); view_195 = None\0A transpose_73 = torch.ops.aten.transpose(transpose_71, 3, 2); transpose_71 = None\0A expand_56 = torch.ops.aten.expand(transpose_70, [1, 8, 4, 64]); transpose_70 = None\0A view_196 = torch.ops.aten.view(expand_56, [8, 4, 64]); expand_56 = None\0A expand_57 = torch.ops.aten.expand(transpose_73, [1, 8, 64, 4]); transpose_73 = None\0A view_197 = torch.ops.aten.view(expand_57, [8, 64, 4]); expand_57 = None\0A bmm_28 = torch.ops.aten.bmm(view_196, view_197); view_196 = view_197 = None\0A _unsafe_view_107 = torch.ops.aten._unsafe_view(bmm_28, [1, 8, 4, 4]); bmm_28 = None\0A add__15 = torch.ops.aten.add_(_unsafe_view_107, add_31); _unsafe_view_107 = None\0A amax_14 = torch.ops.aten.amax(add__15, [-1], True)\0A sub_16 = torch.ops.aten.sub(add__15, amax_14); add__15 = amax_14 = None\0A exp_14 = torch.ops.aten.exp(sub_16); sub_16 = None\0A sum_15 = torch.ops.aten.sum(exp_14, [-1], True)\0A div_18 = torch.ops.aten.div(exp_14, sum_15); exp_14 = sum_15 = None\0A detach_50 = torch.ops.aten.detach(div_18)\0A expand_58 = torch.ops.aten.expand(div_18, [1, 8, 4, 4]); div_18 = None\0A view_198 = torch.ops.aten.view(expand_58, [8, 4, 4]); expand_58 = None\0A expand_59 = torch.ops.aten.expand(transpose_72, [1, 8, 4, 64]); transpose_72 = None\0A view_199 = torch.ops.aten.view(expand_59, [8, 4, 64]); expand_59 = None\0A bmm_29 = torch.ops.aten.bmm(view_198, view_199); view_198 = view_199 = None\0A _unsafe_view_108 = torch.ops.aten._unsafe_view(bmm_29, [1, 8, 4, 64]); bmm_29 = None\0A transpose_74 = torch.ops.aten.transpose(_unsafe_view_108, 1, 2); _unsafe_view_108 = None\0A clone_15 = torch.ops.aten.clone(transpose_74, memory_format = torch.contiguous_format); transpose_74 = None\0A view_200 = torch.ops.aten.view(clone_15, [1, -1, 512]); clone_15 = None\0A _param_constant108 = self._param_constant108\0A t_79 = torch.ops.aten.t(_param_constant108); _param_constant108 = None\0A view_201 = torch.ops.aten.view(view_200, [4, 512]); view_200 = None\0A mm_79 = torch.ops.aten.mm(view_201, t_79); view_201 = t_79 = None\0A _unsafe_view_109 = torch.ops.aten._unsafe_view(mm_79, [1, 4, 512]); mm_79 = None\0A add_57 = torch.ops.aten.add(add_55, _unsafe_view_109); add_55 = _unsafe_view_109 = None\0A pow_27 = torch.ops.aten.pow(add_57, 2)\0A mean_26 = torch.ops.aten.mean(pow_27, [-1], True); pow_27 = None\0A add_58 = torch.ops.aten.add(mean_26, 1e-06); mean_26 = None\0A rsqrt_26 = torch.ops.aten.rsqrt(add_58); add_58 = None\0A detach_51 = torch.ops.aten.detach(rsqrt_26)\0A mul_59 = torch.ops.aten.mul(add_57, rsqrt_26); rsqrt_26 = None\0A _param_constant109 = self._param_constant109\0A mul_60 = torch.ops.aten.mul(_param_constant109, mul_59); _param_constant109 = mul_59 = None\0A _param_constant110 = self._param_constant110\0A t_80 = torch.ops.aten.t(_param_constant110); _param_constant110 = None\0A view_202 = torch.ops.aten.view(mul_60, [4, 512]); mul_60 = None\0A mm_80 = torch.ops.aten.mm(view_202, t_80); view_202 = t_80 = None\0A _unsafe_view_110 = torch.ops.aten._unsafe_view(mm_80, [1, 4, 512]); mm_80 = None\0A view_203 = torch.ops.aten.view(_unsafe_view_110, [1, -1, 8, 64]); _unsafe_view_110 = None\0A transpose_75 = torch.ops.aten.transpose(view_203, 1, 2); view_203 = None\0A _param_constant111 = self._param_constant111\0A t_81 = torch.ops.aten.t(_param_constant111); _param_constant111 = None\0A view_204 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_81 = torch.ops.aten.mm(view_204, t_81); view_204 = t_81 = None\0A _unsafe_view_111 = torch.ops.aten._unsafe_view(mm_81, [1, 15, 512]); mm_81 = None\0A view_205 = torch.ops.aten.view(_unsafe_view_111, [1, -1, 8, 64]); _unsafe_view_111 = None\0A transpose_76 = torch.ops.aten.transpose(view_205, 1, 2); view_205 = None\0A _param_constant112 = self._param_constant112\0A t_82 = torch.ops.aten.t(_param_constant112); _param_constant112 = None\0A view_206 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_82 = torch.ops.aten.mm(view_206, t_82); view_206 = t_82 = None\0A _unsafe_view_112 = torch.ops.aten._unsafe_view(mm_82, [1, 15, 512]); mm_82 = None\0A view_207 = torch.ops.aten.view(_unsafe_view_112, [1, -1, 8, 64]); _unsafe_view_112 = None\0A transpose_77 = torch.ops.aten.transpose(view_207, 1, 2); view_207 = None\0A transpose_78 = torch.ops.aten.transpose(transpose_76, 3, 2); transpose_76 = None\0A expand_60 = torch.ops.aten.expand(transpose_75, [1, 8, 4, 64]); transpose_75 = None\0A view_208 = torch.ops.aten.view(expand_60, [8, 4, 64]); expand_60 = None\0A expand_61 = torch.ops.aten.expand(transpose_78, [1, 8, 64, 15]); transpose_78 = None\0A view_209 = torch.ops.aten.view(expand_61, [8, 64, 15]); expand_61 = None\0A bmm_30 = torch.ops.aten.bmm(view_208, view_209); view_208 = view_209 = None\0A _unsafe_view_113 = torch.ops.aten._unsafe_view(bmm_30, [1, 8, 4, 15]); bmm_30 = None\0A add__16 = torch.ops.aten.add_(_unsafe_view_113, add_34); _unsafe_view_113 = None\0A amax_15 = torch.ops.aten.amax(add__16, [-1], True)\0A sub_17 = torch.ops.aten.sub(add__16, amax_15); add__16 = amax_15 = None\0A exp_15 = torch.ops.aten.exp(sub_17); sub_17 = None\0A sum_16 = torch.ops.aten.sum(exp_15, [-1], True)\0A div_19 = torch.ops.aten.div(exp_15, sum_16); exp_15 = sum_16 = None\0A detach_52 = torch.ops.aten.detach(div_19)\0A expand_62 = torch.ops.aten.expand(div_19, [1, 8, 4, 15]); div_19 = None\0A view_210 = torch.ops.aten.view(expand_62, [8, 4, 15]); expand_62 = None\0A expand_63 = torch.ops.aten.expand(transpose_77, [1, 8, 15, 64]); transpose_77 = None\0A view_211 = torch.ops.aten.view(expand_63, [8, 15, 64]); expand_63 = None\0A bmm_31 = torch.ops.aten.bmm(view_210, view_211); view_210 = view_211 = None\0A _unsafe_view_114 = torch.ops.aten._unsafe_view(bmm_31, [1, 8, 4, 64]); bmm_31 = None\0A transpose_79 = torch.ops.aten.transpose(_unsafe_view_114, 1, 2); _unsafe_view_114 = None\0A clone_16 = torch.ops.aten.clone(transpose_79, memory_format = torch.contiguous_format); transpose_79 = None\0A view_212 = torch.ops.aten.view(clone_16, [1, -1, 512]); clone_16 = None\0A _param_constant113 = self._param_constant113\0A t_83 = torch.ops.aten.t(_param_constant113); _param_constant113 = None\0A view_213 = torch.ops.aten.view(view_212, [4, 512]); view_212 = None\0A mm_83 = torch.ops.aten.mm(view_213, t_83); view_213 = t_83 = None\0A _unsafe_view_115 = torch.ops.aten._unsafe_view(mm_83, [1, 4, 512]); mm_83 = None\0A add_59 = torch.ops.aten.add(add_57, _unsafe_view_115); add_57 = _unsafe_view_115 = None\0A pow_28 = torch.ops.aten.pow(add_59, 2)\0A mean_27 = torch.ops.aten.mean(pow_28, [-1], True); pow_28 = None\0A add_60 = torch.ops.aten.add(mean_27, 1e-06); mean_27 = None\0A rsqrt_27 = torch.ops.aten.rsqrt(add_60); add_60 = None\0A detach_53 = torch.ops.aten.detach(rsqrt_27)\0A mul_61 = torch.ops.aten.mul(add_59, rsqrt_27); rsqrt_27 = None\0A _param_constant114 = self._param_constant114\0A mul_62 = torch.ops.aten.mul(_param_constant114, mul_61); _param_constant114 = mul_61 = None\0A _param_constant115 = self._param_constant115\0A t_84 = torch.ops.aten.t(_param_constant115); _param_constant115 = None\0A view_214 = torch.ops.aten.view(mul_62, [4, 512]); mul_62 = None\0A mm_84 = torch.ops.aten.mm(view_214, t_84); view_214 = t_84 = None\0A _unsafe_view_116 = torch.ops.aten._unsafe_view(mm_84, [1, 4, 2048]); mm_84 = None\0A relu_10 = torch.ops.aten.relu(_unsafe_view_116); _unsafe_view_116 = None\0A detach_54 = torch.ops.aten.detach(relu_10)\0A _param_constant116 = self._param_constant116\0A t_85 = torch.ops.aten.t(_param_constant116); _param_constant116 = None\0A view_215 = torch.ops.aten.view(relu_10, [4, 2048]); relu_10 = None\0A mm_85 = torch.ops.aten.mm(view_215, t_85); view_215 = t_85 = None\0A _unsafe_view_117 = torch.ops.aten._unsafe_view(mm_85, [1, 4, 512]); mm_85 = None\0A add_61 = torch.ops.aten.add(add_59, _unsafe_view_117); add_59 = _unsafe_view_117 = None\0A pow_29 = torch.ops.aten.pow(add_61, 2)\0A mean_28 = torch.ops.aten.mean(pow_29, [-1], True); pow_29 = None\0A add_62 = torch.ops.aten.add(mean_28, 1e-06); mean_28 = None\0A rsqrt_28 = torch.ops.aten.rsqrt(add_62); add_62 = None\0A detach_55 = torch.ops.aten.detach(rsqrt_28)\0A mul_63 = torch.ops.aten.mul(add_61, rsqrt_28); rsqrt_28 = None\0A _param_constant117 = self._param_constant117\0A mul_64 = torch.ops.aten.mul(_param_constant117, mul_63); _param_constant117 = mul_63 = None\0A _param_constant118 = self._param_constant118\0A t_86 = torch.ops.aten.t(_param_constant118); _param_constant118 = None\0A view_216 = torch.ops.aten.view(mul_64, [4, 512])\0A mm_86 = torch.ops.aten.mm(view_216, t_86); view_216 = t_86 = None\0A _unsafe_view_118 = torch.ops.aten._unsafe_view(mm_86, [1, 4, 512]); mm_86 = None\0A view_217 = torch.ops.aten.view(_unsafe_view_118, [1, -1, 8, 64]); _unsafe_view_118 = None\0A transpose_80 = torch.ops.aten.transpose(view_217, 1, 2); view_217 = None\0A _param_constant119 = self._param_constant119\0A t_87 = torch.ops.aten.t(_param_constant119); _param_constant119 = None\0A view_218 = torch.ops.aten.view(mul_64, [4, 512])\0A mm_87 = torch.ops.aten.mm(view_218, t_87); view_218 = t_87 = None\0A _unsafe_view_119 = torch.ops.aten._unsafe_view(mm_87, [1, 4, 512]); mm_87 = None\0A view_219 = torch.ops.aten.view(_unsafe_view_119, [1, -1, 8, 64]); _unsafe_view_119 = None\0A transpose_81 = torch.ops.aten.transpose(view_219, 1, 2); view_219 = None\0A _param_constant120 = self._param_constant120\0A t_88 = torch.ops.aten.t(_param_constant120); _param_constant120 = None\0A view_220 = torch.ops.aten.view(mul_64, [4, 512]); mul_64 = None\0A mm_88 = torch.ops.aten.mm(view_220, t_88); view_220 = t_88 = None\0A _unsafe_view_120 = torch.ops.aten._unsafe_view(mm_88, [1, 4, 512]); mm_88 = None\0A view_221 = torch.ops.aten.view(_unsafe_view_120, [1, -1, 8, 64]); _unsafe_view_120 = None\0A transpose_82 = torch.ops.aten.transpose(view_221, 1, 2); view_221 = None\0A transpose_83 = torch.ops.aten.transpose(transpose_81, 3, 2); transpose_81 = None\0A expand_64 = torch.ops.aten.expand(transpose_80, [1, 8, 4, 64]); transpose_80 = None\0A view_222 = torch.ops.aten.view(expand_64, [8, 4, 64]); expand_64 = None\0A expand_65 = torch.ops.aten.expand(transpose_83, [1, 8, 64, 4]); transpose_83 = None\0A view_223 = torch.ops.aten.view(expand_65, [8, 64, 4]); expand_65 = None\0A bmm_32 = torch.ops.aten.bmm(view_222, view_223); view_222 = view_223 = None\0A _unsafe_view_121 = torch.ops.aten._unsafe_view(bmm_32, [1, 8, 4, 4]); bmm_32 = None\0A add__17 = torch.ops.aten.add_(_unsafe_view_121, add_31); _unsafe_view_121 = add_31 = None\0A amax_16 = torch.ops.aten.amax(add__17, [-1], True)\0A sub_18 = torch.ops.aten.sub(add__17, amax_16); add__17 = amax_16 = None\0A exp_16 = torch.ops.aten.exp(sub_18); sub_18 = None\0A sum_17 = torch.ops.aten.sum(exp_16, [-1], True)\0A div_20 = torch.ops.aten.div(exp_16, sum_17); exp_16 = sum_17 = None\0A detach_56 = torch.ops.aten.detach(div_20)\0A expand_66 = torch.ops.aten.expand(div_20, [1, 8, 4, 4]); div_20 = None\0A view_224 = torch.ops.aten.view(expand_66, [8, 4, 4]); expand_66 = None\0A expand_67 = torch.ops.aten.expand(transpose_82, [1, 8, 4, 64]); transpose_82 = None\0A view_225 = torch.ops.aten.view(expand_67, [8, 4, 64]); expand_67 = None\0A bmm_33 = torch.ops.aten.bmm(view_224, view_225); view_224 = view_225 = None\0A _unsafe_view_122 = torch.ops.aten._unsafe_view(bmm_33, [1, 8, 4, 64]); bmm_33 = None\0A transpose_84 = torch.ops.aten.transpose(_unsafe_view_122, 1, 2); _unsafe_view_122 = None\0A clone_17 = torch.ops.aten.clone(transpose_84, memory_format = torch.contiguous_format); transpose_84 = None\0A view_226 = torch.ops.aten.view(clone_17, [1, -1, 512]); clone_17 = None\0A _param_constant121 = self._param_constant121\0A t_89 = torch.ops.aten.t(_param_constant121); _param_constant121 = None\0A view_227 = torch.ops.aten.view(view_226, [4, 512]); view_226 = None\0A mm_89 = torch.ops.aten.mm(view_227, t_89); view_227 = t_89 = None\0A _unsafe_view_123 = torch.ops.aten._unsafe_view(mm_89, [1, 4, 512]); mm_89 = None\0A add_63 = torch.ops.aten.add(add_61, _unsafe_view_123); add_61 = _unsafe_view_123 = None\0A pow_30 = torch.ops.aten.pow(add_63, 2)\0A mean_29 = torch.ops.aten.mean(pow_30, [-1], True); pow_30 = None\0A add_64 = torch.ops.aten.add(mean_29, 1e-06); mean_29 = None\0A rsqrt_29 = torch.ops.aten.rsqrt(add_64); add_64 = None\0A detach_57 = torch.ops.aten.detach(rsqrt_29)\0A mul_65 = torch.ops.aten.mul(add_63, rsqrt_29); rsqrt_29 = None\0A _param_constant122 = self._param_constant122\0A mul_66 = torch.ops.aten.mul(_param_constant122, mul_65); _param_constant122 = mul_65 = None\0A _param_constant123 = self._param_constant123\0A t_90 = torch.ops.aten.t(_param_constant123); _param_constant123 = None\0A view_228 = torch.ops.aten.view(mul_66, [4, 512]); mul_66 = None\0A mm_90 = torch.ops.aten.mm(view_228, t_90); view_228 = t_90 = None\0A _unsafe_view_124 = torch.ops.aten._unsafe_view(mm_90, [1, 4, 512]); mm_90 = None\0A view_229 = torch.ops.aten.view(_unsafe_view_124, [1, -1, 8, 64]); _unsafe_view_124 = None\0A transpose_85 = torch.ops.aten.transpose(view_229, 1, 2); view_229 = None\0A _param_constant124 = self._param_constant124\0A t_91 = torch.ops.aten.t(_param_constant124); _param_constant124 = None\0A view_230 = torch.ops.aten.view(mul_28, [15, 512])\0A mm_91 = torch.ops.aten.mm(view_230, t_91); view_230 = t_91 = None\0A _unsafe_view_125 = torch.ops.aten._unsafe_view(mm_91, [1, 15, 512]); mm_91 = None\0A view_231 = torch.ops.aten.view(_unsafe_view_125, [1, -1, 8, 64]); _unsafe_view_125 = None\0A transpose_86 = torch.ops.aten.transpose(view_231, 1, 2); view_231 = None\0A _param_constant125 = self._param_constant125\0A t_92 = torch.ops.aten.t(_param_constant125); _param_constant125 = None\0A view_232 = torch.ops.aten.view(mul_28, [15, 512]); mul_28 = None\0A mm_92 = torch.ops.aten.mm(view_232, t_92); view_232 = t_92 = None\0A _unsafe_view_126 = torch.ops.aten._unsafe_view(mm_92, [1, 15, 512]); mm_92 = None\0A view_233 = torch.ops.aten.view(_unsafe_view_126, [1, -1, 8, 64]); _unsafe_view_126 = None\0A transpose_87 = torch.ops.aten.transpose(view_233, 1, 2); view_233 = None\0A transpose_88 = torch.ops.aten.transpose(transpose_86, 3, 2); transpose_86 = None\0A expand_68 = torch.ops.aten.expand(transpose_85, [1, 8, 4, 64]); transpose_85 = None\0A view_234 = torch.ops.aten.view(expand_68, [8, 4, 64]); expand_68 = None\0A expand_69 = torch.ops.aten.expand(transpose_88, [1, 8, 64, 15]); transpose_88 = None\0A view_235 = torch.ops.aten.view(expand_69, [8, 64, 15]); expand_69 = None\0A bmm_34 = torch.ops.aten.bmm(view_234, view_235); view_234 = view_235 = None\0A _unsafe_view_127 = torch.ops.aten._unsafe_view(bmm_34, [1, 8, 4, 15]); bmm_34 = None\0A add__18 = torch.ops.aten.add_(_unsafe_view_127, add_34); _unsafe_view_127 = add_34 = None\0A amax_17 = torch.ops.aten.amax(add__18, [-1], True)\0A sub_19 = torch.ops.aten.sub(add__18, amax_17); add__18 = amax_17 = None\0A exp_17 = torch.ops.aten.exp(sub_19); sub_19 = None\0A sum_18 = torch.ops.aten.sum(exp_17, [-1], True)\0A div_21 = torch.ops.aten.div(exp_17, sum_18); exp_17 = sum_18 = None\0A detach_58 = torch.ops.aten.detach(div_21)\0A expand_70 = torch.ops.aten.expand(div_21, [1, 8, 4, 15]); div_21 = None\0A view_236 = torch.ops.aten.view(expand_70, [8, 4, 15]); expand_70 = None\0A expand_71 = torch.ops.aten.expand(transpose_87, [1, 8, 15, 64]); transpose_87 = None\0A view_237 = torch.ops.aten.view(expand_71, [8, 15, 64]); expand_71 = None\0A bmm_35 = torch.ops.aten.bmm(view_236, view_237); view_236 = view_237 = None\0A _unsafe_view_128 = torch.ops.aten._unsafe_view(bmm_35, [1, 8, 4, 64]); bmm_35 = None\0A transpose_89 = torch.ops.aten.transpose(_unsafe_view_128, 1, 2); _unsafe_view_128 = None\0A clone_18 = torch.ops.aten.clone(transpose_89, memory_format = torch.contiguous_format); transpose_89 = None\0A view_238 = torch.ops.aten.view(clone_18, [1, -1, 512]); clone_18 = None\0A _param_constant126 = self._param_constant126\0A t_93 = torch.ops.aten.t(_param_constant126); _param_constant126 = None\0A view_239 = torch.ops.aten.view(view_238, [4, 512]); view_238 = None\0A mm_93 = torch.ops.aten.mm(view_239, t_93); view_239 = t_93 = None\0A _unsafe_view_129 = torch.ops.aten._unsafe_view(mm_93, [1, 4, 512]); mm_93 = None\0A add_65 = torch.ops.aten.add(add_63, _unsafe_view_129); add_63 = _unsafe_view_129 = None\0A pow_31 = torch.ops.aten.pow(add_65, 2)\0A mean_30 = torch.ops.aten.mean(pow_31, [-1], True); pow_31 = None\0A add_66 = torch.ops.aten.add(mean_30, 1e-06); mean_30 = None\0A rsqrt_30 = torch.ops.aten.rsqrt(add_66); add_66 = None\0A detach_59 = torch.ops.aten.detach(rsqrt_30)\0A mul_67 = torch.ops.aten.mul(add_65, rsqrt_30); rsqrt_30 = None\0A _param_constant127 = self._param_constant127\0A mul_68 = torch.ops.aten.mul(_param_constant127, mul_67); _param_constant127 = mul_67 = None\0A _param_constant128 = self._param_constant128\0A t_94 = torch.ops.aten.t(_param_constant128); _param_constant128 = None\0A view_240 = torch.ops.aten.view(mul_68, [4, 512]); mul_68 = None\0A mm_94 = torch.ops.aten.mm(view_240, t_94); view_240 = t_94 = None\0A _unsafe_view_130 = torch.ops.aten._unsafe_view(mm_94, [1, 4, 2048]); mm_94 = None\0A relu_11 = torch.ops.aten.relu(_unsafe_view_130); _unsafe_view_130 = None\0A detach_60 = torch.ops.aten.detach(relu_11)\0A _param_constant129 = self._param_constant129\0A t_95 = torch.ops.aten.t(_param_constant129); _param_constant129 = None\0A view_241 = torch.ops.aten.view(relu_11, [4, 2048]); relu_11 = None\0A mm_95 = torch.ops.aten.mm(view_241, t_95); view_241 = t_95 = None\0A _unsafe_view_131 = torch.ops.aten._unsafe_view(mm_95, [1, 4, 512]); mm_95 = None\0A add_67 = torch.ops.aten.add(add_65, _unsafe_view_131); add_65 = _unsafe_view_131 = None\0A pow_32 = torch.ops.aten.pow(add_67, 2)\0A mean_31 = torch.ops.aten.mean(pow_32, [-1], True); pow_32 = None\0A add_68 = torch.ops.aten.add(mean_31, 1e-06); mean_31 = None\0A rsqrt_31 = torch.ops.aten.rsqrt(add_68); add_68 = None\0A detach_61 = torch.ops.aten.detach(rsqrt_31)\0A mul_69 = torch.ops.aten.mul(add_67, rsqrt_31); add_67 = rsqrt_31 = None\0A _param_constant130 = self._param_constant130\0A mul_70 = torch.ops.aten.mul(_param_constant130, mul_69); _param_constant130 = mul_69 = None\0A mul_71 = torch.ops.aten.mul(mul_70, 0.04419417382415922); mul_70 = None\0A _param_constant0_2 = self._param_constant0\0A t_96 = torch.ops.aten.t(_param_constant0_2); _param_constant0_2 = None\0A view_242 = torch.ops.aten.view(mul_71, [4, 512]); mul_71 = None\0A mm_96 = torch.ops.aten.mm(view_242, t_96); view_242 = t_96 = None\0A _unsafe_view_132 = torch.ops.aten._unsafe_view(mm_96, [1, 4, 32128]); mm_96 = None\0A return _unsafe_view_132\0A " loc(#loc) | |
| %132 = torch.nn_module { | |
| torch.slot "_param_constant0", %0 : !torch.tensor<[32128,512],f32> loc(#loc) | |
| torch.slot "_param_constant1", %1 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant2", %2 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant3", %3 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant4", %4 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant5", %5 : !torch.tensor<[32,8],f32> loc(#loc) | |
| torch.slot "_param_constant6", %6 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant7", %7 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant8", %8 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant9", %9 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant10", %10 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant11", %11 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant12", %12 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant13", %13 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant14", %14 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant15", %15 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant16", %16 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant17", %17 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant18", %18 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant19", %19 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant20", %20 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant21", %21 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant22", %22 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant23", %23 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant24", %24 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant25", %25 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant26", %26 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant27", %27 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant28", %28 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant29", %29 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant30", %30 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant31", %31 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant32", %32 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant33", %33 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant34", %34 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant35", %35 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant36", %36 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant37", %37 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant38", %38 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant39", %39 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant40", %40 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant41", %41 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant42", %42 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant43", %43 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant44", %44 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant45", %45 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant46", %46 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant47", %47 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant48", %48 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant49", %49 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant50", %50 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant51", %51 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant52", %52 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant53", %53 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant54", %54 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant55", %55 : !torch.tensor<[32,8],f32> loc(#loc) | |
| torch.slot "_param_constant56", %56 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant57", %57 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant58", %58 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant59", %59 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant60", %60 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant61", %61 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant62", %62 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant63", %63 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant64", %64 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant65", %65 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant66", %66 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant67", %67 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant68", %68 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant69", %69 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant70", %70 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant71", %71 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant72", %72 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant73", %73 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant74", %74 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant75", %75 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant76", %76 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant77", %77 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant78", %78 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant79", %79 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant80", %80 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant81", %81 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant82", %82 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant83", %83 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant84", %84 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant85", %85 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant86", %86 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant87", %87 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant88", %88 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant89", %89 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant90", %90 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant91", %91 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant92", %92 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant93", %93 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant94", %94 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant95", %95 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant96", %96 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant97", %97 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant98", %98 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant99", %99 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant100", %100 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant101", %101 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant102", %102 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant103", %103 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant104", %104 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant105", %105 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant106", %106 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant107", %107 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant108", %108 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant109", %109 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant110", %110 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant111", %111 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant112", %112 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant113", %113 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant114", %114 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant115", %115 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant116", %116 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant117", %117 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant118", %118 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant119", %119 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant120", %120 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant121", %121 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant122", %122 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant123", %123 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant124", %124 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant125", %125 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant126", %126 : !torch.tensor<[512,512],f32> loc(#loc) | |
| torch.slot "_param_constant127", %127 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_param_constant128", %128 : !torch.tensor<[2048,512],f32> loc(#loc) | |
| torch.slot "_param_constant129", %129 : !torch.tensor<[512,2048],f32> loc(#loc) | |
| torch.slot "_param_constant130", %130 : !torch.tensor<[512],f32> loc(#loc) | |
| torch.slot "_tensor_constant0", %131 : !torch.tensor<[],si64> loc(#loc) | |
| torch.slot "training", %true : !torch.bool loc(#loc) | |
| torch.slot "_is_full_backward_hook", %none : !torch.none loc(#loc) | |
| torch.slot "_code", %str : !torch.str loc(#loc) | |
| } : !torch.nn.Module<"__torch__.torch.fx.graph_module._lambda"> loc(#loc) | |
| } loc(#loc) | |
| #loc1 = loc("<eval_with_key>.2":75:73) | |
| #loc2 = loc("<eval_with_key>.2":27:44) | |
| #loc3 = loc("<eval_with_key>.2":25:35) | |
| #loc4 = loc("<eval_with_key>.2":14:38) | |
| #loc5 = loc("<eval_with_key>.2":6:49) | |
| #loc6 = loc("<eval_with_key>.2":5:144) | |
| #loc7 = loc("<eval_with_key>.2":5:50) | |
| #loc8 = loc("<eval_with_key>.2":5:53) | |
| #loc9 = loc("<eval_with_key>.2":6:46) | |
| #loc10 = loc("<eval_with_key>.2":8:52) | |
| #loc11 = loc("<eval_with_key>.2":16:44) | |
| #loc12 = loc("<eval_with_key>.2":22:54) | |
| #loc13 = loc("<eval_with_key>.2":23:48) | |
| #loc14 = loc("<eval_with_key>.2":24:40) | |
| #loc15 = loc("<eval_with_key>.2":28:35) | |
| #loc16 = loc("<eval_with_key>.2":36:45) | |
| #loc17 = loc("<eval_with_key>.2":39:55) | |
| #loc18 = loc("<eval_with_key>.2":39:58) | |
| #loc19 = loc("<eval_with_key>.2":71:53) | |
| #loc20 = loc("<eval_with_key>.2":78:36) | |
| #loc21 = loc("<eval_with_key>.2":125:63) | |
| #loc22 = loc("<eval_with_key>.2":620:38) | |
| #loc23 = loc("<eval_with_key>.2":624:51) | |
| #loc24 = loc("<eval_with_key>.2":1429:40) | |
| #loc25 = loc("<eval_with_key>.2":1434:65) | |
| #loc26 = loc("<eval_with_key>.2":5:16) | |
| #loc27 = loc("<eval_with_key>.2":6:14) | |
| #loc28 = loc("<eval_with_key>.2":7:12) | |
| #loc29 = loc("<eval_with_key>.2":8:14) | |
| #loc30 = loc("<eval_with_key>.2":9:12) | |
| #loc31 = loc("<eval_with_key>.2":11:22) | |
| #loc32 = loc("<eval_with_key>.2":12:13) | |
| #loc33 = loc("<eval_with_key>.2":13:12) | |
| #loc34 = loc("<eval_with_key>.2":14:9) | |
| #loc35 = loc("<eval_with_key>.2":15:19) | |
| #loc36 = loc("<eval_with_key>.2":16:11) | |
| #loc37 = loc("<eval_with_key>.2":18:16) | |
| #loc38 = loc("<eval_with_key>.2":19:11) | |
| #loc39 = loc("<eval_with_key>.2":20:14) | |
| #loc40 = loc("<eval_with_key>.2":21:16) | |
| #loc41 = loc("<eval_with_key>.2":22:18) | |
| #loc42 = loc("<eval_with_key>.2":23:14) | |
| #loc43 = loc("<eval_with_key>.2":24:11) | |
| #loc44 = loc("<eval_with_key>.2":25:10) | |
| #loc45 = loc("<eval_with_key>.2":26:12) | |
| #loc46 = loc("<eval_with_key>.2":27:11) | |
| #loc47 = loc("<eval_with_key>.2":28:10) | |
| #loc48 = loc("<eval_with_key>.2":29:12) | |
| #loc49 = loc("<eval_with_key>.2":31:12) | |
| #loc50 = loc("<eval_with_key>.2":33:12) | |
| #loc51 = loc("<eval_with_key>.2":35:8) | |
| #loc52 = loc("<eval_with_key>.2":36:13) | |
| #loc53 = loc("<eval_with_key>.2":37:9) | |
| #loc54 = loc("<eval_with_key>.2":38:19) | |
| #loc55 = loc("<eval_with_key>.2":39:13) | |
| #loc56 = loc("<eval_with_key>.2":40:16) | |
| #loc57 = loc("<eval_with_key>.2":42:10) | |
| #loc58 = loc("<eval_with_key>.2":43:13) | |
| #loc59 = loc("<eval_with_key>.2":44:11) | |
| #loc60 = loc("<eval_with_key>.2":45:21) | |
| #loc61 = loc("<eval_with_key>.2":46:13) | |
| #loc62 = loc("<eval_with_key>.2":47:18) | |
| #loc63 = loc("<eval_with_key>.2":49:10) | |
| #loc64 = loc("<eval_with_key>.2":50:13) | |
| #loc65 = loc("<eval_with_key>.2":51:11) | |
| #loc66 = loc("<eval_with_key>.2":52:21) | |
| #loc67 = loc("<eval_with_key>.2":53:13) | |
| #loc68 = loc("<eval_with_key>.2":54:18) | |
| #loc69 = loc("<eval_with_key>.2":55:18) | |
| #loc70 = loc("<eval_with_key>.2":56:13) | |
| #loc71 = loc("<eval_with_key>.2":57:13) | |
| #loc72 = loc("<eval_with_key>.2":58:15) | |
| #loc73 = loc("<eval_with_key>.2":59:13) | |
| #loc74 = loc("<eval_with_key>.2":60:10) | |
| #loc75 = loc("<eval_with_key>.2":61:21) | |
| #loc76 = loc("<eval_with_key>.2":62:13) | |
| #loc77 = loc("<eval_with_key>.2":63:14) | |
| #loc78 = loc("<eval_with_key>.2":64:18) | |
| #loc79 = loc("<eval_with_key>.2":65:15) | |
| #loc80 = loc("<eval_with_key>.2":66:18) | |
| #loc81 = loc("<eval_with_key>.2":67:14) | |
| #loc82 = loc("<eval_with_key>.2":68:10) | |
| #loc83 = loc("<eval_with_key>.2":69:9) | |
| #loc84 = loc("<eval_with_key>.2":70:27) | |
| #loc85 = loc("<eval_with_key>.2":71:12) | |
| #loc86 = loc("<eval_with_key>.2":72:12) | |
| #loc87 = loc("<eval_with_key>.2":73:12) | |
| #loc88 = loc("<eval_with_key>.2":74:9) | |
| #loc89 = loc("<eval_with_key>.2":75:29) | |
| #loc90 = loc("<eval_with_key>.2":76:10) | |
| #loc91 = loc("<eval_with_key>.2":77:10) | |
| #loc92 = loc("<eval_with_key>.2":78:12) | |
| #loc93 = loc("<eval_with_key>.2":79:12) | |
| #loc94 = loc("<eval_with_key>.2":80:29) | |
| #loc95 = loc("<eval_with_key>.2":81:12) | |
| #loc96 = loc("<eval_with_key>.2":82:16) | |
| #loc97 = loc("<eval_with_key>.2":83:14) | |
| #loc98 = loc("<eval_with_key>.2":84:12) | |
| #loc99 = loc("<eval_with_key>.2":85:11) | |
| #loc100 = loc("<eval_with_key>.2":87:18) | |
| #loc101 = loc("<eval_with_key>.2":88:14) | |
| #loc102 = loc("<eval_with_key>.2":89:18) | |
| #loc103 = loc("<eval_with_key>.2":90:12) | |
| #loc104 = loc("<eval_with_key>.2":91:13) | |
| #loc105 = loc("<eval_with_key>.2":92:11) | |
| #loc106 = loc("<eval_with_key>.2":93:12) | |
| #loc107 = loc("<eval_with_key>.2":94:10) | |
| #loc108 = loc("<eval_with_key>.2":95:12) | |
| #loc109 = loc("<eval_with_key>.2":96:12) | |
| #loc110 = loc("<eval_with_key>.2":98:15) | |
| #loc111 = loc("<eval_with_key>.2":99:13) | |
| #loc112 = loc("<eval_with_key>.2":100:15) | |
| #loc113 = loc("<eval_with_key>.2":101:14) | |
| #loc114 = loc("<eval_with_key>.2":102:12) | |
| #loc115 = loc("<eval_with_key>.2":103:21) | |
| #loc116 = loc("<eval_with_key>.2":104:18) | |
| #loc117 = loc("<eval_with_key>.2":105:14) | |
| #loc118 = loc("<eval_with_key>.2":106:14) | |
| #loc119 = loc("<eval_with_key>.2":108:10) | |
| #loc120 = loc("<eval_with_key>.2":109:14) | |
| #loc121 = loc("<eval_with_key>.2":110:11) | |
| #loc122 = loc("<eval_with_key>.2":111:21) | |
| #loc123 = loc("<eval_with_key>.2":112:12) | |
| #loc124 = loc("<eval_with_key>.2":113:12) | |
| #loc125 = loc("<eval_with_key>.2":114:13) | |
| #loc126 = loc("<eval_with_key>.2":115:12) | |
| #loc127 = loc("<eval_with_key>.2":116:14) | |
| #loc128 = loc("<eval_with_key>.2":118:12) | |
| #loc129 = loc("<eval_with_key>.2":120:12) | |
| #loc130 = loc("<eval_with_key>.2":122:10) | |
| #loc131 = loc("<eval_with_key>.2":123:14) | |
| #loc132 = loc("<eval_with_key>.2":124:11) | |
| #loc133 = loc("<eval_with_key>.2":125:21) | |
| #loc134 = loc("<eval_with_key>.2":126:11) | |
| #loc135 = loc("<eval_with_key>.2":129:10) | |
| #loc136 = loc("<eval_with_key>.2":130:14) | |
| #loc137 = loc("<eval_with_key>.2":131:11) | |
| #loc138 = loc("<eval_with_key>.2":132:21) | |
| #loc139 = loc("<eval_with_key>.2":133:12) | |
| #loc140 = loc("<eval_with_key>.2":134:12) | |
| #loc141 = loc("<eval_with_key>.2":135:13) | |
| #loc142 = loc("<eval_with_key>.2":136:12) | |
| #loc143 = loc("<eval_with_key>.2":137:14) | |
| #loc144 = loc("<eval_with_key>.2":139:12) | |
| #loc145 = loc("<eval_with_key>.2":141:12) | |
| #loc146 = loc("<eval_with_key>.2":143:10) | |
| #loc147 = loc("<eval_with_key>.2":144:14) | |
| #loc148 = loc("<eval_with_key>.2":145:11) | |
| #loc149 = loc("<eval_with_key>.2":146:21) | |
| #loc150 = loc("<eval_with_key>.2":147:14) | |
| #loc151 = loc("<eval_with_key>.2":148:18) | |
| #loc152 = loc("<eval_with_key>.2":150:10) | |
| #loc153 = loc("<eval_with_key>.2":151:14) | |
| #loc154 = loc("<eval_with_key>.2":152:11) | |
| #loc155 = loc("<eval_with_key>.2":153:21) | |
| #loc156 = loc("<eval_with_key>.2":154:14) | |
| #loc157 = loc("<eval_with_key>.2":155:18) | |
| #loc158 = loc("<eval_with_key>.2":157:10) | |
| #loc159 = loc("<eval_with_key>.2":158:14) | |
| #loc160 = loc("<eval_with_key>.2":159:11) | |
| #loc161 = loc("<eval_with_key>.2":160:22) | |
| #loc162 = loc("<eval_with_key>.2":161:14) | |
| #loc163 = loc("<eval_with_key>.2":162:18) | |
| #loc164 = loc("<eval_with_key>.2":163:18) | |
| #loc165 = loc("<eval_with_key>.2":164:15) | |
| #loc166 = loc("<eval_with_key>.2":165:14) | |
| #loc167 = loc("<eval_with_key>.2":166:15) | |
| #loc168 = loc("<eval_with_key>.2":167:14) | |
| #loc169 = loc("<eval_with_key>.2":168:12) | |
| #loc170 = loc("<eval_with_key>.2":169:22) | |
| #loc171 = loc("<eval_with_key>.2":170:13) | |
| #loc172 = loc("<eval_with_key>.2":171:13) | |
| #loc173 = loc("<eval_with_key>.2":172:12) | |
| #loc174 = loc("<eval_with_key>.2":173:12) | |
| #loc175 = loc("<eval_with_key>.2":174:12) | |
| #loc176 = loc("<eval_with_key>.2":175:12) | |
| #loc177 = loc("<eval_with_key>.2":177:15) | |
| #loc178 = loc("<eval_with_key>.2":178:14) | |
| #loc179 = loc("<eval_with_key>.2":179:15) | |
| #loc180 = loc("<eval_with_key>.2":180:14) | |
| #loc181 = loc("<eval_with_key>.2":181:12) | |
| #loc182 = loc("<eval_with_key>.2":182:22) | |
| #loc183 = loc("<eval_with_key>.2":183:18) | |
| #loc184 = loc("<eval_with_key>.2":184:14) | |
| #loc185 = loc("<eval_with_key>.2":185:14) | |
| #loc186 = loc("<eval_with_key>.2":187:10) | |
| #loc187 = loc("<eval_with_key>.2":188:14) | |
| #loc188 = loc("<eval_with_key>.2":189:11) | |
| #loc189 = loc("<eval_with_key>.2":190:22) | |
| #loc190 = loc("<eval_with_key>.2":191:12) | |
| #loc191 = loc("<eval_with_key>.2":192:12) | |
| #loc192 = loc("<eval_with_key>.2":193:13) | |
| #loc193 = loc("<eval_with_key>.2":194:12) | |
| #loc194 = loc("<eval_with_key>.2":195:14) | |
| #loc195 = loc("<eval_with_key>.2":197:12) | |
| #loc196 = loc("<eval_with_key>.2":199:13) | |
| #loc197 = loc("<eval_with_key>.2":201:11) | |
| #loc198 = loc("<eval_with_key>.2":202:14) | |
| #loc199 = loc("<eval_with_key>.2":203:12) | |
| #loc200 = loc("<eval_with_key>.2":204:22) | |
| #loc201 = loc("<eval_with_key>.2":205:13) | |
| #loc202 = loc("<eval_with_key>.2":208:11) | |
| #loc203 = loc("<eval_with_key>.2":209:14) | |
| #loc204 = loc("<eval_with_key>.2":210:12) | |
| #loc205 = loc("<eval_with_key>.2":211:22) | |
| #loc206 = loc("<eval_with_key>.2":212:13) | |
| #loc207 = loc("<eval_with_key>.2":213:12) | |
| #loc208 = loc("<eval_with_key>.2":214:13) | |
| #loc209 = loc("<eval_with_key>.2":215:13) | |
| #loc210 = loc("<eval_with_key>.2":216:14) | |
| #loc211 = loc("<eval_with_key>.2":218:13) | |
| #loc212 = loc("<eval_with_key>.2":220:13) | |
| #loc213 = loc("<eval_with_key>.2":222:11) | |
| #loc214 = loc("<eval_with_key>.2":223:14) | |
| #loc215 = loc("<eval_with_key>.2":224:12) | |
| #loc216 = loc("<eval_with_key>.2":225:22) | |
| #loc217 = loc("<eval_with_key>.2":226:14) | |
| #loc218 = loc("<eval_with_key>.2":227:19) | |
| #loc219 = loc("<eval_with_key>.2":229:11) | |
| #loc220 = loc("<eval_with_key>.2":230:14) | |
| #loc221 = loc("<eval_with_key>.2":231:12) | |
| #loc222 = loc("<eval_with_key>.2":232:22) | |
| #loc223 = loc("<eval_with_key>.2":233:14) | |
| #loc224 = loc("<eval_with_key>.2":234:19) | |
| #loc225 = loc("<eval_with_key>.2":236:11) | |
| #loc226 = loc("<eval_with_key>.2":237:14) | |
| #loc227 = loc("<eval_with_key>.2":238:12) | |
| #loc228 = loc("<eval_with_key>.2":239:22) | |
| #loc229 = loc("<eval_with_key>.2":240:14) | |
| #loc230 = loc("<eval_with_key>.2":241:19) | |
| #loc231 = loc("<eval_with_key>.2":242:19) | |
| #loc232 = loc("<eval_with_key>.2":243:15) | |
| #loc233 = loc("<eval_with_key>.2":244:14) | |
| #loc234 = loc("<eval_with_key>.2":245:15) | |
| #loc235 = loc("<eval_with_key>.2":246:14) | |
| #loc236 = loc("<eval_with_key>.2":247:12) | |
| #loc237 = loc("<eval_with_key>.2":248:22) | |
| #loc238 = loc("<eval_with_key>.2":249:13) | |
| #loc239 = loc("<eval_with_key>.2":250:13) | |
| #loc240 = loc("<eval_with_key>.2":251:12) | |
| #loc241 = loc("<eval_with_key>.2":252:12) | |
| #loc242 = loc("<eval_with_key>.2":253:12) | |
| #loc243 = loc("<eval_with_key>.2":254:12) | |
| #loc244 = loc("<eval_with_key>.2":256:16) | |
| #loc245 = loc("<eval_with_key>.2":257:14) | |
| #loc246 = loc("<eval_with_key>.2":258:16) | |
| #loc247 = loc("<eval_with_key>.2":259:14) | |
| #loc248 = loc("<eval_with_key>.2":260:12) | |
| #loc249 = loc("<eval_with_key>.2":261:22) | |
| #loc250 = loc("<eval_with_key>.2":262:19) | |
| #loc251 = loc("<eval_with_key>.2":263:14) | |
| #loc252 = loc("<eval_with_key>.2":264:14) | |
| #loc253 = loc("<eval_with_key>.2":266:11) | |
| #loc254 = loc("<eval_with_key>.2":267:14) | |
| #loc255 = loc("<eval_with_key>.2":268:12) | |
| #loc256 = loc("<eval_with_key>.2":269:22) | |
| #loc257 = loc("<eval_with_key>.2":270:13) | |
| #loc258 = loc("<eval_with_key>.2":271:12) | |
| #loc259 = loc("<eval_with_key>.2":272:13) | |
| #loc260 = loc("<eval_with_key>.2":273:13) | |
| #loc261 = loc("<eval_with_key>.2":274:14) | |
| #loc262 = loc("<eval_with_key>.2":276:13) | |
| #loc263 = loc("<eval_with_key>.2":278:13) | |
| #loc264 = loc("<eval_with_key>.2":280:11) | |
| #loc265 = loc("<eval_with_key>.2":281:14) | |
| #loc266 = loc("<eval_with_key>.2":282:12) | |
| #loc267 = loc("<eval_with_key>.2":283:22) | |
| #loc268 = loc("<eval_with_key>.2":284:13) | |
| #loc269 = loc("<eval_with_key>.2":287:11) | |
| #loc270 = loc("<eval_with_key>.2":288:14) | |
| #loc271 = loc("<eval_with_key>.2":289:12) | |
| #loc272 = loc("<eval_with_key>.2":290:22) | |
| #loc273 = loc("<eval_with_key>.2":291:13) | |
| #loc274 = loc("<eval_with_key>.2":292:12) | |
| #loc275 = loc("<eval_with_key>.2":293:13) | |
| #loc276 = loc("<eval_with_key>.2":294:13) | |
| #loc277 = loc("<eval_with_key>.2":295:14) | |
| #loc278 = loc("<eval_with_key>.2":297:13) | |
| #loc279 = loc("<eval_with_key>.2":299:13) | |
| #loc280 = loc("<eval_with_key>.2":301:11) | |
| #loc281 = loc("<eval_with_key>.2":302:14) | |
| #loc282 = loc("<eval_with_key>.2":303:12) | |
| #loc283 = loc("<eval_with_key>.2":304:22) | |
| #loc284 = loc("<eval_with_key>.2":305:14) | |
| #loc285 = loc("<eval_with_key>.2":306:19) | |
| #loc286 = loc("<eval_with_key>.2":308:11) | |
| #loc287 = loc("<eval_with_key>.2":309:14) | |
| #loc288 = loc("<eval_with_key>.2":310:12) | |
| #loc289 = loc("<eval_with_key>.2":311:22) | |
| #loc290 = loc("<eval_with_key>.2":312:14) | |
| #loc291 = loc("<eval_with_key>.2":313:19) | |
| #loc292 = loc("<eval_with_key>.2":315:11) | |
| #loc293 = loc("<eval_with_key>.2":316:14) | |
| #loc294 = loc("<eval_with_key>.2":317:12) | |
| #loc295 = loc("<eval_with_key>.2":318:22) | |
| #loc296 = loc("<eval_with_key>.2":319:14) | |
| #loc297 = loc("<eval_with_key>.2":320:19) | |
| #loc298 = loc("<eval_with_key>.2":321:19) | |
| #loc299 = loc("<eval_with_key>.2":322:16) | |
| #loc300 = loc("<eval_with_key>.2":323:14) | |
| #loc301 = loc("<eval_with_key>.2":324:16) | |
| #loc302 = loc("<eval_with_key>.2":325:14) | |
| #loc303 = loc("<eval_with_key>.2":326:12) | |
| #loc304 = loc("<eval_with_key>.2":327:22) | |
| #loc305 = loc("<eval_with_key>.2":328:13) | |
| #loc306 = loc("<eval_with_key>.2":329:13) | |
| #loc307 = loc("<eval_with_key>.2":330:12) | |
| #loc308 = loc("<eval_with_key>.2":331:12) | |
| #loc309 = loc("<eval_with_key>.2":332:12) | |
| #loc310 = loc("<eval_with_key>.2":333:12) | |
| #loc311 = loc("<eval_with_key>.2":335:16) | |
| #loc312 = loc("<eval_with_key>.2":336:14) | |
| #loc313 = loc("<eval_with_key>.2":337:16) | |
| #loc314 = loc("<eval_with_key>.2":338:14) | |
| #loc315 = loc("<eval_with_key>.2":339:12) | |
| #loc316 = loc("<eval_with_key>.2":340:22) | |
| #loc317 = loc("<eval_with_key>.2":341:19) | |
| #loc318 = loc("<eval_with_key>.2":342:14) | |
| #loc319 = loc("<eval_with_key>.2":343:14) | |
| #loc320 = loc("<eval_with_key>.2":345:11) | |
| #loc321 = loc("<eval_with_key>.2":346:14) | |
| #loc322 = loc("<eval_with_key>.2":347:12) | |
| #loc323 = loc("<eval_with_key>.2":348:22) | |
| #loc324 = loc("<eval_with_key>.2":349:13) | |
| #loc325 = loc("<eval_with_key>.2":350:12) | |
| #loc326 = loc("<eval_with_key>.2":351:13) | |
| #loc327 = loc("<eval_with_key>.2":352:13) | |
| #loc328 = loc("<eval_with_key>.2":353:14) | |
| #loc329 = loc("<eval_with_key>.2":355:13) | |
| #loc330 = loc("<eval_with_key>.2":357:13) | |
| #loc331 = loc("<eval_with_key>.2":359:11) | |
| #loc332 = loc("<eval_with_key>.2":360:14) | |
| #loc333 = loc("<eval_with_key>.2":361:12) | |
| #loc334 = loc("<eval_with_key>.2":362:22) | |
| #loc335 = loc("<eval_with_key>.2":363:13) | |
| #loc336 = loc("<eval_with_key>.2":366:11) | |
| #loc337 = loc("<eval_with_key>.2":367:14) | |
| #loc338 = loc("<eval_with_key>.2":368:12) | |
| #loc339 = loc("<eval_with_key>.2":369:22) | |
| #loc340 = loc("<eval_with_key>.2":370:13) | |
| #loc341 = loc("<eval_with_key>.2":371:12) | |
| #loc342 = loc("<eval_with_key>.2":372:13) | |
| #loc343 = loc("<eval_with_key>.2":373:13) | |
| #loc344 = loc("<eval_with_key>.2":374:14) | |
| #loc345 = loc("<eval_with_key>.2":376:13) | |
| #loc346 = loc("<eval_with_key>.2":378:13) | |
| #loc347 = loc("<eval_with_key>.2":380:11) | |
| #loc348 = loc("<eval_with_key>.2":381:14) | |
| #loc349 = loc("<eval_with_key>.2":382:12) | |
| #loc350 = loc("<eval_with_key>.2":383:22) | |
| #loc351 = loc("<eval_with_key>.2":384:14) | |
| #loc352 = loc("<eval_with_key>.2":385:19) | |
| #loc353 = loc("<eval_with_key>.2":387:11) | |
| #loc354 = loc("<eval_with_key>.2":388:14) | |
| #loc355 = loc("<eval_with_key>.2":389:12) | |
| #loc356 = loc("<eval_with_key>.2":390:22) | |
| #loc357 = loc("<eval_with_key>.2":391:14) | |
| #loc358 = loc("<eval_with_key>.2":392:19) | |
| #loc359 = loc("<eval_with_key>.2":394:11) | |
| #loc360 = loc("<eval_with_key>.2":395:14) | |
| #loc361 = loc("<eval_with_key>.2":396:12) | |
| #loc362 = loc("<eval_with_key>.2":397:22) | |
| #loc363 = loc("<eval_with_key>.2":398:14) | |
| #loc364 = loc("<eval_with_key>.2":399:19) | |
| #loc365 = loc("<eval_with_key>.2":400:19) | |
| #loc366 = loc("<eval_with_key>.2":401:16) | |
| #loc367 = loc("<eval_with_key>.2":402:14) | |
| #loc368 = loc("<eval_with_key>.2":403:16) | |
| #loc369 = loc("<eval_with_key>.2":404:14) | |
| #loc370 = loc("<eval_with_key>.2":405:12) | |
| #loc371 = loc("<eval_with_key>.2":406:22) | |
| #loc372 = loc("<eval_with_key>.2":407:13) | |
| #loc373 = loc("<eval_with_key>.2":408:13) | |
| #loc374 = loc("<eval_with_key>.2":409:12) | |
| #loc375 = loc("<eval_with_key>.2":410:12) | |
| #loc376 = loc("<eval_with_key>.2":411:12) | |
| #loc377 = loc("<eval_with_key>.2":412:12) | |
| #loc378 = loc("<eval_with_key>.2":414:16) | |
| #loc379 = loc("<eval_with_key>.2":415:14) | |
| #loc380 = loc("<eval_with_key>.2":416:16) | |
| #loc381 = loc("<eval_with_key>.2":417:14) | |
| #loc382 = loc("<eval_with_key>.2":418:12) | |
| #loc383 = loc("<eval_with_key>.2":419:22) | |
| #loc384 = loc("<eval_with_key>.2":420:19) | |
| #loc385 = loc("<eval_with_key>.2":421:14) | |
| #loc386 = loc("<eval_with_key>.2":422:14) | |
| #loc387 = loc("<eval_with_key>.2":424:11) | |
| #loc388 = loc("<eval_with_key>.2":425:14) | |
| #loc389 = loc("<eval_with_key>.2":426:12) | |
| #loc390 = loc("<eval_with_key>.2":427:22) | |
| #loc391 = loc("<eval_with_key>.2":428:13) | |
| #loc392 = loc("<eval_with_key>.2":429:13) | |
| #loc393 = loc("<eval_with_key>.2":430:13) | |
| #loc394 = loc("<eval_with_key>.2":431:13) | |
| #loc395 = loc("<eval_with_key>.2":432:14) | |
| #loc396 = loc("<eval_with_key>.2":434:13) | |
| #loc397 = loc("<eval_with_key>.2":436:13) | |
| #loc398 = loc("<eval_with_key>.2":438:11) | |
| #loc399 = loc("<eval_with_key>.2":439:14) | |
| #loc400 = loc("<eval_with_key>.2":440:12) | |
| #loc401 = loc("<eval_with_key>.2":441:22) | |
| #loc402 = loc("<eval_with_key>.2":442:13) | |
| #loc403 = loc("<eval_with_key>.2":445:11) | |
| #loc404 = loc("<eval_with_key>.2":446:14) | |
| #loc405 = loc("<eval_with_key>.2":447:12) | |
| #loc406 = loc("<eval_with_key>.2":448:22) | |
| #loc407 = loc("<eval_with_key>.2":449:13) | |
| #loc408 = loc("<eval_with_key>.2":450:13) | |
| #loc409 = loc("<eval_with_key>.2":451:14) | |
| #loc410 = loc("<eval_with_key>.2":452:13) | |
| #loc411 = loc("<eval_with_key>.2":453:15) | |
| #loc412 = loc("<eval_with_key>.2":455:13) | |
| #loc413 = loc("<eval_with_key>.2":457:13) | |
| #loc414 = loc("<eval_with_key>.2":459:11) | |
| #loc415 = loc("<eval_with_key>.2":460:14) | |
| #loc416 = loc("<eval_with_key>.2":461:12) | |
| #loc417 = loc("<eval_with_key>.2":462:22) | |
| #loc418 = loc("<eval_with_key>.2":463:14) | |
| #loc419 = loc("<eval_with_key>.2":464:19) | |
| #loc420 = loc("<eval_with_key>.2":466:11) | |
| #loc421 = loc("<eval_with_key>.2":467:14) | |
| #loc422 = loc("<eval_with_key>.2":468:12) | |
| #loc423 = loc("<eval_with_key>.2":469:22) | |
| #loc424 = loc("<eval_with_key>.2":470:14) | |
| #loc425 = loc("<eval_with_key>.2":471:19) | |
| #loc426 = loc("<eval_with_key>.2":473:11) | |
| #loc427 = loc("<eval_with_key>.2":474:14) | |
| #loc428 = loc("<eval_with_key>.2":475:12) | |
| #loc429 = loc("<eval_with_key>.2":476:22) | |
| #loc430 = loc("<eval_with_key>.2":477:14) | |
| #loc431 = loc("<eval_with_key>.2":478:19) | |
| #loc432 = loc("<eval_with_key>.2":479:19) | |
| #loc433 = loc("<eval_with_key>.2":480:16) | |
| #loc434 = loc("<eval_with_key>.2":481:14) | |
| #loc435 = loc("<eval_with_key>.2":482:16) | |
| #loc436 = loc("<eval_with_key>.2":483:14) | |
| #loc437 = loc("<eval_with_key>.2":484:13) | |
| #loc438 = loc("<eval_with_key>.2":485:22) | |
| #loc439 = loc("<eval_with_key>.2":486:13) | |
| #loc440 = loc("<eval_with_key>.2":487:13) | |
| #loc441 = loc("<eval_with_key>.2":488:12) | |
| #loc442 = loc("<eval_with_key>.2":489:12) | |
| #loc443 = loc("<eval_with_key>.2":490:12) | |
| #loc444 = loc("<eval_with_key>.2":491:12) | |
| #loc445 = loc("<eval_with_key>.2":493:16) | |
| #loc446 = loc("<eval_with_key>.2":494:14) | |
| #loc447 = loc("<eval_with_key>.2":495:16) | |
| #loc448 = loc("<eval_with_key>.2":496:14) | |
| #loc449 = loc("<eval_with_key>.2":497:13) | |
| #loc450 = loc("<eval_with_key>.2":498:22) | |
| #loc451 = loc("<eval_with_key>.2":499:19) | |
| #loc452 = loc("<eval_with_key>.2":500:14) | |
| #loc453 = loc("<eval_with_key>.2":501:14) | |
| #loc454 = loc("<eval_with_key>.2":503:11) | |
| #loc455 = loc("<eval_with_key>.2":504:14) | |
| #loc456 = loc("<eval_with_key>.2":505:12) | |
| #loc457 = loc("<eval_with_key>.2":506:22) | |
| #loc458 = loc("<eval_with_key>.2":507:13) | |
| #loc459 = loc("<eval_with_key>.2":508:13) | |
| #loc460 = loc("<eval_with_key>.2":509:14) | |
| #loc461 = loc("<eval_with_key>.2":510:13) | |
| #loc462 = loc("<eval_with_key>.2":511:15) | |
| #loc463 = loc("<eval_with_key>.2":513:13) | |
| #loc464 = loc("<eval_with_key>.2":515:13) | |
| #loc465 = loc("<eval_with_key>.2":517:11) | |
| #loc466 = loc("<eval_with_key>.2":518:14) | |
| #loc467 = loc("<eval_with_key>.2":519:12) | |
| #loc468 = loc("<eval_with_key>.2":520:22) | |
| #loc469 = loc("<eval_with_key>.2":521:13) | |
| #loc470 = loc("<eval_with_key>.2":524:11) | |
| #loc471 = loc("<eval_with_key>.2":525:14) | |
| #loc472 = loc("<eval_with_key>.2":526:12) | |
| #loc473 = loc("<eval_with_key>.2":527:22) | |
| #loc474 = loc("<eval_with_key>.2":528:13) | |
| #loc475 = loc("<eval_with_key>.2":529:13) | |
| #loc476 = loc("<eval_with_key>.2":530:14) | |
| #loc477 = loc("<eval_with_key>.2":531:13) | |
| #loc478 = loc("<eval_with_key>.2":532:15) | |
| #loc479 = loc("<eval_with_key>.2":534:13) | |
| #loc480 = loc("<eval_with_key>.2":536:13) | |
| #loc481 = loc("<eval_with_key>.2":537:14) | |
| #loc482 = loc("<eval_with_key>.2":539:18) | |
| #loc483 = loc("<eval_with_key>.2":540:13) | |
| #loc484 = loc("<eval_with_key>.2":541:13) | |
| #loc485 = loc("<eval_with_key>.2":542:15) | |
| #loc486 = loc("<eval_with_key>.2":543:18) | |
| #loc487 = loc("<eval_with_key>.2":544:18) | |
| #loc488 = loc("<eval_with_key>.2":545:14) | |
| #loc489 = loc("<eval_with_key>.2":546:13) | |
| #loc490 = loc("<eval_with_key>.2":547:18) | |
| #loc491 = loc("<eval_with_key>.2":548:14) | |
| #loc492 = loc("<eval_with_key>.2":549:18) | |
| #loc493 = loc("<eval_with_key>.2":550:9) | |
| #loc494 = loc("<eval_with_key>.2":551:29) | |
| #loc495 = loc("<eval_with_key>.2":552:14) | |
| #loc496 = loc("<eval_with_key>.2":553:18) | |
| #loc497 = loc("<eval_with_key>.2":554:15) | |
| #loc498 = loc("<eval_with_key>.2":555:15) | |
| #loc499 = loc("<eval_with_key>.2":556:15) | |
| #loc500 = loc("<eval_with_key>.2":557:19) | |
| #loc501 = loc("<eval_with_key>.2":558:19) | |
| #loc502 = loc("<eval_with_key>.2":559:15) | |
| #loc503 = loc("<eval_with_key>.2":560:13) | |
| #loc504 = loc("<eval_with_key>.2":561:13) | |
| #loc505 = loc("<eval_with_key>.2":562:13) | |
| #loc506 = loc("<eval_with_key>.2":563:15) | |
| #loc507 = loc("<eval_with_key>.2":564:19) | |
| #loc508 = loc("<eval_with_key>.2":565:19) | |
| #loc509 = loc("<eval_with_key>.2":566:15) | |
| #loc510 = loc("<eval_with_key>.2":567:29) | |
| #loc511 = loc("<eval_with_key>.2":568:13) | |
| #loc512 = loc("<eval_with_key>.2":569:13) | |
| #loc513 = loc("<eval_with_key>.2":570:13) | |
| #loc514 = loc("<eval_with_key>.2":571:14) | |
| #loc515 = loc("<eval_with_key>.2":572:13) | |
| #loc516 = loc("<eval_with_key>.2":573:15) | |
| #loc517 = loc("<eval_with_key>.2":575:13) | |
| #loc518 = loc("<eval_with_key>.2":577:13) | |
| #loc519 = loc("<eval_with_key>.2":579:11) | |
| #loc520 = loc("<eval_with_key>.2":580:14) | |
| #loc521 = loc("<eval_with_key>.2":581:12) | |
| #loc522 = loc("<eval_with_key>.2":582:22) | |
| #loc523 = loc("<eval_with_key>.2":583:14) | |
| #loc524 = loc("<eval_with_key>.2":584:19) | |
| #loc525 = loc("<eval_with_key>.2":586:11) | |
| #loc526 = loc("<eval_with_key>.2":587:14) | |
| #loc527 = loc("<eval_with_key>.2":588:12) | |
| #loc528 = loc("<eval_with_key>.2":589:22) | |
| #loc529 = loc("<eval_with_key>.2":590:14) | |
| #loc530 = loc("<eval_with_key>.2":591:19) | |
| #loc531 = loc("<eval_with_key>.2":593:11) | |
| #loc532 = loc("<eval_with_key>.2":594:14) | |
| #loc533 = loc("<eval_with_key>.2":595:12) | |
| #loc534 = loc("<eval_with_key>.2":596:22) | |
| #loc535 = loc("<eval_with_key>.2":597:14) | |
| #loc536 = loc("<eval_with_key>.2":598:19) | |
| #loc537 = loc("<eval_with_key>.2":599:19) | |
| #loc538 = loc("<eval_with_key>.2":600:16) | |
| #loc539 = loc("<eval_with_key>.2":601:14) | |
| #loc540 = loc("<eval_with_key>.2":602:16) | |
| #loc541 = loc("<eval_with_key>.2":603:14) | |
| #loc542 = loc("<eval_with_key>.2":604:13) | |
| #loc543 = loc("<eval_with_key>.2":605:22) | |
| #loc544 = loc("<eval_with_key>.2":606:15) | |
| #loc545 = loc("<eval_with_key>.2":607:15) | |
| #loc546 = loc("<eval_with_key>.2":608:19) | |
| #loc547 = loc("<eval_with_key>.2":609:15) | |
| #loc548 = loc("<eval_with_key>.2":610:19) | |
| #loc549 = loc("<eval_with_key>.2":611:15) | |
| #loc550 = loc("<eval_with_key>.2":612:12) | |
| #loc551 = loc("<eval_with_key>.2":613:17) | |
| #loc552 = loc("<eval_with_key>.2":614:16) | |
| #loc553 = loc("<eval_with_key>.2":615:10) | |
| #loc554 = loc("<eval_with_key>.2":616:11) | |
| #loc555 = loc("<eval_with_key>.2":617:29) | |
| #loc556 = loc("<eval_with_key>.2":618:12) | |
| #loc557 = loc("<eval_with_key>.2":619:12) | |
| #loc558 = loc("<eval_with_key>.2":620:12) | |
| #loc559 = loc("<eval_with_key>.2":621:13) | |
| #loc560 = loc("<eval_with_key>.2":622:29) | |
| #loc561 = loc("<eval_with_key>.2":623:13) | |
| #loc562 = loc("<eval_with_key>.2":624:18) | |
| #loc563 = loc("<eval_with_key>.2":625:16) | |
| #loc564 = loc("<eval_with_key>.2":626:14) | |
| #loc565 = loc("<eval_with_key>.2":627:13) | |
| #loc566 = loc("<eval_with_key>.2":629:18) | |
| #loc567 = loc("<eval_with_key>.2":630:16) | |
| #loc568 = loc("<eval_with_key>.2":631:19) | |
| #loc569 = loc("<eval_with_key>.2":632:13) | |
| #loc570 = loc("<eval_with_key>.2":633:13) | |
| #loc571 = loc("<eval_with_key>.2":634:13) | |
| #loc572 = loc("<eval_with_key>.2":635:12) | |
| #loc573 = loc("<eval_with_key>.2":636:12) | |
| #loc574 = loc("<eval_with_key>.2":637:12) | |
| #loc575 = loc("<eval_with_key>.2":638:13) | |
| #loc576 = loc("<eval_with_key>.2":640:16) | |
| #loc577 = loc("<eval_with_key>.2":641:14) | |
| #loc578 = loc("<eval_with_key>.2":642:16) | |
| #loc579 = loc("<eval_with_key>.2":643:14) | |
| #loc580 = loc("<eval_with_key>.2":644:13) | |
| #loc581 = loc("<eval_with_key>.2":645:22) | |
| #loc582 = loc("<eval_with_key>.2":646:19) | |
| #loc583 = loc("<eval_with_key>.2":647:14) | |
| #loc584 = loc("<eval_with_key>.2":648:14) | |
| #loc585 = loc("<eval_with_key>.2":650:11) | |
| #loc586 = loc("<eval_with_key>.2":651:14) | |
| #loc587 = loc("<eval_with_key>.2":652:12) | |
| #loc588 = loc("<eval_with_key>.2":653:22) | |
| #loc589 = loc("<eval_with_key>.2":654:13) | |
| #loc590 = loc("<eval_with_key>.2":655:13) | |
| #loc591 = loc("<eval_with_key>.2":656:14) | |
| #loc592 = loc("<eval_with_key>.2":657:13) | |
| #loc593 = loc("<eval_with_key>.2":658:15) | |
| #loc594 = loc("<eval_with_key>.2":660:13) | |
| #loc595 = loc("<eval_with_key>.2":662:13) | |
| #loc596 = loc("<eval_with_key>.2":664:11) | |
| #loc597 = loc("<eval_with_key>.2":665:14) | |
| #loc598 = loc("<eval_with_key>.2":666:12) | |
| #loc599 = loc("<eval_with_key>.2":667:22) | |
| #loc600 = loc("<eval_with_key>.2":668:14) | |
| #loc601 = loc("<eval_with_key>.2":669:19) | |
| #loc602 = loc("<eval_with_key>.2":671:11) | |
| #loc603 = loc("<eval_with_key>.2":672:15) | |
| #loc604 = loc("<eval_with_key>.2":673:12) | |
| #loc605 = loc("<eval_with_key>.2":674:22) | |
| #loc606 = loc("<eval_with_key>.2":675:15) | |
| #loc607 = loc("<eval_with_key>.2":676:19) | |
| #loc608 = loc("<eval_with_key>.2":678:11) | |
| #loc609 = loc("<eval_with_key>.2":679:15) | |
| #loc610 = loc("<eval_with_key>.2":680:12) | |
| #loc611 = loc("<eval_with_key>.2":681:22) | |
| #loc612 = loc("<eval_with_key>.2":682:15) | |
| #loc613 = loc("<eval_with_key>.2":683:19) | |
| #loc614 = loc("<eval_with_key>.2":684:19) | |
| #loc615 = loc("<eval_with_key>.2":685:16) | |
| #loc616 = loc("<eval_with_key>.2":686:15) | |
| #loc617 = loc("<eval_with_key>.2":687:16) | |
| #loc618 = loc("<eval_with_key>.2":688:15) | |
| #loc619 = loc("<eval_with_key>.2":689:13) | |
| #loc620 = loc("<eval_with_key>.2":690:22) | |
| #loc621 = loc("<eval_with_key>.2":691:12) | |
| #loc622 = loc("<eval_with_key>.2":692:13) | |
| #loc623 = loc("<eval_with_key>.2":693:13) | |
| #loc624 = loc("<eval_with_key>.2":694:13) | |
| #loc625 = loc("<eval_with_key>.2":695:12) | |
| #loc626 = loc("<eval_with_key>.2":696:12) | |
| #loc627 = loc("<eval_with_key>.2":697:12) | |
| #loc628 = loc("<eval_with_key>.2":698:13) | |
| #loc629 = loc("<eval_with_key>.2":700:16) | |
| #loc630 = loc("<eval_with_key>.2":701:15) | |
| #loc631 = loc("<eval_with_key>.2":702:16) | |
| #loc632 = loc("<eval_with_key>.2":703:15) | |
| #loc633 = loc("<eval_with_key>.2":704:13) | |
| #loc634 = loc("<eval_with_key>.2":705:22) | |
| #loc635 = loc("<eval_with_key>.2":706:19) | |
| #loc636 = loc("<eval_with_key>.2":707:14) | |
| #loc637 = loc("<eval_with_key>.2":708:15) | |
| #loc638 = loc("<eval_with_key>.2":710:11) | |
| #loc639 = loc("<eval_with_key>.2":711:15) | |
| #loc640 = loc("<eval_with_key>.2":712:12) | |
| #loc641 = loc("<eval_with_key>.2":713:22) | |
| #loc642 = loc("<eval_with_key>.2":714:13) | |
| #loc643 = loc("<eval_with_key>.2":715:13) | |
| #loc644 = loc("<eval_with_key>.2":716:14) | |
| #loc645 = loc("<eval_with_key>.2":717:13) | |
| #loc646 = loc("<eval_with_key>.2":718:15) | |
| #loc647 = loc("<eval_with_key>.2":720:13) | |
| #loc648 = loc("<eval_with_key>.2":722:13) | |
| #loc649 = loc("<eval_with_key>.2":724:11) | |
| #loc650 = loc("<eval_with_key>.2":725:15) | |
| #loc651 = loc("<eval_with_key>.2":726:12) | |
| #loc652 = loc("<eval_with_key>.2":727:22) | |
| #loc653 = loc("<eval_with_key>.2":728:13) | |
| #loc654 = loc("<eval_with_key>.2":731:11) | |
| #loc655 = loc("<eval_with_key>.2":732:15) | |
| #loc656 = loc("<eval_with_key>.2":733:12) | |
| #loc657 = loc("<eval_with_key>.2":734:22) | |
| #loc658 = loc("<eval_with_key>.2":735:13) | |
| #loc659 = loc("<eval_with_key>.2":736:13) | |
| #loc660 = loc("<eval_with_key>.2":737:14) | |
| #loc661 = loc("<eval_with_key>.2":738:13) | |
| #loc662 = loc("<eval_with_key>.2":739:15) | |
| #loc663 = loc("<eval_with_key>.2":741:13) | |
| #loc664 = loc("<eval_with_key>.2":743:13) | |
| #loc665 = loc("<eval_with_key>.2":745:11) | |
| #loc666 = loc("<eval_with_key>.2":746:15) | |
| #loc667 = loc("<eval_with_key>.2":747:12) | |
| #loc668 = loc("<eval_with_key>.2":748:22) | |
| #loc669 = loc("<eval_with_key>.2":749:15) | |
| #loc670 = loc("<eval_with_key>.2":750:19) | |
| #loc671 = loc("<eval_with_key>.2":752:11) | |
| #loc672 = loc("<eval_with_key>.2":753:15) | |
| #loc673 = loc("<eval_with_key>.2":754:12) | |
| #loc674 = loc("<eval_with_key>.2":755:22) | |
| #loc675 = loc("<eval_with_key>.2":756:15) | |
| #loc676 = loc("<eval_with_key>.2":757:19) | |
| #loc677 = loc("<eval_with_key>.2":759:11) | |
| #loc678 = loc("<eval_with_key>.2":760:15) | |
| #loc679 = loc("<eval_with_key>.2":761:12) | |
| #loc680 = loc("<eval_with_key>.2":762:22) | |
| #loc681 = loc("<eval_with_key>.2":763:15) | |
| #loc682 = loc("<eval_with_key>.2":764:19) | |
| #loc683 = loc("<eval_with_key>.2":765:19) | |
| #loc684 = loc("<eval_with_key>.2":766:16) | |
| #loc685 = loc("<eval_with_key>.2":767:15) | |
| #loc686 = loc("<eval_with_key>.2":768:16) | |
| #loc687 = loc("<eval_with_key>.2":769:15) | |
| #loc688 = loc("<eval_with_key>.2":770:13) | |
| #loc689 = loc("<eval_with_key>.2":771:22) | |
| #loc690 = loc("<eval_with_key>.2":772:13) | |
| #loc691 = loc("<eval_with_key>.2":773:13) | |
| #loc692 = loc("<eval_with_key>.2":774:13) | |
| #loc693 = loc("<eval_with_key>.2":775:12) | |
| #loc694 = loc("<eval_with_key>.2":776:12) | |
| #loc695 = loc("<eval_with_key>.2":777:13) | |
| #loc696 = loc("<eval_with_key>.2":779:16) | |
| #loc697 = loc("<eval_with_key>.2":780:15) | |
| #loc698 = loc("<eval_with_key>.2":781:16) | |
| #loc699 = loc("<eval_with_key>.2":782:15) | |
| #loc700 = loc("<eval_with_key>.2":783:13) | |
| #loc701 = loc("<eval_with_key>.2":784:22) | |
| #loc702 = loc("<eval_with_key>.2":785:19) | |
| #loc703 = loc("<eval_with_key>.2":786:14) | |
| #loc704 = loc("<eval_with_key>.2":787:15) | |
| #loc705 = loc("<eval_with_key>.2":789:11) | |
| #loc706 = loc("<eval_with_key>.2":790:15) | |
| #loc707 = loc("<eval_with_key>.2":791:12) | |
| #loc708 = loc("<eval_with_key>.2":792:22) | |
| #loc709 = loc("<eval_with_key>.2":793:13) | |
| #loc710 = loc("<eval_with_key>.2":794:13) | |
| #loc711 = loc("<eval_with_key>.2":795:14) | |
| #loc712 = loc("<eval_with_key>.2":796:13) | |
| #loc713 = loc("<eval_with_key>.2":797:15) | |
| #loc714 = loc("<eval_with_key>.2":799:13) | |
| #loc715 = loc("<eval_with_key>.2":801:13) | |
| #loc716 = loc("<eval_with_key>.2":803:11) | |
| #loc717 = loc("<eval_with_key>.2":804:15) | |
| #loc718 = loc("<eval_with_key>.2":805:12) | |
| #loc719 = loc("<eval_with_key>.2":806:22) | |
| #loc720 = loc("<eval_with_key>.2":807:15) | |
| #loc721 = loc("<eval_with_key>.2":808:19) | |
| #loc722 = loc("<eval_with_key>.2":810:11) | |
| #loc723 = loc("<eval_with_key>.2":811:15) | |
| #loc724 = loc("<eval_with_key>.2":812:12) | |
| #loc725 = loc("<eval_with_key>.2":813:22) | |
| #loc726 = loc("<eval_with_key>.2":814:15) | |
| #loc727 = loc("<eval_with_key>.2":815:19) | |
| #loc728 = loc("<eval_with_key>.2":817:11) | |
| #loc729 = loc("<eval_with_key>.2":818:15) | |
| #loc730 = loc("<eval_with_key>.2":819:12) | |
| #loc731 = loc("<eval_with_key>.2":820:22) | |
| #loc732 = loc("<eval_with_key>.2":821:15) | |
| #loc733 = loc("<eval_with_key>.2":822:19) | |
| #loc734 = loc("<eval_with_key>.2":823:19) | |
| #loc735 = loc("<eval_with_key>.2":824:16) | |
| #loc736 = loc("<eval_with_key>.2":825:15) | |
| #loc737 = loc("<eval_with_key>.2":826:16) | |
| #loc738 = loc("<eval_with_key>.2":827:15) | |
| #loc739 = loc("<eval_with_key>.2":828:13) | |
| #loc740 = loc("<eval_with_key>.2":829:22) | |
| #loc741 = loc("<eval_with_key>.2":830:14) | |
| #loc742 = loc("<eval_with_key>.2":831:13) | |
| #loc743 = loc("<eval_with_key>.2":832:13) | |
| #loc744 = loc("<eval_with_key>.2":833:12) | |
| #loc745 = loc("<eval_with_key>.2":834:13) | |
| #loc746 = loc("<eval_with_key>.2":835:13) | |
| #loc747 = loc("<eval_with_key>.2":837:16) | |
| #loc748 = loc("<eval_with_key>.2":838:15) | |
| #loc749 = loc("<eval_with_key>.2":839:16) | |
| #loc750 = loc("<eval_with_key>.2":840:15) | |
| #loc751 = loc("<eval_with_key>.2":841:13) | |
| #loc752 = loc("<eval_with_key>.2":842:22) | |
| #loc753 = loc("<eval_with_key>.2":843:19) | |
| #loc754 = loc("<eval_with_key>.2":844:15) | |
| #loc755 = loc("<eval_with_key>.2":845:15) | |
| #loc756 = loc("<eval_with_key>.2":847:11) | |
| #loc757 = loc("<eval_with_key>.2":848:15) | |
| #loc758 = loc("<eval_with_key>.2":849:12) | |
| #loc759 = loc("<eval_with_key>.2":850:22) | |
| #loc760 = loc("<eval_with_key>.2":851:13) | |
| #loc761 = loc("<eval_with_key>.2":852:13) | |
| #loc762 = loc("<eval_with_key>.2":853:14) | |
| #loc763 = loc("<eval_with_key>.2":854:13) | |
| #loc764 = loc("<eval_with_key>.2":855:15) | |
| #loc765 = loc("<eval_with_key>.2":857:13) | |
| #loc766 = loc("<eval_with_key>.2":859:13) | |
| #loc767 = loc("<eval_with_key>.2":861:11) | |
| #loc768 = loc("<eval_with_key>.2":862:15) | |
| #loc769 = loc("<eval_with_key>.2":863:12) | |
| #loc770 = loc("<eval_with_key>.2":864:22) | |
| #loc771 = loc("<eval_with_key>.2":865:13) | |
| #loc772 = loc("<eval_with_key>.2":868:11) | |
| #loc773 = loc("<eval_with_key>.2":869:15) | |
| #loc774 = loc("<eval_with_key>.2":870:12) | |
| #loc775 = loc("<eval_with_key>.2":871:22) | |
| #loc776 = loc("<eval_with_key>.2":872:13) | |
| #loc777 = loc("<eval_with_key>.2":873:13) | |
| #loc778 = loc("<eval_with_key>.2":874:14) | |
| #loc779 = loc("<eval_with_key>.2":875:13) | |
| #loc780 = loc("<eval_with_key>.2":876:15) | |
| #loc781 = loc("<eval_with_key>.2":878:13) | |
| #loc782 = loc("<eval_with_key>.2":880:13) | |
| #loc783 = loc("<eval_with_key>.2":882:11) | |
| #loc784 = loc("<eval_with_key>.2":883:15) | |
| #loc785 = loc("<eval_with_key>.2":884:12) | |
| #loc786 = loc("<eval_with_key>.2":885:22) | |
| #loc787 = loc("<eval_with_key>.2":886:15) | |
| #loc788 = loc("<eval_with_key>.2":887:19) | |
| #loc789 = loc("<eval_with_key>.2":889:11) | |
| #loc790 = loc("<eval_with_key>.2":890:15) | |
| #loc791 = loc("<eval_with_key>.2":891:12) | |
| #loc792 = loc("<eval_with_key>.2":892:22) | |
| #loc793 = loc("<eval_with_key>.2":893:15) | |
| #loc794 = loc("<eval_with_key>.2":894:19) | |
| #loc795 = loc("<eval_with_key>.2":896:11) | |
| #loc796 = loc("<eval_with_key>.2":897:15) | |
| #loc797 = loc("<eval_with_key>.2":898:12) | |
| #loc798 = loc("<eval_with_key>.2":899:22) | |
| #loc799 = loc("<eval_with_key>.2":900:15) | |
| #loc800 = loc("<eval_with_key>.2":901:19) | |
| #loc801 = loc("<eval_with_key>.2":902:19) | |
| #loc802 = loc("<eval_with_key>.2":903:16) | |
| #loc803 = loc("<eval_with_key>.2":904:15) | |
| #loc804 = loc("<eval_with_key>.2":905:16) | |
| #loc805 = loc("<eval_with_key>.2":906:15) | |
| #loc806 = loc("<eval_with_key>.2":907:13) | |
| #loc807 = loc("<eval_with_key>.2":908:22) | |
| #loc808 = loc("<eval_with_key>.2":909:14) | |
| #loc809 = loc("<eval_with_key>.2":910:14) | |
| #loc810 = loc("<eval_with_key>.2":911:13) | |
| #loc811 = loc("<eval_with_key>.2":912:13) | |
| #loc812 = loc("<eval_with_key>.2":913:13) | |
| #loc813 = loc("<eval_with_key>.2":914:13) | |
| #loc814 = loc("<eval_with_key>.2":916:16) | |
| #loc815 = loc("<eval_with_key>.2":917:15) | |
| #loc816 = loc("<eval_with_key>.2":918:16) | |
| #loc817 = loc("<eval_with_key>.2":919:15) | |
| #loc818 = loc("<eval_with_key>.2":920:13) | |
| #loc819 = loc("<eval_with_key>.2":921:22) | |
| #loc820 = loc("<eval_with_key>.2":922:19) | |
| #loc821 = loc("<eval_with_key>.2":923:15) | |
| #loc822 = loc("<eval_with_key>.2":924:15) | |
| #loc823 = loc("<eval_with_key>.2":926:11) | |
| #loc824 = loc("<eval_with_key>.2":927:15) | |
| #loc825 = loc("<eval_with_key>.2":928:12) | |
| #loc826 = loc("<eval_with_key>.2":929:22) | |
| #loc827 = loc("<eval_with_key>.2":930:13) | |
| #loc828 = loc("<eval_with_key>.2":931:13) | |
| #loc829 = loc("<eval_with_key>.2":932:14) | |
| #loc830 = loc("<eval_with_key>.2":933:13) | |
| #loc831 = loc("<eval_with_key>.2":934:15) | |
| #loc832 = loc("<eval_with_key>.2":936:13) | |
| #loc833 = loc("<eval_with_key>.2":938:13) | |
| #loc834 = loc("<eval_with_key>.2":940:11) | |
| #loc835 = loc("<eval_with_key>.2":941:15) | |
| #loc836 = loc("<eval_with_key>.2":942:12) | |
| #loc837 = loc("<eval_with_key>.2":943:22) | |
| #loc838 = loc("<eval_with_key>.2":944:15) | |
| #loc839 = loc("<eval_with_key>.2":945:19) | |
| #loc840 = loc("<eval_with_key>.2":947:11) | |
| #loc841 = loc("<eval_with_key>.2":948:15) | |
| #loc842 = loc("<eval_with_key>.2":949:12) | |
| #loc843 = loc("<eval_with_key>.2":950:22) | |
| #loc844 = loc("<eval_with_key>.2":951:15) | |
| #loc845 = loc("<eval_with_key>.2":952:19) | |
| #loc846 = loc("<eval_with_key>.2":954:11) | |
| #loc847 = loc("<eval_with_key>.2":955:15) | |
| #loc848 = loc("<eval_with_key>.2":956:12) | |
| #loc849 = loc("<eval_with_key>.2":957:22) | |
| #loc850 = loc("<eval_with_key>.2":958:15) | |
| #loc851 = loc("<eval_with_key>.2":959:19) | |
| #loc852 = loc("<eval_with_key>.2":960:19) | |
| #loc853 = loc("<eval_with_key>.2":961:16) | |
| #loc854 = loc("<eval_with_key>.2":962:15) | |
| #loc855 = loc("<eval_with_key>.2":963:16) | |
| #loc856 = loc("<eval_with_key>.2":964:15) | |
| #loc857 = loc("<eval_with_key>.2":965:13) | |
| #loc858 = loc("<eval_with_key>.2":966:22) | |
| #loc859 = loc("<eval_with_key>.2":967:14) | |
| #loc860 = loc("<eval_with_key>.2":968:14) | |
| #loc861 = loc("<eval_with_key>.2":969:13) | |
| #loc862 = loc("<eval_with_key>.2":970:13) | |
| #loc863 = loc("<eval_with_key>.2":971:13) | |
| #loc864 = loc("<eval_with_key>.2":972:13) | |
| #loc865 = loc("<eval_with_key>.2":974:16) | |
| #loc866 = loc("<eval_with_key>.2":975:15) | |
| #loc867 = loc("<eval_with_key>.2":976:16) | |
| #loc868 = loc("<eval_with_key>.2":977:15) | |
| #loc869 = loc("<eval_with_key>.2":978:13) | |
| #loc870 = loc("<eval_with_key>.2":979:22) | |
| #loc871 = loc("<eval_with_key>.2":980:19) | |
| #loc872 = loc("<eval_with_key>.2":981:15) | |
| #loc873 = loc("<eval_with_key>.2":982:15) | |
| #loc874 = loc("<eval_with_key>.2":984:11) | |
| #loc875 = loc("<eval_with_key>.2":985:15) | |
| #loc876 = loc("<eval_with_key>.2":986:12) | |
| #loc877 = loc("<eval_with_key>.2":987:22) | |
| #loc878 = loc("<eval_with_key>.2":988:13) | |
| #loc879 = loc("<eval_with_key>.2":989:13) | |
| #loc880 = loc("<eval_with_key>.2":990:14) | |
| #loc881 = loc("<eval_with_key>.2":991:13) | |
| #loc882 = loc("<eval_with_key>.2":992:15) | |
| #loc883 = loc("<eval_with_key>.2":994:13) | |
| #loc884 = loc("<eval_with_key>.2":996:13) | |
| #loc885 = loc("<eval_with_key>.2":998:11) | |
| #loc886 = loc("<eval_with_key>.2":999:15) | |
| #loc887 = loc("<eval_with_key>.2":1000:12) | |
| #loc888 = loc("<eval_with_key>.2":1001:22) | |
| #loc889 = loc("<eval_with_key>.2":1002:13) | |
| #loc890 = loc("<eval_with_key>.2":1005:11) | |
| #loc891 = loc("<eval_with_key>.2":1006:15) | |
| #loc892 = loc("<eval_with_key>.2":1007:12) | |
| #loc893 = loc("<eval_with_key>.2":1008:22) | |
| #loc894 = loc("<eval_with_key>.2":1009:13) | |
| #loc895 = loc("<eval_with_key>.2":1010:13) | |
| #loc896 = loc("<eval_with_key>.2":1011:14) | |
| #loc897 = loc("<eval_with_key>.2":1012:13) | |
| #loc898 = loc("<eval_with_key>.2":1013:15) | |
| #loc899 = loc("<eval_with_key>.2":1015:13) | |
| #loc900 = loc("<eval_with_key>.2":1017:13) | |
| #loc901 = loc("<eval_with_key>.2":1019:11) | |
| #loc902 = loc("<eval_with_key>.2":1020:15) | |
| #loc903 = loc("<eval_with_key>.2":1021:12) | |
| #loc904 = loc("<eval_with_key>.2":1022:22) | |
| #loc905 = loc("<eval_with_key>.2":1023:15) | |
| #loc906 = loc("<eval_with_key>.2":1024:19) | |
| #loc907 = loc("<eval_with_key>.2":1026:11) | |
| #loc908 = loc("<eval_with_key>.2":1027:15) | |
| #loc909 = loc("<eval_with_key>.2":1028:12) | |
| #loc910 = loc("<eval_with_key>.2":1029:22) | |
| #loc911 = loc("<eval_with_key>.2":1030:15) | |
| #loc912 = loc("<eval_with_key>.2":1031:19) | |
| #loc913 = loc("<eval_with_key>.2":1033:11) | |
| #loc914 = loc("<eval_with_key>.2":1034:15) | |
| #loc915 = loc("<eval_with_key>.2":1035:12) | |
| #loc916 = loc("<eval_with_key>.2":1036:22) | |
| #loc917 = loc("<eval_with_key>.2":1037:15) | |
| #loc918 = loc("<eval_with_key>.2":1038:19) | |
| #loc919 = loc("<eval_with_key>.2":1039:19) | |
| #loc920 = loc("<eval_with_key>.2":1040:16) | |
| #loc921 = loc("<eval_with_key>.2":1041:15) | |
| #loc922 = loc("<eval_with_key>.2":1042:16) | |
| #loc923 = loc("<eval_with_key>.2":1043:15) | |
| #loc924 = loc("<eval_with_key>.2":1044:13) | |
| #loc925 = loc("<eval_with_key>.2":1045:22) | |
| #loc926 = loc("<eval_with_key>.2":1046:14) | |
| #loc927 = loc("<eval_with_key>.2":1047:14) | |
| #loc928 = loc("<eval_with_key>.2":1048:13) | |
| #loc929 = loc("<eval_with_key>.2":1049:13) | |
| #loc930 = loc("<eval_with_key>.2":1050:13) | |
| #loc931 = loc("<eval_with_key>.2":1051:13) | |
| #loc932 = loc("<eval_with_key>.2":1053:16) | |
| #loc933 = loc("<eval_with_key>.2":1054:15) | |
| #loc934 = loc("<eval_with_key>.2":1055:16) | |
| #loc935 = loc("<eval_with_key>.2":1056:15) | |
| #loc936 = loc("<eval_with_key>.2":1057:13) | |
| #loc937 = loc("<eval_with_key>.2":1058:22) | |
| #loc938 = loc("<eval_with_key>.2":1059:19) | |
| #loc939 = loc("<eval_with_key>.2":1060:15) | |
| #loc940 = loc("<eval_with_key>.2":1061:15) | |
| #loc941 = loc("<eval_with_key>.2":1063:11) | |
| #loc942 = loc("<eval_with_key>.2":1064:15) | |
| #loc943 = loc("<eval_with_key>.2":1065:12) | |
| #loc944 = loc("<eval_with_key>.2":1066:22) | |
| #loc945 = loc("<eval_with_key>.2":1067:13) | |
| #loc946 = loc("<eval_with_key>.2":1068:13) | |
| #loc947 = loc("<eval_with_key>.2":1069:14) | |
| #loc948 = loc("<eval_with_key>.2":1070:13) | |
| #loc949 = loc("<eval_with_key>.2":1071:15) | |
| #loc950 = loc("<eval_with_key>.2":1073:13) | |
| #loc951 = loc("<eval_with_key>.2":1075:13) | |
| #loc952 = loc("<eval_with_key>.2":1077:11) | |
| #loc953 = loc("<eval_with_key>.2":1078:15) | |
| #loc954 = loc("<eval_with_key>.2":1079:12) | |
| #loc955 = loc("<eval_with_key>.2":1080:22) | |
| #loc956 = loc("<eval_with_key>.2":1081:15) | |
| #loc957 = loc("<eval_with_key>.2":1082:19) | |
| #loc958 = loc("<eval_with_key>.2":1084:11) | |
| #loc959 = loc("<eval_with_key>.2":1085:15) | |
| #loc960 = loc("<eval_with_key>.2":1086:12) | |
| #loc961 = loc("<eval_with_key>.2":1087:22) | |
| #loc962 = loc("<eval_with_key>.2":1088:15) | |
| #loc963 = loc("<eval_with_key>.2":1089:19) | |
| #loc964 = loc("<eval_with_key>.2":1091:11) | |
| #loc965 = loc("<eval_with_key>.2":1092:15) | |
| #loc966 = loc("<eval_with_key>.2":1093:12) | |
| #loc967 = loc("<eval_with_key>.2":1094:22) | |
| #loc968 = loc("<eval_with_key>.2":1095:15) | |
| #loc969 = loc("<eval_with_key>.2":1096:19) | |
| #loc970 = loc("<eval_with_key>.2":1097:19) | |
| #loc971 = loc("<eval_with_key>.2":1098:16) | |
| #loc972 = loc("<eval_with_key>.2":1099:15) | |
| #loc973 = loc("<eval_with_key>.2":1100:16) | |
| #loc974 = loc("<eval_with_key>.2":1101:15) | |
| #loc975 = loc("<eval_with_key>.2":1102:13) | |
| #loc976 = loc("<eval_with_key>.2":1103:22) | |
| #loc977 = loc("<eval_with_key>.2":1104:14) | |
| #loc978 = loc("<eval_with_key>.2":1105:14) | |
| #loc979 = loc("<eval_with_key>.2":1106:13) | |
| #loc980 = loc("<eval_with_key>.2":1107:13) | |
| #loc981 = loc("<eval_with_key>.2":1108:13) | |
| #loc982 = loc("<eval_with_key>.2":1109:13) | |
| #loc983 = loc("<eval_with_key>.2":1111:16) | |
| #loc984 = loc("<eval_with_key>.2":1112:15) | |
| #loc985 = loc("<eval_with_key>.2":1113:16) | |
| #loc986 = loc("<eval_with_key>.2":1114:15) | |
| #loc987 = loc("<eval_with_key>.2":1115:13) | |
| #loc988 = loc("<eval_with_key>.2":1116:23) | |
| #loc989 = loc("<eval_with_key>.2":1117:19) | |
| #loc990 = loc("<eval_with_key>.2":1118:15) | |
| #loc991 = loc("<eval_with_key>.2":1119:15) | |
| #loc992 = loc("<eval_with_key>.2":1121:11) | |
| #loc993 = loc("<eval_with_key>.2":1122:15) | |
| #loc994 = loc("<eval_with_key>.2":1123:12) | |
| #loc995 = loc("<eval_with_key>.2":1124:23) | |
| #loc996 = loc("<eval_with_key>.2":1125:13) | |
| #loc997 = loc("<eval_with_key>.2":1126:13) | |
| #loc998 = loc("<eval_with_key>.2":1127:14) | |
| #loc999 = loc("<eval_with_key>.2":1128:13) | |
| #loc1000 = loc("<eval_with_key>.2":1129:15) | |
| #loc1001 = loc("<eval_with_key>.2":1131:13) | |
| #loc1002 = loc("<eval_with_key>.2":1133:13) | |
| #loc1003 = loc("<eval_with_key>.2":1135:11) | |
| #loc1004 = loc("<eval_with_key>.2":1136:15) | |
| #loc1005 = loc("<eval_with_key>.2":1137:12) | |
| #loc1006 = loc("<eval_with_key>.2":1138:23) | |
| #loc1007 = loc("<eval_with_key>.2":1139:13) | |
| #loc1008 = loc("<eval_with_key>.2":1142:11) | |
| #loc1009 = loc("<eval_with_key>.2":1143:15) | |
| #loc1010 = loc("<eval_with_key>.2":1144:12) | |
| #loc1011 = loc("<eval_with_key>.2":1145:23) | |
| #loc1012 = loc("<eval_with_key>.2":1146:13) | |
| #loc1013 = loc("<eval_with_key>.2":1147:13) | |
| #loc1014 = loc("<eval_with_key>.2":1148:14) | |
| #loc1015 = loc("<eval_with_key>.2":1149:13) | |
| #loc1016 = loc("<eval_with_key>.2":1150:15) | |
| #loc1017 = loc("<eval_with_key>.2":1152:13) | |
| #loc1018 = loc("<eval_with_key>.2":1154:13) | |
| #loc1019 = loc("<eval_with_key>.2":1156:11) | |
| #loc1020 = loc("<eval_with_key>.2":1157:15) | |
| #loc1021 = loc("<eval_with_key>.2":1158:12) | |
| #loc1022 = loc("<eval_with_key>.2":1159:23) | |
| #loc1023 = loc("<eval_with_key>.2":1160:15) | |
| #loc1024 = loc("<eval_with_key>.2":1161:19) | |
| #loc1025 = loc("<eval_with_key>.2":1163:11) | |
| #loc1026 = loc("<eval_with_key>.2":1164:15) | |
| #loc1027 = loc("<eval_with_key>.2":1165:12) | |
| #loc1028 = loc("<eval_with_key>.2":1166:23) | |
| #loc1029 = loc("<eval_with_key>.2":1167:15) | |
| #loc1030 = loc("<eval_with_key>.2":1168:19) | |
| #loc1031 = loc("<eval_with_key>.2":1170:11) | |
| #loc1032 = loc("<eval_with_key>.2":1171:15) | |
| #loc1033 = loc("<eval_with_key>.2":1172:12) | |
| #loc1034 = loc("<eval_with_key>.2":1173:23) | |
| #loc1035 = loc("<eval_with_key>.2":1174:15) | |
| #loc1036 = loc("<eval_with_key>.2":1175:19) | |
| #loc1037 = loc("<eval_with_key>.2":1176:19) | |
| #loc1038 = loc("<eval_with_key>.2":1177:16) | |
| #loc1039 = loc("<eval_with_key>.2":1178:15) | |
| #loc1040 = loc("<eval_with_key>.2":1179:16) | |
| #loc1041 = loc("<eval_with_key>.2":1180:15) | |
| #loc1042 = loc("<eval_with_key>.2":1181:13) | |
| #loc1043 = loc("<eval_with_key>.2":1182:23) | |
| #loc1044 = loc("<eval_with_key>.2":1183:14) | |
| #loc1045 = loc("<eval_with_key>.2":1184:14) | |
| #loc1046 = loc("<eval_with_key>.2":1185:13) | |
| #loc1047 = loc("<eval_with_key>.2":1186:13) | |
| #loc1048 = loc("<eval_with_key>.2":1187:13) | |
| #loc1049 = loc("<eval_with_key>.2":1188:13) | |
| #loc1050 = loc("<eval_with_key>.2":1190:16) | |
| #loc1051 = loc("<eval_with_key>.2":1191:15) | |
| #loc1052 = loc("<eval_with_key>.2":1192:16) | |
| #loc1053 = loc("<eval_with_key>.2":1193:15) | |
| #loc1054 = loc("<eval_with_key>.2":1194:13) | |
| #loc1055 = loc("<eval_with_key>.2":1195:23) | |
| #loc1056 = loc("<eval_with_key>.2":1196:19) | |
| #loc1057 = loc("<eval_with_key>.2":1197:15) | |
| #loc1058 = loc("<eval_with_key>.2":1198:15) | |
| #loc1059 = loc("<eval_with_key>.2":1200:11) | |
| #loc1060 = loc("<eval_with_key>.2":1201:15) | |
| #loc1061 = loc("<eval_with_key>.2":1202:12) | |
| #loc1062 = loc("<eval_with_key>.2":1203:23) | |
| #loc1063 = loc("<eval_with_key>.2":1204:13) | |
| #loc1064 = loc("<eval_with_key>.2":1205:13) | |
| #loc1065 = loc("<eval_with_key>.2":1206:14) | |
| #loc1066 = loc("<eval_with_key>.2":1207:13) | |
| #loc1067 = loc("<eval_with_key>.2":1208:15) | |
| #loc1068 = loc("<eval_with_key>.2":1210:13) | |
| #loc1069 = loc("<eval_with_key>.2":1212:13) | |
| #loc1070 = loc("<eval_with_key>.2":1214:11) | |
| #loc1071 = loc("<eval_with_key>.2":1215:15) | |
| #loc1072 = loc("<eval_with_key>.2":1216:12) | |
| #loc1073 = loc("<eval_with_key>.2":1217:23) | |
| #loc1074 = loc("<eval_with_key>.2":1218:15) | |
| #loc1075 = loc("<eval_with_key>.2":1219:19) | |
| #loc1076 = loc("<eval_with_key>.2":1221:11) | |
| #loc1077 = loc("<eval_with_key>.2":1222:15) | |
| #loc1078 = loc("<eval_with_key>.2":1223:12) | |
| #loc1079 = loc("<eval_with_key>.2":1224:23) | |
| #loc1080 = loc("<eval_with_key>.2":1225:15) | |
| #loc1081 = loc("<eval_with_key>.2":1226:19) | |
| #loc1082 = loc("<eval_with_key>.2":1228:11) | |
| #loc1083 = loc("<eval_with_key>.2":1229:15) | |
| #loc1084 = loc("<eval_with_key>.2":1230:12) | |
| #loc1085 = loc("<eval_with_key>.2":1231:23) | |
| #loc1086 = loc("<eval_with_key>.2":1232:15) | |
| #loc1087 = loc("<eval_with_key>.2":1233:19) | |
| #loc1088 = loc("<eval_with_key>.2":1234:19) | |
| #loc1089 = loc("<eval_with_key>.2":1235:16) | |
| #loc1090 = loc("<eval_with_key>.2":1236:15) | |
| #loc1091 = loc("<eval_with_key>.2":1237:16) | |
| #loc1092 = loc("<eval_with_key>.2":1238:15) | |
| #loc1093 = loc("<eval_with_key>.2":1239:13) | |
| #loc1094 = loc("<eval_with_key>.2":1240:23) | |
| #loc1095 = loc("<eval_with_key>.2":1241:14) | |
| #loc1096 = loc("<eval_with_key>.2":1242:14) | |
| #loc1097 = loc("<eval_with_key>.2":1243:13) | |
| #loc1098 = loc("<eval_with_key>.2":1244:13) | |
| #loc1099 = loc("<eval_with_key>.2":1245:13) | |
| #loc1100 = loc("<eval_with_key>.2":1246:13) | |
| #loc1101 = loc("<eval_with_key>.2":1248:16) | |
| #loc1102 = loc("<eval_with_key>.2":1249:15) | |
| #loc1103 = loc("<eval_with_key>.2":1250:16) | |
| #loc1104 = loc("<eval_with_key>.2":1251:15) | |
| #loc1105 = loc("<eval_with_key>.2":1252:13) | |
| #loc1106 = loc("<eval_with_key>.2":1253:23) | |
| #loc1107 = loc("<eval_with_key>.2":1254:19) | |
| #loc1108 = loc("<eval_with_key>.2":1255:15) | |
| #loc1109 = loc("<eval_with_key>.2":1256:15) | |
| #loc1110 = loc("<eval_with_key>.2":1258:11) | |
| #loc1111 = loc("<eval_with_key>.2":1259:15) | |
| #loc1112 = loc("<eval_with_key>.2":1260:12) | |
| #loc1113 = loc("<eval_with_key>.2":1261:23) | |
| #loc1114 = loc("<eval_with_key>.2":1262:13) | |
| #loc1115 = loc("<eval_with_key>.2":1263:13) | |
| #loc1116 = loc("<eval_with_key>.2":1264:14) | |
| #loc1117 = loc("<eval_with_key>.2":1265:13) | |
| #loc1118 = loc("<eval_with_key>.2":1266:15) | |
| #loc1119 = loc("<eval_with_key>.2":1268:13) | |
| #loc1120 = loc("<eval_with_key>.2":1270:13) | |
| #loc1121 = loc("<eval_with_key>.2":1272:11) | |
| #loc1122 = loc("<eval_with_key>.2":1273:15) | |
| #loc1123 = loc("<eval_with_key>.2":1274:12) | |
| #loc1124 = loc("<eval_with_key>.2":1275:23) | |
| #loc1125 = loc("<eval_with_key>.2":1276:14) | |
| #loc1126 = loc("<eval_with_key>.2":1279:11) | |
| #loc1127 = loc("<eval_with_key>.2":1280:15) | |
| #loc1128 = loc("<eval_with_key>.2":1281:12) | |
| #loc1129 = loc("<eval_with_key>.2":1282:23) | |
| #loc1130 = loc("<eval_with_key>.2":1283:13) | |
| #loc1131 = loc("<eval_with_key>.2":1284:13) | |
| #loc1132 = loc("<eval_with_key>.2":1285:14) | |
| #loc1133 = loc("<eval_with_key>.2":1286:13) | |
| #loc1134 = loc("<eval_with_key>.2":1287:15) | |
| #loc1135 = loc("<eval_with_key>.2":1289:13) | |
| #loc1136 = loc("<eval_with_key>.2":1291:13) | |
| #loc1137 = loc("<eval_with_key>.2":1293:11) | |
| #loc1138 = loc("<eval_with_key>.2":1294:15) | |
| #loc1139 = loc("<eval_with_key>.2":1295:12) | |
| #loc1140 = loc("<eval_with_key>.2":1296:23) | |
| #loc1141 = loc("<eval_with_key>.2":1297:15) | |
| #loc1142 = loc("<eval_with_key>.2":1298:19) | |
| #loc1143 = loc("<eval_with_key>.2":1300:11) | |
| #loc1144 = loc("<eval_with_key>.2":1301:15) | |
| #loc1145 = loc("<eval_with_key>.2":1302:12) | |
| #loc1146 = loc("<eval_with_key>.2":1303:23) | |
| #loc1147 = loc("<eval_with_key>.2":1304:15) | |
| #loc1148 = loc("<eval_with_key>.2":1305:19) | |
| #loc1149 = loc("<eval_with_key>.2":1307:11) | |
| #loc1150 = loc("<eval_with_key>.2":1308:15) | |
| #loc1151 = loc("<eval_with_key>.2":1309:12) | |
| #loc1152 = loc("<eval_with_key>.2":1310:23) | |
| #loc1153 = loc("<eval_with_key>.2":1311:15) | |
| #loc1154 = loc("<eval_with_key>.2":1312:19) | |
| #loc1155 = loc("<eval_with_key>.2":1313:19) | |
| #loc1156 = loc("<eval_with_key>.2":1314:16) | |
| #loc1157 = loc("<eval_with_key>.2":1315:15) | |
| #loc1158 = loc("<eval_with_key>.2":1316:16) | |
| #loc1159 = loc("<eval_with_key>.2":1317:15) | |
| #loc1160 = loc("<eval_with_key>.2":1318:13) | |
| #loc1161 = loc("<eval_with_key>.2":1319:23) | |
| #loc1162 = loc("<eval_with_key>.2":1320:14) | |
| #loc1163 = loc("<eval_with_key>.2":1321:14) | |
| #loc1164 = loc("<eval_with_key>.2":1322:13) | |
| #loc1165 = loc("<eval_with_key>.2":1323:13) | |
| #loc1166 = loc("<eval_with_key>.2":1324:13) | |
| #loc1167 = loc("<eval_with_key>.2":1325:13) | |
| #loc1168 = loc("<eval_with_key>.2":1327:16) | |
| #loc1169 = loc("<eval_with_key>.2":1328:15) | |
| #loc1170 = loc("<eval_with_key>.2":1329:16) | |
| #loc1171 = loc("<eval_with_key>.2":1330:15) | |
| #loc1172 = loc("<eval_with_key>.2":1331:13) | |
| #loc1173 = loc("<eval_with_key>.2":1332:23) | |
| #loc1174 = loc("<eval_with_key>.2":1333:19) | |
| #loc1175 = loc("<eval_with_key>.2":1334:15) | |
| #loc1176 = loc("<eval_with_key>.2":1335:15) | |
| #loc1177 = loc("<eval_with_key>.2":1337:11) | |
| #loc1178 = loc("<eval_with_key>.2":1338:15) | |
| #loc1179 = loc("<eval_with_key>.2":1339:12) | |
| #loc1180 = loc("<eval_with_key>.2":1340:23) | |
| #loc1181 = loc("<eval_with_key>.2":1341:13) | |
| #loc1182 = loc("<eval_with_key>.2":1342:13) | |
| #loc1183 = loc("<eval_with_key>.2":1343:14) | |
| #loc1184 = loc("<eval_with_key>.2":1344:13) | |
| #loc1185 = loc("<eval_with_key>.2":1345:15) | |
| #loc1186 = loc("<eval_with_key>.2":1347:13) | |
| #loc1187 = loc("<eval_with_key>.2":1349:13) | |
| #loc1188 = loc("<eval_with_key>.2":1351:11) | |
| #loc1189 = loc("<eval_with_key>.2":1352:15) | |
| #loc1190 = loc("<eval_with_key>.2":1353:12) | |
| #loc1191 = loc("<eval_with_key>.2":1354:23) | |
| #loc1192 = loc("<eval_with_key>.2":1355:15) | |
| #loc1193 = loc("<eval_with_key>.2":1356:19) | |
| #loc1194 = loc("<eval_with_key>.2":1358:11) | |
| #loc1195 = loc("<eval_with_key>.2":1359:15) | |
| #loc1196 = loc("<eval_with_key>.2":1360:12) | |
| #loc1197 = loc("<eval_with_key>.2":1361:23) | |
| #loc1198 = loc("<eval_with_key>.2":1362:15) | |
| #loc1199 = loc("<eval_with_key>.2":1363:19) | |
| #loc1200 = loc("<eval_with_key>.2":1365:11) | |
| #loc1201 = loc("<eval_with_key>.2":1366:15) | |
| #loc1202 = loc("<eval_with_key>.2":1367:12) | |
| #loc1203 = loc("<eval_with_key>.2":1368:23) | |
| #loc1204 = loc("<eval_with_key>.2":1369:15) | |
| #loc1205 = loc("<eval_with_key>.2":1370:19) | |
| #loc1206 = loc("<eval_with_key>.2":1371:19) | |
| #loc1207 = loc("<eval_with_key>.2":1372:16) | |
| #loc1208 = loc("<eval_with_key>.2":1373:15) | |
| #loc1209 = loc("<eval_with_key>.2":1374:16) | |
| #loc1210 = loc("<eval_with_key>.2":1375:15) | |
| #loc1211 = loc("<eval_with_key>.2":1376:13) | |
| #loc1212 = loc("<eval_with_key>.2":1377:23) | |
| #loc1213 = loc("<eval_with_key>.2":1378:14) | |
| #loc1214 = loc("<eval_with_key>.2":1379:14) | |
| #loc1215 = loc("<eval_with_key>.2":1380:13) | |
| #loc1216 = loc("<eval_with_key>.2":1381:13) | |
| #loc1217 = loc("<eval_with_key>.2":1382:13) | |
| #loc1218 = loc("<eval_with_key>.2":1383:13) | |
| #loc1219 = loc("<eval_with_key>.2":1385:16) | |
| #loc1220 = loc("<eval_with_key>.2":1386:15) | |
| #loc1221 = loc("<eval_with_key>.2":1387:16) | |
| #loc1222 = loc("<eval_with_key>.2":1388:15) | |
| #loc1223 = loc("<eval_with_key>.2":1389:13) | |
| #loc1224 = loc("<eval_with_key>.2":1390:23) | |
| #loc1225 = loc("<eval_with_key>.2":1391:19) | |
| #loc1226 = loc("<eval_with_key>.2":1392:15) | |
| #loc1227 = loc("<eval_with_key>.2":1393:15) | |
| #loc1228 = loc("<eval_with_key>.2":1395:11) | |
| #loc1229 = loc("<eval_with_key>.2":1396:15) | |
| #loc1230 = loc("<eval_with_key>.2":1397:12) | |
| #loc1231 = loc("<eval_with_key>.2":1398:23) | |
| #loc1232 = loc("<eval_with_key>.2":1399:13) | |
| #loc1233 = loc("<eval_with_key>.2":1400:13) | |
| #loc1234 = loc("<eval_with_key>.2":1401:14) | |
| #loc1235 = loc("<eval_with_key>.2":1402:13) | |
| #loc1236 = loc("<eval_with_key>.2":1403:15) | |
| #loc1237 = loc("<eval_with_key>.2":1405:13) | |
| #loc1238 = loc("<eval_with_key>.2":1407:13) | |
| #loc1239 = loc("<eval_with_key>.2":1409:11) | |
| #loc1240 = loc("<eval_with_key>.2":1410:15) | |
| #loc1241 = loc("<eval_with_key>.2":1411:12) | |
| #loc1242 = loc("<eval_with_key>.2":1412:23) | |
| #loc1243 = loc("<eval_with_key>.2":1413:14) | |
| #loc1244 = loc("<eval_with_key>.2":1416:11) | |
| #loc1245 = loc("<eval_with_key>.2":1417:15) | |
| #loc1246 = loc("<eval_with_key>.2":1418:12) | |
| #loc1247 = loc("<eval_with_key>.2":1419:23) | |
| #loc1248 = loc("<eval_with_key>.2":1420:13) | |
| #loc1249 = loc("<eval_with_key>.2":1421:13) | |
| #loc1250 = loc("<eval_with_key>.2":1422:14) | |
| #loc1251 = loc("<eval_with_key>.2":1423:13) | |
| #loc1252 = loc("<eval_with_key>.2":1424:15) | |
| #loc1253 = loc("<eval_with_key>.2":1426:13) | |
| #loc1254 = loc("<eval_with_key>.2":1428:13) | |
| #loc1255 = loc("<eval_with_key>.2":1429:13) | |
| #loc1256 = loc("<eval_with_key>.2":1431:11) | |
| #loc1257 = loc("<eval_with_key>.2":1432:15) | |
| #loc1258 = loc("<eval_with_key>.2":1433:12) | |
| #loc1259 = loc("<eval_with_key>.2":1434:23) |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment