Created
November 28, 2022 16:29
-
-
Save pashu123/dc21e4608733d5fe81f6aa562f4bdacb to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
func.func @forward(%arg0: !torch.vtensor<[1,77],si64>) -> !torch.vtensor<[1,77,1024],f16> { | |
%0 = torch.vtensor.literal(dense<1.250000e-01> : tensor<f64>) : !torch.vtensor<[],f64> | |
%1 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1x77xsi64>) : !torch.vtensor<[1,77],si64> | |
%2 = torch.vtensor.literal(dense_resource<__elided__> : tensor<49408x1024xf16>) : !torch.vtensor<[49408,1024],f16> | |
%3 = torch.vtensor.literal(dense_resource<__elided__> : tensor<77x1024xf16>) : !torch.vtensor<[77,1024],f16> | |
%4 = torch.vtensor.literal(dense<-6.550400e+04> : tensor<f32>) : !torch.vtensor<[],f32> | |
%5 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1024x1024xf16>) : !torch.vtensor<[1024,1024],f16> | |
%6 = torch.vtensor.literal(dense_resource<__elided__> : tensor<4096x1024xf16>) : !torch.vtensor<[4096,1024],f16> | |
%7 = torch.vtensor.literal(dense_resource<__elided__> : tensor<4096xf16>) : !torch.vtensor<[4096],f16> | |
%8 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1024x4096xf16>) : !torch.vtensor<[1024,4096],f16> | |
%9 = torch.vtensor.literal(dense_resource<__elided__> : tensor<1024xf16>) : !torch.vtensor<[1024],f16> | |
%int-1 = torch.constant.int -1 | |
%int77 = torch.constant.int 77 | |
%int0 = torch.constant.int 0 | |
%int9223372036854775807 = torch.constant.int 9223372036854775807 | |
%int1 = torch.constant.int 1 | |
%false = torch.constant.bool false | |
%int5 = torch.constant.int 5 | |
%none = torch.constant.none | |
%int1024 = torch.constant.int 1024 | |
%float1.000000e-05 = torch.constant.float 1.000000e-05 | |
%int16 = torch.constant.int 16 | |
%int64 = torch.constant.int 64 | |
%int2 = torch.constant.int 2 | |
%int4096 = torch.constant.int 4096 | |
%str = torch.constant.str "none" | |
%10 = torch.prim.ListConstruct %int-1, %int77 : (!torch.int, !torch.int) -> !torch.list<int> | |
%11 = torch.aten.view %arg0, %10 : !torch.vtensor<[1,77],si64>, !torch.list<int> -> !torch.vtensor<[1,77],si64> | |
%12 = torch.aten.slice.Tensor %1, %int0, %int0, %int9223372036854775807, %int1 : !torch.vtensor<[1,77],si64>, !torch.int, !torch.int, !torch.int, !torch.int -> !torch.vtensor<[1,77],si64> | |
%13 = torch.aten.embedding %2, %11, %int-1, %false, %false : !torch.vtensor<[49408,1024],f16>, !torch.vtensor<[1,77],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[1,77,1024],f16> | |
%14 = torch.aten.embedding %3, %12, %int-1, %false, %false : !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1,77],si64>, !torch.int, !torch.bool, !torch.bool -> !torch.vtensor<[1,77,1024],f16> | |
%15 = torch.aten.add.Tensor %13, %14, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%16 = torch.prim.ListConstruct %int1, %int77, %int77 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%cpu = torch.constant.device "cpu" | |
%17 = torch.aten.empty.memory_format %16, %int5, %none, %cpu, %false, %none : !torch.list<int>, !torch.int, !torch.none, !torch.Device, !torch.bool, !torch.none -> !torch.vtensor<[1,77,77],f16> | |
%18 = torch.aten.lift_fresh_copy %4 : !torch.vtensor<[],f32> -> !torch.vtensor<[],f32> | |
%19 = torch.aten.fill.Tensor %17, %18 : !torch.vtensor<[1,77,77],f16>, !torch.vtensor<[],f32> -> !torch.vtensor<[1,77,77],f16> | |
%20 = torch.aten.triu %19, %int1 : !torch.vtensor<[1,77,77],f16>, !torch.int -> !torch.vtensor<[1,77,77],f16> | |
%21 = torch.aten.unsqueeze %20, %int1 : !torch.vtensor<[1,77,77],f16>, !torch.int -> !torch.vtensor<[1,1,77,77],f16> | |
%cuda3A0 = torch.constant.device "cuda:0" | |
%22 = torch.aten._to_copy %21, %int5, %int0, %cuda3A0, %none, %false, %none : !torch.vtensor<[1,1,77,77],f16>, !torch.int, !torch.int, !torch.Device, !torch.none, !torch.bool, !torch.none -> !torch.vtensor<[1,1,77,77],f16> | |
%23 = torch.prim.ListConstruct %int1024 : (!torch.int) -> !torch.list<int> | |
%result0, %result1, %result2 = torch.aten.native_layer_norm %15, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%24 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%25 = torch.prim.ListConstruct %int77, %int1024 : (!torch.int, !torch.int) -> !torch.list<int> | |
%26 = torch.aten.view %result0, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%27 = torch.aten.addmm %9, %26, %24, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%28 = torch.prim.ListConstruct %int1, %int77, %int1024 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%29 = torch.aten.view %27, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%30 = torch.aten.mul.Tensor %29, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%31 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%32 = torch.aten.view %result0, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%33 = torch.aten.addmm %9, %32, %31, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%34 = torch.aten.view %33, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%35 = torch.prim.ListConstruct %int1, %int-1, %int16, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%36 = torch.aten.view %34, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%37 = torch.aten.transpose.int %36, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%38 = torch.aten.clone %37, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%39 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%40 = torch.aten.view %result0, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%41 = torch.aten.addmm %9, %40, %39, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%42 = torch.aten.view %41, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%43 = torch.aten.view %42, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%44 = torch.aten.transpose.int %43, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%45 = torch.aten.clone %44, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%46 = torch.prim.ListConstruct %int1, %int77, %int16, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%47 = torch.aten.view %30, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%48 = torch.aten.transpose.int %47, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%49 = torch.aten.clone %48, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%50 = torch.prim.ListConstruct %int16, %int-1, %int64 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%51 = torch.aten.view %49, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%52 = torch.aten.view %38, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%53 = torch.aten.view %45, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%54 = torch.aten.transpose.int %52, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%55 = torch.aten.bmm %51, %54 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%56 = torch.prim.ListConstruct %int1, %int16, %int77, %int77 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%57 = torch.aten.view %55, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%58 = torch.aten.add.Tensor %57, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%59 = torch.prim.ListConstruct %int16, %int77, %int77 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%60 = torch.aten.view %58, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%61 = torch.aten._softmax %60, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%62 = torch.aten.bmm %61, %53 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%63 = torch.prim.ListConstruct %int1, %int16, %int77, %int64 : (!torch.int, !torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%64 = torch.aten.view %62, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%65 = torch.aten.transpose.int %64, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%66 = torch.aten.clone %65, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%67 = torch.aten._unsafe_view %66, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%68 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%69 = torch.aten.view %67, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%70 = torch.aten.addmm %9, %69, %68, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%71 = torch.aten.view %70, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%72 = torch.aten.add.Tensor %15, %71, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_0, %result1_1, %result2_2 = torch.aten.native_layer_norm %72, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%73 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%74 = torch.aten.view %result0_0, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%75 = torch.aten.addmm %7, %74, %73, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%76 = torch.prim.ListConstruct %int1, %int77, %int4096 : (!torch.int, !torch.int, !torch.int) -> !torch.list<int> | |
%77 = torch.aten.view %75, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%78 = torch.aten.gelu %77, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%79 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%80 = torch.prim.ListConstruct %int77, %int4096 : (!torch.int, !torch.int) -> !torch.list<int> | |
%81 = torch.aten.view %78, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%82 = torch.aten.addmm %9, %81, %79, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%83 = torch.aten.view %82, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%84 = torch.aten.add.Tensor %72, %83, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_3, %result1_4, %result2_5 = torch.aten.native_layer_norm %84, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%85 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%86 = torch.aten.view %result0_3, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%87 = torch.aten.addmm %9, %86, %85, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%88 = torch.aten.view %87, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%89 = torch.aten.mul.Tensor %88, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%90 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%91 = torch.aten.view %result0_3, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%92 = torch.aten.addmm %9, %91, %90, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%93 = torch.aten.view %92, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%94 = torch.aten.view %93, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%95 = torch.aten.transpose.int %94, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%96 = torch.aten.clone %95, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%97 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%98 = torch.aten.view %result0_3, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%99 = torch.aten.addmm %9, %98, %97, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%100 = torch.aten.view %99, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%101 = torch.aten.view %100, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%102 = torch.aten.transpose.int %101, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%103 = torch.aten.clone %102, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%104 = torch.aten.view %89, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%105 = torch.aten.transpose.int %104, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%106 = torch.aten.clone %105, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%107 = torch.aten.view %106, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%108 = torch.aten.view %96, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%109 = torch.aten.view %103, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%110 = torch.aten.transpose.int %108, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%111 = torch.aten.bmm %107, %110 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%112 = torch.aten.view %111, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%113 = torch.aten.add.Tensor %112, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%114 = torch.aten.view %113, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%115 = torch.aten._softmax %114, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%116 = torch.aten.bmm %115, %109 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%117 = torch.aten.view %116, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%118 = torch.aten.transpose.int %117, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%119 = torch.aten.clone %118, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%120 = torch.aten._unsafe_view %119, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%121 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%122 = torch.aten.view %120, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%123 = torch.aten.addmm %9, %122, %121, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%124 = torch.aten.view %123, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%125 = torch.aten.add.Tensor %84, %124, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_6, %result1_7, %result2_8 = torch.aten.native_layer_norm %125, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%126 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%127 = torch.aten.view %result0_6, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%128 = torch.aten.addmm %7, %127, %126, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%129 = torch.aten.view %128, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%130 = torch.aten.gelu %129, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%131 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%132 = torch.aten.view %130, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%133 = torch.aten.addmm %9, %132, %131, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%134 = torch.aten.view %133, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%135 = torch.aten.add.Tensor %125, %134, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_9, %result1_10, %result2_11 = torch.aten.native_layer_norm %135, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%136 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%137 = torch.aten.view %result0_9, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%138 = torch.aten.addmm %9, %137, %136, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%139 = torch.aten.view %138, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%140 = torch.aten.mul.Tensor %139, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%141 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%142 = torch.aten.view %result0_9, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%143 = torch.aten.addmm %9, %142, %141, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%144 = torch.aten.view %143, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%145 = torch.aten.view %144, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%146 = torch.aten.transpose.int %145, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%147 = torch.aten.clone %146, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%148 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%149 = torch.aten.view %result0_9, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%150 = torch.aten.addmm %9, %149, %148, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%151 = torch.aten.view %150, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%152 = torch.aten.view %151, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%153 = torch.aten.transpose.int %152, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%154 = torch.aten.clone %153, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%155 = torch.aten.view %140, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%156 = torch.aten.transpose.int %155, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%157 = torch.aten.clone %156, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%158 = torch.aten.view %157, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%159 = torch.aten.view %147, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%160 = torch.aten.view %154, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%161 = torch.aten.transpose.int %159, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%162 = torch.aten.bmm %158, %161 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%163 = torch.aten.view %162, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%164 = torch.aten.add.Tensor %163, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%165 = torch.aten.view %164, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%166 = torch.aten._softmax %165, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%167 = torch.aten.bmm %166, %160 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%168 = torch.aten.view %167, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%169 = torch.aten.transpose.int %168, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%170 = torch.aten.clone %169, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%171 = torch.aten._unsafe_view %170, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%172 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%173 = torch.aten.view %171, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%174 = torch.aten.addmm %9, %173, %172, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%175 = torch.aten.view %174, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%176 = torch.aten.add.Tensor %135, %175, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_12, %result1_13, %result2_14 = torch.aten.native_layer_norm %176, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%177 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%178 = torch.aten.view %result0_12, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%179 = torch.aten.addmm %7, %178, %177, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%180 = torch.aten.view %179, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%181 = torch.aten.gelu %180, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%182 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%183 = torch.aten.view %181, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%184 = torch.aten.addmm %9, %183, %182, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%185 = torch.aten.view %184, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%186 = torch.aten.add.Tensor %176, %185, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_15, %result1_16, %result2_17 = torch.aten.native_layer_norm %186, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%187 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%188 = torch.aten.view %result0_15, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%189 = torch.aten.addmm %9, %188, %187, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%190 = torch.aten.view %189, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%191 = torch.aten.mul.Tensor %190, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%192 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%193 = torch.aten.view %result0_15, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%194 = torch.aten.addmm %9, %193, %192, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%195 = torch.aten.view %194, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%196 = torch.aten.view %195, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%197 = torch.aten.transpose.int %196, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%198 = torch.aten.clone %197, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%199 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%200 = torch.aten.view %result0_15, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%201 = torch.aten.addmm %9, %200, %199, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%202 = torch.aten.view %201, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%203 = torch.aten.view %202, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%204 = torch.aten.transpose.int %203, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%205 = torch.aten.clone %204, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%206 = torch.aten.view %191, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%207 = torch.aten.transpose.int %206, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%208 = torch.aten.clone %207, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%209 = torch.aten.view %208, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%210 = torch.aten.view %198, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%211 = torch.aten.view %205, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%212 = torch.aten.transpose.int %210, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%213 = torch.aten.bmm %209, %212 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%214 = torch.aten.view %213, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%215 = torch.aten.add.Tensor %214, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%216 = torch.aten.view %215, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%217 = torch.aten._softmax %216, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%218 = torch.aten.bmm %217, %211 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%219 = torch.aten.view %218, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%220 = torch.aten.transpose.int %219, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%221 = torch.aten.clone %220, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%222 = torch.aten._unsafe_view %221, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%223 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%224 = torch.aten.view %222, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%225 = torch.aten.addmm %9, %224, %223, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%226 = torch.aten.view %225, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%227 = torch.aten.add.Tensor %186, %226, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_18, %result1_19, %result2_20 = torch.aten.native_layer_norm %227, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%228 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%229 = torch.aten.view %result0_18, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%230 = torch.aten.addmm %7, %229, %228, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%231 = torch.aten.view %230, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%232 = torch.aten.gelu %231, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%233 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%234 = torch.aten.view %232, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%235 = torch.aten.addmm %9, %234, %233, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%236 = torch.aten.view %235, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%237 = torch.aten.add.Tensor %227, %236, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_21, %result1_22, %result2_23 = torch.aten.native_layer_norm %237, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%238 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%239 = torch.aten.view %result0_21, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%240 = torch.aten.addmm %9, %239, %238, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%241 = torch.aten.view %240, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%242 = torch.aten.mul.Tensor %241, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%243 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%244 = torch.aten.view %result0_21, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%245 = torch.aten.addmm %9, %244, %243, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%246 = torch.aten.view %245, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%247 = torch.aten.view %246, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%248 = torch.aten.transpose.int %247, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%249 = torch.aten.clone %248, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%250 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%251 = torch.aten.view %result0_21, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%252 = torch.aten.addmm %9, %251, %250, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%253 = torch.aten.view %252, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%254 = torch.aten.view %253, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%255 = torch.aten.transpose.int %254, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%256 = torch.aten.clone %255, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%257 = torch.aten.view %242, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%258 = torch.aten.transpose.int %257, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%259 = torch.aten.clone %258, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%260 = torch.aten.view %259, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%261 = torch.aten.view %249, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%262 = torch.aten.view %256, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%263 = torch.aten.transpose.int %261, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%264 = torch.aten.bmm %260, %263 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%265 = torch.aten.view %264, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%266 = torch.aten.add.Tensor %265, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%267 = torch.aten.view %266, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%268 = torch.aten._softmax %267, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%269 = torch.aten.bmm %268, %262 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%270 = torch.aten.view %269, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%271 = torch.aten.transpose.int %270, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%272 = torch.aten.clone %271, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%273 = torch.aten._unsafe_view %272, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%274 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%275 = torch.aten.view %273, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%276 = torch.aten.addmm %9, %275, %274, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%277 = torch.aten.view %276, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%278 = torch.aten.add.Tensor %237, %277, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_24, %result1_25, %result2_26 = torch.aten.native_layer_norm %278, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%279 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%280 = torch.aten.view %result0_24, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%281 = torch.aten.addmm %7, %280, %279, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%282 = torch.aten.view %281, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%283 = torch.aten.gelu %282, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%284 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%285 = torch.aten.view %283, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%286 = torch.aten.addmm %9, %285, %284, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%287 = torch.aten.view %286, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%288 = torch.aten.add.Tensor %278, %287, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_27, %result1_28, %result2_29 = torch.aten.native_layer_norm %288, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%289 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%290 = torch.aten.view %result0_27, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%291 = torch.aten.addmm %9, %290, %289, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%292 = torch.aten.view %291, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%293 = torch.aten.mul.Tensor %292, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%294 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%295 = torch.aten.view %result0_27, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%296 = torch.aten.addmm %9, %295, %294, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%297 = torch.aten.view %296, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%298 = torch.aten.view %297, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%299 = torch.aten.transpose.int %298, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%300 = torch.aten.clone %299, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%301 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%302 = torch.aten.view %result0_27, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%303 = torch.aten.addmm %9, %302, %301, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%304 = torch.aten.view %303, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%305 = torch.aten.view %304, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%306 = torch.aten.transpose.int %305, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%307 = torch.aten.clone %306, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%308 = torch.aten.view %293, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%309 = torch.aten.transpose.int %308, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%310 = torch.aten.clone %309, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%311 = torch.aten.view %310, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%312 = torch.aten.view %300, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%313 = torch.aten.view %307, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%314 = torch.aten.transpose.int %312, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%315 = torch.aten.bmm %311, %314 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%316 = torch.aten.view %315, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%317 = torch.aten.add.Tensor %316, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%318 = torch.aten.view %317, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%319 = torch.aten._softmax %318, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%320 = torch.aten.bmm %319, %313 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%321 = torch.aten.view %320, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%322 = torch.aten.transpose.int %321, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%323 = torch.aten.clone %322, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%324 = torch.aten._unsafe_view %323, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%325 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%326 = torch.aten.view %324, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%327 = torch.aten.addmm %9, %326, %325, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%328 = torch.aten.view %327, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%329 = torch.aten.add.Tensor %288, %328, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_30, %result1_31, %result2_32 = torch.aten.native_layer_norm %329, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%330 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%331 = torch.aten.view %result0_30, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%332 = torch.aten.addmm %7, %331, %330, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%333 = torch.aten.view %332, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%334 = torch.aten.gelu %333, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%335 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%336 = torch.aten.view %334, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%337 = torch.aten.addmm %9, %336, %335, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%338 = torch.aten.view %337, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%339 = torch.aten.add.Tensor %329, %338, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_33, %result1_34, %result2_35 = torch.aten.native_layer_norm %339, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%340 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%341 = torch.aten.view %result0_33, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%342 = torch.aten.addmm %9, %341, %340, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%343 = torch.aten.view %342, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%344 = torch.aten.mul.Tensor %343, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%345 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%346 = torch.aten.view %result0_33, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%347 = torch.aten.addmm %9, %346, %345, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%348 = torch.aten.view %347, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%349 = torch.aten.view %348, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%350 = torch.aten.transpose.int %349, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%351 = torch.aten.clone %350, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%352 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%353 = torch.aten.view %result0_33, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%354 = torch.aten.addmm %9, %353, %352, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%355 = torch.aten.view %354, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%356 = torch.aten.view %355, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%357 = torch.aten.transpose.int %356, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%358 = torch.aten.clone %357, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%359 = torch.aten.view %344, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%360 = torch.aten.transpose.int %359, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%361 = torch.aten.clone %360, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%362 = torch.aten.view %361, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%363 = torch.aten.view %351, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%364 = torch.aten.view %358, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%365 = torch.aten.transpose.int %363, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%366 = torch.aten.bmm %362, %365 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%367 = torch.aten.view %366, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%368 = torch.aten.add.Tensor %367, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%369 = torch.aten.view %368, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%370 = torch.aten._softmax %369, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%371 = torch.aten.bmm %370, %364 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%372 = torch.aten.view %371, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%373 = torch.aten.transpose.int %372, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%374 = torch.aten.clone %373, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%375 = torch.aten._unsafe_view %374, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%376 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%377 = torch.aten.view %375, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%378 = torch.aten.addmm %9, %377, %376, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%379 = torch.aten.view %378, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%380 = torch.aten.add.Tensor %339, %379, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_36, %result1_37, %result2_38 = torch.aten.native_layer_norm %380, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%381 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%382 = torch.aten.view %result0_36, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%383 = torch.aten.addmm %7, %382, %381, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%384 = torch.aten.view %383, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%385 = torch.aten.gelu %384, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%386 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%387 = torch.aten.view %385, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%388 = torch.aten.addmm %9, %387, %386, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%389 = torch.aten.view %388, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%390 = torch.aten.add.Tensor %380, %389, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_39, %result1_40, %result2_41 = torch.aten.native_layer_norm %390, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%391 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%392 = torch.aten.view %result0_39, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%393 = torch.aten.addmm %9, %392, %391, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%394 = torch.aten.view %393, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%395 = torch.aten.mul.Tensor %394, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%396 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%397 = torch.aten.view %result0_39, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%398 = torch.aten.addmm %9, %397, %396, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%399 = torch.aten.view %398, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%400 = torch.aten.view %399, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%401 = torch.aten.transpose.int %400, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%402 = torch.aten.clone %401, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%403 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%404 = torch.aten.view %result0_39, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%405 = torch.aten.addmm %9, %404, %403, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%406 = torch.aten.view %405, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%407 = torch.aten.view %406, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%408 = torch.aten.transpose.int %407, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%409 = torch.aten.clone %408, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%410 = torch.aten.view %395, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%411 = torch.aten.transpose.int %410, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%412 = torch.aten.clone %411, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%413 = torch.aten.view %412, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%414 = torch.aten.view %402, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%415 = torch.aten.view %409, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%416 = torch.aten.transpose.int %414, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%417 = torch.aten.bmm %413, %416 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%418 = torch.aten.view %417, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%419 = torch.aten.add.Tensor %418, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%420 = torch.aten.view %419, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%421 = torch.aten._softmax %420, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%422 = torch.aten.bmm %421, %415 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%423 = torch.aten.view %422, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%424 = torch.aten.transpose.int %423, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%425 = torch.aten.clone %424, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%426 = torch.aten._unsafe_view %425, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%427 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%428 = torch.aten.view %426, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%429 = torch.aten.addmm %9, %428, %427, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%430 = torch.aten.view %429, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%431 = torch.aten.add.Tensor %390, %430, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_42, %result1_43, %result2_44 = torch.aten.native_layer_norm %431, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%432 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%433 = torch.aten.view %result0_42, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%434 = torch.aten.addmm %7, %433, %432, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%435 = torch.aten.view %434, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%436 = torch.aten.gelu %435, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%437 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%438 = torch.aten.view %436, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%439 = torch.aten.addmm %9, %438, %437, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%440 = torch.aten.view %439, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%441 = torch.aten.add.Tensor %431, %440, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_45, %result1_46, %result2_47 = torch.aten.native_layer_norm %441, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%442 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%443 = torch.aten.view %result0_45, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%444 = torch.aten.addmm %9, %443, %442, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%445 = torch.aten.view %444, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%446 = torch.aten.mul.Tensor %445, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%447 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%448 = torch.aten.view %result0_45, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%449 = torch.aten.addmm %9, %448, %447, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%450 = torch.aten.view %449, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%451 = torch.aten.view %450, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%452 = torch.aten.transpose.int %451, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%453 = torch.aten.clone %452, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%454 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%455 = torch.aten.view %result0_45, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%456 = torch.aten.addmm %9, %455, %454, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%457 = torch.aten.view %456, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%458 = torch.aten.view %457, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%459 = torch.aten.transpose.int %458, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%460 = torch.aten.clone %459, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%461 = torch.aten.view %446, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%462 = torch.aten.transpose.int %461, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%463 = torch.aten.clone %462, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%464 = torch.aten.view %463, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%465 = torch.aten.view %453, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%466 = torch.aten.view %460, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%467 = torch.aten.transpose.int %465, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%468 = torch.aten.bmm %464, %467 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%469 = torch.aten.view %468, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%470 = torch.aten.add.Tensor %469, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%471 = torch.aten.view %470, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%472 = torch.aten._softmax %471, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%473 = torch.aten.bmm %472, %466 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%474 = torch.aten.view %473, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%475 = torch.aten.transpose.int %474, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%476 = torch.aten.clone %475, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%477 = torch.aten._unsafe_view %476, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%478 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%479 = torch.aten.view %477, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%480 = torch.aten.addmm %9, %479, %478, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%481 = torch.aten.view %480, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%482 = torch.aten.add.Tensor %441, %481, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_48, %result1_49, %result2_50 = torch.aten.native_layer_norm %482, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%483 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%484 = torch.aten.view %result0_48, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%485 = torch.aten.addmm %7, %484, %483, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%486 = torch.aten.view %485, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%487 = torch.aten.gelu %486, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%488 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%489 = torch.aten.view %487, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%490 = torch.aten.addmm %9, %489, %488, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%491 = torch.aten.view %490, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%492 = torch.aten.add.Tensor %482, %491, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_51, %result1_52, %result2_53 = torch.aten.native_layer_norm %492, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%493 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%494 = torch.aten.view %result0_51, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%495 = torch.aten.addmm %9, %494, %493, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%496 = torch.aten.view %495, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%497 = torch.aten.mul.Tensor %496, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%498 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%499 = torch.aten.view %result0_51, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%500 = torch.aten.addmm %9, %499, %498, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%501 = torch.aten.view %500, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%502 = torch.aten.view %501, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%503 = torch.aten.transpose.int %502, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%504 = torch.aten.clone %503, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%505 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%506 = torch.aten.view %result0_51, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%507 = torch.aten.addmm %9, %506, %505, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%508 = torch.aten.view %507, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%509 = torch.aten.view %508, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%510 = torch.aten.transpose.int %509, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%511 = torch.aten.clone %510, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%512 = torch.aten.view %497, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%513 = torch.aten.transpose.int %512, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%514 = torch.aten.clone %513, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%515 = torch.aten.view %514, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%516 = torch.aten.view %504, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%517 = torch.aten.view %511, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%518 = torch.aten.transpose.int %516, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%519 = torch.aten.bmm %515, %518 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%520 = torch.aten.view %519, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%521 = torch.aten.add.Tensor %520, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%522 = torch.aten.view %521, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%523 = torch.aten._softmax %522, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%524 = torch.aten.bmm %523, %517 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%525 = torch.aten.view %524, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%526 = torch.aten.transpose.int %525, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%527 = torch.aten.clone %526, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%528 = torch.aten._unsafe_view %527, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%529 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%530 = torch.aten.view %528, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%531 = torch.aten.addmm %9, %530, %529, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%532 = torch.aten.view %531, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%533 = torch.aten.add.Tensor %492, %532, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_54, %result1_55, %result2_56 = torch.aten.native_layer_norm %533, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%534 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%535 = torch.aten.view %result0_54, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%536 = torch.aten.addmm %7, %535, %534, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%537 = torch.aten.view %536, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%538 = torch.aten.gelu %537, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%539 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%540 = torch.aten.view %538, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%541 = torch.aten.addmm %9, %540, %539, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%542 = torch.aten.view %541, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%543 = torch.aten.add.Tensor %533, %542, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_57, %result1_58, %result2_59 = torch.aten.native_layer_norm %543, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%544 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%545 = torch.aten.view %result0_57, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%546 = torch.aten.addmm %9, %545, %544, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%547 = torch.aten.view %546, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%548 = torch.aten.mul.Tensor %547, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%549 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%550 = torch.aten.view %result0_57, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%551 = torch.aten.addmm %9, %550, %549, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%552 = torch.aten.view %551, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%553 = torch.aten.view %552, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%554 = torch.aten.transpose.int %553, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%555 = torch.aten.clone %554, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%556 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%557 = torch.aten.view %result0_57, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%558 = torch.aten.addmm %9, %557, %556, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%559 = torch.aten.view %558, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%560 = torch.aten.view %559, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%561 = torch.aten.transpose.int %560, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%562 = torch.aten.clone %561, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%563 = torch.aten.view %548, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%564 = torch.aten.transpose.int %563, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%565 = torch.aten.clone %564, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%566 = torch.aten.view %565, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%567 = torch.aten.view %555, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%568 = torch.aten.view %562, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%569 = torch.aten.transpose.int %567, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%570 = torch.aten.bmm %566, %569 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%571 = torch.aten.view %570, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%572 = torch.aten.add.Tensor %571, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%573 = torch.aten.view %572, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%574 = torch.aten._softmax %573, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%575 = torch.aten.bmm %574, %568 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%576 = torch.aten.view %575, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%577 = torch.aten.transpose.int %576, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%578 = torch.aten.clone %577, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%579 = torch.aten._unsafe_view %578, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%580 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%581 = torch.aten.view %579, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%582 = torch.aten.addmm %9, %581, %580, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%583 = torch.aten.view %582, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%584 = torch.aten.add.Tensor %543, %583, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_60, %result1_61, %result2_62 = torch.aten.native_layer_norm %584, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%585 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%586 = torch.aten.view %result0_60, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%587 = torch.aten.addmm %7, %586, %585, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%588 = torch.aten.view %587, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%589 = torch.aten.gelu %588, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%590 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%591 = torch.aten.view %589, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%592 = torch.aten.addmm %9, %591, %590, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%593 = torch.aten.view %592, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%594 = torch.aten.add.Tensor %584, %593, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_63, %result1_64, %result2_65 = torch.aten.native_layer_norm %594, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%595 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%596 = torch.aten.view %result0_63, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%597 = torch.aten.addmm %9, %596, %595, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%598 = torch.aten.view %597, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%599 = torch.aten.mul.Tensor %598, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%600 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%601 = torch.aten.view %result0_63, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%602 = torch.aten.addmm %9, %601, %600, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%603 = torch.aten.view %602, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%604 = torch.aten.view %603, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%605 = torch.aten.transpose.int %604, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%606 = torch.aten.clone %605, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%607 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%608 = torch.aten.view %result0_63, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%609 = torch.aten.addmm %9, %608, %607, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%610 = torch.aten.view %609, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%611 = torch.aten.view %610, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%612 = torch.aten.transpose.int %611, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%613 = torch.aten.clone %612, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%614 = torch.aten.view %599, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%615 = torch.aten.transpose.int %614, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%616 = torch.aten.clone %615, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%617 = torch.aten.view %616, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%618 = torch.aten.view %606, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%619 = torch.aten.view %613, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%620 = torch.aten.transpose.int %618, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%621 = torch.aten.bmm %617, %620 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%622 = torch.aten.view %621, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%623 = torch.aten.add.Tensor %622, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%624 = torch.aten.view %623, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%625 = torch.aten._softmax %624, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%626 = torch.aten.bmm %625, %619 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%627 = torch.aten.view %626, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%628 = torch.aten.transpose.int %627, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%629 = torch.aten.clone %628, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%630 = torch.aten._unsafe_view %629, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%631 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%632 = torch.aten.view %630, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%633 = torch.aten.addmm %9, %632, %631, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%634 = torch.aten.view %633, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%635 = torch.aten.add.Tensor %594, %634, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_66, %result1_67, %result2_68 = torch.aten.native_layer_norm %635, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%636 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%637 = torch.aten.view %result0_66, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%638 = torch.aten.addmm %7, %637, %636, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%639 = torch.aten.view %638, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%640 = torch.aten.gelu %639, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%641 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%642 = torch.aten.view %640, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%643 = torch.aten.addmm %9, %642, %641, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%644 = torch.aten.view %643, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%645 = torch.aten.add.Tensor %635, %644, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_69, %result1_70, %result2_71 = torch.aten.native_layer_norm %645, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%646 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%647 = torch.aten.view %result0_69, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%648 = torch.aten.addmm %9, %647, %646, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%649 = torch.aten.view %648, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%650 = torch.aten.mul.Tensor %649, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%651 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%652 = torch.aten.view %result0_69, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%653 = torch.aten.addmm %9, %652, %651, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%654 = torch.aten.view %653, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%655 = torch.aten.view %654, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%656 = torch.aten.transpose.int %655, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%657 = torch.aten.clone %656, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%658 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%659 = torch.aten.view %result0_69, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%660 = torch.aten.addmm %9, %659, %658, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%661 = torch.aten.view %660, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%662 = torch.aten.view %661, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%663 = torch.aten.transpose.int %662, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%664 = torch.aten.clone %663, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%665 = torch.aten.view %650, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%666 = torch.aten.transpose.int %665, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%667 = torch.aten.clone %666, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%668 = torch.aten.view %667, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%669 = torch.aten.view %657, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%670 = torch.aten.view %664, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%671 = torch.aten.transpose.int %669, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%672 = torch.aten.bmm %668, %671 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%673 = torch.aten.view %672, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%674 = torch.aten.add.Tensor %673, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%675 = torch.aten.view %674, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%676 = torch.aten._softmax %675, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%677 = torch.aten.bmm %676, %670 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%678 = torch.aten.view %677, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%679 = torch.aten.transpose.int %678, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%680 = torch.aten.clone %679, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%681 = torch.aten._unsafe_view %680, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%682 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%683 = torch.aten.view %681, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%684 = torch.aten.addmm %9, %683, %682, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%685 = torch.aten.view %684, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%686 = torch.aten.add.Tensor %645, %685, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_72, %result1_73, %result2_74 = torch.aten.native_layer_norm %686, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%687 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%688 = torch.aten.view %result0_72, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%689 = torch.aten.addmm %7, %688, %687, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%690 = torch.aten.view %689, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%691 = torch.aten.gelu %690, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%692 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%693 = torch.aten.view %691, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%694 = torch.aten.addmm %9, %693, %692, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%695 = torch.aten.view %694, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%696 = torch.aten.add.Tensor %686, %695, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_75, %result1_76, %result2_77 = torch.aten.native_layer_norm %696, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%697 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%698 = torch.aten.view %result0_75, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%699 = torch.aten.addmm %9, %698, %697, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%700 = torch.aten.view %699, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%701 = torch.aten.mul.Tensor %700, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%702 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%703 = torch.aten.view %result0_75, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%704 = torch.aten.addmm %9, %703, %702, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%705 = torch.aten.view %704, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%706 = torch.aten.view %705, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%707 = torch.aten.transpose.int %706, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%708 = torch.aten.clone %707, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%709 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%710 = torch.aten.view %result0_75, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%711 = torch.aten.addmm %9, %710, %709, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%712 = torch.aten.view %711, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%713 = torch.aten.view %712, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%714 = torch.aten.transpose.int %713, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%715 = torch.aten.clone %714, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%716 = torch.aten.view %701, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%717 = torch.aten.transpose.int %716, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%718 = torch.aten.clone %717, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%719 = torch.aten.view %718, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%720 = torch.aten.view %708, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%721 = torch.aten.view %715, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%722 = torch.aten.transpose.int %720, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%723 = torch.aten.bmm %719, %722 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%724 = torch.aten.view %723, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%725 = torch.aten.add.Tensor %724, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%726 = torch.aten.view %725, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%727 = torch.aten._softmax %726, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%728 = torch.aten.bmm %727, %721 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%729 = torch.aten.view %728, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%730 = torch.aten.transpose.int %729, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%731 = torch.aten.clone %730, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%732 = torch.aten._unsafe_view %731, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%733 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%734 = torch.aten.view %732, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%735 = torch.aten.addmm %9, %734, %733, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%736 = torch.aten.view %735, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%737 = torch.aten.add.Tensor %696, %736, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_78, %result1_79, %result2_80 = torch.aten.native_layer_norm %737, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%738 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%739 = torch.aten.view %result0_78, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%740 = torch.aten.addmm %7, %739, %738, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%741 = torch.aten.view %740, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%742 = torch.aten.gelu %741, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%743 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%744 = torch.aten.view %742, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%745 = torch.aten.addmm %9, %744, %743, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%746 = torch.aten.view %745, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%747 = torch.aten.add.Tensor %737, %746, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_81, %result1_82, %result2_83 = torch.aten.native_layer_norm %747, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%748 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%749 = torch.aten.view %result0_81, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%750 = torch.aten.addmm %9, %749, %748, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%751 = torch.aten.view %750, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%752 = torch.aten.mul.Tensor %751, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%753 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%754 = torch.aten.view %result0_81, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%755 = torch.aten.addmm %9, %754, %753, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%756 = torch.aten.view %755, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%757 = torch.aten.view %756, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%758 = torch.aten.transpose.int %757, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%759 = torch.aten.clone %758, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%760 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%761 = torch.aten.view %result0_81, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%762 = torch.aten.addmm %9, %761, %760, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%763 = torch.aten.view %762, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%764 = torch.aten.view %763, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%765 = torch.aten.transpose.int %764, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%766 = torch.aten.clone %765, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%767 = torch.aten.view %752, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%768 = torch.aten.transpose.int %767, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%769 = torch.aten.clone %768, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%770 = torch.aten.view %769, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%771 = torch.aten.view %759, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%772 = torch.aten.view %766, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%773 = torch.aten.transpose.int %771, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%774 = torch.aten.bmm %770, %773 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%775 = torch.aten.view %774, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%776 = torch.aten.add.Tensor %775, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%777 = torch.aten.view %776, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%778 = torch.aten._softmax %777, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%779 = torch.aten.bmm %778, %772 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%780 = torch.aten.view %779, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%781 = torch.aten.transpose.int %780, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%782 = torch.aten.clone %781, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%783 = torch.aten._unsafe_view %782, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%784 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%785 = torch.aten.view %783, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%786 = torch.aten.addmm %9, %785, %784, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%787 = torch.aten.view %786, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%788 = torch.aten.add.Tensor %747, %787, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_84, %result1_85, %result2_86 = torch.aten.native_layer_norm %788, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%789 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%790 = torch.aten.view %result0_84, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%791 = torch.aten.addmm %7, %790, %789, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%792 = torch.aten.view %791, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%793 = torch.aten.gelu %792, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%794 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%795 = torch.aten.view %793, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%796 = torch.aten.addmm %9, %795, %794, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%797 = torch.aten.view %796, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%798 = torch.aten.add.Tensor %788, %797, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_87, %result1_88, %result2_89 = torch.aten.native_layer_norm %798, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%799 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%800 = torch.aten.view %result0_87, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%801 = torch.aten.addmm %9, %800, %799, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%802 = torch.aten.view %801, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%803 = torch.aten.mul.Tensor %802, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%804 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%805 = torch.aten.view %result0_87, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%806 = torch.aten.addmm %9, %805, %804, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%807 = torch.aten.view %806, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%808 = torch.aten.view %807, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%809 = torch.aten.transpose.int %808, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%810 = torch.aten.clone %809, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%811 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%812 = torch.aten.view %result0_87, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%813 = torch.aten.addmm %9, %812, %811, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%814 = torch.aten.view %813, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%815 = torch.aten.view %814, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%816 = torch.aten.transpose.int %815, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%817 = torch.aten.clone %816, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%818 = torch.aten.view %803, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%819 = torch.aten.transpose.int %818, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%820 = torch.aten.clone %819, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%821 = torch.aten.view %820, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%822 = torch.aten.view %810, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%823 = torch.aten.view %817, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%824 = torch.aten.transpose.int %822, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%825 = torch.aten.bmm %821, %824 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%826 = torch.aten.view %825, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%827 = torch.aten.add.Tensor %826, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%828 = torch.aten.view %827, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%829 = torch.aten._softmax %828, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%830 = torch.aten.bmm %829, %823 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%831 = torch.aten.view %830, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%832 = torch.aten.transpose.int %831, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%833 = torch.aten.clone %832, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%834 = torch.aten._unsafe_view %833, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%835 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%836 = torch.aten.view %834, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%837 = torch.aten.addmm %9, %836, %835, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%838 = torch.aten.view %837, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%839 = torch.aten.add.Tensor %798, %838, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_90, %result1_91, %result2_92 = torch.aten.native_layer_norm %839, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%840 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%841 = torch.aten.view %result0_90, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%842 = torch.aten.addmm %7, %841, %840, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%843 = torch.aten.view %842, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%844 = torch.aten.gelu %843, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%845 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%846 = torch.aten.view %844, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%847 = torch.aten.addmm %9, %846, %845, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%848 = torch.aten.view %847, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%849 = torch.aten.add.Tensor %839, %848, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_93, %result1_94, %result2_95 = torch.aten.native_layer_norm %849, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%850 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%851 = torch.aten.view %result0_93, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%852 = torch.aten.addmm %9, %851, %850, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%853 = torch.aten.view %852, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%854 = torch.aten.mul.Tensor %853, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%855 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%856 = torch.aten.view %result0_93, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%857 = torch.aten.addmm %9, %856, %855, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%858 = torch.aten.view %857, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%859 = torch.aten.view %858, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%860 = torch.aten.transpose.int %859, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%861 = torch.aten.clone %860, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%862 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%863 = torch.aten.view %result0_93, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%864 = torch.aten.addmm %9, %863, %862, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%865 = torch.aten.view %864, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%866 = torch.aten.view %865, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%867 = torch.aten.transpose.int %866, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%868 = torch.aten.clone %867, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%869 = torch.aten.view %854, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%870 = torch.aten.transpose.int %869, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%871 = torch.aten.clone %870, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%872 = torch.aten.view %871, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%873 = torch.aten.view %861, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%874 = torch.aten.view %868, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%875 = torch.aten.transpose.int %873, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%876 = torch.aten.bmm %872, %875 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%877 = torch.aten.view %876, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%878 = torch.aten.add.Tensor %877, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%879 = torch.aten.view %878, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%880 = torch.aten._softmax %879, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%881 = torch.aten.bmm %880, %874 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%882 = torch.aten.view %881, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%883 = torch.aten.transpose.int %882, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%884 = torch.aten.clone %883, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%885 = torch.aten._unsafe_view %884, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%886 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%887 = torch.aten.view %885, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%888 = torch.aten.addmm %9, %887, %886, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%889 = torch.aten.view %888, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%890 = torch.aten.add.Tensor %849, %889, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_96, %result1_97, %result2_98 = torch.aten.native_layer_norm %890, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%891 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%892 = torch.aten.view %result0_96, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%893 = torch.aten.addmm %7, %892, %891, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%894 = torch.aten.view %893, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%895 = torch.aten.gelu %894, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%896 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%897 = torch.aten.view %895, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%898 = torch.aten.addmm %9, %897, %896, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%899 = torch.aten.view %898, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%900 = torch.aten.add.Tensor %890, %899, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_99, %result1_100, %result2_101 = torch.aten.native_layer_norm %900, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%901 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%902 = torch.aten.view %result0_99, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%903 = torch.aten.addmm %9, %902, %901, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%904 = torch.aten.view %903, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%905 = torch.aten.mul.Tensor %904, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%906 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%907 = torch.aten.view %result0_99, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%908 = torch.aten.addmm %9, %907, %906, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%909 = torch.aten.view %908, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%910 = torch.aten.view %909, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%911 = torch.aten.transpose.int %910, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%912 = torch.aten.clone %911, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%913 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%914 = torch.aten.view %result0_99, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%915 = torch.aten.addmm %9, %914, %913, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%916 = torch.aten.view %915, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%917 = torch.aten.view %916, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%918 = torch.aten.transpose.int %917, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%919 = torch.aten.clone %918, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%920 = torch.aten.view %905, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%921 = torch.aten.transpose.int %920, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%922 = torch.aten.clone %921, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%923 = torch.aten.view %922, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%924 = torch.aten.view %912, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%925 = torch.aten.view %919, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%926 = torch.aten.transpose.int %924, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%927 = torch.aten.bmm %923, %926 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%928 = torch.aten.view %927, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%929 = torch.aten.add.Tensor %928, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%930 = torch.aten.view %929, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%931 = torch.aten._softmax %930, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%932 = torch.aten.bmm %931, %925 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%933 = torch.aten.view %932, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%934 = torch.aten.transpose.int %933, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%935 = torch.aten.clone %934, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%936 = torch.aten._unsafe_view %935, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%937 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%938 = torch.aten.view %936, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%939 = torch.aten.addmm %9, %938, %937, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%940 = torch.aten.view %939, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%941 = torch.aten.add.Tensor %900, %940, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_102, %result1_103, %result2_104 = torch.aten.native_layer_norm %941, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%942 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%943 = torch.aten.view %result0_102, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%944 = torch.aten.addmm %7, %943, %942, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%945 = torch.aten.view %944, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%946 = torch.aten.gelu %945, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%947 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%948 = torch.aten.view %946, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%949 = torch.aten.addmm %9, %948, %947, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%950 = torch.aten.view %949, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%951 = torch.aten.add.Tensor %941, %950, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_105, %result1_106, %result2_107 = torch.aten.native_layer_norm %951, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%952 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%953 = torch.aten.view %result0_105, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%954 = torch.aten.addmm %9, %953, %952, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%955 = torch.aten.view %954, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%956 = torch.aten.mul.Tensor %955, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%957 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%958 = torch.aten.view %result0_105, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%959 = torch.aten.addmm %9, %958, %957, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%960 = torch.aten.view %959, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%961 = torch.aten.view %960, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%962 = torch.aten.transpose.int %961, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%963 = torch.aten.clone %962, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%964 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%965 = torch.aten.view %result0_105, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%966 = torch.aten.addmm %9, %965, %964, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%967 = torch.aten.view %966, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%968 = torch.aten.view %967, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%969 = torch.aten.transpose.int %968, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%970 = torch.aten.clone %969, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%971 = torch.aten.view %956, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%972 = torch.aten.transpose.int %971, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%973 = torch.aten.clone %972, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%974 = torch.aten.view %973, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%975 = torch.aten.view %963, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%976 = torch.aten.view %970, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%977 = torch.aten.transpose.int %975, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%978 = torch.aten.bmm %974, %977 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%979 = torch.aten.view %978, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%980 = torch.aten.add.Tensor %979, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%981 = torch.aten.view %980, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%982 = torch.aten._softmax %981, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%983 = torch.aten.bmm %982, %976 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%984 = torch.aten.view %983, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%985 = torch.aten.transpose.int %984, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%986 = torch.aten.clone %985, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%987 = torch.aten._unsafe_view %986, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%988 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%989 = torch.aten.view %987, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%990 = torch.aten.addmm %9, %989, %988, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%991 = torch.aten.view %990, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%992 = torch.aten.add.Tensor %951, %991, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_108, %result1_109, %result2_110 = torch.aten.native_layer_norm %992, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%993 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%994 = torch.aten.view %result0_108, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%995 = torch.aten.addmm %7, %994, %993, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%996 = torch.aten.view %995, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%997 = torch.aten.gelu %996, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%998 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%999 = torch.aten.view %997, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%1000 = torch.aten.addmm %9, %999, %998, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1001 = torch.aten.view %1000, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1002 = torch.aten.add.Tensor %992, %1001, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_111, %result1_112, %result2_113 = torch.aten.native_layer_norm %1002, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1003 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1004 = torch.aten.view %result0_111, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1005 = torch.aten.addmm %9, %1004, %1003, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1006 = torch.aten.view %1005, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1007 = torch.aten.mul.Tensor %1006, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%1008 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1009 = torch.aten.view %result0_111, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1010 = torch.aten.addmm %9, %1009, %1008, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1011 = torch.aten.view %1010, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1012 = torch.aten.view %1011, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1013 = torch.aten.transpose.int %1012, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1014 = torch.aten.clone %1013, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1015 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1016 = torch.aten.view %result0_111, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1017 = torch.aten.addmm %9, %1016, %1015, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1018 = torch.aten.view %1017, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1019 = torch.aten.view %1018, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1020 = torch.aten.transpose.int %1019, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1021 = torch.aten.clone %1020, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1022 = torch.aten.view %1007, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1023 = torch.aten.transpose.int %1022, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1024 = torch.aten.clone %1023, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1025 = torch.aten.view %1024, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1026 = torch.aten.view %1014, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1027 = torch.aten.view %1021, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1028 = torch.aten.transpose.int %1026, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%1029 = torch.aten.bmm %1025, %1028 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%1030 = torch.aten.view %1029, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%1031 = torch.aten.add.Tensor %1030, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%1032 = torch.aten.view %1031, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%1033 = torch.aten._softmax %1032, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%1034 = torch.aten.bmm %1033, %1027 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%1035 = torch.aten.view %1034, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%1036 = torch.aten.transpose.int %1035, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1037 = torch.aten.clone %1036, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1038 = torch.aten._unsafe_view %1037, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1039 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1040 = torch.aten.view %1038, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1041 = torch.aten.addmm %9, %1040, %1039, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1042 = torch.aten.view %1041, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1043 = torch.aten.add.Tensor %1002, %1042, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_114, %result1_115, %result2_116 = torch.aten.native_layer_norm %1043, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1044 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%1045 = torch.aten.view %result0_114, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1046 = torch.aten.addmm %7, %1045, %1044, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%1047 = torch.aten.view %1046, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%1048 = torch.aten.gelu %1047, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%1049 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%1050 = torch.aten.view %1048, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%1051 = torch.aten.addmm %9, %1050, %1049, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1052 = torch.aten.view %1051, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1053 = torch.aten.add.Tensor %1043, %1052, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_117, %result1_118, %result2_119 = torch.aten.native_layer_norm %1053, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1054 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1055 = torch.aten.view %result0_117, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1056 = torch.aten.addmm %9, %1055, %1054, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1057 = torch.aten.view %1056, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1058 = torch.aten.mul.Tensor %1057, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%1059 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1060 = torch.aten.view %result0_117, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1061 = torch.aten.addmm %9, %1060, %1059, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1062 = torch.aten.view %1061, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1063 = torch.aten.view %1062, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1064 = torch.aten.transpose.int %1063, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1065 = torch.aten.clone %1064, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1066 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1067 = torch.aten.view %result0_117, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1068 = torch.aten.addmm %9, %1067, %1066, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1069 = torch.aten.view %1068, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1070 = torch.aten.view %1069, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1071 = torch.aten.transpose.int %1070, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1072 = torch.aten.clone %1071, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1073 = torch.aten.view %1058, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1074 = torch.aten.transpose.int %1073, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1075 = torch.aten.clone %1074, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1076 = torch.aten.view %1075, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1077 = torch.aten.view %1065, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1078 = torch.aten.view %1072, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1079 = torch.aten.transpose.int %1077, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%1080 = torch.aten.bmm %1076, %1079 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%1081 = torch.aten.view %1080, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%1082 = torch.aten.add.Tensor %1081, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%1083 = torch.aten.view %1082, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%1084 = torch.aten._softmax %1083, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%1085 = torch.aten.bmm %1084, %1078 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%1086 = torch.aten.view %1085, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%1087 = torch.aten.transpose.int %1086, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1088 = torch.aten.clone %1087, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1089 = torch.aten._unsafe_view %1088, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1090 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1091 = torch.aten.view %1089, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1092 = torch.aten.addmm %9, %1091, %1090, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1093 = torch.aten.view %1092, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1094 = torch.aten.add.Tensor %1053, %1093, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_120, %result1_121, %result2_122 = torch.aten.native_layer_norm %1094, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1095 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%1096 = torch.aten.view %result0_120, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1097 = torch.aten.addmm %7, %1096, %1095, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%1098 = torch.aten.view %1097, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%1099 = torch.aten.gelu %1098, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%1100 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%1101 = torch.aten.view %1099, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%1102 = torch.aten.addmm %9, %1101, %1100, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1103 = torch.aten.view %1102, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1104 = torch.aten.add.Tensor %1094, %1103, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_123, %result1_124, %result2_125 = torch.aten.native_layer_norm %1104, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1105 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1106 = torch.aten.view %result0_123, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1107 = torch.aten.addmm %9, %1106, %1105, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1108 = torch.aten.view %1107, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1109 = torch.aten.mul.Tensor %1108, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%1110 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1111 = torch.aten.view %result0_123, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1112 = torch.aten.addmm %9, %1111, %1110, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1113 = torch.aten.view %1112, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1114 = torch.aten.view %1113, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1115 = torch.aten.transpose.int %1114, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1116 = torch.aten.clone %1115, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1117 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1118 = torch.aten.view %result0_123, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1119 = torch.aten.addmm %9, %1118, %1117, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1120 = torch.aten.view %1119, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1121 = torch.aten.view %1120, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1122 = torch.aten.transpose.int %1121, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1123 = torch.aten.clone %1122, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1124 = torch.aten.view %1109, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1125 = torch.aten.transpose.int %1124, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1126 = torch.aten.clone %1125, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1127 = torch.aten.view %1126, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1128 = torch.aten.view %1116, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1129 = torch.aten.view %1123, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1130 = torch.aten.transpose.int %1128, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%1131 = torch.aten.bmm %1127, %1130 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%1132 = torch.aten.view %1131, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%1133 = torch.aten.add.Tensor %1132, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%1134 = torch.aten.view %1133, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%1135 = torch.aten._softmax %1134, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%1136 = torch.aten.bmm %1135, %1129 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%1137 = torch.aten.view %1136, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%1138 = torch.aten.transpose.int %1137, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1139 = torch.aten.clone %1138, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1140 = torch.aten._unsafe_view %1139, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1141 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1142 = torch.aten.view %1140, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1143 = torch.aten.addmm %9, %1142, %1141, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1144 = torch.aten.view %1143, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1145 = torch.aten.add.Tensor %1104, %1144, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_126, %result1_127, %result2_128 = torch.aten.native_layer_norm %1145, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1146 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%1147 = torch.aten.view %result0_126, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1148 = torch.aten.addmm %7, %1147, %1146, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%1149 = torch.aten.view %1148, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%1150 = torch.aten.gelu %1149, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%1151 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%1152 = torch.aten.view %1150, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%1153 = torch.aten.addmm %9, %1152, %1151, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1154 = torch.aten.view %1153, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1155 = torch.aten.add.Tensor %1145, %1154, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_129, %result1_130, %result2_131 = torch.aten.native_layer_norm %1155, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1156 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1157 = torch.aten.view %result0_129, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1158 = torch.aten.addmm %9, %1157, %1156, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1159 = torch.aten.view %1158, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1160 = torch.aten.mul.Tensor %1159, %0 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[],f64> -> !torch.vtensor<[1,77,1024],f16> | |
%1161 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1162 = torch.aten.view %result0_129, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1163 = torch.aten.addmm %9, %1162, %1161, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1164 = torch.aten.view %1163, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1165 = torch.aten.view %1164, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1166 = torch.aten.transpose.int %1165, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1167 = torch.aten.clone %1166, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1168 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1169 = torch.aten.view %result0_129, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1170 = torch.aten.addmm %9, %1169, %1168, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1171 = torch.aten.view %1170, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1172 = torch.aten.view %1171, %35 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1173 = torch.aten.transpose.int %1172, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1174 = torch.aten.clone %1173, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1175 = torch.aten.view %1160, %46 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,16,64],f16> | |
%1176 = torch.aten.transpose.int %1175, %int1, %int2 : !torch.vtensor<[1,77,16,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1177 = torch.aten.clone %1176, %int0 : !torch.vtensor<[1,16,77,64],f16>, !torch.int -> !torch.vtensor<[1,16,77,64],f16> | |
%1178 = torch.aten.view %1177, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1179 = torch.aten.view %1167, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1180 = torch.aten.view %1174, %50 : !torch.vtensor<[1,16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[16,77,64],f16> | |
%1181 = torch.aten.transpose.int %1179, %int1, %int2 : !torch.vtensor<[16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[16,64,77],f16> | |
%1182 = torch.aten.bmm %1178, %1181 : !torch.vtensor<[16,77,64],f16>, !torch.vtensor<[16,64,77],f16> -> !torch.vtensor<[16,77,77],f16> | |
%1183 = torch.aten.view %1182, %56 : !torch.vtensor<[16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,77],f16> | |
%1184 = torch.aten.add.Tensor %1183, %22, %int1 : !torch.vtensor<[1,16,77,77],f16>, !torch.vtensor<[1,1,77,77],f16>, !torch.int -> !torch.vtensor<[1,16,77,77],f16> | |
%1185 = torch.aten.view %1184, %59 : !torch.vtensor<[1,16,77,77],f16>, !torch.list<int> -> !torch.vtensor<[16,77,77],f16> | |
%1186 = torch.aten._softmax %1185, %int-1, %false : !torch.vtensor<[16,77,77],f16>, !torch.int, !torch.bool -> !torch.vtensor<[16,77,77],f16> | |
%1187 = torch.aten.bmm %1186, %1180 : !torch.vtensor<[16,77,77],f16>, !torch.vtensor<[16,77,64],f16> -> !torch.vtensor<[16,77,64],f16> | |
%1188 = torch.aten.view %1187, %63 : !torch.vtensor<[16,77,64],f16>, !torch.list<int> -> !torch.vtensor<[1,16,77,64],f16> | |
%1189 = torch.aten.transpose.int %1188, %int1, %int2 : !torch.vtensor<[1,16,77,64],f16>, !torch.int, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1190 = torch.aten.clone %1189, %int0 : !torch.vtensor<[1,77,16,64],f16>, !torch.int -> !torch.vtensor<[1,77,16,64],f16> | |
%1191 = torch.aten._unsafe_view %1190, %28 : !torch.vtensor<[1,77,16,64],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1192 = torch.aten.t %5 : !torch.vtensor<[1024,1024],f16> -> !torch.vtensor<[1024,1024],f16> | |
%1193 = torch.aten.view %1191, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1194 = torch.aten.addmm %9, %1193, %1192, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1195 = torch.aten.view %1194, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1196 = torch.aten.add.Tensor %1155, %1195, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_132, %result1_133, %result2_134 = torch.aten.native_layer_norm %1196, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
%1197 = torch.aten.t %6 : !torch.vtensor<[4096,1024],f16> -> !torch.vtensor<[1024,4096],f16> | |
%1198 = torch.aten.view %result0_132, %25 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int> -> !torch.vtensor<[77,1024],f16> | |
%1199 = torch.aten.addmm %7, %1198, %1197, %int1, %int1 : !torch.vtensor<[4096],f16>, !torch.vtensor<[77,1024],f16>, !torch.vtensor<[1024,4096],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,4096],f16> | |
%1200 = torch.aten.view %1199, %76 : !torch.vtensor<[77,4096],f16>, !torch.list<int> -> !torch.vtensor<[1,77,4096],f16> | |
%1201 = torch.aten.gelu %1200, %str : !torch.vtensor<[1,77,4096],f16>, !torch.str -> !torch.vtensor<[1,77,4096],f16> | |
%1202 = torch.aten.t %8 : !torch.vtensor<[1024,4096],f16> -> !torch.vtensor<[4096,1024],f16> | |
%1203 = torch.aten.view %1201, %80 : !torch.vtensor<[1,77,4096],f16>, !torch.list<int> -> !torch.vtensor<[77,4096],f16> | |
%1204 = torch.aten.addmm %9, %1203, %1202, %int1, %int1 : !torch.vtensor<[1024],f16>, !torch.vtensor<[77,4096],f16>, !torch.vtensor<[4096,1024],f16>, !torch.int, !torch.int -> !torch.vtensor<[77,1024],f16> | |
%1205 = torch.aten.view %1204, %28 : !torch.vtensor<[77,1024],f16>, !torch.list<int> -> !torch.vtensor<[1,77,1024],f16> | |
%1206 = torch.aten.add.Tensor %1196, %1205, %int1 : !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1024],f16>, !torch.int -> !torch.vtensor<[1,77,1024],f16> | |
%result0_135, %result1_136, %result2_137 = torch.aten.native_layer_norm %1206, %23, %9, %9, %float1.000000e-05 : !torch.vtensor<[1,77,1024],f16>, !torch.list<int>, !torch.vtensor<[1024],f16>, !torch.vtensor<[1024],f16>, !torch.float -> !torch.vtensor<[1,77,1024],f16>, !torch.vtensor<[1,77,1],f32>, !torch.vtensor<[1,77,1],f32> | |
return %result0_135 : !torch.vtensor<[1,77,1024],f16> | |
} | |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment