Last active
June 28, 2022 22:30
-
-
Save davidberard98/56d3cc5afe750e04dab7399794733119 to your computer and use it in GitHub Desktop.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
====== resnext50_32x4d_forward_0 ====== | |
Generating testing data... | |
...Aten2aten called ... | |
['getattr', '_operator.getitem', 'torch.ops.aten.abs.default', 'torch.ops.aten.abs.out', 'torch.ops.aten.acos.default', 'torch.ops.aten.acos.out', 'torch.ops.aten.acos.int', 'torch.ops.aten.acos.float', 'torch.ops.aten.acos.complex', 'torch.ops.aten.acos.Scalar', 'torch.ops.aten.acosh.default', 'torch.ops.aten.acosh.out', 'torch.ops.aten.acosh.int', 'torch.ops.aten.acosh.float', 'torch.ops.aten.acosh.complex', 'torch.ops.aten.acosh.Scalar', 'torch.ops.aten.asin.default', 'torch.ops.aten.asin.out', 'torch.ops.aten.asin.int', 'torch.ops.aten.asin.float', 'torch.ops.aten.asin.complex', 'torch.ops.aten.asin.Scalar', 'torch.ops.aten.atan.default', 'torch.ops.aten.atan.out', 'torch.ops.aten.atan.int', 'torch.ops.aten.atan.float', 'torch.ops.aten.atan.complex', 'torch.ops.aten.atan.Scalar', 'torch.ops.aten.bitwise_not.default', 'torch.ops.aten.bitwise_not.out', 'torch.ops.aten.ceil.default', 'torch.ops.aten.ceil.out', 'torch.ops.aten.ceil.int', 'torch.ops.aten.ceil.float', 'torch.ops.aten.ceil.Scalar', 'torch.ops.aten.cos.default', 'torch.ops.aten.cos.out', 'torch.ops.aten.cos.int', 'torch.ops.aten.cos.float', 'torch.ops.aten.cos.complex', 'torch.ops.aten.cos.Scalar', 'torch.ops.aten.cosh.default', 'torch.ops.aten.cosh.out', 'torch.ops.aten.cosh.int', 'torch.ops.aten.cosh.float', 'torch.ops.aten.cosh.complex', 'torch.ops.aten.cosh.Scalar', 'torch.ops.aten.digamma.default', 'torch.ops.aten.digamma.out', 'torch.ops.aten.erf.default', 'torch.ops.aten.erf.out', 'torch.ops.aten.erf.int', 'torch.ops.aten.erf.float', 'torch.ops.aten.erf.Scalar', 'torch.ops.aten.erfinv.default', 'torch.ops.aten.erfinv.out', 'torch.ops.aten.erfc.default', 'torch.ops.aten.erfc.out', 'torch.ops.aten.erfc.int', 'torch.ops.aten.erfc.float', 'torch.ops.aten.erfc.Scalar', 'torch.ops.aten.exp.default', 'torch.ops.aten.exp.out', 'torch.ops.aten.exp.int', 'torch.ops.aten.exp.float', 'torch.ops.aten.exp.complex', 'torch.ops.aten.exp.Scalar', 'torch.ops.aten.expm1.default', 'torch.ops.aten.expm1.out', 'torch.ops.aten.expm1.int', 'torch.ops.aten.expm1.float', 'torch.ops.aten.expm1.Scalar', 'torch.ops.aten.exp2.default', 'torch.ops.aten.exp2.out', 'torch.ops.aten.floor.default', 'torch.ops.aten.floor.out', 'torch.ops.aten.floor.int', 'torch.ops.aten.floor.float', 'torch.ops.aten.floor.Scalar', 'torch.ops.aten.frac.default', 'torch.ops.aten.frac.out', 'torch.ops.aten.isinf.default', 'torch.ops.aten.isinf.float', 'torch.ops.aten.isinf.complex', 'torch.ops.aten.isnan.default', 'torch.ops.aten.isnan.float', 'torch.ops.aten.isnan.complex', 'torch.ops.aten.special_i0.default', 'torch.ops.aten.special_i0.out', 'torch.ops.aten.lgamma.default', 'torch.ops.aten.lgamma.out', 'torch.ops.aten.lgamma.int', 'torch.ops.aten.lgamma.float', 'torch.ops.aten.lgamma.Scalar', 'torch.ops.aten.log.default', 'torch.ops.aten.log.out', 'torch.ops.aten.log.int', 'torch.ops.aten.log.float', 'torch.ops.aten.log.complex', 'torch.ops.aten.log.Scalar', 'torch.ops.aten.log.int_int', 'torch.ops.aten.log.float_float', 'torch.ops.aten.log.complex_complex', 'torch.ops.aten.log.int_float', 'torch.ops.aten.log.float_int', 'torch.ops.aten.log.int_complex', 'torch.ops.aten.log.complex_int', 'torch.ops.aten.log.float_complex', 'torch.ops.aten.log.complex_float', 'torch.ops.aten.log.Scalar_Scalar', 'torch.ops.aten.log1p.default', 'torch.ops.aten.log1p.out', 'torch.ops.aten.log1p.int', 'torch.ops.aten.log1p.float', 'torch.ops.aten.log1p.Scalar', 'torch.ops.aten.log2.default', 'torch.ops.aten.log2.out', 'torch.ops.aten.log10.default', 'torch.ops.aten.log10.out', 'torch.ops.aten.log10.int', 'torch.ops.aten.log10.float', 'torch.ops.aten.log10.complex', 'torch.ops.aten.log10.Scalar', 'torch.ops.aten.nan_to_num.out', 'torch.ops.aten.nan_to_num.default', 'torch.ops.aten.neg.default', 'torch.ops.aten.neg.out', 'torch.ops.aten.neg.int', 'torch.ops.aten.neg.float', 'torch.ops.aten.neg.complex', 'torch.ops.aten.neg.Scalar', 'torch.ops.aten.reciprocal.default', 'torch.ops.aten.reciprocal.out', 'torch.ops.aten.rsqrt.default', 'torch.ops.aten.rsqrt.out', 'torch.ops.aten.sigmoid.default', 'torch.ops.aten.sigmoid.out', 'torch.ops.aten.sign.default', 'torch.ops.aten.sign.out', 'torch.ops.aten.signbit.default', 'torch.ops.aten.signbit.out', 'torch.ops.aten.sin.default', 'torch.ops.aten.sin.out', 'torch.ops.aten.sin.int', 'torch.ops.aten.sin.float', 'torch.ops.aten.sin.complex', 'torch.ops.aten.sin.Scalar', 'torch.ops.aten.sinh.default', 'torch.ops.aten.sinh.out', 'torch.ops.aten.sinh.int', 'torch.ops.aten.sinh.float', 'torch.ops.aten.sinh.complex', 'torch.ops.aten.sinh.Scalar', 'torch.ops.aten.sqrt.default', 'torch.ops.aten.sqrt.out', 'torch.ops.aten.sqrt.int', 'torch.ops.aten.sqrt.float', 'torch.ops.aten.sqrt.complex', 'torch.ops.aten.sqrt.Scalar', 'torch.ops.aten.tan.default', 'torch.ops.aten.tan.out', 'torch.ops.aten.tan.int', 'torch.ops.aten.tan.float', 'torch.ops.aten.tan.complex', 'torch.ops.aten.tan.Scalar', 'torch.ops.aten.tanh.default', 'torch.ops.aten.tanh.out', 'torch.ops.aten.tanh.int', 'torch.ops.aten.tanh.float', 'torch.ops.aten.tanh.complex', 'torch.ops.aten.tanh.Scalar', 'torch.ops.aten.add.Tensor', 'torch.ops.aten.add.Scalar', 'torch.ops.aten.add.out', 'torch.ops.aten.add.Scalar_out', 'torch.ops.aten.add.t', 'torch.ops.aten.add.str', 'torch.ops.aten.add.int', 'torch.ops.aten.add.complex', 'torch.ops.aten.add.float', 'torch.ops.aten.add.int_complex', 'torch.ops.aten.add.complex_int', 'torch.ops.aten.add.float_complex', 'torch.ops.aten.add.complex_float', 'torch.ops.aten.add.int_float', 'torch.ops.aten.add.float_int', 'torch.ops.aten.add.default', 'torch.ops.aten.atan2.default', 'torch.ops.aten.atan2.out', 'torch.ops.aten.atan2.int', 'torch.ops.aten.atan2.float', 'torch.ops.aten.atan2.int_float', 'torch.ops.aten.atan2.float_int', 'torch.ops.aten.atan2.Scalar_Scalar', 'torch.ops.aten.bitwise_and.Tensor', 'torch.ops.aten.bitwise_and.Tensor_out', 'torch.ops.aten.bitwise_and.Scalar', 'torch.ops.aten.bitwise_and.Scalar_out', 'torch.ops.aten.bitwise_and.Scalar_Tensor', 'torch.ops.aten.bitwise_left_shift.Tensor', 'torch.ops.aten.bitwise_left_shift.Tensor_out', 'torch.ops.aten.bitwise_left_shift.Tensor_Scalar', 'torch.ops.aten.bitwise_left_shift.Tensor_Scalar_out', 'torch.ops.aten.bitwise_left_shift.Scalar_Tensor', 'torch.ops.aten.bitwise_or.Tensor', 'torch.ops.aten.bitwise_or.Tensor_out', 'torch.ops.aten.bitwise_or.Scalar_out', 'torch.ops.aten.bitwise_or.Scalar_Tensor', 'torch.ops.aten.bitwise_or.Scalar', 'torch.ops.aten.bitwise_right_shift.Tensor', 'torch.ops.aten.bitwise_right_shift.Tensor_out', 'torch.ops.aten.bitwise_right_shift.Tensor_Scalar', 'torch.ops.aten.bitwise_right_shift.Tensor_Scalar_out', 'torch.ops.aten.bitwise_right_shift.Scalar_Tensor', 'torch.ops.aten.bitwise_xor.Tensor', 'torch.ops.aten.bitwise_xor.Tensor_out', 'torch.ops.aten.bitwise_xor.Scalar_out', 'torch.ops.aten.bitwise_xor.Scalar_Tensor', 'torch.ops.aten.bitwise_xor.Scalar', 'torch.ops.aten.eq.Tensor', 'torch.ops.aten.eq.Scalar', 'torch.ops.aten.eq.Scalar_out', 'torch.ops.aten.eq.Tensor_out', 'torch.ops.aten.eq.int_list', 'torch.ops.aten.eq.device', 'torch.ops.aten.eq.bool', 'torch.ops.aten.eq.enum', 'torch.ops.aten.eq.int', 'torch.ops.aten.eq.complex', 'torch.ops.aten.eq.float', 'torch.ops.aten.eq.int_float', 'torch.ops.aten.eq.float_int', 'torch.ops.aten.eq.float_complex', 'torch.ops.aten.eq.complex_float', 'torch.ops.aten.eq.default', 'torch.ops.aten.eq.str', 'torch.ops.aten.eq.float_list', 'torch.ops.aten.eq.Tensor_list', 'torch.ops.aten.eq.bool_list', 'torch.ops.aten.eq.str_list', 'torch.ops.aten.fmax.default', 'torch.ops.aten.fmax.out', 'torch.ops.aten.fmin.default', 'torch.ops.aten.fmin.out', 'torch.ops.aten.fmod.Tensor', 'torch.ops.aten.fmod.Tensor_out', 'torch.ops.aten.fmod.Scalar', 'torch.ops.aten.fmod.Scalar_out', 'torch.ops.aten.fmod.int', 'torch.ops.aten.fmod.float', 'torch.ops.aten.fmod.int_float', 'torch.ops.aten.fmod.float_int', 'torch.ops.aten.fmod.default', 'torch.ops.aten.gcd.default', 'torch.ops.aten.gcd.out', 'torch.ops.aten.gcd.int', 'torch.ops.aten.ge.Tensor', 'torch.ops.aten.ge.Scalar', 'torch.ops.aten.ge.Scalar_out', 'torch.ops.aten.ge.Tensor_out', 'torch.ops.aten.ge.int', 'torch.ops.aten.ge.float', 'torch.ops.aten.ge.int_float', 'torch.ops.aten.ge.float_int', 'torch.ops.aten.ge.default', 'torch.ops.aten.ge.str', 'torch.ops.aten.gt.Tensor', 'torch.ops.aten.gt.Scalar', 'torch.ops.aten.gt.Scalar_out', 'torch.ops.aten.gt.Tensor_out', 'torch.ops.aten.gt.int', 'torch.ops.aten.gt.float', 'torch.ops.aten.gt.int_float', 'torch.ops.aten.gt.float_int', 'torch.ops.aten.gt.default', 'torch.ops.aten.gt.str', 'torch.ops.aten.igamma.default', 'torch.ops.aten.igamma.out', 'torch.ops.aten.igammac.default', 'torch.ops.aten.igammac.out', 'torch.ops.aten.lcm.default', 'torch.ops.aten.lcm.out', 'torch.ops.aten.le.Tensor', 'torch.ops.aten.le.Scalar', 'torch.ops.aten.le.Scalar_out', 'torch.ops.aten.le.Tensor_out', 'torch.ops.aten.le.int', 'torch.ops.aten.le.float', 'torch.ops.aten.le.int_float', 'torch.ops.aten.le.float_int', 'torch.ops.aten.le.default', 'torch.ops.aten.le.str', 'torch.ops.aten.logical_and.default', 'torch.ops.aten.logical_and.out', 'torch.ops.aten.logical_not.default', 'torch.ops.aten.logical_not.out', 'torch.ops.aten.logical_or.default', 'torch.ops.aten.logical_or.out', 'torch.ops.aten.logical_xor.default', 'torch.ops.aten.logical_xor.out', 'torch.ops.aten.lt.Tensor', 'torch.ops.aten.lt.Scalar', 'torch.ops.aten.lt.Scalar_out', 'torch.ops.aten.lt.Tensor_out', 'torch.ops.aten.lt.int', 'torch.ops.aten.lt.float', 'torch.ops.aten.lt.int_float', 'torch.ops.aten.lt.float_int', 'torch.ops.aten.lt.default', 'torch.ops.aten.lt.str', 'torch.ops.aten.maximum.default', 'torch.ops.aten.maximum.out', 'torch.ops.aten.minimum.default', 'torch.ops.aten.minimum.out', 'torch.ops.aten.mul.Tensor', 'torch.ops.aten.mul.Scalar', 'torch.ops.aten.mul.out', 'torch.ops.aten.mul.Scalar_out', 'torch.ops.aten.mul.left_t', 'torch.ops.aten.mul.right_', 'torch.ops.aten.mul.int', 'torch.ops.aten.mul.complex', 'torch.ops.aten.mul.float', 'torch.ops.aten.mul.int_complex', 'torch.ops.aten.mul.complex_int', 'torch.ops.aten.mul.float_complex', 'torch.ops.aten.mul.complex_float', 'torch.ops.aten.mul.int_float', 'torch.ops.aten.mul.float_int', 'torch.ops.aten.mul.default', 'torch.ops.aten.ne.Tensor', 'torch.ops.aten.ne.Scalar', 'torch.ops.aten.ne.Scalar_out', 'torch.ops.aten.ne.Tensor_out', 'torch.ops.aten.ne.int_list', 'torch.ops.aten.ne.device', 'torch.ops.aten.ne.bool', 'torch.ops.aten.ne.enum', 'torch.ops.aten.ne.int', 'torch.ops.aten.ne.complex', 'torch.ops.aten.ne.float', 'torch.ops.aten.ne.int_float', 'torch.ops.aten.ne.float_int', 'torch.ops.aten.ne.float_complex', 'torch.ops.aten.ne.complex_float', 'torch.ops.aten.ne.default', 'torch.ops.aten.ne.str', 'torch.ops.aten.ne.float_list', 'torch.ops.aten.ne.Tensor_list', 'torch.ops.aten.ne.bool_list', 'torch.ops.aten.ne.str_list', 'torch.ops.aten.nextafter.default', 'torch.ops.aten.nextafter.out', 'torch.ops.aten.pow.Tensor_Tensor', 'torch.ops.aten.pow.Tensor_Scalar', 'torch.ops.aten.pow.Scalar', 'torch.ops.aten.pow.Scalar_out', 'torch.ops.aten.pow.Tensor_Scalar_out', 'torch.ops.aten.pow.Tensor_Tensor_out', 'torch.ops.aten.pow.int', 'torch.ops.aten.pow.complex', 'torch.ops.aten.pow.float', 'torch.ops.aten.pow.int_float', 'torch.ops.aten.pow.float_int', 'torch.ops.aten.pow.float_complex', 'torch.ops.aten.pow.complex_float', 'torch.ops.aten.pow.Scalar_Scalar', 'torch.ops.aten.pow.int_to_int', 'torch.ops.aten.remainder.Tensor', 'torch.ops.aten.remainder.Tensor_out', 'torch.ops.aten.remainder.Scalar_Tensor', 'torch.ops.aten.remainder.Scalar', 'torch.ops.aten.remainder.Scalar_out', 'torch.ops.aten.remainder.int', 'torch.ops.aten.remainder.float', 'torch.ops.aten.remainder.int_float', 'torch.ops.aten.remainder.float_int', 'torch.ops.aten.remainder.default', 'torch.ops.aten.sub.Tensor', 'torch.ops.aten.sub.Scalar', 'torch.ops.aten.sub.out', 'torch.ops.aten.sub.Scalar_out', 'torch.ops.aten.sub.int', 'torch.ops.aten.sub.complex', 'torch.ops.aten.sub.float', 'torch.ops.aten.sub.int_complex', 'torch.ops.aten.sub.complex_int', 'torch.ops.aten.sub.float_complex', 'torch.ops.aten.sub.complex_float', 'torch.ops.aten.sub.int_float', 'torch.ops.aten.sub.float_int', 'torch.ops.aten.sub.default', 'torch.ops.aten.where.self', 'torch.ops.aten.where.self_out', 'torch.ops.aten.where.ScalarSelf', 'torch.ops.aten.where.ScalarOther', 'torch.ops.aten.where.Scalar', 'torch.ops.aten.where.default', 'torch.ops.aten.sum.dim_IntList', 'torch.ops.aten.sum.default', 'torch.ops.aten.sum.dim_DimnameList', 'torch.ops.aten.sum.DimnameList_out', 'torch.ops.aten.sum.IntList_out', 'torch.ops.aten.sum.int', 'torch.ops.aten.sum.float', 'torch.ops.aten.sum.complex', 'torch.ops.aten.sum.bool', 'torch.ops.aten.prod.default', 'torch.ops.aten.prod.dim_int', 'torch.ops.aten.prod.dim_Dimname', 'torch.ops.aten.prod.Dimname_out', 'torch.ops.aten.prod.int_out', 'torch.ops.aten.std_mean.correction', 'torch.ops.aten.addr.default', 'torch.ops.aten.addr.out', 'torch.ops.aten.cat.default', 'torch.ops.aten.cat.names', 'torch.ops.aten.cat.names_out', 'torch.ops.aten.cat.out', 'torch.ops.aten.flip.default', 'torch.ops.aten.native_layer_norm.default', 'torch.ops.aten.permute.default', 'torch.ops.aten.roll.default', 'torch.ops.aten.rot90.default', 'torch.ops.aten.stack.default', 'torch.ops.aten.stack.out', 'torch.ops.aten.t.default', 'torch.ops.aten.unsqueeze.default', 'torch.ops.aten.trace.default', 'torch.ops.aten.celu.default', 'torch.ops.aten.celu.out', 'torch.ops.aten.relu.default', 'torch.ops.aten.relu.out', 'torch.ops.aten.leaky_relu.default', 'torch.ops.aten.leaky_relu.out', 'torch.ops.aten.mish.default', 'torch.ops.aten.mish.out', 'torch.ops.aten.softplus.default', 'torch.ops.aten.softplus.out', 'torch.ops.aten.hardtanh.default', 'torch.ops.aten.hardtanh.out', 'torch.ops.aten.gelu.default', 'torch.ops.aten.gelu.out', 'torch.ops.aten.special_i0e.default', 'torch.ops.aten.special_i0e.out', 'torch.ops.aten.special_i1.default', 'torch.ops.aten.special_i1.out', 'torch.ops.aten.special_i1e.default', 'torch.ops.aten.special_i1e.out', 'torch.ops.aten.logit.default', 'torch.ops.aten.logit.out', 'torch.ops.aten.special_zeta.default', 'torch.ops.aten.special_zeta.out', 'torch.ops.aten.special_zeta.self_scalar', 'torch.ops.aten.special_zeta.self_scalar_out', 'torch.ops.aten.special_zeta.other_scalar', 'torch.ops.aten.special_zeta.other_scalar_out'] |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
====== resnext50_32x4d_forward_0 ====== | |
Generating testing data... | |
...Aten2aten called ... | |
['getattr', '_operator.getitem', 'torch.ops.aten.abs', 'torch.ops.aten.acos', 'torch.ops.aten.acosh', 'torch.ops.aten.asin', 'torch.ops.aten.atan', 'torch.ops.aten.bitwise_not', 'torch.ops.aten.ceil', 'torch.ops.aten.cos', 'torch.ops.aten.cosh', 'torch.ops.aten.digamma', 'torch.ops.aten.erf', 'torch.ops.aten.erfinv', 'torch.ops.aten.erfc', 'torch.ops.aten.exp', 'torch.ops.aten.expm1', 'torch.ops.aten.exp2', 'torch.ops.aten.floor', 'torch.ops.aten.frac', 'torch.ops.aten.isinf', 'torch.ops.aten.isnan', 'torch.ops.aten.special_i0', 'torch.ops.aten.lgamma', 'torch.ops.aten.log', 'torch.ops.aten.log1p', 'torch.ops.aten.log2', 'torch.ops.aten.log10', 'torch.ops.aten.nan_to_num', 'torch.ops.aten.neg', 'torch.ops.aten.reciprocal', 'torch.ops.aten.rsqrt', 'torch.ops.aten.sigmoid', 'torch.ops.aten.sign', 'torch.ops.aten.signbit', 'torch.ops.aten.sin', 'torch.ops.aten.sinh', 'torch.ops.aten.sqrt', 'torch.ops.aten.tan', 'torch.ops.aten.tanh', 'torch.ops.aten.add', 'torch.ops.aten.atan2', 'torch.ops.aten.bitwise_and', 'torch.ops.aten.bitwise_left_shift', 'torch.ops.aten.bitwise_or', 'torch.ops.aten.bitwise_right_shift', 'torch.ops.aten.bitwise_xor', 'torch.ops.aten.eq', 'torch.ops.aten.fmax', 'torch.ops.aten.fmin', 'torch.ops.aten.fmod', 'torch.ops.aten.gcd', 'torch.ops.aten.ge', 'torch.ops.aten.gt', 'torch.ops.aten.igamma', 'torch.ops.aten.igammac', 'torch.ops.aten.lcm', 'torch.ops.aten.le', 'torch.ops.aten.logical_and', 'torch.ops.aten.logical_not', 'torch.ops.aten.logical_or', 'torch.ops.aten.logical_xor', 'torch.ops.aten.lt', 'torch.ops.aten.maximum', 'torch.ops.aten.minimum', 'torch.ops.aten.mul', 'torch.ops.aten.ne', 'torch.ops.aten.nextafter', 'torch.ops.aten.pow', 'torch.ops.aten.remainder', 'torch.ops.aten.sub', 'torch.ops.aten.where', 'torch.ops.aten.sum', 'torch.ops.aten.prod', 'torch.ops.aten.std_mean', 'torch.ops.aten.addr', 'torch.ops.aten.cat', 'torch.ops.aten.flip', 'torch.ops.aten.native_layer_norm', 'torch.ops.aten.permute', 'torch.ops.aten.roll', 'torch.ops.aten.rot90', 'torch.ops.aten.stack', 'torch.ops.aten.t', 'torch.ops.aten.unsqueeze', 'torch.ops.aten.trace', 'torch.ops.aten.celu', 'torch.ops.aten.relu', 'torch.ops.aten.leaky_relu', 'torch.ops.aten.mish', 'torch.ops.aten.softplus', 'torch.ops.aten.hardtanh', 'torch.ops.aten.gelu', 'torch.ops.aten.special_i0e', 'torch.ops.aten.special_i1', 'torch.ops.aten.special_i1e', 'torch.ops.aten.logit', 'torch.ops.aten.special_zeta'] | |
Num of partitions proposed: 53 | |
partition #2 ['mul_157', 'mul_158'] | |
partition #4 ['mul_154', 'mul_155'] | |
partition #6 ['mul_151', 'mul_152'] | |
partition #8 ['mul_149', 'mul_148'] | |
partition #10 ['mul_145', 'mul_146'] | |
partition #12 ['mul_143', 'mul_142'] | |
partition #14 ['mul_139', 'mul_140'] | |
partition #16 ['mul_137', 'mul_136'] | |
partition #18 ['mul_134', 'mul_133'] | |
partition #20 ['mul_130', 'mul_131'] | |
partition #22 ['mul_128', 'mul_127'] | |
partition #24 ['mul_124', 'mul_125'] | |
partition #26 ['mul_121', 'mul_122'] | |
partition #28 ['mul_118', 'mul_119'] | |
partition #30 ['mul_116', 'mul_115'] | |
partition #32 ['mul_112', 'mul_113'] | |
partition #34 ['mul_110', 'mul_109'] | |
partition #36 ['mul_107', 'mul_106'] | |
partition #38 ['mul_104', 'mul_103'] | |
partition #40 ['mul_101', 'mul_100'] | |
partition #42 ['mul_98', 'mul_97'] | |
partition #44 ['mul_94', 'mul_95'] | |
partition #46 ['mul_91', 'mul_92'] | |
partition #48 ['mul_89', 'mul_88'] | |
partition #50 ['mul_86', 'mul_85'] | |
partition #52 ['mul_83', 'mul_82'] | |
partition #54 ['mul_80', 'mul_79'] | |
partition #56 ['mul_77', 'mul_76'] | |
partition #58 ['mul_73', 'mul_74'] | |
partition #60 ['mul_70', 'mul_71'] | |
partition #62 ['mul_68', 'mul_67'] | |
partition #64 ['mul_64', 'mul_65'] | |
partition #66 ['mul_62', 'mul_61'] | |
partition #68 ['mul_59', 'mul_58'] | |
partition #70 ['mul_56', 'mul_55'] | |
partition #72 ['mul_53', 'mul_52'] | |
partition #74 ['mul_49', 'mul_50'] | |
partition #76 ['mul_47', 'mul_46'] | |
partition #78 ['mul_43', 'mul_44'] | |
partition #80 ['mul_41', 'mul_40'] | |
partition #82 ['mul_38', 'mul_37'] | |
partition #84 ['mul_35', 'mul_34'] | |
partition #86 ['mul_31', 'mul_32'] | |
partition #88 ['mul_29', 'mul_28'] | |
partition #90 ['mul_26', 'mul_25'] | |
partition #92 ['mul_23', 'mul_22'] | |
partition #94 ['mul_20', 'mul_19'] | |
partition #96 ['mul_17', 'mul_16'] | |
partition #98 ['mul_13', 'mul_14'] | |
partition #100 ['mul_11', 'mul_10'] | |
partition #102 ['mul_8', 'mul_7'] | |
partition #104 ['mul_5', 'mul_4'] | |
partition #108 ['mul_1', 'mul_2'] | |
Partioning time: 167.612 ms | |
Eager execution time: 4233.338 ms | |
nvFuser 1st call execution time: 28851.011 ms | |
nvFuser 2nd call execution time: 4614.273 ms | |
resnext50_32x4d_forward_0 failed! Tensor-likes are not close! | |
Mismatched elements: 32410 / 3211264 (1.0%) | |
Greatest absolute difference: 1.5019504757280604e+18 at index (6, 268, 22, 21) (up to 1e-05 allowed) | |
Greatest relative difference: 0.00010452292165946674 at index (6, 328, 22, 21) (up to 1e-05 allowed) | |
The failure occurred for item [19] | |
====== resnext50_32x4d_backward_0 ====== | |
Generating testing data... | |
...Aten2aten called ... | |
['getattr', '_operator.getitem', 'torch.ops.aten.abs', 'torch.ops.aten.acos', 'torch.ops.aten.acosh', 'torch.ops.aten.asin', 'torch.ops.aten.atan', 'torch.ops.aten.bitwise_not', 'torch.ops.aten.ceil', 'torch.ops.aten.cos', 'torch.ops.aten.cosh', 'torch.ops.aten.digamma', 'torch.ops.aten.erf', 'torch.ops.aten.erfinv', 'torch.ops.aten.erfc', 'torch.ops.aten.exp', 'torch.ops.aten.expm1', 'torch.ops.aten.exp2', 'torch.ops.aten.floor', 'torch.ops.aten.frac', 'torch.ops.aten.isinf', 'torch.ops.aten.isnan', 'torch.ops.aten.special_i0', 'torch.ops.aten.lgamma', 'torch.ops.aten.log', 'torch.ops.aten.log1p', 'torch.ops.aten.log2', 'torch.ops.aten.log10', 'torch.ops.aten.nan_to_num', 'torch.ops.aten.neg', 'torch.ops.aten.reciprocal', 'torch.ops.aten.rsqrt', 'torch.ops.aten.sigmoid', 'torch.ops.aten.sign', 'torch.ops.aten.signbit', 'torch.ops.aten.sin', 'torch.ops.aten.sinh', 'torch.ops.aten.sqrt', 'torch.ops.aten.tan', 'torch.ops.aten.tanh', 'torch.ops.aten.add', 'torch.ops.aten.atan2', 'torch.ops.aten.bitwise_and', 'torch.ops.aten.bitwise_left_shift', 'torch.ops.aten.bitwise_or', 'torch.ops.aten.bitwise_right_shift', 'torch.ops.aten.bitwise_xor', 'torch.ops.aten.eq', 'torch.ops.aten.fmax', 'torch.ops.aten.fmin', 'torch.ops.aten.fmod', 'torch.ops.aten.gcd', 'torch.ops.aten.ge', 'torch.ops.aten.gt', 'torch.ops.aten.igamma', 'torch.ops.aten.igammac', 'torch.ops.aten.lcm', 'torch.ops.aten.le', 'torch.ops.aten.logical_and', 'torch.ops.aten.logical_not', 'torch.ops.aten.logical_or', 'torch.ops.aten.logical_xor', 'torch.ops.aten.lt', 'torch.ops.aten.maximum', 'torch.ops.aten.minimum', 'torch.ops.aten.mul', 'torch.ops.aten.ne', 'torch.ops.aten.nextafter', 'torch.ops.aten.pow', 'torch.ops.aten.remainder', 'torch.ops.aten.sub', 'torch.ops.aten.where', 'torch.ops.aten.sum', 'torch.ops.aten.prod', 'torch.ops.aten.std_mean', 'torch.ops.aten.addr', 'torch.ops.aten.cat', 'torch.ops.aten.flip', 'torch.ops.aten.native_layer_norm', 'torch.ops.aten.permute', 'torch.ops.aten.roll', 'torch.ops.aten.rot90', 'torch.ops.aten.stack', 'torch.ops.aten.t', 'torch.ops.aten.unsqueeze', 'torch.ops.aten.trace', 'torch.ops.aten.celu', 'torch.ops.aten.relu', 'torch.ops.aten.leaky_relu', 'torch.ops.aten.mish', 'torch.ops.aten.softplus', 'torch.ops.aten.hardtanh', 'torch.ops.aten.gelu', 'torch.ops.aten.special_i0e', 'torch.ops.aten.special_i1', 'torch.ops.aten.special_i1e', 'torch.ops.aten.logit', 'torch.ops.aten.special_zeta'] | |
Num of partitions proposed: 0 | |
Partioning time: 72.669 ms | |
Eager execution time: 40.734 ms | |
nvFuser 1st call execution time: 147.460 ms | |
nvFuser 2nd call execution time: 144.853 ms | |
resnext50_32x4d_backward_0 has 0 fusion groups, Passed! | |
====== nvidia_deeprecommender_backward_0 ====== | |
Generating testing data... | |
...Aten2aten called ... | |
['getattr', '_operator.getitem', 'torch.ops.aten.abs', 'torch.ops.aten.acos', 'torch.ops.aten.acosh', 'torch.ops.aten.asin', 'torch.ops.aten.atan', 'torch.ops.aten.bitwise_not', 'torch.ops.aten.ceil', 'torch.ops.aten.cos', 'torch.ops.aten.cosh', 'torch.ops.aten.digamma', 'torch.ops.aten.erf', 'torch.ops.aten.erfinv', 'torch.ops.aten.erfc', 'torch.ops.aten.exp', 'torch.ops.aten.expm1', 'torch.ops.aten.exp2', 'torch.ops.aten.floor', 'torch.ops.aten.frac', 'torch.ops.aten.isinf', 'torch.ops.aten.isnan', 'torch.ops.aten.special_i0', 'torch.ops.aten.lgamma', 'torch.ops.aten.log', 'torch.ops.aten.log1p', 'torch.ops.aten.log2', 'torch.ops.aten.log10', 'torch.ops.aten.nan_to_num', 'torch.ops.aten.neg', 'torch.ops.aten.reciprocal', 'torch.ops.aten.rsqrt', 'torch.ops.aten.sigmoid', 'torch.ops.aten.sign', 'torch.ops.aten.signbit', 'torch.ops.aten.sin', 'torch.ops.aten.sinh', 'torch.ops.aten.sqrt', 'torch.ops.aten.tan', 'torch.ops.aten.tanh', 'torch.ops.aten.add', 'torch.ops.aten.atan2', 'torch.ops.aten.bitwise_and', 'torch.ops.aten.bitwise_left_shift', 'torch.ops.aten.bitwise_or', 'torch.ops.aten.bitwise_right_shift', 'torch.ops.aten.bitwise_xor', 'torch.ops.aten.eq', 'torch.ops.aten.fmax', 'torch.ops.aten.fmin', 'torch.ops.aten.fmod', 'torch.ops.aten.gcd', 'torch.ops.aten.ge', 'torch.ops.aten.gt', 'torch.ops.aten.igamma', 'torch.ops.aten.igammac', 'torch.ops.aten.lcm', 'torch.ops.aten.le', 'torch.ops.aten.logical_and', 'torch.ops.aten.logical_not', 'torch.ops.aten.logical_or', 'torch.ops.aten.logical_xor', 'torch.ops.aten.lt', 'torch.ops.aten.maximum', 'torch.ops.aten.minimum', 'torch.ops.aten.mul', 'torch.ops.aten.ne', 'torch.ops.aten.nextafter', 'torch.ops.aten.pow', 'torch.ops.aten.remainder', 'torch.ops.aten.sub', 'torch.ops.aten.where', 'torch.ops.aten.sum', 'torch.ops.aten.prod', 'torch.ops.aten.std_mean', 'torch.ops.aten.addr', 'torch.ops.aten.cat', 'torch.ops.aten.flip', 'torch.ops.aten.native_layer_norm', 'torch.ops.aten.permute', 'torch.ops.aten.roll', 'torch.ops.aten.rot90', 'torch.ops.aten.stack', 'torch.ops.aten.t', 'torch.ops.aten.unsqueeze', 'torch.ops.aten.trace', 'torch.ops.aten.celu', 'torch.ops.aten.relu', 'torch.ops.aten.leaky_relu', 'torch.ops.aten.mish', 'torch.ops.aten.softplus', 'torch.ops.aten.hardtanh', 'torch.ops.aten.gelu', 'torch.ops.aten.special_i0e', 'torch.ops.aten.special_i1', 'torch.ops.aten.special_i1e', 'torch.ops.aten.logit', 'torch.ops.aten.special_zeta'] | |
Num of partitions proposed: 6 | |
partition #1 ['mul_25', 'mul_28', 'mul_26'] | |
partition #5 ['mul_21', 'mul_20', 'mul_23'] | |
partition #9 ['mul_16', 'mul_18', 'mul_15'] | |
partition #13 ['mul_13', 'mul_10', 'mul_11'] | |
partition #17 ['mul_8', 'mul_6', 'mul_5'] | |
partition #21 ['mul_1', 'mul_3', 'mul'] | |
Partioning time: 10.758 ms | |
Eager execution time: 19.699 ms | |
nvFuser 1st call execution time: 2188.614 ms | |
nvFuser 2nd call execution time: 35.050 ms | |
nvidia_deeprecommender_backward_0 has 6 fusion groups, Passed! | |
====== nvidia_deeprecommender_forward_0 ====== | |
Generating testing data... | |
...Aten2aten called ... | |
['getattr', '_operator.getitem', 'torch.ops.aten.abs', 'torch.ops.aten.acos', 'torch.ops.aten.acosh', 'torch.ops.aten.asin', 'torch.ops.aten.atan', 'torch.ops.aten.bitwise_not', 'torch.ops.aten.ceil', 'torch.ops.aten.cos', 'torch.ops.aten.cosh', 'torch.ops.aten.digamma', 'torch.ops.aten.erf', 'torch.ops.aten.erfinv', 'torch.ops.aten.erfc', 'torch.ops.aten.exp', 'torch.ops.aten.expm1', 'torch.ops.aten.exp2', 'torch.ops.aten.floor', 'torch.ops.aten.frac', 'torch.ops.aten.isinf', 'torch.ops.aten.isnan', 'torch.ops.aten.special_i0', 'torch.ops.aten.lgamma', 'torch.ops.aten.log', 'torch.ops.aten.log1p', 'torch.ops.aten.log2', 'torch.ops.aten.log10', 'torch.ops.aten.nan_to_num', 'torch.ops.aten.neg', 'torch.ops.aten.reciprocal', 'torch.ops.aten.rsqrt', 'torch.ops.aten.sigmoid', 'torch.ops.aten.sign', 'torch.ops.aten.signbit', 'torch.ops.aten.sin', 'torch.ops.aten.sinh', 'torch.ops.aten.sqrt', 'torch.ops.aten.tan', 'torch.ops.aten.tanh', 'torch.ops.aten.add', 'torch.ops.aten.atan2', 'torch.ops.aten.bitwise_and', 'torch.ops.aten.bitwise_left_shift', 'torch.ops.aten.bitwise_or', 'torch.ops.aten.bitwise_right_shift', 'torch.ops.aten.bitwise_xor', 'torch.ops.aten.eq', 'torch.ops.aten.fmax', 'torch.ops.aten.fmin', 'torch.ops.aten.fmod', 'torch.ops.aten.gcd', 'torch.ops.aten.ge', 'torch.ops.aten.gt', 'torch.ops.aten.igamma', 'torch.ops.aten.igammac', 'torch.ops.aten.lcm', 'torch.ops.aten.le', 'torch.ops.aten.logical_and', 'torch.ops.aten.logical_not', 'torch.ops.aten.logical_or', 'torch.ops.aten.logical_xor', 'torch.ops.aten.lt', 'torch.ops.aten.maximum', 'torch.ops.aten.minimum', 'torch.ops.aten.mul', 'torch.ops.aten.ne', 'torch.ops.aten.nextafter', 'torch.ops.aten.pow', 'torch.ops.aten.remainder', 'torch.ops.aten.sub', 'torch.ops.aten.where', 'torch.ops.aten.sum', 'torch.ops.aten.prod', 'torch.ops.aten.std_mean', 'torch.ops.aten.addr', 'torch.ops.aten.cat', 'torch.ops.aten.flip', 'torch.ops.aten.native_layer_norm', 'torch.ops.aten.permute', 'torch.ops.aten.roll', 'torch.ops.aten.rot90', 'torch.ops.aten.stack', 'torch.ops.aten.t', 'torch.ops.aten.unsqueeze', 'torch.ops.aten.trace', 'torch.ops.aten.celu', 'torch.ops.aten.relu', 'torch.ops.aten.leaky_relu', 'torch.ops.aten.mish', 'torch.ops.aten.softplus', 'torch.ops.aten.hardtanh', 'torch.ops.aten.gelu', 'torch.ops.aten.special_i0e', 'torch.ops.aten.special_i1', 'torch.ops.aten.special_i1e', 'torch.ops.aten.logit', 'torch.ops.aten.special_zeta'] | |
Num of partitions proposed: 0 | |
Partioning time: 7.311 ms | |
Eager execution time: 14.359 ms | |
nvFuser 1st call execution time: 30.493 ms | |
nvFuser 2nd call execution time: 30.498 ms | |
nvidia_deeprecommender_forward_0 has 0 fusion groups, Passed! | |
====== moco_forward_4 ====== | |
Generating testing data... | |
...Aten2aten called ... | |
['getattr', '_operator.getitem', 'torch.ops.aten.abs', 'torch.ops.aten.acos', 'torch.ops.aten.acosh', 'torch.ops.aten.asin', 'torch.ops.aten.atan', 'torch.ops.aten.bitwise_not', 'torch.ops.aten.ceil', 'torch.ops.aten.cos', 'torch.ops.aten.cosh', 'torch.ops.aten.digamma', 'torch.ops.aten.erf', 'torch.ops.aten.erfinv', 'torch.ops.aten.erfc', 'torch.ops.aten.exp', 'torch.ops.aten.expm1', 'torch.ops.aten.exp2', 'torch.ops.aten.floor', 'torch.ops.aten.frac', 'torch.ops.aten.isinf', 'torch.ops.aten.isnan', 'torch.ops.aten.special_i0', 'torch.ops.aten.lgamma', 'torch.ops.aten.log', 'torch.ops.aten.log1p', 'torch.ops.aten.log2', 'torch.ops.aten.log10', 'torch.ops.aten.nan_to_num', 'torch.ops.aten.neg', 'torch.ops.aten.reciprocal', 'torch.ops.aten.rsqrt', 'torch.ops.aten.sigmoid', 'torch.ops.aten.sign', 'torch.ops.aten.signbit', 'torch.ops.aten.sin', 'torch.ops.aten.sinh', 'torch.ops.aten.sqrt', 'torch.ops.aten.tan', 'torch.ops.aten.tanh', 'torch.ops.aten.add', 'torch.ops.aten.atan2', 'torch.ops.aten.bitwise_and', 'torch.ops.aten.bitwise_left_shift', 'torch.ops.aten.bitwise_or', 'torch.ops.aten.bitwise_right_shift', 'torch.ops.aten.bitwise_xor', 'torch.ops.aten.eq', 'torch.ops.aten.fmax', 'torch.ops.aten.fmin', 'torch.ops.aten.fmod', 'torch.ops.aten.gcd', 'torch.ops.aten.ge', 'torch.ops.aten.gt', 'torch.ops.aten.igamma', 'torch.ops.aten.igammac', 'torch.ops.aten.lcm', 'torch.ops.aten.le', 'torch.ops.aten.logical_and', 'torch.ops.aten.logical_not', 'torch.ops.aten.logical_or', 'torch.ops.aten.logical_xor', 'torch.ops.aten.lt', 'torch.ops.aten.maximum', 'torch.ops.aten.minimum', 'torch.ops.aten.mul', 'torch.ops.aten.ne', 'torch.ops.aten.nextafter', 'torch.ops.aten.pow', 'torch.ops.aten.remainder', 'torch.ops.aten.sub', 'torch.ops.aten.where', 'torch.ops.aten.sum', 'torch.ops.aten.prod', 'torch.ops.aten.std_mean', 'torch.ops.aten.addr', 'torch.ops.aten.cat', 'torch.ops.aten.flip', 'torch.ops.aten.native_layer_norm', 'torch.ops.aten.permute', 'torch.ops.aten.roll', 'torch.ops.aten.rot90', 'torch.ops.aten.stack', 'torch.ops.aten.t', 'torch.ops.aten.unsqueeze', 'torch.ops.aten.trace', 'torch.ops.aten.celu', 'torch.ops.aten.relu', 'torch.ops.aten.leaky_relu', 'torch.ops.aten.mish', 'torch.ops.aten.softplus', 'torch.ops.aten.hardtanh', 'torch.ops.aten.gelu', 'torch.ops.aten.special_i0e', 'torch.ops.aten.special_i1', 'torch.ops.aten.special_i1e', 'torch.ops.aten.logit', 'torch.ops.aten.special_zeta'] |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment