@@ -400,8 +400,6 @@ def run_meta_crossref(
400
400
torch .mode : {b8 , bf16 , f16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 }, # aten::mode
401
401
torch .multinomial : {bf16 , f32 , f64 }, # aten::multinomial, aten::multinomial.out
402
402
torch .mvlgamma : {bf16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 }, # aten::_local_scalar_dense, aten::mvlgamma.out
403
- torch .nanmean : {bf16 , f16 , f32 , f64 },
404
- torch .nanquantile : {f32 , f64 },
405
403
torch .nn .functional .conv1d : {bf16 , f32 , f64 , i64 },
406
404
torch .nn .functional .conv2d : {bf16 , f32 , f64 , i64 },
407
405
torch .nn .functional .conv_transpose1d : {f32 , f64 , i64 },
@@ -465,9 +463,9 @@ def run_meta_crossref(
465
463
torch .functional .cdist : {f32 , f64 },
466
464
torch .functional .tensordot : {bf16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
467
465
torch .inner : {bf16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
468
- torch .logical_not : {b8 , bf16 , f16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
469
466
torch .nn .functional .cross_entropy : {bf16 , f32 , f64 },
470
467
torch .nn .functional .interpolate : {bf16 , f32 , f64 , u8 },
468
+ torch .nanmean : {bf16 , f16 , f32 , f64 }, # TODO(chilli): Doesn't seem to work for some reason?
471
469
torch .nn .functional .nll_loss : {bf16 , f32 , f64 }, # TODO
472
470
torch .linalg .pinv : {f32 , f64 },
473
471
torch .empty : {b8 , bf16 , c128 , c64 , c32 , f16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
@@ -627,8 +625,6 @@ def __torch_function__(self, func, types, args=(), kwargs=None):
627
625
aten .log_sigmoid_forward .output : {bf16 , f64 , f32 },
628
626
aten .logcumsumexp .default : {bf16 , f64 , f32 },
629
627
aten .logcumsumexp .out : {bf16 , f64 , f32 },
630
- aten .logical_not .out : {i64 , bf16 , f16 , u8 , b8 , f32 , i8 , f64 , i16 , i32 },
631
- aten .logical_not_ .default : {bf16 , f16 , f64 , f32 },
632
628
aten .masked_select .default : {i64 , bf16 , f16 , u8 , b8 , f32 , i8 , f64 , i16 , i32 },
633
629
aten .masked_select .out : {i64 , bf16 , f16 , u8 , b8 , f32 , i8 , f64 , i16 , i32 },
634
630
aten .max_pool3d_with_indices .default : {f64 , f32 },
0 commit comments