@@ -400,6 +400,8 @@ def run_meta_crossref(
400
400
torch .mode : {b8 , bf16 , f16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 }, # aten::mode
401
401
torch .multinomial : {bf16 , f32 , f64 }, # aten::multinomial, aten::multinomial.out
402
402
torch .mvlgamma : {bf16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 }, # aten::_local_scalar_dense, aten::mvlgamma.out
403
+ torch .nanmean : {bf16 , f16 , f32 , f64 },
404
+ torch .nanquantile : {f32 , f64 },
403
405
torch .nn .functional .conv1d : {bf16 , f32 , f64 , i64 },
404
406
torch .nn .functional .conv2d : {bf16 , f32 , f64 , i64 },
405
407
torch .nn .functional .conv_transpose1d : {f32 , f64 , i64 },
@@ -463,9 +465,9 @@ def run_meta_crossref(
463
465
torch .functional .cdist : {f32 , f64 },
464
466
torch .functional .tensordot : {bf16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
465
467
torch .inner : {bf16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
468
+ torch .logical_not : {b8 , bf16 , f16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
466
469
torch .nn .functional .cross_entropy : {bf16 , f32 , f64 },
467
470
torch .nn .functional .interpolate : {bf16 , f32 , f64 , u8 },
468
- torch .nanmean : {bf16 , f16 , f32 , f64 }, # TODO(chilli): Doesn't seem to work for some reason?
469
471
torch .nn .functional .nll_loss : {bf16 , f32 , f64 }, # TODO
470
472
torch .linalg .pinv : {f32 , f64 },
471
473
torch .empty : {b8 , bf16 , c128 , c64 , c32 , f16 , f32 , f64 , i16 , i32 , i64 , i8 , u8 },
@@ -625,6 +627,8 @@ def __torch_function__(self, func, types, args=(), kwargs=None):
625
627
aten .log_sigmoid_forward .output : {bf16 , f64 , f32 },
626
628
aten .logcumsumexp .default : {bf16 , f64 , f32 },
627
629
aten .logcumsumexp .out : {bf16 , f64 , f32 },
630
+ aten .logical_not .out : {i64 , bf16 , f16 , u8 , b8 , f32 , i8 , f64 , i16 , i32 },
631
+ aten .logical_not_ .default : {bf16 , f16 , f64 , f32 },
628
632
aten .masked_select .default : {i64 , bf16 , f16 , u8 , b8 , f32 , i8 , f64 , i16 , i32 },
629
633
aten .masked_select .out : {i64 , bf16 , f16 , u8 , b8 , f32 , i8 , f64 , i16 , i32 },
630
634
aten .max_pool3d_with_indices .default : {f64 , f32 },
0 commit comments