@@ -1360,7 +1360,7 @@ def _linear_f16_act_floatx_weight_impl(input_tensor, weight_tensor, bias):
1360
1360
1361
1361
return out .view (* act .shape [:- 1 ], out_dim ).to (act .dtype )
1362
1362
1363
- def _linear_fp_act_fp8_weight_check (
1363
+ def _linear_fp8_act_fp8_weight_check (
1364
1364
input_tensor : Union [torch .Tensor , AffineQuantizedTensor ],
1365
1365
weight_tensor : Union [torch .Tensor , AffineQuantizedTensor ],
1366
1366
bias : Optional [torch .Tensor ],
@@ -1384,7 +1384,7 @@ def preprocess_scale(input_scale: torch.Tensor, input_shape: Tuple[int]):
1384
1384
1385
1385
return input_scale
1386
1386
1387
- def _linear_fp_act_fp8_weight_impl (
1387
+ def _linear_fp8_act_fp8_weight_impl (
1388
1388
input_tensor : AffineQuantizedTensor ,
1389
1389
weight_tensor : AffineQuantizedTensor ,
1390
1390
bias : Optional [torch .Tensor ],
@@ -1473,7 +1473,7 @@ def _register_aqt_quantized_linear_dispatches():
1473
1473
for dispatch_condition , impl in [
1474
1474
(_linear_int8_act_int8_weight_check , _linear_int8_act_int8_weight_impl ),
1475
1475
(_linear_int8_act_int8_weight_semi_structured_sparse_check , _linear_int8_act_int8_weight_semi_structured_sparse_impl ),
1476
- (_linear_fp_act_fp8_weight_check , _linear_fp_act_fp8_weight_impl ),
1476
+ (_linear_fp8_act_fp8_weight_check , _linear_fp8_act_fp8_weight_impl ),
1477
1477
(_linear_bf16_act_uint4_weight_check , _linear_bf16_act_uint4_weight_impl ),
1478
1478
(_linear_fp_act_int8_weight_check , _linear_fp_act_int8_weight_impl ),
1479
1479
(_linear_f16_act_floatx_weight_check , _linear_f16_act_floatx_weight_impl ),
0 commit comments