We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
per_block_cast_to_fp8
1 parent 42337d2 commit fd9423aCopy full SHA for fd9423a
vllm/utils/deep_gemm.py
@@ -135,7 +135,7 @@ def _align(x: int, y: int) -> int:
135
136
137
# Taken from https://github.com/deepseek-ai/DeepGEMM/blob/dd6ed14acbc7445dcef224248a77ab4d22b5f240/deep_gemm/utils/math.py#L38
138
-# TODO(wentao): optimize this function, using triton or cuda kernel
+@torch.compile(dynamic=True, backend=current_platform.simple_compile_backend)
139
def per_block_cast_to_fp8(
140
x: torch.Tensor,
141
block_size: list[int] = DEFAULT_BLOCK_SIZE,
@@ -187,4 +187,4 @@ def should_use_deepgemm_for_fp8_linear(output_dtype: torch.dtype,
187
"is_deep_gemm_e8m0_used",
188
"is_deep_gemm_supported",
189
"should_use_deepgemm_for_fp8_linear",
190
-]
+]
0 commit comments