We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
torch.backends.cuda.matmul.fp32_precision
1 parent a154c20 commit 117f11aCopy full SHA for 117f11a
torch/_inductor/kernel/flex/flex_attention.py
@@ -54,6 +54,8 @@ def flex_attention_grid(batch_size, q_heads, num_queries, d_model, meta, *, cdiv
54
def get_float32_precision():
55
if (
56
torch.backends.cuda.matmul.fp32_precision == "ieee"
57
+ if torch.backends.cuda.matmul.fp32_precision != "none"
58
+ else torch.get_float32_matmul_precision() == "highest"
59
or torch.version.hip
60
or torch.mtia.is_available()
61
):
0 commit comments