Commit 401c2b7
committed
Makes fallback float8 1x128 by 128x128 gemm output bfloat16
Summary:
For now, we just care about bf16 output. We can add fp32 and a flag to
control it later, if needed.
Test Plan:
```
pytest test/quantization/quantize_/workflows/float8/test_float8_tensor.py -s -k fp8_linear_variants -x
```
Reviewers:
Subscribers:
Tasks:
Tags:
ghstack-source-id: f00cd47
ghstack-comment-id: 3469836810
Pull-Request: #32651 parent a0ee87e commit 401c2b7
File tree
2 files changed
+3
-2
lines changed- test/kernel
- torchao/kernel
2 files changed
+3
-2
lines changed| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
66 | 66 | | |
67 | 67 | | |
68 | 68 | | |
| 69 | + | |
69 | 70 | | |
70 | 71 | | |
71 | 72 | | |
| |||
| Original file line number | Diff line number | Diff line change | |
|---|---|---|---|
| |||
92 | 92 | | |
93 | 93 | | |
94 | 94 | | |
95 | | - | |
| 95 | + | |
96 | 96 | | |
97 | 97 | | |
98 | 98 | | |
| |||
105 | 105 | | |
106 | 106 | | |
107 | 107 | | |
108 | | - | |
| 108 | + | |
109 | 109 | | |
110 | 110 | | |
111 | 111 | | |
| |||
0 commit comments