Skip to content

Commit fa233ec

Browse files
committed
Skip int4 QAT tests for nightly for now
int4 tinygemm quantization is currently broken in master and being fixed in #517. Let's skip these tests for now until that is fixed.
1 parent f8789f7 commit fa233ec

File tree

1 file changed

+6
-0
lines changed

1 file changed

+6
-0
lines changed

test/quantization/test_qat.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -366,6 +366,8 @@ def _assert_close_4w(self, val, ref):
366366

367367
@unittest.skipIf(not TORCH_VERSION_AFTER_2_4, "skipping when torch version is 2.4 or lower")
368368
@unittest.skipIf(not _CUDA_IS_AVAILABLE, "skipping when cuda is not available")
369+
# TODO: remove once we fix int4 error: https://github.com/pytorch/ao/pull/517
370+
@unittest.skipIf(TORCH_VERSION_AFTER_2_5, "int4 doesn't work for 2.5+ right now")
369371
def test_qat_4w_primitives(self):
370372
n_bit = 4
371373
group_size = 32
@@ -411,6 +413,8 @@ def test_qat_4w_primitives(self):
411413

412414
@unittest.skipIf(not TORCH_VERSION_AFTER_2_4, "skipping when torch version is 2.4 or lower")
413415
@unittest.skipIf(not _CUDA_IS_AVAILABLE, "skipping when cuda is not available")
416+
# TODO: remove once we fix int4 error: https://github.com/pytorch/ao/pull/517
417+
@unittest.skipIf(TORCH_VERSION_AFTER_2_5, "int4 doesn't work for 2.5+ right now")
414418
def test_qat_4w_linear(self):
415419
from torchao.quantization.prototype.qat import Int4WeightOnlyQATLinear
416420
from torchao.quantization.GPTQ import WeightOnlyInt4Linear
@@ -439,6 +443,8 @@ def test_qat_4w_linear(self):
439443

440444
@unittest.skipIf(not TORCH_VERSION_AFTER_2_4, "skipping when torch version is 2.4 or lower")
441445
@unittest.skipIf(not _CUDA_IS_AVAILABLE, "skipping when cuda is not available")
446+
# TODO: remove once we fix int4 error: https://github.com/pytorch/ao/pull/517
447+
@unittest.skipIf(TORCH_VERSION_AFTER_2_5, "int4 doesn't work for 2.5+ right now")
442448
def test_qat_4w_quantizer(self):
443449
from torchao.quantization.prototype.qat import Int4WeightOnlyQATQuantizer
444450
from torchao.quantization.GPTQ import Int4WeightOnlyQuantizer

0 commit comments

Comments
 (0)