Skip to content

Commit e6b16a3

Browse files
committed
Update
[ghstack-poisoned]
2 parents 31e3c5b + cafe668 commit e6b16a3

File tree

2 files changed

+1
-2
lines changed

2 files changed

+1
-2
lines changed

torchao/_models/llama/eval.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -174,6 +174,7 @@ def run_evaluation(
174174
if quantization == "float8_a1x128_w128x128":
175175
config = Float8DynamicActivationFloat8WeightConfig(
176176
granularity=(PerBlock([1, 128]), PerBlock([128, 128])),
177+
activation_value_lb=1e-12,
177178
)
178179
# TODO(future): all workflows in this file should be skipping quantization
179180
# of `lm_head`

torchao/quantization/quant_api.py

Lines changed: 0 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1778,8 +1778,6 @@ def __post_init__(self):
17781778

17791779
default_use_fast_accum = True
17801780
if _granularity_is_a_1_128_w_128_128(self.granularity):
1781-
assert self.activation_value_lb is None, "unimplemented"
1782-
assert self.activation_value_ub is None, "unimplemented"
17831781
assert self.kernel_preference in (
17841782
KernelPreference.AUTO,
17851783
KernelPreference.TORCH,

0 commit comments

Comments
 (0)