We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
There was an error while loading. Please reload this page.
1 parent ae9d934 commit a5b6ef8Copy full SHA for a5b6ef8
neural_compressor/torch/algorithms/fp8_quant/fp8_quant.py
@@ -38,8 +38,8 @@ def prepare(self, model):
38
return model
39
40
def convert(self, model):
41
- if getattr(model, "prepared", False) and with_patched_module(model): # if model was calibrated on hpu
42
- finish_measurements(model)
+ if with_patched_module(model): # if model was calibrated on hpu
+ finish_measurements(model) # dump the measurements into files to be loaded in _convert
43
# for INC flow, it calls `prepare` and then `convert` user-facing API in one run
44
restore_patched_module(model)
45
_convert(model, self.quant_config)
0 commit comments