Skip to content

Commit 3b5dbf6

Browse files
authored
Set low_gpu_mem_usage=False for AutoRound
Signed-off-by: Kaihui-intel <[email protected]>
1 parent 41244d3 commit 3b5dbf6

File tree

1 file changed

+2
-2
lines changed
  • neural_compressor/torch/quantization

1 file changed

+2
-2
lines changed

neural_compressor/torch/quantization/config.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -905,7 +905,7 @@ def __init__(
905905
enable_minmax_tuning: bool = True,
906906
lr: float = None,
907907
minmax_lr: float = None,
908-
low_gpu_mem_usage: bool = True,
908+
low_gpu_mem_usage: bool = False,
909909
iters: int = 200,
910910
seqlen: int = 2048,
911911
nsamples: int = 128,
@@ -938,7 +938,7 @@ def __init__(
938938
enable_minmax_tuning (bool): Whether to enable min-max tuning (default is True).
939939
lr (float): The learning rate (default is 0.005).
940940
minmax_lr (float): The learning rate for min-max tuning (default is None).
941-
low_gpu_mem_usage (bool): Whether to use low GPU memory (default is True).
941+
low_gpu_mem_usage (bool): Whether to use low GPU memory (default is False).
942942
iters (int): Number of iterations (default is 200).
943943
seqlen (int): Length of the sequence.
944944
nsamples (int): Number of samples (default is 512).

0 commit comments

Comments
 (0)