Skip to content

Commit 21b1759

Browse files
authored
More log fixes (#1328)
* due to how the pb loops, call pb.iter() to get the current iter step Signed-off-by: Qubitium <[email protected]> * rename Signed-off-by: Qubitium <[email protected]> * fix pb usage Signed-off-by: Qubitium <[email protected]> --------- Signed-off-by: Qubitium <[email protected]>
1 parent 2ff2ddf commit 21b1759

File tree

3 files changed

+18
-15
lines changed

3 files changed

+18
-15
lines changed

gptqmodel/utils/model.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -516,7 +516,7 @@ def pack_module(name, qModules, quant_result, layers, pbar=None):
516516
qModules[name].pack(linear=layers[name], scales=scale, zeros=zero, g_idx=g_idx)
517517
qModules[name].to(layer_device)
518518
if pbar:
519-
pbar.iter()
519+
pbar.next()
520520
pbar.progress()
521521

522522

gptqmodel/utils/progress.py

Lines changed: 10 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -63,7 +63,7 @@ def __init__(self,
6363
self.bar_length = bar_length
6464
self.fill = fill
6565
self.info_text = info
66-
self.current_iteration = 0
66+
self.current_iter_step = 0
6767
self.time = time.time()
6868

6969
def info(self, info:str):
@@ -74,7 +74,7 @@ def info(self, info:str):
7474

7575
def progress(self, iteration:int = None):
7676
if not iteration:
77-
iteration = self.current_iteration
77+
iteration = self.current_iter_step
7878

7979
columns, _ = terminal_size()
8080
bar_length = columns
@@ -98,7 +98,7 @@ def progress(self, iteration:int = None):
9898

9999
filled_length = int(bar_length * iteration // len(self))
100100
bar = self.fill * filled_length + '-' * (bar_length - filled_length)
101-
self.log(bar=bar, log=log, padding=padding, end='\n' if percent_num >= 1.0 else '')
101+
self.log(bar=bar, log=log, padding=padding, end='') # '\n' if percent_num >= 1.0 else ''
102102

103103
def calc_time(self, iteration):
104104
used_time = int(time.time() - self.time)
@@ -169,18 +169,21 @@ def _comparable(self):
169169
def __hash__(self):
170170
return id(self)
171171

172-
def iter(self):
173-
self.current_iteration += 1
172+
def step(self) -> int:
173+
return self.current_iter_step
174+
175+
def next(self):
176+
self.current_iter_step += 1
174177

175178
def __iter__(self):
176179
iterable = self.iterable
177180

178181
for obj in iterable:
179-
self.iter()
182+
self.next()
180183
self.progress()
181184
yield obj
182185

183-
self.progress()
186+
# self.progress()
184187
self.close()
185188
return
186189

tests/inference_speed.py

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -32,7 +32,7 @@
3232
class InferenceSpeed(unittest.TestCase):
3333
NATIVE_MODEL_ID = "/monster/data/model/DeepSeek-R1-Distill-Qwen-7B-gptqmodel-4bit-vortex-v2"
3434
BITBLAS_NATIVE_MODEL_ID = "/monster/data/model/opt-125M-autoround-lm_head-false-symTrue"
35-
MAX_NEW_TOEKNS = 10
35+
MAX_NEW_TOKENS = 10
3636
NUM_RUNS = 20
3737
PROMPTS = [
3838
"I am in Paris and I",
@@ -69,11 +69,11 @@ def inference(self, model_path, backend, tokens_per_second, assert_result=True,
6969
# compile kernels need JIT compile (Bitblas, IPEX, Triton) so we should do some warmup before actual speed run
7070
if warmup_runs > 0:
7171
pb = ProgressBar(range(warmup_runs))
72-
for i in pb:
73-
pb.info(f"warmup run index {i} of {warmup_runs - 1}")
72+
for _ in pb:
73+
pb.info(f"warmup run index {pb.step()} of {len(pb)}")
7474
pb.progress()
7575
start_time = time.time()
76-
result = model.generate(**inp, max_new_tokens=self.MAX_NEW_TOEKNS, pad_token_id=tokenizer.pad_token_id)
76+
result = model.generate(**inp, max_new_tokens=self.MAX_NEW_TOKENS, pad_token_id=tokenizer.pad_token_id)
7777
end_time = time.time()
7878
elapsed_time = end_time - start_time
7979
times.append(elapsed_time)
@@ -97,10 +97,10 @@ def inference(self, model_path, backend, tokens_per_second, assert_result=True,
9797
print(f"**************** {backend} Warm-up Result Info End****************")
9898

9999
pb = ProgressBar(range(self.NUM_RUNS))
100-
for i in pb:
101-
pb.info(f"run index {i} of {self.NUM_RUNS - 1}")
100+
for _ in pb:
101+
pb.info(f"run index {pb.step()} of {len(pb)}")
102102
start_time = time.time()
103-
result = model.generate(**inp, max_new_tokens=self.MAX_NEW_TOEKNS, pad_token_id=tokenizer.pad_token_id)
103+
result = model.generate(**inp, max_new_tokens=self.MAX_NEW_TOKENS, pad_token_id=tokenizer.pad_token_id)
104104
end_time = time.time()
105105
elapsed_time = end_time - start_time
106106
times.append(elapsed_time)

0 commit comments

Comments
 (0)