@@ -574,8 +574,6 @@ def _get_stats(self,
574
574
num_waiting = len (self .scheduler .waiting )
575
575
576
576
# Iteration stats if we have scheduler output.
577
- num_prompt_tokens = 0
578
- num_generation_tokens = 0
579
577
num_prompt_tokens_lst = []
580
578
num_generation_tokens_lst = []
581
579
request_n = []
@@ -594,13 +592,7 @@ def _get_stats(self,
594
592
for scheduled_seq_group in
595
593
scheduler_outputs .scheduled_seq_groups
596
594
]
597
- num_prompt_tokens = sum (num_prompt_tokens_lst )
598
- num_generation_tokens = sum (
599
- scheduled_seq_group .seq_group .num_seqs ()
600
- for scheduled_seq_group in
601
- scheduler_outputs .scheduled_seq_groups )
602
595
else :
603
- num_generation_tokens = scheduler_outputs .num_batched_tokens
604
596
num_generation_tokens_lst = [
605
597
seq .get_output_len () for scheduled_seq_group in
606
598
scheduler_outputs .scheduled_seq_groups for seq in
@@ -651,8 +643,6 @@ def _get_stats(self,
651
643
gpu_cache_usage = gpu_cache_usage ,
652
644
cpu_cache_usage = cpu_cache_usage ,
653
645
finished_reason_lst = finished_reason_lst ,
654
- num_prompt_tokens = num_prompt_tokens ,
655
- num_generation_tokens = num_generation_tokens ,
656
646
num_prompt_tokens_lst = num_prompt_tokens_lst ,
657
647
num_generation_tokens_lst = num_generation_tokens_lst ,
658
648
request_n = request_n ,
0 commit comments