@@ -1517,9 +1517,12 @@ int main(int argc, char ** argv) {
1517
1517
const cmd_params_instance * prev_inst = nullptr ;
1518
1518
1519
1519
int params_idx = 0 ;
1520
+ auto params_count = params_instances.size ();
1520
1521
for (const auto & inst : params_instances) {
1521
1522
params_idx ++;
1522
- LOG_TEE (" llama-bench: benchmark %d/%ld: starting\n " , params_idx, params_instances.size ());
1523
+ if (params.verbose ) {
1524
+ LOG_TEE (" llama-bench: benchmark %d/%ld: starting\n " , params_idx, params_count);
1525
+ }
1523
1526
// keep the same model between tests when possible
1524
1527
if (!lmodel || !prev_inst || !inst.equal_mparams (*prev_inst)) {
1525
1528
if (lmodel) {
@@ -1569,12 +1572,16 @@ int main(int argc, char ** argv) {
1569
1572
1570
1573
// warmup run
1571
1574
if (t.n_prompt > 0 ) {
1572
- LOG_TEE (" llama-bench: benchmark %d/%ld: warmup prompt run\n " , params_idx, params_instances.size ());
1575
+ if (params.verbose ) {
1576
+ LOG_TEE (" llama-bench: benchmark %d/%ld: warmup prompt run\n " , params_idx, params_count);
1577
+ }
1573
1578
// test_prompt(ctx, std::min(t.n_batch, std::min(t.n_prompt, 32)), 0, t.n_batch, t.n_threads);
1574
1579
test_prompt (ctx, t.n_prompt , 0 , t.n_batch , t.n_threads );
1575
1580
}
1576
1581
if (t.n_gen > 0 ) {
1577
- LOG_TEE (" llama-bench: benchmark %d/%ld: warmup generation run\n " , params_idx, params_instances.size ());
1582
+ if (params.verbose ) {
1583
+ LOG_TEE (" llama-bench: benchmark %d/%ld: warmup generation run\n " , params_idx, params_count);
1584
+ }
1578
1585
test_gen (ctx, 1 , 0 , t.n_threads );
1579
1586
}
1580
1587
@@ -1584,11 +1591,15 @@ int main(int argc, char ** argv) {
1584
1591
uint64_t t_start = get_time_ns ();
1585
1592
1586
1593
if (t.n_prompt > 0 ) {
1587
- LOG_TEE (" llama-bench: benchmark %d/%ld: prompt run %d/%d\n " , params_idx, params_instances.size (), i + 1 , params.reps );
1594
+ if (params.verbose ) {
1595
+ LOG_TEE (" llama-bench: benchmark %d/%ld: prompt run %d/%d\n " , params_idx, params_count, i + 1 , params.reps );
1596
+ }
1588
1597
test_prompt (ctx, t.n_prompt , 0 , t.n_batch , t.n_threads );
1589
1598
}
1590
1599
if (t.n_gen > 0 ) {
1591
- LOG_TEE (" llama-bench: benchmark %d/%ld: generation run %d/%d\n " , params_idx, params_instances.size (), i + 1 , params.reps );
1600
+ if (params.verbose ) {
1601
+ LOG_TEE (" llama-bench: benchmark %d/%ld: generation run %d/%d\n " , params_idx, params_count, i + 1 , params.reps );
1602
+ }
1592
1603
test_gen (ctx, t.n_gen , t.n_prompt , t.n_threads );
1593
1604
}
1594
1605
0 commit comments