File tree Expand file tree Collapse file tree 2 files changed +1
-6
lines changed Expand file tree Collapse file tree 2 files changed +1
-6
lines changed Original file line number Diff line number Diff line change @@ -412,7 +412,6 @@ void llama_model::load_hparams(llama_model_loader & ml) {
412
412
ml.get_key (LLM_KV_BLOCK_COUNT, hparams.n_layer );
413
413
ml.get_key (LLM_KV_EXPERT_COUNT, hparams.n_expert , false );
414
414
ml.get_key (LLM_KV_EXPERT_USED_COUNT, hparams.n_expert_used , false );
415
- ml.get_key (LLM_KV_EXPERT_WEIGHTS_SCALE, hparams.expert_weights_scale , false );
416
415
417
416
if (arch == LLM_ARCH_WAVTOKENIZER_DEC) {
418
417
ml.get_key (LLM_KV_FEATURES_LENGTH, hparams.n_embd_features );
@@ -3702,10 +3701,6 @@ void llama_model::print_info() const {
3702
3701
LLAMA_LOG_INFO (" %s: f_attention_scale = %f\n " , __func__, hparams.f_attention_scale );
3703
3702
}
3704
3703
3705
- if (arch == LLM_ARCH_LLAMA) {
3706
- LLAMA_LOG_INFO (" %s: expert_weights_scale = %.1f\n " , __func__, hparams.expert_weights_scale );
3707
- }
3708
-
3709
3704
vocab.print_info ();
3710
3705
}
3711
3706
Original file line number Diff line number Diff line change @@ -1596,7 +1596,7 @@ struct llm_build_context {
1596
1596
nullptr ,
1597
1597
n_expert, n_expert_used,
1598
1598
LLM_FFN_SILU, true ,
1599
- hparams. expert_weights_scale > 0 , hparams. expert_weights_scale ,
1599
+ false , 0.0 ,
1600
1600
LLAMA_EXPERT_GATING_FUNC_TYPE_SOFTMAX,
1601
1601
cb, il);
1602
1602
cb (cur, " ffn_moe_out" , il);
You can’t perform that action at this time.
0 commit comments