Skip to content

Commit d29b6d5

Browse files
authored
Merge pull request ggml-org#12 from anon998/clear-logit-bias
Clear logit bias between requests.
2 parents af71126 + 8cbc4be commit d29b6d5

File tree

1 file changed

+7
-4
lines changed

1 file changed

+7
-4
lines changed

examples/server/server.cpp

Lines changed: 7 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -559,6 +559,7 @@ json format_generation_settings(llama_server_context &llama) {
559559
{ "n_keep", llama.params.n_keep },
560560
{ "ignore_eos", ignore_eos },
561561
{ "stream", llama.stream },
562+
{ "logit_bias", llama.params.logit_bias },
562563
};
563564
}
564565

@@ -638,7 +639,7 @@ bool parse_options_completion(json body, llama_server_context& llama, Response &
638639
if (!body["penalize_nl"].is_null()) {
639640
llama.params.penalize_nl = body["penalize_nl"].get<float>();
640641
} else {
641-
llama.params.penalize_nl = false;
642+
llama.params.penalize_nl = default_params.penalize_nl;
642643
}
643644
if (!body["n_keep"].is_null()) {
644645
llama.params.n_keep = body["n_keep"].get<int>();
@@ -650,10 +651,10 @@ bool parse_options_completion(json body, llama_server_context& llama, Response &
650651
} else {
651652
llama.params.seed = time(NULL);
652653
}
654+
655+
llama.params.logit_bias.clear();
653656
if (!body["ignore_eos"].is_null() && body["ignore_eos"].get<bool>()) {
654657
llama.params.logit_bias[llama_token_eos()] = -INFINITY;
655-
} else {
656-
llama.params.logit_bias.erase(llama_token_eos());
657658
}
658659
if (body["logit_bias"].is_array()) {
659660
int n_vocab = llama_n_vocab(llama.ctx);
@@ -665,6 +666,7 @@ bool parse_options_completion(json body, llama_server_context& llama, Response &
665666
}
666667
}
667668
}
669+
668670
if (!body["prompt"].is_null()) {
669671
llama.params.prompt = body["prompt"].get<std::string>();
670672
} else {
@@ -673,6 +675,7 @@ bool parse_options_completion(json body, llama_server_context& llama, Response &
673675
res.status = 400;
674676
return false;
675677
}
678+
676679
llama.params.antiprompt.clear();
677680
if (!body["stop"].is_null()) {
678681
const auto stop = body["stop"].get<std::vector<std::string>>();
@@ -888,7 +891,7 @@ int main(int argc, char **argv)
888891
}
889892
});
890893

891-
svr.Options(R"(/.*)", [&llama](const Request &req, Response &res)
894+
svr.Options(R"(/.*)", [&llama](const Request &, Response &res)
892895
{
893896
return res.set_content("", "application/json");
894897
});

0 commit comments

Comments
 (0)