Skip to content

Commit 3989b29

Browse files
authored
examples : fix n_gpu_layers usage in talk-llama (#1441)
1 parent 0463028 commit 3989b29

File tree

1 file changed

+3
-3
lines changed

1 file changed

+3
-3
lines changed

examples/talk-llama/talk-llama.cpp

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -266,6 +266,9 @@ int main(int argc, char ** argv) {
266266
llama_backend_init(true);
267267

268268
auto lmparams = llama_model_default_params();
269+
if (!params.use_gpu) {
270+
lcparams.lmparams = 0;
271+
}
269272

270273
struct llama_model * model_llama = llama_load_model_from_file(params.model_llama.c_str(), lmparams);
271274

@@ -276,9 +279,6 @@ int main(int argc, char ** argv) {
276279
lcparams.seed = 1;
277280
lcparams.f16_kv = true;
278281
lcparams.n_threads = params.n_threads;
279-
if (!params.use_gpu) {
280-
lcparams.n_gpu_layers = 0;
281-
}
282282

283283
struct llama_context * ctx_llama = llama_new_context_with_model(model_llama, lcparams);
284284

0 commit comments

Comments
 (0)