From: Jhen-Jie Hong Date: Tue, 7 Nov 2023 01:36:23 +0000 (+0800) Subject: examples : fix n_gpu_layers usage in talk-llama (#1441) X-Git-Tag: upstream/1.7.4~1287 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=3989b29a9b0a2c0de92e848957d9a47635957f84;p=pkg%2Fggml%2Fsources%2Fwhisper.cpp examples : fix n_gpu_layers usage in talk-llama (#1441) --- diff --git a/examples/talk-llama/talk-llama.cpp b/examples/talk-llama/talk-llama.cpp index 6cc30c16..bc0119a7 100644 --- a/examples/talk-llama/talk-llama.cpp +++ b/examples/talk-llama/talk-llama.cpp @@ -266,6 +266,9 @@ int main(int argc, char ** argv) { llama_backend_init(true); auto lmparams = llama_model_default_params(); + if (!params.use_gpu) { + lcparams.lmparams = 0; + } struct llama_model * model_llama = llama_load_model_from_file(params.model_llama.c_str(), lmparams); @@ -276,9 +279,6 @@ int main(int argc, char ** argv) { lcparams.seed = 1; lcparams.f16_kv = true; lcparams.n_threads = params.n_threads; - if (!params.use_gpu) { - lcparams.n_gpu_layers = 0; - } struct llama_context * ctx_llama = llama_new_context_with_model(model_llama, lcparams);