From: Jhen-Jie Hong Date: Tue, 7 Nov 2023 08:51:27 +0000 (+0800) Subject: talk-llama : fix n_gpu_layers usage again (#1442) X-Git-Tag: upstream/1.7.4~1285 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=75dc800d212bdffe7f6ac7e4bbb5cb311e9889fb;p=pkg%2Fggml%2Fsources%2Fwhisper.cpp talk-llama : fix n_gpu_layers usage again (#1442) --- diff --git a/examples/talk-llama/talk-llama.cpp b/examples/talk-llama/talk-llama.cpp index bc0119a7..8c41ef54 100644 --- a/examples/talk-llama/talk-llama.cpp +++ b/examples/talk-llama/talk-llama.cpp @@ -267,7 +267,7 @@ int main(int argc, char ** argv) { auto lmparams = llama_model_default_params(); if (!params.use_gpu) { - lcparams.lmparams = 0; + lmparams.n_gpu_layers = 0; } struct llama_model * model_llama = llama_load_model_from_file(params.model_llama.c_str(), lmparams);