From: Daniel Bevenius Date: Thu, 13 Feb 2025 06:07:51 +0000 (+0100) Subject: llama : update llama_decode_internal ref [no ci] (#11840) X-Git-Tag: upstream/0.0.4719~16 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=3e693197724c31d53a9b69018c2f1bd0b93ebab2;p=pkg%2Fggml%2Fsources%2Fllama.cpp llama : update llama_decode_internal ref [no ci] (#11840) This commit updates the comment in llama_kv_cache.h to reflect the change of the function name from llama_decode_internal to llama_decode_impl. --- diff --git a/src/llama-kv-cache.h b/src/llama-kv-cache.h index dca6f399..1ed688e3 100644 --- a/src/llama-kv-cache.h +++ b/src/llama-kv-cache.h @@ -37,7 +37,7 @@ struct llama_kv_cache { bool can_shift = false; // Note: The value of head isn't only used to optimize searching - // for a free KV slot. llama_decode_internal also uses it, so it + // for a free KV slot. llama_decode_impl also uses it, so it // cannot be freely changed after a slot has been allocated. uint32_t head = 0; uint32_t size = 0;