From: slaren Date: Wed, 24 Jan 2024 11:48:14 +0000 (+0100) Subject: llama : pre-allocate input tensors in a separate buffer (llama/5100) X-Git-Tag: upstream/1.7.4~1080 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=94c1ae86689bd8861382b17deacaa54cfdaf09d5;p=pkg%2Fggml%2Fsources%2Fwhisper.cpp llama : pre-allocate input tensors in a separate buffer (llama/5100) --- diff --git a/ggml-alloc.c b/ggml-alloc.c index 89b85d34..60141a34 100644 --- a/ggml-alloc.c +++ b/ggml-alloc.c @@ -109,8 +109,8 @@ void ggml_tallocr_alloc(ggml_tallocr_t alloc, struct ggml_tensor * tensor) { if (block->size >= size) { best_fit_block = alloc->n_free_blocks - 1; } else { - fprintf(stderr, "%s: not enough space in the buffer (needed %zu, largest block available %zu)\n", - __func__, size, max_avail); + fprintf(stderr, "%s: not enough space in the buffer to allocate %s (needed %zu, largest block available %zu)\n", + __func__, tensor->name, size, max_avail); GGML_ASSERT(!"not enough space in the buffer"); return; }