From: Aaron Teo Date: Thu, 8 Jan 2026 21:34:56 +0000 (+0800) Subject: llama: use host memory if device reports 0 memory (llama/18587) X-Git-Tag: upstream/1.8.3~16 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=fff3ebd93d3a5f3679e26a19ee6acdd40d568c3b;p=pkg%2Fggml%2Fsources%2Fwhisper.cpp llama: use host memory if device reports 0 memory (llama/18587) --- diff --git a/ggml/src/ggml-backend-impl.h b/ggml/src/ggml-backend-impl.h index 6792ba98..59190b7c 100644 --- a/ggml/src/ggml-backend-impl.h +++ b/ggml/src/ggml-backend-impl.h @@ -144,7 +144,7 @@ extern "C" { // device description: short informative description of the device, could be the model name const char * (*get_description)(ggml_backend_dev_t dev); - // device memory in bytes + // device memory in bytes: 0 bytes to indicate no memory to report void (*get_memory)(ggml_backend_dev_t dev, size_t * free, size_t * total); // device type diff --git a/ggml/src/ggml-opencl/ggml-opencl.cpp b/ggml/src/ggml-opencl/ggml-opencl.cpp index 472e2df5..e50ca8e0 100644 --- a/ggml/src/ggml-opencl/ggml-opencl.cpp +++ b/ggml/src/ggml-opencl/ggml-opencl.cpp @@ -4287,8 +4287,8 @@ static const char * ggml_backend_opencl_device_get_description(ggml_backend_dev_ } static void ggml_backend_opencl_device_get_memory(ggml_backend_dev_t dev, size_t * free, size_t * total) { - *free = 1; - *total = 1; + *free = 0; + *total = 0; GGML_UNUSED(dev); }