From: MasterYi1024 Date: Fri, 12 Apr 2024 08:28:12 +0000 (+0800) Subject: Correct free memory and total memory. (#6630) X-Git-Tag: upstream/0.0.4488~1830 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=dee7f8d6928cc680cc969f7d93f98c3e24dcad41;p=pkg%2Fggml%2Fsources%2Fllama.cpp Correct free memory and total memory. (#6630) Co-authored-by: MasterYi --- diff --git a/llama.cpp b/llama.cpp index 73ff6070..dad2c4fb 100644 --- a/llama.cpp +++ b/llama.cpp @@ -1638,17 +1638,17 @@ static size_t llama_get_device_memory(int device) { #if defined(GGML_USE_CUDA) size_t total; size_t free; - ggml_backend_cuda_get_device_memory(device, &total, &free); + ggml_backend_cuda_get_device_memory(device, &free, &total); return free; #elif defined(GGML_USE_SYCL) size_t total; size_t free; - ggml_backend_sycl_get_device_memory(device, &total, &free); + ggml_backend_sycl_get_device_memory(device, &free, &total); return free; #elif defined(GGML_USE_VULKAN) size_t total; size_t free; - ggml_backend_vk_get_device_memory(device, &total, &free); + ggml_backend_vk_get_device_memory(device, &free, &total); return free; #else return 1;