From: Neo Zhang Jianyu Date: Thu, 1 Feb 2024 19:48:53 +0000 (+0800) Subject: add --no-mmap in llama-bench (llama/5257) X-Git-Tag: upstream/0.0.1642~999 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=34916bd3d9aadd96ea618e7f153cbc8be520e86b;p=pkg%2Fggml%2Fsources%2Fggml add --no-mmap in llama-bench (llama/5257) * add --no-mmap, show sycl backend * fix conflict * fix code format, change print for --no-mmap * ren no_mmap to mmap, show mmap when not default value in printer * update guide for mmap * mv position to reduce model reload --- diff --git a/ggml-sycl.cpp b/ggml-sycl.cpp index 1cc55ef5..e8ba4835 100644 --- a/ggml-sycl.cpp +++ b/ggml-sycl.cpp @@ -2928,7 +2928,6 @@ void ggml_sycl_set_main_device(int main_device); void ggml_sycl_set_mul_mat_q(bool mul_mat_q); void ggml_sycl_set_scratch_size(size_t scratch_size); void ggml_sycl_free_scratch(void); -int ggml_sycl_get_device_count(void); void ggml_sycl_get_device_description(int device, char * description, size_t description_size); bool ggml_backend_is_sycl(ggml_backend_t backend); int ggml_backend_sycl_get_device(ggml_backend_t backend); @@ -14493,6 +14492,37 @@ bool ggml_sycl_compute_forward(struct ggml_compute_params * params, struct ggml_ return true; } +GGML_API GGML_CALL void ggml_sycl_get_gpu_list(int *id_list, int max_len) try { + int max_compute_units = -1; + for(int i=0;i