From: Borislav Stanimirov Date: Mon, 9 Oct 2023 15:15:20 +0000 (+0300) Subject: cuda : int counters for device, fix fprintf warning (#560) X-Git-Tag: upstream/0.0.1642~1224 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=faebeb9ed96a80a65bc8b33decf60eff3589d518;p=pkg%2Fggml%2Fsources%2Fggml cuda : int counters for device, fix fprintf warning (#560) --- diff --git a/src/ggml-cuda.cu b/src/ggml-cuda.cu index 7e92c519..5bd83bb5 100644 --- a/src/ggml-cuda.cu +++ b/src/ggml-cuda.cu @@ -5647,10 +5647,10 @@ void ggml_init_cublas() { GGML_ASSERT(g_device_count <= GGML_CUDA_MAX_DEVICES); int64_t total_vram = 0; fprintf(stderr, "%s: found %d " GGML_CUDA_NAME " devices:\n", __func__, g_device_count); - for (int64_t id = 0; id < g_device_count; ++id) { + for (int id = 0; id < g_device_count; ++id) { cudaDeviceProp prop; CUDA_CHECK(cudaGetDeviceProperties(&prop, id)); - fprintf(stderr, " Device %ld: %s, compute capability %d.%d\n", id, prop.name, prop.major, prop.minor); + fprintf(stderr, " Device %d: %s, compute capability %d.%d\n", id, prop.name, prop.major, prop.minor); g_tensor_split[id] = total_vram; total_vram += prop.totalGlobalMem; @@ -5660,15 +5660,15 @@ void ggml_init_cublas() { g_compute_capabilities[id] = 100*prop.major + 10*prop.minor; #endif // defined(GGML_USE_HIPBLAS) && defined(__HIP_PLATFORM_AMD__) } - for (int64_t id = 0; id < g_device_count; ++id) { + for (int id = 0; id < g_device_count; ++id) { g_tensor_split[id] /= total_vram; } - for (int64_t id = 0; id < g_device_count; ++id) { + for (int id = 0; id < g_device_count; ++id) { CUDA_CHECK(ggml_cuda_set_device(id)); // create cuda streams - for (int64_t is = 0; is < MAX_STREAMS; ++is) { + for (int is = 0; is < MAX_STREAMS; ++is) { CUDA_CHECK(cudaStreamCreateWithFlags(&g_cudaStreams[id][is], cudaStreamNonBlocking)); }