From: ddh0 Date: Thu, 11 Sep 2025 17:12:34 +0000 (-0500) Subject: nitpick : correct MB to MiB (#15934) X-Git-Tag: upstream/0.0.6527~78 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=df082f56309073ecf885eceaa21b86e8a487e61b;p=pkg%2Fggml%2Fsources%2Fllama.cpp nitpick : correct MB to MiB (#15934) MB was incorrectly used for 1024 x 1024 bytes instead of MiB --- diff --git a/src/llama-quant.cpp b/src/llama-quant.cpp index 1d0361cc..c93e8065 100644 --- a/src/llama-quant.cpp +++ b/src/llama-quant.cpp @@ -920,7 +920,7 @@ static void llama_model_quantize_impl(const std::string & fname_inp, const std:: new_type = tensor->type; new_data = tensor->data; new_size = ggml_nbytes(tensor); - LLAMA_LOG_INFO("size = %8.3f MB\n", ggml_nbytes(tensor)/1024.0/1024.0); + LLAMA_LOG_INFO("size = %8.3f MiB\n", ggml_nbytes(tensor)/1024.0/1024.0); } else { const int64_t nelements = ggml_nelements(tensor); @@ -1037,8 +1037,8 @@ static void llama_model_quantize_impl(const std::string & fname_inp, const std:: } close_ofstream(); - LLAMA_LOG_INFO("%s: model size = %8.2f MB\n", __func__, total_size_org/1024.0/1024.0); - LLAMA_LOG_INFO("%s: quant size = %8.2f MB\n", __func__, total_size_new/1024.0/1024.0); + LLAMA_LOG_INFO("%s: model size = %8.2f MiB\n", __func__, total_size_org/1024.0/1024.0); + LLAMA_LOG_INFO("%s: quant size = %8.2f MiB\n", __func__, total_size_new/1024.0/1024.0); if (qs.n_fallback > 0) { LLAMA_LOG_WARN("%s: WARNING: %d of %d tensor(s) required fallback quantization\n",