]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commitdiff
Temporary bump the memory buffer size - hopefully fix issues from 483bab2e
authorGeorgi Gerganov <redacted>
Fri, 24 Mar 2023 16:23:56 +0000 (18:23 +0200)
committerGeorgi Gerganov <redacted>
Fri, 24 Mar 2023 16:23:56 +0000 (18:23 +0200)
llama.cpp

index cdb8628285c1520a8d4eeddaeaa0978f57344373..9a93409cccb4cb530a09e76716d49df2428678c5 100644 (file)
--- a/llama.cpp
+++ b/llama.cpp
@@ -632,7 +632,7 @@ static bool llama_eval_internal(
     auto & mem_per_token = lctx.mem_per_token;
 
     // TODO: fix this hardcoded size
-    static size_t buf_size = 512u*1024*1024;
+    static size_t buf_size = 2048u*1024*1024; // TMP !!!
     static void * buf = malloc(buf_size);
 
     if (mem_per_token > 0 && mem_per_token*N > buf_size) {