]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
common: llama_load_model_from_url split support (#6192)
authorPierrick Hymbert <redacted>
Sat, 23 Mar 2024 17:07:00 +0000 (18:07 +0100)
committerGitHub <redacted>
Sat, 23 Mar 2024 17:07:00 +0000 (18:07 +0100)
commitf482bb2e4920e544651fb832f2e0bcb4d2ff69ab
tree9fabefd6f3b34aef6bf13a8469c7cdf363cc88cb
parent1997577d5e121568ae39f538021733ccd4278c23
common: llama_load_model_from_url split support  (#6192)

* llama: llama_split_prefix fix strncpy does not include string termination
common: llama_load_model_from_url:
 - fix header name case sensitive
 - support downloading additional split in parallel
 - hide password in url

* common: EOL EOF

* common: remove redundant LLAMA_CURL_MAX_PATH_LENGTH definition

* common: change max url max length

* common: minor comment

* server: support HF URL options

* llama: llama_model_loader fix log

* common: use a constant for max url length

* common: clean up curl if file cannot be loaded in gguf

* server: tests: add split tests, and HF options params

* common: move llama_download_hide_password_in_url inside llama_download_file as a lambda

* server: tests: enable back Release test on PR

* spacing

Co-authored-by: Georgi Gerganov <redacted>
* spacing

Co-authored-by: Georgi Gerganov <redacted>
* spacing

Co-authored-by: Georgi Gerganov <redacted>
---------

Co-authored-by: Georgi Gerganov <redacted>
.github/workflows/server.yml
common/common.cpp
common/common.h
examples/gguf-split/gguf-split.cpp
examples/server/README.md
examples/server/server.cpp
examples/server/tests/features/parallel.feature
examples/server/tests/features/server.feature
examples/server/tests/features/steps/steps.py
llama.cpp