From: Mathieu Baudier Date: Sat, 18 Oct 2025 08:57:19 +0000 (+0200) Subject: Align package names with Debian official X-Git-Tag: debian/0.0.7011-1~2 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=b075147195839aa19a670ce06e44dccb094edd11;p=pkg%2Fggml%2Fsources%2Fllama.cpp Align package names with Debian official --- diff --git a/debian/changelog b/debian/changelog index aa0e2218..2715de05 100644 --- a/debian/changelog +++ b/debian/changelog @@ -1,12 +1,7 @@ -llama-cpp (0.0.6641-1) unstable; urgency=medium +llama.cpp (0.0.6764-1) unstable; urgency=medium * Update upstream + * Align package names with Debian official - -- Mathieu Baudier Fri, 10 Oct 2025 14:03:01 +0000 - -llama-cpp (0.0.6527-1) unstable; urgency=medium - - * Update upstream - - -- Mathieu Baudier Mon, 22 Sep 2025 12:46:43 +0000 + -- Mathieu Baudier Sat, 18 Oct 2025 08:51:56 +0000 \ No newline at end of file diff --git a/debian/control b/debian/control index 3d580627..4e5acb6f 100644 --- a/debian/control +++ b/debian/control @@ -1,4 +1,4 @@ -Source: llama-cpp +Source: llama.cpp Section: science Priority: optional Maintainer: Mathieu Baudier @@ -43,7 +43,7 @@ Description: Inference of large language models in pure C/C++ (multimodal librar mtmd provides multimodal inference. # We only distribute a few useful tools, with stable CLI options -Package: llama-cpp-tools +Package: llama.cpp-tools Architecture: any Depends: libllama0 (= ${binary:Version}), curl, @@ -61,7 +61,7 @@ Description: Inference of large language models in pure C/C++ (tools) llama-bench: benchmarking of large language models or ggml backends. -Package: llama-cpp-tools-multimodal +Package: llama.cpp-tools-multimodal Architecture: any Depends: libmtmd0 (= ${binary:Version}), curl, diff --git a/debian/llama-cpp-tools-multimodal.bash-completion b/debian/llama-cpp-tools-multimodal.bash-completion deleted file mode 100644 index 3ae0a65d..00000000 --- a/debian/llama-cpp-tools-multimodal.bash-completion +++ /dev/null @@ -1 +0,0 @@ -completions/llama-mtmd-cli diff --git a/debian/llama-cpp-tools-multimodal.install b/debian/llama-cpp-tools-multimodal.install deleted file mode 100644 index 5aae3775..00000000 --- a/debian/llama-cpp-tools-multimodal.install +++ /dev/null @@ -1,2 +0,0 @@ -/usr/bin/llama-mtmd-cli -/usr/libexec/*/ggml/llama-mtmd-cli diff --git a/debian/llama-cpp-tools.bash-completion b/debian/llama-cpp-tools.bash-completion deleted file mode 100644 index 432aa940..00000000 --- a/debian/llama-cpp-tools.bash-completion +++ /dev/null @@ -1,3 +0,0 @@ -completions/llama-cli -# completions/llama-quantize does not actually use these options -completions/llama-bench diff --git a/debian/llama-cpp-tools.install b/debian/llama-cpp-tools.install deleted file mode 100644 index a028e995..00000000 --- a/debian/llama-cpp-tools.install +++ /dev/null @@ -1,8 +0,0 @@ -/usr/bin/llama-cli -/usr/libexec/*/ggml/llama-cli - -/usr/bin/llama-quantize -/usr/libexec/*/ggml/llama-quantize - -/usr/bin/llama-bench -/usr/libexec/*/ggml/llama-bench diff --git a/debian/llama.cpp-tools-multimodal.bash-completion b/debian/llama.cpp-tools-multimodal.bash-completion new file mode 100644 index 00000000..3ae0a65d --- /dev/null +++ b/debian/llama.cpp-tools-multimodal.bash-completion @@ -0,0 +1 @@ +completions/llama-mtmd-cli diff --git a/debian/llama.cpp-tools-multimodal.install b/debian/llama.cpp-tools-multimodal.install new file mode 100644 index 00000000..5aae3775 --- /dev/null +++ b/debian/llama.cpp-tools-multimodal.install @@ -0,0 +1,2 @@ +/usr/bin/llama-mtmd-cli +/usr/libexec/*/ggml/llama-mtmd-cli diff --git a/debian/llama.cpp-tools.bash-completion b/debian/llama.cpp-tools.bash-completion new file mode 100644 index 00000000..432aa940 --- /dev/null +++ b/debian/llama.cpp-tools.bash-completion @@ -0,0 +1,3 @@ +completions/llama-cli +# completions/llama-quantize does not actually use these options +completions/llama-bench diff --git a/debian/llama.cpp-tools.install b/debian/llama.cpp-tools.install new file mode 100644 index 00000000..a028e995 --- /dev/null +++ b/debian/llama.cpp-tools.install @@ -0,0 +1,8 @@ +/usr/bin/llama-cli +/usr/libexec/*/ggml/llama-cli + +/usr/bin/llama-quantize +/usr/libexec/*/ggml/llama-quantize + +/usr/bin/llama-bench +/usr/libexec/*/ggml/llama-bench diff --git a/debian/source/lintian-overrides b/debian/source/lintian-overrides index 1304de99..30edff27 100644 --- a/debian/source/lintian-overrides +++ b/debian/source/lintian-overrides @@ -1,11 +1,11 @@ -llama-cpp source: source-is-missing [examples/server/public_legacy/index-new.html] -llama-cpp source: source-is-missing [examples/server/public_legacy/index.js] -llama-cpp source: source-is-missing [examples/server/public_legacy/system-prompts.js] -llama-cpp source: source-contains-prebuilt-java-object [examples/llama.android/gradle/wrapper/gradle-wrapper.jar] -llama-cpp source: source-contains-prebuilt-javascript-object [examples/server/public_legacy/index.js] -llama-cpp source: source-contains-prebuilt-javascript-object [examples/server/public_legacy/system-prompts.js] -llama-cpp source: missing-prerequisite-for-pyproject-backend poetry.core.masonry.api (does not satisfy python3-poetry-core:any, pybuild-plugin-pyproject:any) [pyproject.toml:39] +llama.cpp source: source-is-missing [examples/server/public_legacy/index-new.html] +llama.cpp source: source-is-missing [examples/server/public_legacy/index.js] +llama.cpp source: source-is-missing [examples/server/public_legacy/system-prompts.js] +llama.cpp source: source-contains-prebuilt-java-object [examples/llama.android/gradle/wrapper/gradle-wrapper.jar] +llama.cpp source: source-contains-prebuilt-javascript-object [examples/server/public_legacy/index.js] +llama.cpp source: source-contains-prebuilt-javascript-object [examples/server/public_legacy/system-prompts.js] +llama.cpp source: missing-prerequisite-for-pyproject-backend poetry.core.masonry.api (does not satisfy python3-poetry-core:any, pybuild-plugin-pyproject:any) [pyproject.toml:39] # some of the "examples" are actually source code form this packaging perspective -llama-cpp source: package-does-not-install-examples [examples/] -llama-cpp source: package-does-not-install-examples [ggml/src/ggml-kompute/kompute/examples/] -llama-cpp source: package-does-not-install-examples [gguf-py/examples/] \ No newline at end of file +llama.cpp source: package-does-not-install-examples [examples/] +llama.cpp source: package-does-not-install-examples [ggml/src/ggml-kompute/kompute/examples/] +llama.cpp source: package-does-not-install-examples [gguf-py/examples/] \ No newline at end of file