Architecture: any
Multi-Arch: same
Pre-Depends: ${misc:Pre-Depends}
-Depends: libggml0,
- libggml0-backend-cpu,
- ${misc:Depends},
+Depends: ${misc:Depends},
${shlibs:Depends},
Description: Inference of large language models in pure C/C++ (shared library)
llama.cpp leverages the ggml tensor library in order to run
# We only distribute a few useful tools, with stable CLI options
Package: llama.cpp-tools
Architecture: any
-Depends: libllama0 (= ${binary:Version}),
+Depends: libggml0-backend-cpu,
+ libllama0 (= ${binary:Version}),
libmtmd0 (= ${binary:Version}),
${misc:Depends},
${shlibs:Depends},
-DLLAMA_BUILD_TESTS=OFF \
-DLLAMA_BUILD_SERVER=ON \
-ifeq ($(DEB_TARGET_ARCH),arm64)
FLAGS_ARGEO=-O3 -g -DNDEBUG
-ifeq ($(distribution),Ubuntu)
-# Optimize for Raspberry Pi 5 on Ubuntu
-#FLAGS_ARGEO += -mcpu=cortex-a76+crc+crypto
-CMAKE_FLAGS += -DGGML_CPU_ARM_ARCH=armv8-a
-#CMAKE_FLAGS += -DGGML_NATIVE=ON
-else
-#CMAKE_FLAGS += -DGGML_CPU_ARM_ARCH=armv8-a
-CMAKE_FLAGS += -DGGML_CPU_ARM_ARCH=armv8-a
-endif
-
-else
-FLAGS_ARGEO=-O3 -g -DNDEBUG
-
-endif
-
CMAKE_FLAGS += -DCMAKE_C_FLAGS_RELWITHDEBINFO="$(FLAGS_ARGEO)" \
-DCMAKE_CXX_FLAGS_RELWITHDEBINFO="$(FLAGS_ARGEO)" \