]> git.djapps.eu Git - pkg/ggml/sources/ggml/commit
llama : offload to RPC in addition to other backends (llama/7640)
authorRadoslav Gerganov <redacted>
Mon, 3 Jun 2024 17:03:26 +0000 (20:03 +0300)
committerGeorgi Gerganov <redacted>
Sat, 15 Jun 2024 19:05:47 +0000 (22:05 +0300)
commit8d8d6b1f922fe3c69b8e7f2a3bfa9ebb2f4c8a0f
tree00a9d61b89cf52009db5f0512af83a0b4574be64
parent703d5e6ad6e109bbfe5f59002c11b2fc4de4dfbb
llama : offload to RPC in addition to other backends (llama/7640)

* llama : offload to RPC in addition to other backends

* - fix copy_tensor being called on the src buffer instead of the dst buffer

- always initialize views in the view_src buffer

- add RPC backend to Makefile build

- add endpoint to all RPC object names

* add rpc-server to Makefile

* Update llama.cpp

Co-authored-by: slaren <redacted>
---------

Co-authored-by: slaren <redacted>
include/ggml/ggml-backend.h
src/ggml-alloc.c
src/ggml-backend.c
src/ggml-rpc.cpp