]> git.djapps.eu Git - pkg/ggml/sources/llama.cpp/commit
model : Nomic Embed Text V2 with Mixture-of-Experts (MoE) architecture (#12466)
authorAT <redacted>
Mon, 28 Apr 2025 19:52:15 +0000 (15:52 -0400)
committerGitHub <redacted>
Mon, 28 Apr 2025 19:52:15 +0000 (22:52 +0300)
commit5f5e39e1ba5dbea814e41f2a15e035d749a520bc
tree7c609747b2e639e490659ba677d1763c98e292f9
parenteaea3253244dc4bbe07f6cd81325847ccc6cf93e
model : Nomic Embed Text V2 with Mixture-of-Experts (MoE) architecture (#12466)

* Nomic Embed Text V2 with Mixture-of-Experts (MoE) architecture

- Adds MoE-based embedding model supporting multilingual embeddings.
- Selects architecture variant based on hyperparameter detection (MoE layers).
- Removes unnecessary subclass initialization checks for clarity.

https://www.nomic.ai/blog/posts/nomic-embed-text-v2

Co-authored-by: Jared Van Bortel <redacted>
* fix tokenizer

* don't rename this tensor

---------

Co-authored-by: Jared Van Bortel <redacted>
convert_hf_to_gguf.py
gguf-py/gguf/constants.py
gguf-py/gguf/gguf_writer.py
gguf-py/gguf/tensor_mapping.py
src/llama-arch.cpp
src/llama-arch.h
src/llama-graph.cpp
src/llama-hparams.h
src/llama-model.cpp