case LLM_ARCH_JINA_BERT_V2:
case LLM_ARCH_NOMIC_BERT:
case LLM_ARCH_NOMIC_BERT_MOE:
+ case LLM_ARCH_WAVTOKENIZER_DEC:
{
res = nullptr;
} break;
params.model = params.vocoder.model;
params.embedding = true;
+ params.ctx_shift = false; // silence warning
+ params.n_ubatch = params.n_batch;
common_init_result llama_init_cts = common_init_from_params(params);
}
GGML_ASSERT(batch.n_tokens == n_codes);
- if (llama_decode(ctx_cts, batch) != 0) {
- LOG_ERR("%s: llama_decode() failed\n", __func__);
+ if (llama_encode(ctx_cts, batch) != 0) {
+ LOG_ERR("%s: llama_encode() failed\n", __func__);
return 1;
}