From: Aman Gupta Date: Mon, 9 Mar 2026 08:15:36 +0000 (+0800) Subject: ggml-cuda: disable gdn for musa (llama/20278) X-Git-Tag: upstream/1.8.4~64 X-Git-Url: https://git.djapps.eu/?a=commitdiff_plain;h=3984ae384d7e0011f0a182693089c3cca67a4a6f;p=pkg%2Fggml%2Fsources%2Fwhisper.cpp ggml-cuda: disable gdn for musa (llama/20278) --- diff --git a/ggml/src/ggml-cuda/ggml-cuda.cu b/ggml/src/ggml-cuda/ggml-cuda.cu index 0fafaf00..cda275b8 100644 --- a/ggml/src/ggml-cuda/ggml-cuda.cu +++ b/ggml/src/ggml-cuda/ggml-cuda.cu @@ -4992,9 +4992,15 @@ static bool ggml_backend_cuda_device_supports_op(ggml_backend_dev_t dev, const g case GGML_OP_LEAKY_RELU: case GGML_OP_RWKV_WKV6: case GGML_OP_GATED_LINEAR_ATTN: - case GGML_OP_GATED_DELTA_NET: case GGML_OP_RWKV_WKV7: return true; + case GGML_OP_GATED_DELTA_NET: + //TODO: enable once MUSA compiler is solved https://github.com/ggml-org/llama.cpp/pull/19504#issuecomment-4018634327 +#ifdef GGML_USE_MUSA + return false; +#else + return true; +#endif // GGML_USE_MUSA case GGML_OP_FLASH_ATTN_EXT: return ggml_cuda_flash_attn_ext_supported(dev_ctx->device, op); case GGML_OP_CROSS_ENTROPY_LOSS: