* add changes from llama upstream
* add sycl abstraction
* add sycl build
* update cmake
* add sycl build config
* fix bug
* fix bug
* refactor build
* fix bug
* update build
* call build
* use sycl header
* add examples
* add target
* fix typecast in quant.c
* readd fp16 and readme
* fix quant typecast
* add sample
* add readme
* remove cxx file check
#include "ggml-cuda.h"
#endif
+#ifdef GGML_USE_SYCL
+#include "ggml-sycl.h"
+#endif
+
#ifdef WHISPER_USE_OPENVINO
#include "openvino/whisper-openvino-encoder.h"
#endif
}
#endif
+#ifdef GGML_USE_SYCL
+ if (params.use_gpu) {
+ WHISPER_LOG_INFO("%s: using SYCL backend\n", __func__);
+ backend_gpu = ggml_backend_sycl_init(params.gpu_device);
+ if (!backend_gpu) {
+ WHISPER_LOG_ERROR("%s: ggml_backend_sycl_init() failed\n", __func__);
+ }
+ }
+#endif
+
if (backend_gpu) {
return backend_gpu;
}