};
struct ggml_context * ctx0 = ggml_init(params);
- struct ggml_cgraph gf = { .n_threads = n_threads };
+ struct ggml_cgraph gf = {};
+ gf.n_threads = n_threads;
struct ggml_tensor * embd = ggml_new_tensor_1d(ctx0, GGML_TYPE_I32, N);
memcpy(embd->data, embd_inp.data(), N*ggml_element_size(embd));
};
struct ggml_context * ctx0 = ggml_init(params);
- struct ggml_cgraph gf = { .n_threads = n_threads };
+ struct ggml_cgraph gf = {};
+ gf.n_threads = n_threads;
struct ggml_tensor * embd = ggml_new_tensor_1d(ctx0, GGML_TYPE_I32, N);
memcpy(embd->data, embd_inp.data(), N*ggml_element_size(embd));
};
struct ggml_context * ctx0 = ggml_init(params);
- struct ggml_cgraph gf = { .n_threads = n_threads };
+ struct ggml_cgraph gf = {};
+ gf.n_threads = n_threads;
struct ggml_tensor * input = ggml_new_tensor_1d(ctx0, GGML_TYPE_F32, hparams.n_input);
memcpy(input->data, digit.data(), ggml_nbytes(input));
};
struct ggml_context * ctx0 = ggml_init(params);
- struct ggml_cgraph gf = { .n_threads = n_threads };
+ struct ggml_cgraph gf = {};
+ gf.n_threads = n_threads;
struct ggml_tensor * embd = ggml_new_tensor_1d(ctx0, GGML_TYPE_I32, N);
memcpy(embd->data, embd_inp.data(), N*ggml_element_size(embd));