return;
}
- cl_mem mem = (cl_mem)tensor->data;
+ cl_mem mem = (cl_mem)tensor->extra;
clReleaseMemObject(mem);
}
size_t d_size;
cl_mem d_X = ggml_cl_pool_malloc(ne0 * sizeof(float), &x_size); // src0
- cl_mem d_Y = (cl_mem) src1->data; // src1 is already on device, broadcasted.
+ cl_mem d_Y = (cl_mem) src1->extra; // src1 is already on device, broadcasted.
cl_mem d_D = ggml_cl_pool_malloc(ne0 * sizeof(float), &d_size); // dst
size_t d_size;
cl_mem d_X;
if (src0->backend == GGML_BACKEND_GPU) { // NOLINT
- d_X = (cl_mem) src0->data;
+ d_X = (cl_mem) src0->extra;
} else {
d_X = ggml_cl_pool_malloc(sizeof(float) * x_ne, &x_size);
}
size_t d_size;
cl_mem d_X;
if (src0->backend == GGML_BACKEND_GPU) { // NOLINT
- d_X = (cl_mem) src0->data;
+ d_X = (cl_mem) src0->extra;
} else {
d_X = ggml_cl_pool_malloc(sizeof(ggml_fp16_t) * x_ne, &x_size);
}
events.emplace_back();
CL_CHECK(ggml_cl_h2d_tensor_2d(queue, d_Q, 0, src0, i03, i02, events.data() + ev_idx++));
} else if (src0->backend == GGML_BACKEND_GPU) {
- d_Q = (cl_mem) src0->data;
+ d_Q = (cl_mem) src0->extra;
} else {
GGML_ASSERT(false);
}
CL_CHECK(clFinish(queue));
- tensor->data = dst;
+ tensor->extra = dst;
GGML_ASSERT(tensor->backend == GGML_BACKEND_GPU);
}