llama_cpp 0.14.2 → 0.14.4

Sign up to get free protection for your applications and to get access to all the features.
@@ -2234,6 +2234,11 @@ static ggml_backend_buffer_type_t ggml_backend_opencl_get_default_buffer_type(gg
2234
2234
  static ggml_status ggml_backend_opencl_graph_compute(ggml_backend_t backend, ggml_cgraph * graph) {
2235
2235
  for (int i = 0; i < graph->n_nodes; ++i) {
2236
2236
  ggml_tensor * node = graph->nodes[i];
2237
+
2238
+ if (ggml_is_empty(node)) {
2239
+ continue;
2240
+ }
2241
+
2237
2242
  switch (node->op) {
2238
2243
  case GGML_OP_MUL_MAT:
2239
2244
  ggml_cl_mul_mat(node->src[0], node->src[1], node, nullptr, 0);