llama_cpp 0.14.2 → 0.14.4

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -2234,6 +2234,11 @@ static ggml_backend_buffer_type_t ggml_backend_opencl_get_default_buffer_type(gg
2234
2234
  static ggml_status ggml_backend_opencl_graph_compute(ggml_backend_t backend, ggml_cgraph * graph) {
2235
2235
  for (int i = 0; i < graph->n_nodes; ++i) {
2236
2236
  ggml_tensor * node = graph->nodes[i];
2237
+
2238
+ if (ggml_is_empty(node)) {
2239
+ continue;
2240
+ }
2241
+
2237
2242
  switch (node->op) {
2238
2243
  case GGML_OP_MUL_MAT:
2239
2244
  ggml_cl_mul_mat(node->src[0], node->src[1], node, nullptr, 0);