llama_cpp 0.15.1 → 0.15.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1835,7 +1835,10 @@ static void ggml_cl_mul_mat_q_f32(const ggml_tensor * src0, const ggml_tensor *
1835
1835
  CL_CHECK(clEnqueueNDRangeKernel(queue, *to_fp32_cl, 1, &offset, &global, local > 0 ? &local : NULL, events.size(), !events.empty() ? events.data() : NULL, NULL));
1836
1836
  }
1837
1837
 
1838
- for (int64_t i12 = i02 * r2, e12 = i12 + r2; i12 < e12; i12++) {
1838
+ int64_t i12 = i02 * r2;
1839
+ int64_t e12 = i12 + r2;
1840
+ events.reserve(e12 - i12);
1841
+ for (; i12 < e12; i12++) {
1839
1842
  if (mul_mat_vec) { // specialized dequantize_mul_mat_vec kernel
1840
1843
  // copy src1 to device
1841
1844
  events.emplace_back();