llama_cpp 0.15.2 → 0.15.3

Sign up to get free protection for your applications and to get access to all the features.
@@ -1835,7 +1835,10 @@ static void ggml_cl_mul_mat_q_f32(const ggml_tensor * src0, const ggml_tensor *
1835
1835
  CL_CHECK(clEnqueueNDRangeKernel(queue, *to_fp32_cl, 1, &offset, &global, local > 0 ? &local : NULL, events.size(), !events.empty() ? events.data() : NULL, NULL));
1836
1836
  }
1837
1837
 
1838
- for (int64_t i12 = i02 * r2, e12 = i12 + r2; i12 < e12; i12++) {
1838
+ int64_t i12 = i02 * r2;
1839
+ int64_t e12 = i12 + r2;
1840
+ events.reserve(e12 - i12);
1841
+ for (; i12 < e12; i12++) {
1839
1842
  if (mul_mat_vec) { // specialized dequantize_mul_mat_vec kernel
1840
1843
  // copy src1 to device
1841
1844
  events.emplace_back();