llama_cpp 0.14.2 → 0.14.4

Sign up to get free protection for your applications and to get access to all the features.
@@ -214,9 +214,10 @@
214
214
  # define GGML_ATTRIBUTE_FORMAT(...) __attribute__((format(printf, __VA_ARGS__)))
215
215
  #endif
216
216
 
217
- #include <stdint.h>
218
- #include <stddef.h>
219
217
  #include <stdbool.h>
218
+ #include <stddef.h>
219
+ #include <stdint.h>
220
+ #include <stdio.h>
220
221
 
221
222
  #define GGML_FILE_MAGIC 0x67676d6c // "ggml"
222
223
  #define GGML_FILE_VERSION 1
@@ -368,6 +369,7 @@ extern "C" {
368
369
  GGML_TYPE_I32 = 26,
369
370
  GGML_TYPE_I64 = 27,
370
371
  GGML_TYPE_F64 = 28,
372
+ GGML_TYPE_IQ1_M = 29,
371
373
  GGML_TYPE_COUNT,
372
374
  };
373
375
 
@@ -407,6 +409,7 @@ extern "C" {
407
409
  GGML_FTYPE_MOSTLY_IQ3_S = 20, // except 1d tensors
408
410
  GGML_FTYPE_MOSTLY_IQ2_S = 21, // except 1d tensors
409
411
  GGML_FTYPE_MOSTLY_IQ4_XS = 22, // except 1d tensors
412
+ GGML_FTYPE_MOSTLY_IQ1_M = 23, // except 1d tensors
410
413
  };
411
414
 
412
415
  // available tensor operations:
@@ -708,6 +711,9 @@ extern "C" {
708
711
 
709
712
  GGML_API void ggml_print_backtrace(void);
710
713
 
714
+ // accepts a UTF-8 path, even on Windows
715
+ GGML_API FILE * ggml_fopen(const char * fname, const char * mode);
716
+
711
717
  GGML_API void ggml_numa_init(enum ggml_numa_strategy numa); // call once for better performance on NUMA systems
712
718
  GGML_API bool ggml_is_numa(void); // true if init detected that system has >1 NUMA node
713
719
 
@@ -744,6 +750,7 @@ extern "C" {
744
750
  GGML_API GGML_CALL bool ggml_is_transposed(const struct ggml_tensor * tensor);
745
751
  GGML_API GGML_CALL bool ggml_is_contiguous(const struct ggml_tensor * tensor);
746
752
  GGML_API GGML_CALL bool ggml_is_permuted (const struct ggml_tensor * tensor);
753
+ GGML_API GGML_CALL bool ggml_is_empty (const struct ggml_tensor * tensor);
747
754
  GGML_API bool ggml_is_scalar (const struct ggml_tensor * tensor);
748
755
  GGML_API bool ggml_is_vector (const struct ggml_tensor * tensor);
749
756
  GGML_API bool ggml_is_matrix (const struct ggml_tensor * tensor);
@@ -1157,8 +1164,7 @@ extern "C" {
1157
1164
  // ggml_mul_mat_id(ctx, as, ids, id, b) ~= ggml_mul_mat(as[ids[id]], b)
1158
1165
  GGML_API struct ggml_tensor * ggml_mul_mat_id(
1159
1166
  struct ggml_context * ctx,
1160
- struct ggml_tensor * const as[],
1161
- int n_as,
1167
+ struct ggml_tensor * as,
1162
1168
  struct ggml_tensor * ids,
1163
1169
  int id,
1164
1170
  struct ggml_tensor * b);
@@ -2350,7 +2356,7 @@ extern "C" {
2350
2356
  GGML_API int ggml_cpu_has_fp16_va (void);
2351
2357
  GGML_API int ggml_cpu_has_wasm_simd (void);
2352
2358
  GGML_API int ggml_cpu_has_blas (void);
2353
- GGML_API int ggml_cpu_has_cublas (void);
2359
+ GGML_API int ggml_cpu_has_cuda (void);
2354
2360
  GGML_API int ggml_cpu_has_clblast (void);
2355
2361
  GGML_API int ggml_cpu_has_vulkan (void);
2356
2362
  GGML_API int ggml_cpu_has_kompute (void);