llama_cpp 0.1.1 → 0.1.3
Sign up to get free protection for your applications and to get access to all the features.
- checksums.yaml +4 -4
- data/CHANGELOG.md +11 -0
- data/ext/llama_cpp/extconf.rb +8 -2
- data/ext/llama_cpp/llama_cpp.cpp +60 -6
- data/ext/llama_cpp/src/ggml-cuda.h +2 -0
- data/ext/llama_cpp/src/ggml-opencl.cpp +1034 -0
- data/ext/llama_cpp/src/ggml-opencl.h +8 -10
- data/ext/llama_cpp/src/ggml.c +398 -184
- data/ext/llama_cpp/src/ggml.h +14 -3
- data/ext/llama_cpp/src/llama-util.h +23 -23
- data/ext/llama_cpp/src/llama.cpp +191 -92
- data/ext/llama_cpp/src/llama.h +30 -17
- data/lib/llama_cpp/version.rb +2 -2
- data/lib/llama_cpp.rb +2 -0
- data/sig/llama_cpp.rbs +1 -0
- metadata +3 -3
- data/ext/llama_cpp/src/ggml-opencl.c +0 -361
@@ -1,23 +1,21 @@
|
|
1
1
|
#pragma once
|
2
2
|
|
3
|
+
#include "ggml.h"
|
4
|
+
|
3
5
|
#ifdef __cplusplus
|
4
6
|
extern "C" {
|
5
7
|
#endif
|
6
8
|
|
7
9
|
void ggml_cl_init(void);
|
8
10
|
|
9
|
-
|
10
|
-
|
11
|
-
|
12
|
-
};
|
11
|
+
bool ggml_cl_can_mul_mat(const struct ggml_tensor * src0, const struct ggml_tensor * src1, struct ggml_tensor * dst);
|
12
|
+
size_t ggml_cl_mul_mat_get_wsize(const struct ggml_tensor * src0, const struct ggml_tensor * src1, struct ggml_tensor * dst);
|
13
|
+
void ggml_cl_mul_mat(const struct ggml_tensor * src0, const struct ggml_tensor * src1, struct ggml_tensor * dst, void * wdata, size_t wsize);
|
13
14
|
|
14
|
-
|
15
|
-
|
16
|
-
GGML_BLAS_OP_T = 112,
|
17
|
-
GGML_BLAS_OP_C = 113,
|
18
|
-
};
|
15
|
+
void * ggml_cl_host_malloc(size_t size);
|
16
|
+
void ggml_cl_host_free(void * ptr);
|
19
17
|
|
20
|
-
void
|
18
|
+
void ggml_cl_transform_tensor(struct ggml_tensor * tensor);
|
21
19
|
|
22
20
|
#ifdef __cplusplus
|
23
21
|
}
|