cui-llama.rn 1.3.6 → 1.4.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +22 -1
- package/android/src/main/CMakeLists.txt +25 -26
- package/android/src/main/java/com/rnllama/LlamaContext.java +31 -9
- package/android/src/main/java/com/rnllama/RNLlama.java +98 -0
- package/android/src/main/jni-utils.h +94 -0
- package/android/src/main/jni.cpp +133 -63
- package/android/src/newarch/java/com/rnllama/RNLlamaModule.java +15 -0
- package/android/src/oldarch/java/com/rnllama/RNLlamaModule.java +15 -0
- package/cpp/common.cpp +2085 -1982
- package/cpp/common.h +696 -664
- package/cpp/ggml-alloc.c +1042 -1037
- package/cpp/ggml-backend-impl.h +255 -256
- package/cpp/ggml-backend-reg.cpp +582 -582
- package/cpp/ggml-backend.cpp +2002 -2002
- package/cpp/ggml-backend.h +354 -352
- package/cpp/ggml-common.h +1853 -1853
- package/cpp/ggml-cpp.h +39 -39
- package/cpp/ggml-cpu-aarch64.cpp +4247 -4247
- package/cpp/ggml-cpu-aarch64.h +8 -8
- package/cpp/ggml-cpu-impl.h +386 -386
- package/cpp/ggml-cpu-quants.c +10920 -10839
- package/cpp/ggml-cpu-traits.cpp +36 -36
- package/cpp/ggml-cpu-traits.h +38 -38
- package/cpp/ggml-cpu.c +14391 -14122
- package/cpp/ggml-cpu.cpp +635 -627
- package/cpp/ggml-cpu.h +135 -135
- package/cpp/ggml-impl.h +567 -567
- package/cpp/ggml-metal-impl.h +288 -0
- package/cpp/ggml-metal.m +4884 -4884
- package/cpp/ggml-opt.cpp +854 -0
- package/cpp/ggml-opt.h +216 -0
- package/cpp/ggml-quants.c +5238 -5238
- package/cpp/ggml-threading.h +14 -14
- package/cpp/ggml.c +6514 -6448
- package/cpp/ggml.h +2194 -2163
- package/cpp/gguf.cpp +1329 -1325
- package/cpp/gguf.h +202 -202
- package/cpp/json-schema-to-grammar.cpp +1045 -1045
- package/cpp/json-schema-to-grammar.h +8 -8
- package/cpp/json.hpp +24766 -24766
- package/cpp/llama-adapter.cpp +347 -346
- package/cpp/llama-adapter.h +74 -73
- package/cpp/llama-arch.cpp +1487 -1434
- package/cpp/llama-arch.h +400 -395
- package/cpp/llama-batch.cpp +368 -368
- package/cpp/llama-batch.h +88 -88
- package/cpp/llama-chat.cpp +578 -567
- package/cpp/llama-chat.h +52 -51
- package/cpp/llama-context.cpp +1775 -1771
- package/cpp/llama-context.h +128 -128
- package/cpp/llama-cparams.cpp +1 -1
- package/cpp/llama-cparams.h +37 -37
- package/cpp/llama-cpp.h +30 -30
- package/cpp/llama-grammar.cpp +1139 -1139
- package/cpp/llama-grammar.h +143 -143
- package/cpp/llama-hparams.cpp +71 -71
- package/cpp/llama-hparams.h +139 -140
- package/cpp/llama-impl.cpp +167 -167
- package/cpp/llama-impl.h +61 -61
- package/cpp/llama-kv-cache.cpp +718 -718
- package/cpp/llama-kv-cache.h +218 -218
- package/cpp/llama-mmap.cpp +590 -589
- package/cpp/llama-mmap.h +67 -67
- package/cpp/llama-model-loader.cpp +1124 -1011
- package/cpp/llama-model-loader.h +167 -158
- package/cpp/llama-model.cpp +3997 -2202
- package/cpp/llama-model.h +370 -391
- package/cpp/llama-sampling.cpp +2408 -2406
- package/cpp/llama-sampling.h +32 -48
- package/cpp/llama-vocab.cpp +3247 -1982
- package/cpp/llama-vocab.h +125 -182
- package/cpp/llama.cpp +10077 -12544
- package/cpp/llama.h +1323 -1285
- package/cpp/log.cpp +401 -401
- package/cpp/log.h +121 -121
- package/cpp/rn-llama.hpp +123 -116
- package/cpp/sampling.cpp +505 -500
- package/cpp/sgemm.cpp +2597 -2597
- package/cpp/sgemm.h +14 -14
- package/cpp/speculative.cpp +277 -274
- package/cpp/speculative.h +28 -28
- package/cpp/unicode.cpp +2 -3
- package/ios/RNLlama.mm +47 -0
- package/ios/RNLlamaContext.h +3 -1
- package/ios/RNLlamaContext.mm +71 -14
- package/jest/mock.js +15 -3
- package/lib/commonjs/NativeRNLlama.js.map +1 -1
- package/lib/commonjs/index.js +33 -37
- package/lib/commonjs/index.js.map +1 -1
- package/lib/module/NativeRNLlama.js.map +1 -1
- package/lib/module/index.js +31 -35
- package/lib/module/index.js.map +1 -1
- package/lib/typescript/NativeRNLlama.d.ts +26 -6
- package/lib/typescript/NativeRNLlama.d.ts.map +1 -1
- package/lib/typescript/index.d.ts +21 -36
- package/lib/typescript/index.d.ts.map +1 -1
- package/llama-rn.podspec +4 -18
- package/package.json +2 -3
- package/src/NativeRNLlama.ts +32 -13
- package/src/index.ts +52 -47
- package/cpp/llama.cpp.rej +0 -23
package/cpp/llama-chat.h
CHANGED
@@ -1,51 +1,52 @@
|
|
1
|
-
#pragma once
|
2
|
-
|
3
|
-
#include <string>
|
4
|
-
#include <vector>
|
5
|
-
#include <cstdint>
|
6
|
-
|
7
|
-
enum llm_chat_template {
|
8
|
-
LLM_CHAT_TEMPLATE_CHATML,
|
9
|
-
LLM_CHAT_TEMPLATE_LLAMA_2,
|
10
|
-
LLM_CHAT_TEMPLATE_LLAMA_2_SYS,
|
11
|
-
LLM_CHAT_TEMPLATE_LLAMA_2_SYS_BOS,
|
12
|
-
LLM_CHAT_TEMPLATE_LLAMA_2_SYS_STRIP,
|
13
|
-
LLM_CHAT_TEMPLATE_MISTRAL_V1,
|
14
|
-
LLM_CHAT_TEMPLATE_MISTRAL_V3,
|
15
|
-
LLM_CHAT_TEMPLATE_MISTRAL_V3_TEKKEN,
|
16
|
-
LLM_CHAT_TEMPLATE_MISTRAL_V7,
|
17
|
-
LLM_CHAT_TEMPLATE_PHI_3,
|
18
|
-
|
19
|
-
|
20
|
-
|
21
|
-
|
22
|
-
|
23
|
-
|
24
|
-
|
25
|
-
|
26
|
-
|
27
|
-
|
28
|
-
|
29
|
-
|
30
|
-
|
31
|
-
|
32
|
-
|
33
|
-
|
34
|
-
|
35
|
-
|
36
|
-
|
37
|
-
|
38
|
-
|
39
|
-
|
40
|
-
|
41
|
-
|
42
|
-
|
43
|
-
|
44
|
-
|
45
|
-
|
46
|
-
|
47
|
-
|
48
|
-
|
49
|
-
|
50
|
-
|
51
|
-
std::
|
1
|
+
#pragma once
|
2
|
+
|
3
|
+
#include <string>
|
4
|
+
#include <vector>
|
5
|
+
#include <cstdint>
|
6
|
+
|
7
|
+
enum llm_chat_template {
|
8
|
+
LLM_CHAT_TEMPLATE_CHATML,
|
9
|
+
LLM_CHAT_TEMPLATE_LLAMA_2,
|
10
|
+
LLM_CHAT_TEMPLATE_LLAMA_2_SYS,
|
11
|
+
LLM_CHAT_TEMPLATE_LLAMA_2_SYS_BOS,
|
12
|
+
LLM_CHAT_TEMPLATE_LLAMA_2_SYS_STRIP,
|
13
|
+
LLM_CHAT_TEMPLATE_MISTRAL_V1,
|
14
|
+
LLM_CHAT_TEMPLATE_MISTRAL_V3,
|
15
|
+
LLM_CHAT_TEMPLATE_MISTRAL_V3_TEKKEN,
|
16
|
+
LLM_CHAT_TEMPLATE_MISTRAL_V7,
|
17
|
+
LLM_CHAT_TEMPLATE_PHI_3,
|
18
|
+
LLM_CHAT_TEMPLATE_PHI_4,
|
19
|
+
LLM_CHAT_TEMPLATE_FALCON_3,
|
20
|
+
LLM_CHAT_TEMPLATE_ZEPHYR,
|
21
|
+
LLM_CHAT_TEMPLATE_MONARCH,
|
22
|
+
LLM_CHAT_TEMPLATE_GEMMA,
|
23
|
+
LLM_CHAT_TEMPLATE_ORION,
|
24
|
+
LLM_CHAT_TEMPLATE_OPENCHAT,
|
25
|
+
LLM_CHAT_TEMPLATE_VICUNA,
|
26
|
+
LLM_CHAT_TEMPLATE_VICUNA_ORCA,
|
27
|
+
LLM_CHAT_TEMPLATE_DEEPSEEK,
|
28
|
+
LLM_CHAT_TEMPLATE_DEEPSEEK_2,
|
29
|
+
LLM_CHAT_TEMPLATE_DEEPSEEK_3,
|
30
|
+
LLM_CHAT_TEMPLATE_COMMAND_R,
|
31
|
+
LLM_CHAT_TEMPLATE_LLAMA_3,
|
32
|
+
LLM_CHAT_TEMPLATE_CHATGML_3,
|
33
|
+
LLM_CHAT_TEMPLATE_CHATGML_4,
|
34
|
+
LLM_CHAT_TEMPLATE_MINICPM,
|
35
|
+
LLM_CHAT_TEMPLATE_EXAONE_3,
|
36
|
+
LLM_CHAT_TEMPLATE_RWKV_WORLD,
|
37
|
+
LLM_CHAT_TEMPLATE_GRANITE,
|
38
|
+
LLM_CHAT_TEMPLATE_GIGACHAT,
|
39
|
+
LLM_CHAT_TEMPLATE_MEGREZ,
|
40
|
+
LLM_CHAT_TEMPLATE_UNKNOWN,
|
41
|
+
};
|
42
|
+
|
43
|
+
struct llama_chat_message;
|
44
|
+
|
45
|
+
llm_chat_template llm_chat_template_from_str(const std::string & name);
|
46
|
+
|
47
|
+
llm_chat_template llm_chat_detect_template(const std::string & tmpl);
|
48
|
+
|
49
|
+
int32_t llm_chat_apply_template(
|
50
|
+
llm_chat_template tmpl,
|
51
|
+
const std::vector<const llama_chat_message *> & chat,
|
52
|
+
std::string & dest, bool add_ass);
|