@fugood/llama.node 1.2.0 → 1.2.2
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/package.json +16 -15
- package/src/llama.cpp/CMakeLists.txt +7 -0
- package/src/llama.cpp/common/arg.cpp +405 -221
- package/src/llama.cpp/common/chat.cpp +27 -15
- package/src/llama.cpp/common/common.h +23 -8
- package/src/llama.cpp/common/json-schema-to-grammar.cpp +7 -6
- package/src/llama.cpp/ggml/CMakeLists.txt +37 -22
- package/src/llama.cpp/ggml/include/ggml-metal.h +1 -0
- package/src/llama.cpp/ggml/include/ggml-zdnn.h +0 -2
- package/src/llama.cpp/ggml/include/ggml.h +4 -4
- package/src/llama.cpp/ggml/src/CMakeLists.txt +3 -0
- package/src/llama.cpp/ggml/src/ggml-cpu/amx/amx.cpp +4 -2
- package/src/llama.cpp/ggml/src/ggml-cpu/common.h +14 -0
- package/src/llama.cpp/ggml/src/ggml-cpu/ggml-cpu.cpp +1 -1
- package/src/llama.cpp/ggml/src/ggml-cpu/ops.cpp +74 -852
- package/src/llama.cpp/src/llama-arch.cpp +43 -10
- package/src/llama.cpp/src/llama-arch.h +8 -0
- package/src/llama.cpp/src/llama-chat.cpp +17 -0
- package/src/llama.cpp/src/llama-chat.h +1 -0
- package/src/llama.cpp/src/llama-context.cpp +8 -8
- package/src/llama.cpp/src/llama-cparams.h +1 -1
- package/src/llama.cpp/src/llama-graph.cpp +3 -3
- package/src/llama.cpp/src/llama-hparams.h +13 -3
- package/src/llama.cpp/src/llama-model.cpp +328 -44
- package/src/llama.cpp/src/llama-model.h +3 -0
- package/src/llama.cpp/src/llama-quant.cpp +3 -1
- package/src/llama.cpp/src/llama-vocab.cpp +13 -1
- package/src/llama.cpp/src/llama-vocab.h +1 -0
package/package.json
CHANGED
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@fugood/llama.node",
|
|
3
3
|
"access": "public",
|
|
4
|
-
"version": "1.2.
|
|
4
|
+
"version": "1.2.2",
|
|
5
5
|
"description": "An another Node binding of llama.cpp",
|
|
6
6
|
"main": "lib/index.js",
|
|
7
7
|
"scripts": {
|
|
@@ -72,19 +72,19 @@
|
|
|
72
72
|
"CMakeLists.txt"
|
|
73
73
|
],
|
|
74
74
|
"optionalDependencies": {
|
|
75
|
-
"@fugood/node-llama-linux-x64": "1.2.
|
|
76
|
-
"@fugood/node-llama-linux-x64-vulkan": "1.2.
|
|
77
|
-
"@fugood/node-llama-linux-x64-cuda": "1.2.
|
|
78
|
-
"@fugood/node-llama-linux-arm64": "1.2.
|
|
79
|
-
"@fugood/node-llama-linux-arm64-vulkan": "1.2.
|
|
80
|
-
"@fugood/node-llama-linux-arm64-cuda": "1.2.
|
|
81
|
-
"@fugood/node-llama-win32-x64": "1.2.
|
|
82
|
-
"@fugood/node-llama-win32-x64-vulkan": "1.2.
|
|
83
|
-
"@fugood/node-llama-win32-x64-cuda": "1.2.
|
|
84
|
-
"@fugood/node-llama-win32-arm64": "1.2.
|
|
85
|
-
"@fugood/node-llama-win32-arm64-vulkan": "1.2.
|
|
86
|
-
"@fugood/node-llama-darwin-x64": "1.2.
|
|
87
|
-
"@fugood/node-llama-darwin-arm64": "1.2.
|
|
75
|
+
"@fugood/node-llama-linux-x64": "1.2.2",
|
|
76
|
+
"@fugood/node-llama-linux-x64-vulkan": "1.2.2",
|
|
77
|
+
"@fugood/node-llama-linux-x64-cuda": "1.2.2",
|
|
78
|
+
"@fugood/node-llama-linux-arm64": "1.2.2",
|
|
79
|
+
"@fugood/node-llama-linux-arm64-vulkan": "1.2.2",
|
|
80
|
+
"@fugood/node-llama-linux-arm64-cuda": "1.2.2",
|
|
81
|
+
"@fugood/node-llama-win32-x64": "1.2.2",
|
|
82
|
+
"@fugood/node-llama-win32-x64-vulkan": "1.2.2",
|
|
83
|
+
"@fugood/node-llama-win32-x64-cuda": "1.2.2",
|
|
84
|
+
"@fugood/node-llama-win32-arm64": "1.2.2",
|
|
85
|
+
"@fugood/node-llama-win32-arm64-vulkan": "1.2.2",
|
|
86
|
+
"@fugood/node-llama-darwin-x64": "1.2.2",
|
|
87
|
+
"@fugood/node-llama-darwin-arm64": "1.2.2"
|
|
88
88
|
},
|
|
89
89
|
"devDependencies": {
|
|
90
90
|
"@babel/preset-env": "^7.24.4",
|
|
@@ -118,7 +118,8 @@
|
|
|
118
118
|
"**/*.test.ts"
|
|
119
119
|
],
|
|
120
120
|
"testPathIgnorePatterns": [
|
|
121
|
-
"<rootDir>/src/llama.rn/"
|
|
121
|
+
"<rootDir>/src/llama.rn/",
|
|
122
|
+
"<rootDir>/src/llama.cpp/"
|
|
122
123
|
]
|
|
123
124
|
},
|
|
124
125
|
"prettier": {
|
|
@@ -58,6 +58,12 @@ if (MSVC)
|
|
|
58
58
|
add_compile_options("$<$<COMPILE_LANGUAGE:CXX>:/bigobj>")
|
|
59
59
|
endif()
|
|
60
60
|
|
|
61
|
+
if (CMAKE_SYSTEM_NAME STREQUAL "iOS")
|
|
62
|
+
set(LLAMA_TOOLS_INSTALL_DEFAULT OFF)
|
|
63
|
+
else()
|
|
64
|
+
set(LLAMA_TOOLS_INSTALL_DEFAULT ${LLAMA_STANDALONE})
|
|
65
|
+
endif()
|
|
66
|
+
|
|
61
67
|
#
|
|
62
68
|
# option list
|
|
63
69
|
#
|
|
@@ -82,6 +88,7 @@ option(LLAMA_BUILD_TESTS "llama: build tests" ${LLAMA_STANDALONE})
|
|
|
82
88
|
option(LLAMA_BUILD_TOOLS "llama: build tools" ${LLAMA_STANDALONE})
|
|
83
89
|
option(LLAMA_BUILD_EXAMPLES "llama: build examples" ${LLAMA_STANDALONE})
|
|
84
90
|
option(LLAMA_BUILD_SERVER "llama: build server example" ${LLAMA_STANDALONE})
|
|
91
|
+
option(LLAMA_TOOLS_INSTALL "llama: install tools" ${LLAMA_TOOLS_INSTALL_DEFAULT})
|
|
85
92
|
|
|
86
93
|
# 3rd party libs
|
|
87
94
|
option(LLAMA_CURL "llama: use libcurl to download model from an URL" ON)
|