node-llama-cpp 3.0.2 → 3.0.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -5,6 +5,7 @@
5
5
  AddonThreadSafeLogCallbackFunction addonThreadSafeLoggerCallback;
6
6
  bool addonJsLoggerCallbackSet = false;
7
7
  int addonLoggerLogLevel = 5;
8
+ int addonLastLoggerLogLevel = 6;
8
9
 
9
10
  static int addonGetGgmlLogLevelNumber(ggml_log_level level) {
10
11
  switch (level) {
@@ -13,6 +14,7 @@ static int addonGetGgmlLogLevelNumber(ggml_log_level level) {
13
14
  case GGML_LOG_LEVEL_INFO: return 4;
14
15
  case GGML_LOG_LEVEL_NONE: return 5;
15
16
  case GGML_LOG_LEVEL_DEBUG: return 6;
17
+ case GGML_LOG_LEVEL_CONT: return addonLastLoggerLogLevel;
16
18
  }
17
19
 
18
20
  return 1;
@@ -53,6 +55,7 @@ void addonCallJsLogCallback(
53
55
 
54
56
  void addonLlamaCppLogCallback(ggml_log_level level, const char* text, void* user_data) {
55
57
  int logLevelNumber = addonGetGgmlLogLevelNumber(level);
58
+ addonLastLoggerLogLevel = logLevelNumber;
56
59
 
57
60
  if (logLevelNumber > addonLoggerLogLevel) {
58
61
  return;
@@ -18,4 +18,4 @@ using AddonThreadSafeLogCallbackFunction =
18
18
  Napi::TypedThreadSafeFunction<AddonThreadSafeLogCallbackFunctionContext, addon_logger_log, addonCallJsLogCallback>;
19
19
 
20
20
  Napi::Value setLogger(const Napi::CallbackInfo& info);
21
- Napi::Value setLoggerLogLevel(const Napi::CallbackInfo& info);
21
+ Napi::Value setLoggerLogLevel(const Napi::CallbackInfo& info);
@@ -1,3 +1,3 @@
1
1
  {
2
- "release": "b3821"
2
+ "release": "b3825"
3
3
  }
Binary file
@@ -1,4 +1,4 @@
1
1
  {
2
- "tag": "b3821",
2
+ "tag": "b3825",
3
3
  "llamaCppGithubRepo": "ggerganov/llama.cpp"
4
4
  }
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "node-llama-cpp",
3
- "version": "3.0.2",
3
+ "version": "3.0.3",
4
4
  "description": "Run AI models locally on your machine with node.js bindings for llama.cpp. Enforce a JSON schema on the model output on the generation level",
5
5
  "main": "./dist/index.js",
6
6
  "type": "module",
@@ -131,13 +131,14 @@
131
131
  "@nolebase/vitepress-plugin-og-image": "^2.5.0",
132
132
  "@resvg/resvg-js": "^2.6.2",
133
133
  "@semantic-release/exec": "^6.0.3",
134
+ "@semantic-release/github": "11.0.0",
134
135
  "@semantic-release/npm": "12.0.1",
135
136
  "@shikijs/vitepress-twoslash": "^1.18.0",
136
137
  "@types/async-retry": "^1.4.8",
137
138
  "@types/bytes": "^3.1.4",
138
139
  "@types/cross-spawn": "^6.0.2",
139
140
  "@types/fs-extra": "^11.0.4",
140
- "@types/node": "^22.5.5",
141
+ "@types/node": "^22.7.0",
141
142
  "@types/proper-lockfile": "^4.1.4",
142
143
  "@types/semver": "^7.5.8",
143
144
  "@types/validate-npm-package-name": "^4.0.2",
@@ -159,8 +160,8 @@
159
160
  "sharp": "^0.33.5",
160
161
  "tslib": "^2.7.0",
161
162
  "typedoc": "^0.26.7",
162
- "typedoc-plugin-markdown": "^4.2.7",
163
- "typedoc-plugin-mdn-links": "^3.3.0",
163
+ "typedoc-plugin-markdown": "^4.2.8",
164
+ "typedoc-plugin-mdn-links": "^3.3.1",
164
165
  "typedoc-vitepress-theme": "^1.0.1",
165
166
  "typescript": "^5.6.2",
166
167
  "vite-node": "^2.1.1",
@@ -181,7 +182,7 @@
181
182
  "filenamify": "^6.0.0",
182
183
  "fs-extra": "^11.2.0",
183
184
  "ignore": "^5.3.2",
184
- "ipull": "^3.7.2",
185
+ "ipull": "^3.7.4",
185
186
  "is-unicode-supported": "^2.1.0",
186
187
  "lifecycle-utils": "^1.7.0",
187
188
  "log-symbols": "^7.0.0",
@@ -209,16 +210,16 @@
209
210
  }
210
211
  },
211
212
  "optionalDependencies": {
212
- "@node-llama-cpp/linux-arm64": "3.0.2",
213
- "@node-llama-cpp/linux-armv7l": "3.0.2",
214
- "@node-llama-cpp/linux-x64": "3.0.2",
215
- "@node-llama-cpp/linux-x64-cuda": "3.0.2",
216
- "@node-llama-cpp/linux-x64-vulkan": "3.0.2",
217
- "@node-llama-cpp/mac-arm64-metal": "3.0.2",
218
- "@node-llama-cpp/mac-x64": "3.0.2",
219
- "@node-llama-cpp/win-arm64": "3.0.2",
220
- "@node-llama-cpp/win-x64": "3.0.2",
221
- "@node-llama-cpp/win-x64-cuda": "3.0.2",
222
- "@node-llama-cpp/win-x64-vulkan": "3.0.2"
213
+ "@node-llama-cpp/linux-arm64": "3.0.3",
214
+ "@node-llama-cpp/linux-armv7l": "3.0.3",
215
+ "@node-llama-cpp/linux-x64": "3.0.3",
216
+ "@node-llama-cpp/linux-x64-cuda": "3.0.3",
217
+ "@node-llama-cpp/linux-x64-vulkan": "3.0.3",
218
+ "@node-llama-cpp/mac-arm64-metal": "3.0.3",
219
+ "@node-llama-cpp/mac-x64": "3.0.3",
220
+ "@node-llama-cpp/win-arm64": "3.0.3",
221
+ "@node-llama-cpp/win-x64": "3.0.3",
222
+ "@node-llama-cpp/win-x64-cuda": "3.0.3",
223
+ "@node-llama-cpp/win-x64-vulkan": "3.0.3"
223
224
  }
224
225
  }