@bike4mind/cli 0.2.33 → 0.2.34-feat-tokenizer-request-logger-context.19947
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/{chunk-YVYGYSOW.js → chunk-4WVX6KBG.js} +1 -1
- package/dist/{chunk-IKM2PF4K.js → chunk-CLVFA6GQ.js} +1 -1
- package/dist/{chunk-5T3QXFEQ.js → chunk-EDBDHBFK.js} +20 -8
- package/dist/{chunk-ZDUDUW2Z.js → chunk-PW7WH4H2.js} +6 -6
- package/dist/{chunk-X3UIP4R3.js → chunk-R7OTXEGV.js} +1 -1
- package/dist/commands/doctorCommand.js +1 -1
- package/dist/commands/updateCommand.js +1 -1
- package/dist/{create-36AXLGLT.js → create-4NLWCBVI.js} +2 -2
- package/dist/index.js +5 -5
- package/dist/{mementoService-SMXC3TKF.js → mementoService-HWVEQT2U.js} +2 -2
- package/dist/{src-JEDN6YEP.js → src-JVSIMKZW.js} +1 -1
- package/dist/{subtractCredits-5BME7E4T.js → subtractCredits-SBR2P3XN.js} +2 -2
- package/package.json +6 -6
|
@@ -12245,11 +12245,11 @@ var TiktokenTokenizer = class {
|
|
|
12245
12245
|
* @param modelId - Model ID to determine encoding (optional)
|
|
12246
12246
|
* @returns Promise<number> - Token count
|
|
12247
12247
|
*/
|
|
12248
|
-
async countTokens(text, modelId) {
|
|
12248
|
+
async countTokens(text, modelId, logger) {
|
|
12249
12249
|
if (this.isShuttingDown) {
|
|
12250
12250
|
throw new Error("TiktokenTokenizer is shutting down");
|
|
12251
12251
|
}
|
|
12252
|
-
const encoder = await this.getEncoder(modelId);
|
|
12252
|
+
const encoder = await this.getEncoder(modelId, logger);
|
|
12253
12253
|
const texts = Array.isArray(text) ? text : [text];
|
|
12254
12254
|
return texts.reduce((sum, t) => sum + encoder.encode(t).length, 0);
|
|
12255
12255
|
}
|
|
@@ -12259,18 +12259,30 @@ var TiktokenTokenizer = class {
|
|
|
12259
12259
|
* @param modelId - Model ID to determine encoding (optional)
|
|
12260
12260
|
* @returns Promise<number[]> - Array of token IDs
|
|
12261
12261
|
*/
|
|
12262
|
-
async encodeTokens(text, modelId) {
|
|
12262
|
+
async encodeTokens(text, modelId, logger) {
|
|
12263
12263
|
if (this.isShuttingDown) {
|
|
12264
12264
|
throw new Error("TiktokenTokenizer is shutting down");
|
|
12265
12265
|
}
|
|
12266
|
-
const encoder = await this.getEncoder(modelId);
|
|
12266
|
+
const encoder = await this.getEncoder(modelId, logger);
|
|
12267
12267
|
return Array.from(encoder.encode(text));
|
|
12268
12268
|
}
|
|
12269
|
+
/**
|
|
12270
|
+
* Returns a lightweight ITokenizer proxy that delegates WASM encoder operations
|
|
12271
|
+
* to this instance (preserving the shared encoder cache) but routes log output
|
|
12272
|
+
* through the provided logger. Useful for attaching per-request context (e.g.
|
|
12273
|
+
* requestId, userId) to tokenizer logs without sacrificing the singleton benefit.
|
|
12274
|
+
*/
|
|
12275
|
+
withLogger(logger) {
|
|
12276
|
+
return {
|
|
12277
|
+
countTokens: (text, modelId) => this.countTokens(text, modelId, logger),
|
|
12278
|
+
encodeTokens: (text, modelId) => this.encodeTokens(text, modelId, logger)
|
|
12279
|
+
};
|
|
12280
|
+
}
|
|
12269
12281
|
/**
|
|
12270
12282
|
* Get or create an encoder for the given model
|
|
12271
12283
|
* @private
|
|
12272
12284
|
*/
|
|
12273
|
-
async getEncoder(modelId) {
|
|
12285
|
+
async getEncoder(modelId, logger = this.logger) {
|
|
12274
12286
|
const { encoding_for_model, get_encoding } = await import("tiktoken");
|
|
12275
12287
|
const cacheKey = modelId || this.fallbackEncoding;
|
|
12276
12288
|
if (this.enableCaching && this.encoderCache.has(cacheKey)) {
|
|
@@ -12280,16 +12292,16 @@ var TiktokenTokenizer = class {
|
|
|
12280
12292
|
try {
|
|
12281
12293
|
if (modelId) {
|
|
12282
12294
|
encoder = encoding_for_model(modelId);
|
|
12283
|
-
|
|
12295
|
+
logger.debug(`Created tiktoken encoder for model: ${modelId}`);
|
|
12284
12296
|
} else {
|
|
12285
12297
|
encoder = get_encoding(this.fallbackEncoding);
|
|
12286
|
-
|
|
12298
|
+
logger.debug(`Created tiktoken encoder with ${this.fallbackEncoding} encoding`);
|
|
12287
12299
|
}
|
|
12288
12300
|
if (this.enableCaching) {
|
|
12289
12301
|
this.encoderCache.set(cacheKey, encoder);
|
|
12290
12302
|
}
|
|
12291
12303
|
} catch (error) {
|
|
12292
|
-
|
|
12304
|
+
logger.warn(`Failed to create encoder for model ${modelId}, falling back to ${this.fallbackEncoding}:`, error);
|
|
12293
12305
|
encoder = get_encoding(this.fallbackEncoding);
|
|
12294
12306
|
if (this.enableCaching) {
|
|
12295
12307
|
this.encoderCache.set(this.fallbackEncoding, encoder);
|
|
@@ -3,7 +3,7 @@
|
|
|
3
3
|
// package.json
|
|
4
4
|
var package_default = {
|
|
5
5
|
name: "@bike4mind/cli",
|
|
6
|
-
version: "0.2.
|
|
6
|
+
version: "0.2.34-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
7
7
|
type: "module",
|
|
8
8
|
description: "Interactive CLI tool for Bike4Mind with ReAct agents",
|
|
9
9
|
license: "UNLICENSED",
|
|
@@ -114,10 +114,10 @@ var package_default = {
|
|
|
114
114
|
},
|
|
115
115
|
devDependencies: {
|
|
116
116
|
"@bike4mind/agents": "0.1.0",
|
|
117
|
-
"@bike4mind/common": "2.55.
|
|
118
|
-
"@bike4mind/mcp": "1.32.
|
|
119
|
-
"@bike4mind/services": "2.52.
|
|
120
|
-
"@bike4mind/utils": "2.9.
|
|
117
|
+
"@bike4mind/common": "2.55.1-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
118
|
+
"@bike4mind/mcp": "1.32.3-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
119
|
+
"@bike4mind/services": "2.52.1-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
120
|
+
"@bike4mind/utils": "2.9.1-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
121
121
|
"@types/better-sqlite3": "^7.6.13",
|
|
122
122
|
"@types/diff": "^5.0.9",
|
|
123
123
|
"@types/jsonwebtoken": "^9.0.4",
|
|
@@ -138,7 +138,7 @@ var package_default = {
|
|
|
138
138
|
optionalDependencies: {
|
|
139
139
|
"@vscode/ripgrep": "^1.17.0"
|
|
140
140
|
},
|
|
141
|
-
gitHead: "
|
|
141
|
+
gitHead: "5fb63440bb63190e2e6af89f54a7904f70082a1f"
|
|
142
142
|
};
|
|
143
143
|
|
|
144
144
|
// src/utils/updateChecker.ts
|
package/dist/index.js
CHANGED
|
@@ -4,10 +4,10 @@ import {
|
|
|
4
4
|
getOpenWeatherKey,
|
|
5
5
|
getSerperKey,
|
|
6
6
|
getWolframAlphaKey
|
|
7
|
-
} from "./chunk-
|
|
7
|
+
} from "./chunk-R7OTXEGV.js";
|
|
8
8
|
import "./chunk-GQGOWACU.js";
|
|
9
|
-
import "./chunk-
|
|
10
|
-
import "./chunk-
|
|
9
|
+
import "./chunk-CLVFA6GQ.js";
|
|
10
|
+
import "./chunk-4WVX6KBG.js";
|
|
11
11
|
import {
|
|
12
12
|
BFLImageService,
|
|
13
13
|
BaseStorage,
|
|
@@ -19,7 +19,7 @@ import {
|
|
|
19
19
|
OpenAIBackend,
|
|
20
20
|
OpenAIImageService,
|
|
21
21
|
XAIImageService
|
|
22
|
-
} from "./chunk-
|
|
22
|
+
} from "./chunk-EDBDHBFK.js";
|
|
23
23
|
import "./chunk-BPFEGDC7.js";
|
|
24
24
|
import "./chunk-BDQBOLYG.js";
|
|
25
25
|
import {
|
|
@@ -29,7 +29,7 @@ import {
|
|
|
29
29
|
import {
|
|
30
30
|
checkForUpdate,
|
|
31
31
|
package_default
|
|
32
|
-
} from "./chunk-
|
|
32
|
+
} from "./chunk-PW7WH4H2.js";
|
|
33
33
|
import {
|
|
34
34
|
selectActiveBackgroundAgents,
|
|
35
35
|
useCliStore
|
package/package.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
|
1
1
|
{
|
|
2
2
|
"name": "@bike4mind/cli",
|
|
3
|
-
"version": "0.2.
|
|
3
|
+
"version": "0.2.34-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
4
4
|
"type": "module",
|
|
5
5
|
"description": "Interactive CLI tool for Bike4Mind with ReAct agents",
|
|
6
6
|
"license": "UNLICENSED",
|
|
@@ -111,10 +111,10 @@
|
|
|
111
111
|
},
|
|
112
112
|
"devDependencies": {
|
|
113
113
|
"@bike4mind/agents": "0.1.0",
|
|
114
|
-
"@bike4mind/common": "2.55.
|
|
115
|
-
"@bike4mind/mcp": "1.32.
|
|
116
|
-
"@bike4mind/services": "2.52.
|
|
117
|
-
"@bike4mind/utils": "2.9.
|
|
114
|
+
"@bike4mind/common": "2.55.1-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
115
|
+
"@bike4mind/mcp": "1.32.3-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
116
|
+
"@bike4mind/services": "2.52.1-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
117
|
+
"@bike4mind/utils": "2.9.1-feat-tokenizer-request-logger-context.19947+5fb63440b",
|
|
118
118
|
"@types/better-sqlite3": "^7.6.13",
|
|
119
119
|
"@types/diff": "^5.0.9",
|
|
120
120
|
"@types/jsonwebtoken": "^9.0.4",
|
|
@@ -135,5 +135,5 @@
|
|
|
135
135
|
"optionalDependencies": {
|
|
136
136
|
"@vscode/ripgrep": "^1.17.0"
|
|
137
137
|
},
|
|
138
|
-
"gitHead": "
|
|
138
|
+
"gitHead": "5fb63440bb63190e2e6af89f54a7904f70082a1f"
|
|
139
139
|
}
|