@bike4mind/cli 0.2.53-refactor-slack-remove-dead-code-simplify-prompts.21212 → 0.2.53

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -1,7 +1,7 @@
1
1
  #!/usr/bin/env node
2
2
  import {
3
3
  CurationArtifactType
4
- } from "./chunk-QXNNUB2O.js";
4
+ } from "./chunk-KHFMOKEM.js";
5
5
 
6
6
  // ../../b4m-core/packages/services/dist/src/notebookCurationService/artifactExtractor.js
7
7
  var ARTIFACT_TAG_REGEX = /<artifact\s+(.*?)>([\s\S]*?)<\/artifact>/gi;
@@ -3,7 +3,7 @@
3
3
  // package.json
4
4
  var package_default = {
5
5
  name: "@bike4mind/cli",
6
- version: "0.2.53-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
6
+ version: "0.2.53",
7
7
  type: "module",
8
8
  description: "Interactive CLI tool for Bike4Mind with ReAct agents",
9
9
  license: "UNLICENSED",
@@ -117,11 +117,11 @@ var package_default = {
117
117
  zustand: "^4.5.4"
118
118
  },
119
119
  devDependencies: {
120
- "@bike4mind/agents": "0.1.4-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
121
- "@bike4mind/common": "2.69.1-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
122
- "@bike4mind/mcp": "1.33.14-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
123
- "@bike4mind/services": "2.64.1-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
124
- "@bike4mind/utils": "2.15.7-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
120
+ "@bike4mind/agents": "0.1.4",
121
+ "@bike4mind/common": "2.69.0",
122
+ "@bike4mind/mcp": "1.33.13",
123
+ "@bike4mind/services": "2.64.1",
124
+ "@bike4mind/utils": "2.15.7",
125
125
  "@types/better-sqlite3": "^7.6.13",
126
126
  "@types/diff": "^8.0.0",
127
127
  "@types/jsonwebtoken": "^9.0.4",
@@ -139,7 +139,7 @@ var package_default = {
139
139
  optionalDependencies: {
140
140
  "@vscode/ripgrep": "^1.17.0"
141
141
  },
142
- gitHead: "a1c57624ba358a7183a96a10c046bd8338a97090"
142
+ gitHead: "5ebeac1702534ebc59e39d53f778f1969fc077af"
143
143
  };
144
144
 
145
145
  // src/utils/updateChecker.ts
@@ -6,12 +6,12 @@ import {
6
6
  getSettingsByNames,
7
7
  obfuscateApiKey,
8
8
  secureParameters
9
- } from "./chunk-ZXBRRAVW.js";
9
+ } from "./chunk-HG3KOSRI.js";
10
10
  import {
11
11
  ApiKeyType,
12
12
  MementoTier,
13
13
  isSupportedEmbeddingModel
14
- } from "./chunk-QXNNUB2O.js";
14
+ } from "./chunk-KHFMOKEM.js";
15
15
 
16
16
  // ../../b4m-core/packages/services/dist/src/apiKeyService/get.js
17
17
  import { z } from "zod";
@@ -7,11 +7,11 @@ import {
7
7
  getSettingsMap,
8
8
  getSettingsValue,
9
9
  secureParameters
10
- } from "./chunk-ZXBRRAVW.js";
10
+ } from "./chunk-HG3KOSRI.js";
11
11
  import {
12
12
  KnowledgeType,
13
13
  SupportedFabFileMimeTypes
14
- } from "./chunk-QXNNUB2O.js";
14
+ } from "./chunk-KHFMOKEM.js";
15
15
 
16
16
  // ../../b4m-core/packages/services/dist/src/fabFileService/create.js
17
17
  import { z } from "zod";
@@ -4,7 +4,7 @@ import {
4
4
  getOpenWeatherKey,
5
5
  getSerperKey,
6
6
  getWolframAlphaKey
7
- } from "./chunk-L46CVMQB.js";
7
+ } from "./chunk-BAUVNB3G.js";
8
8
  import {
9
9
  BFLImageService,
10
10
  BaseStorage,
@@ -16,14 +16,14 @@ import {
16
16
  OpenAIBackend,
17
17
  OpenAIImageService,
18
18
  XAIImageService
19
- } from "./chunk-ZXBRRAVW.js";
19
+ } from "./chunk-HG3KOSRI.js";
20
20
  import {
21
21
  Logger
22
22
  } from "./chunk-PFBYGCOW.js";
23
23
  import {
24
24
  ConfigStore,
25
25
  logger
26
- } from "./chunk-522LM44T.js";
26
+ } from "./chunk-OJSNMX74.js";
27
27
  import {
28
28
  ALERT_THRESHOLDS,
29
29
  AiEvents,
@@ -84,7 +84,7 @@ import {
84
84
  getViewById,
85
85
  resolveNavigationIntents,
86
86
  sanitizeTelemetryError
87
- } from "./chunk-QXNNUB2O.js";
87
+ } from "./chunk-KHFMOKEM.js";
88
88
 
89
89
  // src/utils/fileSearch.ts
90
90
  import * as fs from "fs";
@@ -13971,9 +13971,7 @@ var QuestStartBodySchema = z141.object({
13971
13971
  /** User's timezone (IANA format, e.g., "America/New_York") */
13972
13972
  timezone: z141.string().optional(),
13973
13973
  /** Persona-based sub-agent filter — only these agent names are available for delegation */
13974
- allowedAgents: z141.array(z141.string()).optional(),
13975
- /** When true, Quest Processor injects Slack-specific tool configs (help, notebooks, curated files) */
13976
- enableSlackTools: z141.boolean().optional()
13974
+ allowedAgents: z141.array(z141.string()).optional()
13977
13975
  });
13978
13976
 
13979
13977
  // ../../b4m-core/packages/services/dist/src/llm/StatusManager.js
@@ -20,7 +20,7 @@ import {
20
20
  extractSnippetMeta,
21
21
  isGPTImageModel,
22
22
  settingsMap
23
- } from "./chunk-QXNNUB2O.js";
23
+ } from "./chunk-KHFMOKEM.js";
24
24
 
25
25
  // ../../b4m-core/packages/utils/dist/src/storage/S3Storage.js
26
26
  import { S3Client, PutObjectCommand, DeleteObjectCommand, GetObjectCommand, HeadObjectCommand } from "@aws-sdk/client-s3";
@@ -8969,9 +8969,18 @@ var OPENAI_EMBEDDING_MODEL_MAP = {
8969
8969
  dimensions: [1536]
8970
8970
  }
8971
8971
  };
8972
- var OpenAIEmbeddingService = class {
8972
+ var OpenAIEmbeddingService = class _OpenAIEmbeddingService {
8973
8973
  client;
8974
8974
  model;
8975
+ /** Hard limit imposed by OpenAI's embeddings API. */
8976
+ static MAX_TOKENS_PER_REQUEST = 3e5;
8977
+ /**
8978
+ * Effective token limit with a 10% safety buffer.
8979
+ * The tiktoken fallback (text.length/3) deliberately overestimates to be safe,
8980
+ * but DB token counts may have been produced by a different tokenizer (Bedrock, Voyage)
8981
+ * that underestimates. The buffer keeps us clear of the hard limit under tokenizer variance.
8982
+ */
8983
+ static EFFECTIVE_TOKEN_LIMIT = Math.floor(_OpenAIEmbeddingService.MAX_TOKENS_PER_REQUEST * 0.9);
8975
8984
  constructor(apiKey, model = OpenAIEmbeddingModel.TEXT_EMBEDDING_ADA_002) {
8976
8985
  this.client = new OpenAI3({ apiKey });
8977
8986
  this.validateModel(model);
@@ -8983,19 +8992,14 @@ var OpenAIEmbeddingService = class {
8983
8992
  }
8984
8993
  }
8985
8994
  async generateEmbedding(text) {
8986
- try {
8987
- const response = await this.client.embeddings.create({
8988
- model: this.model,
8989
- input: text
8990
- });
8991
- if (response.data && response.data.length > 0) {
8992
- return response.data[0].embedding;
8993
- } else {
8994
- throw new Error("No embedding data received from OpenAI");
8995
- }
8996
- } catch (error) {
8997
- throw error;
8995
+ const response = await this.client.embeddings.create({
8996
+ model: this.model,
8997
+ input: text
8998
+ });
8999
+ if (response.data && response.data.length > 0) {
9000
+ return response.data[0].embedding;
8998
9001
  }
9002
+ throw new Error("No embedding data received from OpenAI");
8999
9003
  }
9000
9004
  /**
9001
9005
  * Generate embeddings for multiple texts in a single API call.
@@ -9015,7 +9019,6 @@ var OpenAIEmbeddingService = class {
9015
9019
  return [];
9016
9020
  }
9017
9021
  const MAX_INPUTS_PER_REQUEST = 2048;
9018
- const MAX_TOKENS_PER_REQUEST = 3e5;
9019
9022
  const MAX_TOKENS_PER_INPUT = 8192;
9020
9023
  let tokens;
9021
9024
  let needsRecalculation = false;
@@ -9044,8 +9047,8 @@ var OpenAIEmbeddingService = class {
9044
9047
  totalTokens += tokenCount;
9045
9048
  }
9046
9049
  console.debug(`[OpenAI] Batch embedding: ${texts.length} inputs, ${totalTokens} total tokens`);
9047
- const batches = this.createBatches(texts, tokens, MAX_INPUTS_PER_REQUEST, MAX_TOKENS_PER_REQUEST);
9048
- console.debug(`[OpenAI] Split into ${batches.length} batch(es) to respect 300k token limit`);
9050
+ const batches = this.createBatches(texts, tokens, MAX_INPUTS_PER_REQUEST, _OpenAIEmbeddingService.EFFECTIVE_TOKEN_LIMIT);
9051
+ console.debug(`[OpenAI] Split into ${batches.length} batch(es) (effective limit: ${_OpenAIEmbeddingService.EFFECTIVE_TOKEN_LIMIT} tokens)`);
9049
9052
  if (batches.length === 1) {
9050
9053
  return await this.processSingleBatch(batches[0].texts);
9051
9054
  }
@@ -9073,8 +9076,8 @@ var OpenAIEmbeddingService = class {
9073
9076
  encoding.free();
9074
9077
  return counts;
9075
9078
  } catch (error) {
9076
- console.warn("tiktoken not available, using token estimation. Install tiktoken for accurate counts.");
9077
- return texts.map((text) => Math.ceil(text.length / 4));
9079
+ console.warn("tiktoken not available, using conservative token estimation (chars/3). Install tiktoken for accuracy.");
9080
+ return texts.map((text) => Math.ceil(text.length / 3));
9078
9081
  }
9079
9082
  }
9080
9083
  /**
@@ -9115,8 +9118,8 @@ var OpenAIEmbeddingService = class {
9115
9118
  async processSingleBatch(texts, preCalculatedTokens) {
9116
9119
  const tokenCounts = preCalculatedTokens || await this.calculateTokenCounts(texts);
9117
9120
  const batchTokens = tokenCounts.reduce((sum, count) => sum + count, 0);
9118
- if (batchTokens > 3e5) {
9119
- console.warn(`[OpenAI] Batch exceeds 300k token limit (${batchTokens} tokens), splitting recursively for safety`);
9121
+ if (batchTokens > _OpenAIEmbeddingService.EFFECTIVE_TOKEN_LIMIT) {
9122
+ console.warn(`[OpenAI] Batch exceeds effective token limit (${batchTokens}/${_OpenAIEmbeddingService.EFFECTIVE_TOKEN_LIMIT} tokens), splitting recursively`);
9120
9123
  const mid = Math.ceil(texts.length / 2);
9121
9124
  const firstHalf = texts.slice(0, mid);
9122
9125
  const secondHalf = texts.slice(mid);
@@ -9145,6 +9148,18 @@ var OpenAIEmbeddingService = class {
9145
9148
  throw new Error("No embedding data received from OpenAI");
9146
9149
  }
9147
9150
  } catch (error) {
9151
+ if (this.isTokenLimitError(error)) {
9152
+ if (texts.length === 1) {
9153
+ throw error;
9154
+ }
9155
+ console.warn(`[OpenAI] API rejected batch due to token limit, splitting ${texts.length} texts and retrying`);
9156
+ const mid = Math.ceil(texts.length / 2);
9157
+ const [firstEmbeddings, secondEmbeddings] = await Promise.all([
9158
+ this.processSingleBatch(texts.slice(0, mid)),
9159
+ this.processSingleBatch(texts.slice(mid))
9160
+ ]);
9161
+ return [...firstEmbeddings, ...secondEmbeddings];
9162
+ }
9148
9163
  if (this.shouldFallbackToIndividual(error)) {
9149
9164
  console.warn(`Batch embedding failed, falling back to individual calls: ${error}`);
9150
9165
  return await this.generateEmbeddingBatchFallback(texts);
@@ -9152,13 +9167,24 @@ var OpenAIEmbeddingService = class {
9152
9167
  throw error;
9153
9168
  }
9154
9169
  }
9170
+ /**
9171
+ * Check if the error is an OpenAI token limit rejection.
9172
+ */
9173
+ isTokenLimitError(error) {
9174
+ if (error instanceof OpenAI3.BadRequestError) {
9175
+ const body = error.error;
9176
+ return body?.type === "max_tokens_per_request";
9177
+ }
9178
+ return false;
9179
+ }
9155
9180
  /**
9156
9181
  * Check if we should fallback to individual calls on error.
9157
9182
  * Fallback for rate limits (429) and server errors (5xx), but not for auth/validation errors.
9158
9183
  */
9159
9184
  shouldFallbackToIndividual(error) {
9160
- const status = error?.status || error?.response?.status;
9161
- return status === 429 || status >= 500 && status < 600;
9185
+ const err = error;
9186
+ const status = err.status ?? err.response?.status;
9187
+ return status === 429 || status !== void 0 && status >= 500 && status < 600;
9162
9188
  }
9163
9189
  /**
9164
9190
  * Fallback mechanism: Process texts individually if batch fails.
@@ -2,7 +2,7 @@
2
2
  import {
3
3
  BadRequestError,
4
4
  secureParameters
5
- } from "./chunk-ZXBRRAVW.js";
5
+ } from "./chunk-HG3KOSRI.js";
6
6
  import {
7
7
  CompletionApiUsageTransaction,
8
8
  GenericCreditDeductTransaction,
@@ -13,7 +13,7 @@ import {
13
13
  ToolUsageTransaction,
14
14
  TransferCreditTransaction,
15
15
  VideoGenerationUsageTransaction
16
- } from "./chunk-QXNNUB2O.js";
16
+ } from "./chunk-KHFMOKEM.js";
17
17
 
18
18
  // ../../b4m-core/packages/services/dist/src/creditService/subtractCredits.js
19
19
  import { z } from "zod";
@@ -6374,9 +6374,7 @@ var ChatCompletionInvokeParamsSchema = z29.object({
6374
6374
  /** User's timezone (IANA format, e.g., "America/New_York") */
6375
6375
  timezone: z29.string().optional(),
6376
6376
  /** Persona-based sub-agent filter — only these agent names are available for delegation */
6377
- allowedAgents: z29.array(z29.string()).optional(),
6378
- /** When true, Quest Processor injects Slack-specific tool configs (help, notebooks, curated files) */
6379
- enableSlackTools: z29.boolean().optional()
6377
+ allowedAgents: z29.array(z29.string()).optional()
6380
6378
  });
6381
6379
  var LLMApiRequestBodySchema = ChatCompletionInvokeParamsSchema.extend({
6382
6380
  /** Notebook session ID */
@@ -1,7 +1,7 @@
1
1
  #!/usr/bin/env node
2
2
  import {
3
3
  ChatModels
4
- } from "./chunk-QXNNUB2O.js";
4
+ } from "./chunk-KHFMOKEM.js";
5
5
  import {
6
6
  DEFAULT_SANDBOX_CONFIG
7
7
  } from "./chunk-4BIBE3J7.js";
@@ -3,7 +3,7 @@ import {
3
3
  fetchLatestVersion,
4
4
  forceCheckForUpdate,
5
5
  package_default
6
- } from "../chunk-BKE2X66F.js";
6
+ } from "../chunk-46FVYAFI.js";
7
7
 
8
8
  // src/commands/doctorCommand.ts
9
9
  import { execSync } from "child_process";
@@ -36,20 +36,20 @@ import {
36
36
  isReadOnlyTool,
37
37
  loadContextFiles,
38
38
  setWebSocketToolExecutor
39
- } from "../chunk-7MRE7B6A.js";
39
+ } from "../chunk-FUSKVQZR.js";
40
40
  import "../chunk-BDQBOLYG.js";
41
- import "../chunk-L46CVMQB.js";
41
+ import "../chunk-BAUVNB3G.js";
42
42
  import "../chunk-GQGOWACU.js";
43
- import "../chunk-WZAH4RU6.js";
44
- import "../chunk-XHLZ2JNZ.js";
45
- import "../chunk-ZXBRRAVW.js";
43
+ import "../chunk-JULZW2WV.js";
44
+ import "../chunk-FRWP4IOP.js";
45
+ import "../chunk-HG3KOSRI.js";
46
46
  import "../chunk-PFBYGCOW.js";
47
47
  import "../chunk-BPFEGDC7.js";
48
48
  import {
49
49
  ConfigStore,
50
50
  logger
51
- } from "../chunk-522LM44T.js";
52
- import "../chunk-QXNNUB2O.js";
51
+ } from "../chunk-OJSNMX74.js";
52
+ import "../chunk-KHFMOKEM.js";
53
53
  import {
54
54
  DEFAULT_SANDBOX_CONFIG
55
55
  } from "../chunk-4BIBE3J7.js";
@@ -1,8 +1,8 @@
1
1
  #!/usr/bin/env node
2
2
  import {
3
3
  ConfigStore
4
- } from "../chunk-522LM44T.js";
5
- import "../chunk-QXNNUB2O.js";
4
+ } from "../chunk-OJSNMX74.js";
5
+ import "../chunk-KHFMOKEM.js";
6
6
  import "../chunk-4BIBE3J7.js";
7
7
 
8
8
  // src/commands/mcpCommand.ts
@@ -2,7 +2,7 @@
2
2
  import {
3
3
  forceCheckForUpdate,
4
4
  package_default
5
- } from "../chunk-BKE2X66F.js";
5
+ } from "../chunk-46FVYAFI.js";
6
6
 
7
7
  // src/commands/updateCommand.ts
8
8
  import { execSync } from "child_process";
@@ -2,10 +2,10 @@
2
2
  import {
3
3
  createFabFile,
4
4
  createFabFileSchema
5
- } from "./chunk-XHLZ2JNZ.js";
6
- import "./chunk-ZXBRRAVW.js";
5
+ } from "./chunk-FRWP4IOP.js";
6
+ import "./chunk-HG3KOSRI.js";
7
7
  import "./chunk-PFBYGCOW.js";
8
- import "./chunk-QXNNUB2O.js";
8
+ import "./chunk-KHFMOKEM.js";
9
9
  export {
10
10
  createFabFile,
11
11
  createFabFileSchema
package/dist/index.js CHANGED
@@ -46,25 +46,25 @@ import {
46
46
  setWebSocketToolExecutor,
47
47
  substituteArguments,
48
48
  warmFileCache
49
- } from "./chunk-7MRE7B6A.js";
49
+ } from "./chunk-FUSKVQZR.js";
50
50
  import "./chunk-BDQBOLYG.js";
51
- import "./chunk-L46CVMQB.js";
51
+ import "./chunk-BAUVNB3G.js";
52
52
  import "./chunk-GQGOWACU.js";
53
- import "./chunk-WZAH4RU6.js";
54
- import "./chunk-XHLZ2JNZ.js";
53
+ import "./chunk-JULZW2WV.js";
54
+ import "./chunk-FRWP4IOP.js";
55
55
  import {
56
56
  OllamaBackend
57
- } from "./chunk-ZXBRRAVW.js";
57
+ } from "./chunk-HG3KOSRI.js";
58
58
  import "./chunk-PFBYGCOW.js";
59
59
  import "./chunk-BPFEGDC7.js";
60
60
  import {
61
61
  ConfigStore,
62
62
  logger
63
- } from "./chunk-522LM44T.js";
63
+ } from "./chunk-OJSNMX74.js";
64
64
  import {
65
65
  checkForUpdate,
66
66
  package_default
67
- } from "./chunk-BKE2X66F.js";
67
+ } from "./chunk-46FVYAFI.js";
68
68
  import {
69
69
  selectActiveBackgroundAgents,
70
70
  useCliStore
@@ -72,7 +72,7 @@ import {
72
72
  import {
73
73
  CREDIT_DEDUCT_TRANSACTION_TYPES,
74
74
  ChatModels
75
- } from "./chunk-QXNNUB2O.js";
75
+ } from "./chunk-KHFMOKEM.js";
76
76
  import "./chunk-4BIBE3J7.js";
77
77
 
78
78
  // src/index.tsx
@@ -1,7 +1,7 @@
1
1
  #!/usr/bin/env node
2
2
  import {
3
3
  CurationArtifactType
4
- } from "./chunk-QXNNUB2O.js";
4
+ } from "./chunk-KHFMOKEM.js";
5
5
 
6
6
  // ../../b4m-core/packages/services/dist/src/notebookCurationService/llmMarkdownGenerator.js
7
7
  var DEFAULT_OPTIONS = {
@@ -1,7 +1,7 @@
1
1
  #!/usr/bin/env node
2
2
  import {
3
3
  CurationArtifactType
4
- } from "./chunk-QXNNUB2O.js";
4
+ } from "./chunk-KHFMOKEM.js";
5
5
 
6
6
  // ../../b4m-core/packages/services/dist/src/notebookCurationService/markdownGenerator.js
7
7
  var DEFAULT_OPTIONS = {
@@ -2,10 +2,10 @@
2
2
  import {
3
3
  findMostSimilarMemento,
4
4
  getRelevantMementos
5
- } from "./chunk-L46CVMQB.js";
6
- import "./chunk-ZXBRRAVW.js";
5
+ } from "./chunk-BAUVNB3G.js";
6
+ import "./chunk-HG3KOSRI.js";
7
7
  import "./chunk-PFBYGCOW.js";
8
- import "./chunk-QXNNUB2O.js";
8
+ import "./chunk-KHFMOKEM.js";
9
9
  export {
10
10
  findMostSimilarMemento,
11
11
  getRelevantMementos
@@ -143,7 +143,7 @@ import {
143
143
  validateUrlForFetch,
144
144
  warmUpSettingsCache,
145
145
  withRetry
146
- } from "./chunk-ZXBRRAVW.js";
146
+ } from "./chunk-HG3KOSRI.js";
147
147
  import {
148
148
  Logger,
149
149
  NotificationDeduplicator,
@@ -156,7 +156,7 @@ import {
156
156
  buildRateLimitLogEntry,
157
157
  isNearLimit,
158
158
  parseRateLimitHeaders
159
- } from "./chunk-QXNNUB2O.js";
159
+ } from "./chunk-KHFMOKEM.js";
160
160
  export {
161
161
  AIVideoService,
162
162
  AWSBackend,
@@ -537,7 +537,7 @@ import {
537
537
  validateReactArtifactV2,
538
538
  validateSvgArtifactV2,
539
539
  wikiMarkupToAdf
540
- } from "./chunk-QXNNUB2O.js";
540
+ } from "./chunk-KHFMOKEM.js";
541
541
  export {
542
542
  ALERT_THRESHOLDS,
543
543
  ALL_IMAGE_MODELS,
@@ -2,10 +2,10 @@
2
2
  import {
3
3
  SubtractCreditsSchema,
4
4
  subtractCredits
5
- } from "./chunk-WZAH4RU6.js";
6
- import "./chunk-ZXBRRAVW.js";
5
+ } from "./chunk-JULZW2WV.js";
6
+ import "./chunk-HG3KOSRI.js";
7
7
  import "./chunk-PFBYGCOW.js";
8
- import "./chunk-QXNNUB2O.js";
8
+ import "./chunk-KHFMOKEM.js";
9
9
  export {
10
10
  SubtractCreditsSchema,
11
11
  subtractCredits
package/package.json CHANGED
@@ -1,6 +1,6 @@
1
1
  {
2
2
  "name": "@bike4mind/cli",
3
- "version": "0.2.53-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
3
+ "version": "0.2.53",
4
4
  "type": "module",
5
5
  "description": "Interactive CLI tool for Bike4Mind with ReAct agents",
6
6
  "license": "UNLICENSED",
@@ -114,11 +114,11 @@
114
114
  "zustand": "^4.5.4"
115
115
  },
116
116
  "devDependencies": {
117
- "@bike4mind/agents": "0.1.4-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
118
- "@bike4mind/common": "2.69.1-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
119
- "@bike4mind/mcp": "1.33.14-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
120
- "@bike4mind/services": "2.64.1-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
121
- "@bike4mind/utils": "2.15.7-refactor-slack-remove-dead-code-simplify-prompts.21212+a1c57624b",
117
+ "@bike4mind/agents": "0.1.4",
118
+ "@bike4mind/common": "2.69.0",
119
+ "@bike4mind/mcp": "1.33.13",
120
+ "@bike4mind/services": "2.64.1",
121
+ "@bike4mind/utils": "2.15.7",
122
122
  "@types/better-sqlite3": "^7.6.13",
123
123
  "@types/diff": "^8.0.0",
124
124
  "@types/jsonwebtoken": "^9.0.4",
@@ -136,5 +136,5 @@
136
136
  "optionalDependencies": {
137
137
  "@vscode/ripgrep": "^1.17.0"
138
138
  },
139
- "gitHead": "a1c57624ba358a7183a96a10c046bd8338a97090"
139
+ "gitHead": "5ebeac1702534ebc59e39d53f778f1969fc077af"
140
140
  }