@joshuaswarren/openclaw-engram 9.0.96 → 9.0.97

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -3,13 +3,13 @@ import {
3
3
  EngramAccessService,
4
4
  Orchestrator,
5
5
  parseConfig
6
- } from "./chunk-AHYRVKIJ.js";
7
- import "./chunk-R2BJ47PU.js";
6
+ } from "./chunk-6ZL6J6ZZ.js";
7
+ import "./chunk-LTELW5MZ.js";
8
8
  import "./chunk-WSXU2GHQ.js";
9
9
  import "./chunk-H7LMMCVZ.js";
10
10
  import "./chunk-BNBG2XP6.js";
11
11
  import "./chunk-AFMNQR7H.js";
12
- import "./chunk-PW5V7OJH.js";
12
+ import "./chunk-V2E2ORE5.js";
13
13
  import "./chunk-DEIBZP3O.js";
14
14
  import "./chunk-SSIIJJKA.js";
15
15
 
@@ -1,7 +1,7 @@
1
1
  // openclaw-engram: Local-first memory plugin
2
2
  import {
3
3
  FallbackLlmClient
4
- } from "./chunk-PW5V7OJH.js";
4
+ } from "./chunk-V2E2ORE5.js";
5
5
  import {
6
6
  listJsonFiles
7
7
  } from "./chunk-DEIBZP3O.js";
@@ -233,4 +233,4 @@ export {
233
233
  runCalibrationIfEnabled,
234
234
  synthesizeCalibrationRules
235
235
  };
236
- //# sourceMappingURL=calibration-TLGCHI7D.js.map
236
+ //# sourceMappingURL=calibration-LBE4XTN4.js.map
@@ -8,7 +8,7 @@ import {
8
8
  } from "./chunk-AFMNQR7H.js";
9
9
  import {
10
10
  FallbackLlmClient
11
- } from "./chunk-PW5V7OJH.js";
11
+ } from "./chunk-V2E2ORE5.js";
12
12
  import {
13
13
  listJsonFiles,
14
14
  readJsonFile
@@ -203,4 +203,4 @@ export {
203
203
  deriveCausalPromotionCandidates,
204
204
  synthesizeCausalPreferencesViaLlm
205
205
  };
206
- //# sourceMappingURL=causal-consolidation-LN3WQ6OV.js.map
206
+ //# sourceMappingURL=causal-consolidation-INSQMIFR.js.map
@@ -3,7 +3,7 @@ import {
3
3
  CompoundingEngine,
4
4
  SharedContextManager,
5
5
  defaultTierMigrationCycleBudget
6
- } from "./chunk-R2BJ47PU.js";
6
+ } from "./chunk-LTELW5MZ.js";
7
7
  import {
8
8
  searchCausalTrajectories
9
9
  } from "./chunk-WSXU2GHQ.js";
@@ -43,7 +43,7 @@ import {
43
43
  buildChatCompletionTokenLimit,
44
44
  extractJsonCandidates,
45
45
  shouldAssumeOpenAiChatCompletions
46
- } from "./chunk-PW5V7OJH.js";
46
+ } from "./chunk-V2E2ORE5.js";
47
47
  import {
48
48
  listJsonFiles,
49
49
  listNamedFiles,
@@ -3570,7 +3570,7 @@ var LocalLlmClient = class _LocalLlmClient {
3570
3570
  requestBody.chat_template_kwargs = { enable_thinking: false };
3571
3571
  }
3572
3572
  const baseUrl = this.config.localLlmUrl.replace("localhost", "127.0.0.1").replace(/\/+$/, "");
3573
- const chatUrl = `${baseUrl}/chat/completions`;
3573
+ const chatUrl = baseUrl.endsWith("/v1") ? `${baseUrl}/chat/completions` : `${baseUrl}/v1/chat/completions`;
3574
3574
  const requestBodyJson = JSON.stringify(requestBody);
3575
3575
  log.debug(
3576
3576
  `local LLM: sending request to ${chatUrl} with model ${this.config.localLlmModel}`
@@ -22291,7 +22291,7 @@ ${trimmedBody}`;
22291
22291
  return null;
22292
22292
  }
22293
22293
  try {
22294
- const { getCalibrationRulesForRecall, buildCalibrationRecallSection } = await import("./calibration-TLGCHI7D.js");
22294
+ const { getCalibrationRulesForRecall, buildCalibrationRecallSection } = await import("./calibration-LBE4XTN4.js");
22295
22295
  const rules = await getCalibrationRulesForRecall(this.config.memoryDir);
22296
22296
  if (rules.length === 0) {
22297
22297
  timings.calibrationRules = "skip(no-rules)";
@@ -28107,4 +28107,4 @@ export {
28107
28107
  EngramAccessInputError,
28108
28108
  EngramAccessService
28109
28109
  };
28110
- //# sourceMappingURL=chunk-AHYRVKIJ.js.map
28110
+ //# sourceMappingURL=chunk-6ZL6J6ZZ.js.map