@sprucelabs/sprucebot-llm 15.2.0 → 16.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -18,6 +18,7 @@ export default class AnthropicAdapter implements LlmAdapter {
18
18
  private sendHandler;
19
19
  setModel(model: string): void;
20
20
  setReasoningEffort(effort: LllmReasoningEffort): void;
21
+ setMemoryLimit(limit: number): void;
21
22
  }
22
23
  export interface AnthropicAdapterOptions {
23
24
  log?: Log;
@@ -81,6 +81,9 @@ class AnthropicAdapter {
81
81
  setReasoningEffort(effort) {
82
82
  this.isThinkingEnabled = effort !== 'none';
83
83
  }
84
+ setMemoryLimit(limit) {
85
+ this.memoryLimit = limit;
86
+ }
84
87
  }
85
88
  AnthropicAdapter.Anthropic = sdk_1.default;
86
89
  exports.default = AnthropicAdapter;
@@ -10,6 +10,7 @@ export default class OllamaAdapter implements LlmAdapter {
10
10
  sendMessage(bot: SprucebotLlmBot<Schema>, options?: SendMessageOptions): Promise<string>;
11
11
  setModel(model: string): void;
12
12
  setReasoningEffort(effort: LllmReasoningEffort): void;
13
+ setMemoryLimit(_limit: number): void;
13
14
  }
14
15
  export interface OllamaAdapterOptions {
15
16
  log?: Log;
@@ -25,5 +25,8 @@ class OllamaAdapter {
25
25
  setReasoningEffort(effort) {
26
26
  this.think = effort === 'high';
27
27
  }
28
+ setMemoryLimit(_limit) {
29
+ throw new Error('Method not implemented.');
30
+ }
28
31
  }
29
32
  exports.default = OllamaAdapter;
@@ -16,7 +16,7 @@ export default class OpenAiAdapter implements LlmAdapter {
16
16
  private sendHandler;
17
17
  private getReasoningEffort;
18
18
  setModel(model: string): void;
19
- setMessageMemoryLimit(limit: number): void;
19
+ setMemoryLimit(limit: number): void;
20
20
  setReasoningEffort(effort: ReasoningEffort): void;
21
21
  }
22
22
  export declare const MESSAGE_RESPONSE_ERROR_MESSAGE = "Oh no! Something went wrong and I can't talk right now!";
@@ -42,7 +42,7 @@ class OpenAiAdapter {
42
42
  setModel(model) {
43
43
  this.model = model;
44
44
  }
45
- setMessageMemoryLimit(limit) {
45
+ setMemoryLimit(limit) {
46
46
  this.memoryLimit = limit;
47
47
  }
48
48
  setReasoningEffort(effort) {
@@ -18,6 +18,7 @@ export default class AnthropicAdapter implements LlmAdapter {
18
18
  private sendHandler;
19
19
  setModel(model: string): void;
20
20
  setReasoningEffort(effort: LllmReasoningEffort): void;
21
+ setMemoryLimit(limit: number): void;
21
22
  }
22
23
  export interface AnthropicAdapterOptions {
23
24
  log?: Log;
@@ -84,6 +84,9 @@ class AnthropicAdapter {
84
84
  setReasoningEffort(effort) {
85
85
  this.isThinkingEnabled = effort !== 'none';
86
86
  }
87
+ setMemoryLimit(limit) {
88
+ this.memoryLimit = limit;
89
+ }
87
90
  }
88
91
  AnthropicAdapter.Anthropic = Anthropic;
89
92
  export default AnthropicAdapter;
@@ -10,6 +10,7 @@ export default class OllamaAdapter implements LlmAdapter {
10
10
  sendMessage(bot: SprucebotLlmBot<Schema>, options?: SendMessageOptions): Promise<string>;
11
11
  setModel(model: string): void;
12
12
  setReasoningEffort(effort: LllmReasoningEffort): void;
13
+ setMemoryLimit(_limit: number): void;
13
14
  }
14
15
  export interface OllamaAdapterOptions {
15
16
  log?: Log;
@@ -30,4 +30,7 @@ export default class OllamaAdapter {
30
30
  setReasoningEffort(effort) {
31
31
  this.think = effort === 'high';
32
32
  }
33
+ setMemoryLimit(_limit) {
34
+ throw new Error('Method not implemented.');
35
+ }
33
36
  }
@@ -16,7 +16,7 @@ export default class OpenAiAdapter implements LlmAdapter {
16
16
  private sendHandler;
17
17
  private getReasoningEffort;
18
18
  setModel(model: string): void;
19
- setMessageMemoryLimit(limit: number): void;
19
+ setMemoryLimit(limit: number): void;
20
20
  setReasoningEffort(effort: ReasoningEffort): void;
21
21
  }
22
22
  export declare const MESSAGE_RESPONSE_ERROR_MESSAGE = "Oh no! Something went wrong and I can't talk right now!";
@@ -47,7 +47,7 @@ class OpenAiAdapter {
47
47
  setModel(model) {
48
48
  this.model = model;
49
49
  }
50
- setMessageMemoryLimit(limit) {
50
+ setMemoryLimit(limit) {
51
51
  this.memoryLimit = limit;
52
52
  }
53
53
  setReasoningEffort(effort) {
@@ -20,6 +20,7 @@ export interface LlmAdapter {
20
20
  sendMessage(bot: SprucebotLlmBot<Schema>, options?: SendMessageOptions): Promise<string>;
21
21
  setModel(model: string): void;
22
22
  setReasoningEffort(effort: LllmReasoningEffort): void;
23
+ setMemoryLimit(limit: number): void;
23
24
  }
24
25
  export interface PromptOptions<StateSchema extends Schema, State extends SchemaValues<StateSchema> = SchemaValues<StateSchema>> {
25
26
  /**
@@ -15,4 +15,5 @@ export default class SpyLlmAdapter implements LlmAdapter {
15
15
  sendMessage(bot: SprucebotLlmBot, options?: SendMessageOptions): Promise<string>;
16
16
  setModel(model: string): void;
17
17
  setReasoningEffort(_effort: LllmReasoningEffort): void;
18
+ setMemoryLimit(_limit: number): void;
18
19
  }
@@ -33,4 +33,5 @@ export default class SpyLlmAdapter {
33
33
  this.manuallySetModel = model;
34
34
  }
35
35
  setReasoningEffort(_effort) { }
36
+ setMemoryLimit(_limit) { }
36
37
  }
@@ -20,6 +20,7 @@ export interface LlmAdapter {
20
20
  sendMessage(bot: SprucebotLlmBot<Schema>, options?: SendMessageOptions): Promise<string>;
21
21
  setModel(model: string): void;
22
22
  setReasoningEffort(effort: LllmReasoningEffort): void;
23
+ setMemoryLimit(limit: number): void;
23
24
  }
24
25
  export interface PromptOptions<StateSchema extends Schema, State extends SchemaValues<StateSchema> = SchemaValues<StateSchema>> {
25
26
  /**
@@ -15,4 +15,5 @@ export default class SpyLlmAdapter implements LlmAdapter {
15
15
  sendMessage(bot: SprucebotLlmBot, options?: SendMessageOptions): Promise<string>;
16
16
  setModel(model: string): void;
17
17
  setReasoningEffort(_effort: LllmReasoningEffort): void;
18
+ setMemoryLimit(_limit: number): void;
18
19
  }
@@ -24,5 +24,6 @@ class SpyLlmAdapter {
24
24
  this.manuallySetModel = model;
25
25
  }
26
26
  setReasoningEffort(_effort) { }
27
+ setMemoryLimit(_limit) { }
27
28
  }
28
29
  exports.default = SpyLlmAdapter;
package/package.json CHANGED
@@ -8,7 +8,7 @@
8
8
  "eta"
9
9
  ]
10
10
  },
11
- "version": "15.2.0",
11
+ "version": "16.0.1",
12
12
  "files": [
13
13
  "build"
14
14
  ],
@@ -65,14 +65,14 @@
65
65
  "chat.images": "node ./build/chatWithImages.js"
66
66
  },
67
67
  "dependencies": {
68
- "@anthropic-ai/sdk": "^0.78.0",
69
- "@sprucelabs/error": "^8.1.14",
70
- "@sprucelabs/mercury-event-emitter": "^46.1.13",
71
- "@sprucelabs/mercury-types": "^49.1.15",
72
- "@sprucelabs/schema": "^34.0.2",
73
- "@sprucelabs/spruce-skill-utils": "^34.0.21",
68
+ "@anthropic-ai/sdk": "^0.86.1",
69
+ "@sprucelabs/error": "^8.1.16",
70
+ "@sprucelabs/mercury-event-emitter": "^46.1.19",
71
+ "@sprucelabs/mercury-types": "^49.1.20",
72
+ "@sprucelabs/schema": "^34.0.7",
73
+ "@sprucelabs/spruce-skill-utils": "^34.0.26",
74
74
  "eta": "3.5.0",
75
- "openai": "^6.27.0"
75
+ "openai": "^6.34.0"
76
76
  },
77
77
  "resolutions": {
78
78
  "eta": "3.5.0"
@@ -80,23 +80,23 @@
80
80
  "devDependencies": {
81
81
  "@regressionproof/cli": "^0.9.4",
82
82
  "@regressionproof/jest-reporter": "^0.9.4",
83
- "@sprucelabs/esm-postbuild": "^9.0.23",
84
- "@sprucelabs/jest-json-reporter": "^10.0.31",
85
- "@sprucelabs/resolve-path-aliases": "^4.0.22",
83
+ "@sprucelabs/esm-postbuild": "^9.0.26",
84
+ "@sprucelabs/jest-json-reporter": "^10.0.33",
85
+ "@sprucelabs/resolve-path-aliases": "^4.0.26",
86
86
  "@sprucelabs/semantic-release": "^6.0.0",
87
- "@sprucelabs/test": "^11.1.6",
88
- "@sprucelabs/test-utils": "^7.2.17",
89
- "@types/node": "^25.5.0",
87
+ "@sprucelabs/test": "^11.1.7",
88
+ "@sprucelabs/test-utils": "^7.2.20",
89
+ "@types/node": "^25.5.2",
90
90
  "chokidar-cli": "^3.0.0",
91
- "dotenv": "^17.3.1",
92
- "eslint": "^10.0.3",
91
+ "dotenv": "^17.4.1",
92
+ "eslint": "^10.2.0",
93
93
  "eslint-config-spruce": "^11.2.35",
94
94
  "jest": "^30.3.0",
95
95
  "jest-circus": "^30.3.0",
96
96
  "prettier": "^3.8.1",
97
97
  "ts-node": "^10.9.2",
98
98
  "tsc-watch": "^7.2.0",
99
- "typescript": "^5.9.3"
99
+ "typescript": "^6.0.2"
100
100
  },
101
101
  "jest": {
102
102
  "testRunner": "jest-circus/runner",