@librechat/agents 3.0.78 → 3.0.79

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (36) hide show
  1. package/dist/cjs/llm/bedrock/index.cjs +135 -23
  2. package/dist/cjs/llm/bedrock/index.cjs.map +1 -1
  3. package/dist/cjs/llm/bedrock/utils/message_inputs.cjs +465 -0
  4. package/dist/cjs/llm/bedrock/utils/message_inputs.cjs.map +1 -0
  5. package/dist/cjs/llm/bedrock/utils/message_outputs.cjs +238 -0
  6. package/dist/cjs/llm/bedrock/utils/message_outputs.cjs.map +1 -0
  7. package/dist/cjs/messages/core.cjs +1 -1
  8. package/dist/cjs/messages/core.cjs.map +1 -1
  9. package/dist/cjs/stream.cjs +4 -2
  10. package/dist/cjs/stream.cjs.map +1 -1
  11. package/dist/esm/llm/bedrock/index.mjs +134 -22
  12. package/dist/esm/llm/bedrock/index.mjs.map +1 -1
  13. package/dist/esm/llm/bedrock/utils/message_inputs.mjs +460 -0
  14. package/dist/esm/llm/bedrock/utils/message_inputs.mjs.map +1 -0
  15. package/dist/esm/llm/bedrock/utils/message_outputs.mjs +231 -0
  16. package/dist/esm/llm/bedrock/utils/message_outputs.mjs.map +1 -0
  17. package/dist/esm/messages/core.mjs +1 -1
  18. package/dist/esm/messages/core.mjs.map +1 -1
  19. package/dist/esm/stream.mjs +4 -2
  20. package/dist/esm/stream.mjs.map +1 -1
  21. package/dist/types/llm/bedrock/index.d.ts +83 -7
  22. package/dist/types/llm/bedrock/types.d.ts +27 -0
  23. package/dist/types/llm/bedrock/utils/index.d.ts +5 -0
  24. package/dist/types/llm/bedrock/utils/message_inputs.d.ts +31 -0
  25. package/dist/types/llm/bedrock/utils/message_outputs.d.ts +33 -0
  26. package/package.json +4 -3
  27. package/src/llm/bedrock/index.ts +232 -41
  28. package/src/llm/bedrock/llm.spec.ts +616 -0
  29. package/src/llm/bedrock/types.ts +51 -0
  30. package/src/llm/bedrock/utils/index.ts +18 -0
  31. package/src/llm/bedrock/utils/message_inputs.ts +563 -0
  32. package/src/llm/bedrock/utils/message_outputs.ts +296 -0
  33. package/src/messages/core.ts +1 -1
  34. package/src/scripts/thinking.ts +39 -18
  35. package/src/scripts/tools.ts +7 -3
  36. package/src/stream.ts +4 -2
@@ -0,0 +1,296 @@
1
+ /**
2
+ * Utility functions for converting Bedrock Converse responses to LangChain messages.
3
+ * Ported from @langchain/aws common.js
4
+ */
5
+ import { AIMessage, AIMessageChunk } from '@langchain/core/messages';
6
+ import { ChatGenerationChunk } from '@langchain/core/outputs';
7
+ import type {
8
+ BedrockMessage,
9
+ ConverseResponse,
10
+ ContentBlockDeltaEvent,
11
+ ConverseStreamMetadataEvent,
12
+ ContentBlockStartEvent,
13
+ ReasoningContentBlock,
14
+ ReasoningContentBlockDelta,
15
+ MessageContentReasoningBlock,
16
+ MessageContentReasoningBlockReasoningTextPartial,
17
+ MessageContentReasoningBlockRedacted,
18
+ } from '../types';
19
+
20
+ /**
21
+ * Convert a Bedrock reasoning block delta to a LangChain partial reasoning block.
22
+ */
23
+ export function bedrockReasoningDeltaToLangchainPartialReasoningBlock(
24
+ reasoningContent: ReasoningContentBlockDelta
25
+ ):
26
+ | MessageContentReasoningBlockReasoningTextPartial
27
+ | MessageContentReasoningBlockRedacted {
28
+ const { text, redactedContent, signature } =
29
+ reasoningContent as ReasoningContentBlockDelta & {
30
+ text?: string;
31
+ redactedContent?: Uint8Array;
32
+ signature?: string;
33
+ };
34
+
35
+ if (typeof text === 'string') {
36
+ return {
37
+ type: 'reasoning_content',
38
+ reasoningText: { text },
39
+ };
40
+ }
41
+ if (signature) {
42
+ return {
43
+ type: 'reasoning_content',
44
+ reasoningText: { signature },
45
+ };
46
+ }
47
+ if (redactedContent) {
48
+ return {
49
+ type: 'reasoning_content',
50
+ redactedContent: Buffer.from(redactedContent).toString('base64'),
51
+ };
52
+ }
53
+ throw new Error('Invalid reasoning content');
54
+ }
55
+
56
+ /**
57
+ * Convert a Bedrock reasoning block to a LangChain reasoning block.
58
+ */
59
+ export function bedrockReasoningBlockToLangchainReasoningBlock(
60
+ reasoningContent: ReasoningContentBlock
61
+ ): MessageContentReasoningBlock {
62
+ const { reasoningText, redactedContent } =
63
+ reasoningContent as ReasoningContentBlock & {
64
+ reasoningText?: { text?: string; signature?: string };
65
+ redactedContent?: Uint8Array;
66
+ };
67
+
68
+ if (reasoningText) {
69
+ return {
70
+ type: 'reasoning_content',
71
+ reasoningText: reasoningText,
72
+ };
73
+ }
74
+ if (redactedContent) {
75
+ return {
76
+ type: 'reasoning_content',
77
+ redactedContent: Buffer.from(redactedContent).toString('base64'),
78
+ };
79
+ }
80
+ throw new Error('Invalid reasoning content');
81
+ }
82
+
83
+ /**
84
+ * Convert a Bedrock Converse message to a LangChain message.
85
+ */
86
+ export function convertConverseMessageToLangChainMessage(
87
+ message: BedrockMessage,
88
+ responseMetadata: Omit<ConverseResponse, 'output'>
89
+ ): AIMessage {
90
+ if (!message.content) {
91
+ throw new Error('No message content found in response.');
92
+ }
93
+ if (message.role !== 'assistant') {
94
+ throw new Error(
95
+ `Unsupported message role received in ChatBedrockConverse response: ${message.role}`
96
+ );
97
+ }
98
+
99
+ let requestId: string | undefined;
100
+ if (
101
+ '$metadata' in responseMetadata &&
102
+ responseMetadata.$metadata &&
103
+ typeof responseMetadata.$metadata === 'object' &&
104
+ 'requestId' in responseMetadata.$metadata
105
+ ) {
106
+ requestId = responseMetadata.$metadata.requestId as string;
107
+ }
108
+
109
+ let tokenUsage:
110
+ | { input_tokens: number; output_tokens: number; total_tokens: number }
111
+ | undefined;
112
+ if (responseMetadata.usage) {
113
+ const input_tokens = responseMetadata.usage.inputTokens ?? 0;
114
+ const output_tokens = responseMetadata.usage.outputTokens ?? 0;
115
+ tokenUsage = {
116
+ input_tokens,
117
+ output_tokens,
118
+ total_tokens:
119
+ responseMetadata.usage.totalTokens ?? input_tokens + output_tokens,
120
+ };
121
+ }
122
+
123
+ if (
124
+ message.content.length === 1 &&
125
+ 'text' in message.content[0] &&
126
+ typeof message.content[0].text === 'string'
127
+ ) {
128
+ return new AIMessage({
129
+ content: message.content[0].text,
130
+ response_metadata: responseMetadata,
131
+ usage_metadata: tokenUsage,
132
+ id: requestId,
133
+ });
134
+ } else {
135
+ const toolCalls: Array<{
136
+ id?: string;
137
+ name: string;
138
+ args: Record<string, unknown>;
139
+ type: 'tool_call';
140
+ }> = [];
141
+ // eslint-disable-next-line @typescript-eslint/no-explicit-any
142
+ const content: any[] = [];
143
+
144
+ message.content.forEach((c) => {
145
+ if (
146
+ 'toolUse' in c &&
147
+ c.toolUse &&
148
+ c.toolUse.name &&
149
+ c.toolUse.input &&
150
+ typeof c.toolUse.input === 'object'
151
+ ) {
152
+ toolCalls.push({
153
+ id: c.toolUse.toolUseId,
154
+ name: c.toolUse.name,
155
+ args: c.toolUse.input as Record<string, unknown>,
156
+ type: 'tool_call',
157
+ });
158
+ } else if ('text' in c && typeof c.text === 'string') {
159
+ content.push({ type: 'text', text: c.text });
160
+ } else if ('reasoningContent' in c && c.reasoningContent) {
161
+ content.push(
162
+ bedrockReasoningBlockToLangchainReasoningBlock(c.reasoningContent)
163
+ );
164
+ } else {
165
+ content.push(c);
166
+ }
167
+ });
168
+
169
+ return new AIMessage({
170
+ content: content.length ? content : '',
171
+ tool_calls: toolCalls.length ? toolCalls : undefined,
172
+ response_metadata: responseMetadata,
173
+ usage_metadata: tokenUsage,
174
+ id: requestId,
175
+ });
176
+ }
177
+ }
178
+
179
+ /**
180
+ * Handle a content block delta event from Bedrock Converse stream.
181
+ */
182
+ export function handleConverseStreamContentBlockDelta(
183
+ contentBlockDelta: ContentBlockDeltaEvent
184
+ ): ChatGenerationChunk {
185
+ if (!contentBlockDelta.delta) {
186
+ throw new Error('No delta found in content block.');
187
+ }
188
+
189
+ if (typeof contentBlockDelta.delta.text === 'string') {
190
+ return new ChatGenerationChunk({
191
+ text: contentBlockDelta.delta.text,
192
+ message: new AIMessageChunk({
193
+ content: contentBlockDelta.delta.text,
194
+ response_metadata: {
195
+ contentBlockIndex: contentBlockDelta.contentBlockIndex,
196
+ },
197
+ }),
198
+ });
199
+ } else if (contentBlockDelta.delta.toolUse) {
200
+ const index = contentBlockDelta.contentBlockIndex;
201
+ return new ChatGenerationChunk({
202
+ text: '',
203
+ message: new AIMessageChunk({
204
+ content: '',
205
+ tool_call_chunks: [
206
+ {
207
+ args: contentBlockDelta.delta.toolUse.input as string,
208
+ index,
209
+ type: 'tool_call_chunk',
210
+ },
211
+ ],
212
+ response_metadata: {
213
+ contentBlockIndex: contentBlockDelta.contentBlockIndex,
214
+ },
215
+ }),
216
+ });
217
+ } else if (contentBlockDelta.delta.reasoningContent) {
218
+ return new ChatGenerationChunk({
219
+ text: '',
220
+ message: new AIMessageChunk({
221
+ content: [
222
+ bedrockReasoningDeltaToLangchainPartialReasoningBlock(
223
+ contentBlockDelta.delta.reasoningContent
224
+ ),
225
+ ],
226
+ response_metadata: {
227
+ contentBlockIndex: contentBlockDelta.contentBlockIndex,
228
+ },
229
+ }),
230
+ });
231
+ } else {
232
+ throw new Error(
233
+ `Unsupported content block type(s): ${JSON.stringify(contentBlockDelta.delta, null, 2)}`
234
+ );
235
+ }
236
+ }
237
+
238
+ /**
239
+ * Handle a content block start event from Bedrock Converse stream.
240
+ */
241
+ export function handleConverseStreamContentBlockStart(
242
+ contentBlockStart: ContentBlockStartEvent
243
+ ): ChatGenerationChunk | null {
244
+ const index = contentBlockStart.contentBlockIndex;
245
+
246
+ if (contentBlockStart.start?.toolUse) {
247
+ return new ChatGenerationChunk({
248
+ text: '',
249
+ message: new AIMessageChunk({
250
+ content: '',
251
+ tool_call_chunks: [
252
+ {
253
+ name: contentBlockStart.start.toolUse.name,
254
+ id: contentBlockStart.start.toolUse.toolUseId,
255
+ index,
256
+ type: 'tool_call_chunk',
257
+ },
258
+ ],
259
+ response_metadata: {
260
+ contentBlockIndex: index,
261
+ },
262
+ }),
263
+ });
264
+ }
265
+
266
+ // Return null for non-tool content block starts (text blocks don't need special handling)
267
+ return null;
268
+ }
269
+
270
+ /**
271
+ * Handle a metadata event from Bedrock Converse stream.
272
+ */
273
+ export function handleConverseStreamMetadata(
274
+ metadata: ConverseStreamMetadataEvent,
275
+ extra: { streamUsage: boolean }
276
+ ): ChatGenerationChunk {
277
+ const inputTokens = metadata.usage?.inputTokens ?? 0;
278
+ const outputTokens = metadata.usage?.outputTokens ?? 0;
279
+ const usage_metadata = {
280
+ input_tokens: inputTokens,
281
+ output_tokens: outputTokens,
282
+ total_tokens: metadata.usage?.totalTokens ?? inputTokens + outputTokens,
283
+ };
284
+
285
+ return new ChatGenerationChunk({
286
+ text: '',
287
+ message: new AIMessageChunk({
288
+ content: '',
289
+ usage_metadata: extra.streamUsage ? usage_metadata : undefined,
290
+ response_metadata: {
291
+ // Use the same key as returned from the Converse API
292
+ metadata,
293
+ },
294
+ }),
295
+ });
296
+ }
@@ -41,7 +41,7 @@ User: ${userMessage[1]}
41
41
  const _allowedTypes = ['image_url', 'text', 'tool_use', 'tool_result'];
42
42
  const allowedTypesByProvider: Record<string, string[]> = {
43
43
  default: _allowedTypes,
44
- [Providers.ANTHROPIC]: [..._allowedTypes, 'thinking'],
44
+ [Providers.ANTHROPIC]: [..._allowedTypes, 'thinking', 'redacted_thinking'],
45
45
  [Providers.BEDROCK]: [..._allowedTypes, 'reasoning_content'],
46
46
  [Providers.OPENAI]: _allowedTypes,
47
47
  };
@@ -1,7 +1,11 @@
1
1
  // src/scripts/test-thinking.ts
2
2
  import { config } from 'dotenv';
3
3
  config();
4
- import { HumanMessage, SystemMessage, BaseMessage } from '@langchain/core/messages';
4
+ import {
5
+ HumanMessage,
6
+ SystemMessage,
7
+ BaseMessage,
8
+ } from '@langchain/core/messages';
5
9
  import type { UsageMetadata } from '@langchain/core/messages';
6
10
  import * as t from '@/types';
7
11
  import { ChatModelStreamHandler, createContentAggregator } from '@/stream';
@@ -21,17 +25,23 @@ async function testThinking(): Promise<void> {
21
25
  const instructions = `You are a helpful AI assistant for ${userName}. When answering questions, be thorough in your reasoning.`;
22
26
  const { contentParts, aggregateContent } = createContentAggregator();
23
27
  _contentParts = contentParts as t.MessageContentComplex[];
24
-
28
+
25
29
  // Set up event handlers
26
30
  const customHandlers = {
27
31
  [GraphEvents.TOOL_END]: new ToolEndHandler(),
28
32
  [GraphEvents.CHAT_MODEL_END]: new ModelEndHandler(collectedUsage),
29
33
  [GraphEvents.CHAT_MODEL_STREAM]: new ChatModelStreamHandler(),
30
34
  [GraphEvents.ON_RUN_STEP_COMPLETED]: {
31
- handle: (event: GraphEvents.ON_RUN_STEP_COMPLETED, data: t.StreamEventData): void => {
35
+ handle: (
36
+ event: GraphEvents.ON_RUN_STEP_COMPLETED,
37
+ data: t.StreamEventData
38
+ ): void => {
32
39
  console.log('====== ON_RUN_STEP_COMPLETED ======');
33
- aggregateContent({ event, data: data as unknown as { result: t.ToolEndEvent } });
34
- }
40
+ aggregateContent({
41
+ event,
42
+ data: data as unknown as { result: t.ToolEndEvent },
43
+ });
44
+ },
35
45
  },
36
46
  [GraphEvents.ON_RUN_STEP]: {
37
47
  handle: (event: GraphEvents.ON_RUN_STEP, data: t.RunStep) => {
@@ -39,29 +49,38 @@ async function testThinking(): Promise<void> {
39
49
  },
40
50
  },
41
51
  [GraphEvents.ON_RUN_STEP_DELTA]: {
42
- handle: (event: GraphEvents.ON_RUN_STEP_DELTA, data: t.RunStepDeltaEvent) => {
52
+ handle: (
53
+ event: GraphEvents.ON_RUN_STEP_DELTA,
54
+ data: t.RunStepDeltaEvent
55
+ ) => {
43
56
  aggregateContent({ event, data });
44
57
  },
45
58
  },
46
59
  [GraphEvents.ON_MESSAGE_DELTA]: {
47
- handle: (event: GraphEvents.ON_MESSAGE_DELTA, data: t.MessageDeltaEvent) => {
60
+ handle: (
61
+ event: GraphEvents.ON_MESSAGE_DELTA,
62
+ data: t.MessageDeltaEvent
63
+ ) => {
48
64
  aggregateContent({ event, data });
49
65
  },
50
66
  },
51
67
  [GraphEvents.ON_REASONING_DELTA]: {
52
- handle: (event: GraphEvents.ON_REASONING_DELTA, data: t.ReasoningDeltaEvent) => {
68
+ handle: (
69
+ event: GraphEvents.ON_REASONING_DELTA,
70
+ data: t.ReasoningDeltaEvent
71
+ ) => {
53
72
  aggregateContent({ event, data });
54
73
  },
55
74
  },
56
75
  };
57
76
 
58
77
  const baseLlmConfig: t.LLMConfig = getLLMConfig(Providers.ANTHROPIC);
59
-
78
+
60
79
  // Enable thinking with token budget
61
80
  const llmConfig = {
62
81
  ...baseLlmConfig,
63
82
  model: 'claude-3-7-sonnet-latest',
64
- thinking: { type: "enabled", budget_tokens: 2000 }
83
+ thinking: { type: 'enabled', budget_tokens: 2000 },
65
84
  };
66
85
 
67
86
  const run = await Run.create<t.IState>({
@@ -93,7 +112,7 @@ async function testThinking(): Promise<void> {
93
112
  console.log('Running first query with thinking enabled...');
94
113
  const firstInputs = { messages: [...conversationHistory] };
95
114
  await run.processStream(firstInputs, config);
96
-
115
+
97
116
  // Extract and display thinking blocks
98
117
  const finalMessages = run.getRunMessages();
99
118
 
@@ -101,30 +120,32 @@ async function testThinking(): Promise<void> {
101
120
  console.log('\n\nTest 2: Multi-turn conversation with thinking enabled');
102
121
  const userMessage2 = `Given your previous analysis, what would be the most significant technical challenges in making this transition?`;
103
122
  conversationHistory.push(new HumanMessage(userMessage2));
104
-
123
+
105
124
  console.log('Running second query with thinking enabled...');
106
125
  const secondInputs = { messages: [...conversationHistory] };
107
126
  await run.processStream(secondInputs, config);
108
-
127
+
109
128
  // Display thinking blocks for second response
110
129
  const finalMessages2 = run.getRunMessages();
111
130
 
112
131
  // Test 3: Redacted thinking mode
113
132
  console.log('\n\nTest 3: Redacted thinking mode');
114
- const magicString = "ANTHROPIC_MAGIC_STRING_TRIGGER_REDACTED_THINKING_46C9A13E193C177646C7398A98432ECCCE4C1253D5E2D82641AC0E52CC2876CB";
133
+ const magicString =
134
+ 'ANTHROPIC_MAGIC_STRING_TRIGGER_REDACTED_THINKING_46C9A13E193C177646C7398A98432ECCCE4C1253D5E2D82641AC0E52CC2876CB';
115
135
  const userMessage3 = `${magicString}\n\nExplain how quantum computing works in simple terms.`;
116
-
136
+
117
137
  // Reset conversation for clean test
118
138
  conversationHistory.length = 0;
119
139
  conversationHistory.push(new HumanMessage(userMessage3));
120
-
140
+
121
141
  console.log('Running query with redacted thinking...');
122
142
  const thirdInputs = { messages: [...conversationHistory] };
123
143
  await run.processStream(thirdInputs, config);
124
-
144
+
125
145
  // Display redacted thinking blocks
126
146
  const finalMessages3 = run.getRunMessages();
127
147
  console.log('\n\nThinking feature test completed!');
148
+ console.dir(finalMessages3, { depth: null });
128
149
  }
129
150
 
130
151
  process.on('unhandledRejection', (reason, promise) => {
@@ -147,4 +168,4 @@ testThinking().catch((err) => {
147
168
  console.log('Content parts:');
148
169
  console.dir(_contentParts, { depth: null });
149
170
  process.exit(1);
150
- });
171
+ });
@@ -18,9 +18,13 @@ async function testStandardStreaming(): Promise<void> {
18
18
  const { userName, location, provider, currentDate } = await getArgs();
19
19
  const { contentParts, aggregateContent } = createContentAggregator();
20
20
  const customHandlers = {
21
- [GraphEvents.TOOL_END]: new ToolEndHandler(undefined, (name?: string) => {
22
- return true;
23
- }),
21
+ [GraphEvents.TOOL_END]: new ToolEndHandler(
22
+ undefined,
23
+ undefined,
24
+ (name?: string) => {
25
+ return true;
26
+ }
27
+ ),
24
28
  [GraphEvents.CHAT_MODEL_END]: {
25
29
  handle: (
26
30
  _event: string,
package/src/stream.ts CHANGED
@@ -339,7 +339,8 @@ hasToolCallChunks: ${hasToolCallChunks}
339
339
  (c) =>
340
340
  (c.type?.startsWith(ContentTypes.THINKING) ?? false) ||
341
341
  (c.type?.startsWith(ContentTypes.REASONING) ?? false) ||
342
- (c.type?.startsWith(ContentTypes.REASONING_CONTENT) ?? false)
342
+ (c.type?.startsWith(ContentTypes.REASONING_CONTENT) ?? false) ||
343
+ c.type === 'redacted_thinking'
343
344
  )
344
345
  ) {
345
346
  await graph.dispatchReasoningDelta(stepId, {
@@ -365,7 +366,8 @@ hasToolCallChunks: ${hasToolCallChunks}
365
366
  Array.isArray(chunk.content) &&
366
367
  (chunk.content[0]?.type === ContentTypes.THINKING ||
367
368
  chunk.content[0]?.type === ContentTypes.REASONING ||
368
- chunk.content[0]?.type === ContentTypes.REASONING_CONTENT)
369
+ chunk.content[0]?.type === ContentTypes.REASONING_CONTENT ||
370
+ chunk.content[0]?.type === 'redacted_thinking')
369
371
  ) {
370
372
  reasoning_content = 'valid';
371
373
  } else if (