@librechat/agents 3.1.74 → 3.1.75-dev.1
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +66 -0
- package/dist/cjs/agents/AgentContext.cjs +84 -37
- package/dist/cjs/agents/AgentContext.cjs.map +1 -1
- package/dist/cjs/graphs/Graph.cjs +13 -3
- package/dist/cjs/graphs/Graph.cjs.map +1 -1
- package/dist/cjs/langchain/google-common.cjs +3 -0
- package/dist/cjs/langchain/google-common.cjs.map +1 -0
- package/dist/cjs/langchain/index.cjs +86 -0
- package/dist/cjs/langchain/index.cjs.map +1 -0
- package/dist/cjs/langchain/language_models/chat_models.cjs +3 -0
- package/dist/cjs/langchain/language_models/chat_models.cjs.map +1 -0
- package/dist/cjs/langchain/messages/tool.cjs +3 -0
- package/dist/cjs/langchain/messages/tool.cjs.map +1 -0
- package/dist/cjs/langchain/messages.cjs +51 -0
- package/dist/cjs/langchain/messages.cjs.map +1 -0
- package/dist/cjs/langchain/openai.cjs +3 -0
- package/dist/cjs/langchain/openai.cjs.map +1 -0
- package/dist/cjs/langchain/prompts.cjs +11 -0
- package/dist/cjs/langchain/prompts.cjs.map +1 -0
- package/dist/cjs/langchain/runnables.cjs +19 -0
- package/dist/cjs/langchain/runnables.cjs.map +1 -0
- package/dist/cjs/langchain/tools.cjs +23 -0
- package/dist/cjs/langchain/tools.cjs.map +1 -0
- package/dist/cjs/langchain/utils/env.cjs +11 -0
- package/dist/cjs/langchain/utils/env.cjs.map +1 -0
- package/dist/cjs/llm/anthropic/index.cjs +145 -52
- package/dist/cjs/llm/anthropic/index.cjs.map +1 -1
- package/dist/cjs/llm/anthropic/types.cjs.map +1 -1
- package/dist/cjs/llm/anthropic/utils/message_inputs.cjs +25 -15
- package/dist/cjs/llm/anthropic/utils/message_inputs.cjs.map +1 -1
- package/dist/cjs/llm/anthropic/utils/message_outputs.cjs +84 -70
- package/dist/cjs/llm/anthropic/utils/message_outputs.cjs.map +1 -1
- package/dist/cjs/llm/bedrock/index.cjs +1 -1
- package/dist/cjs/llm/bedrock/index.cjs.map +1 -1
- package/dist/cjs/llm/bedrock/utils/message_inputs.cjs +213 -3
- package/dist/cjs/llm/bedrock/utils/message_inputs.cjs.map +1 -1
- package/dist/cjs/llm/bedrock/utils/message_outputs.cjs +2 -1
- package/dist/cjs/llm/bedrock/utils/message_outputs.cjs.map +1 -1
- package/dist/cjs/llm/google/utils/common.cjs +5 -4
- package/dist/cjs/llm/google/utils/common.cjs.map +1 -1
- package/dist/cjs/llm/openai/index.cjs +468 -647
- package/dist/cjs/llm/openai/index.cjs.map +1 -1
- package/dist/cjs/llm/openai/utils/index.cjs +1 -448
- package/dist/cjs/llm/openai/utils/index.cjs.map +1 -1
- package/dist/cjs/llm/openrouter/index.cjs +57 -175
- package/dist/cjs/llm/openrouter/index.cjs.map +1 -1
- package/dist/cjs/llm/vertexai/index.cjs +5 -3
- package/dist/cjs/llm/vertexai/index.cjs.map +1 -1
- package/dist/cjs/main.cjs +83 -3
- package/dist/cjs/main.cjs.map +1 -1
- package/dist/cjs/messages/cache.cjs +39 -4
- package/dist/cjs/messages/cache.cjs.map +1 -1
- package/dist/cjs/messages/core.cjs +7 -6
- package/dist/cjs/messages/core.cjs.map +1 -1
- package/dist/cjs/messages/format.cjs +7 -6
- package/dist/cjs/messages/format.cjs.map +1 -1
- package/dist/cjs/messages/langchain.cjs +26 -0
- package/dist/cjs/messages/langchain.cjs.map +1 -0
- package/dist/cjs/messages/prune.cjs +7 -6
- package/dist/cjs/messages/prune.cjs.map +1 -1
- package/dist/cjs/tools/ToolNode.cjs +5 -1
- package/dist/cjs/tools/ToolNode.cjs.map +1 -1
- package/dist/esm/agents/AgentContext.mjs +85 -38
- package/dist/esm/agents/AgentContext.mjs.map +1 -1
- package/dist/esm/graphs/Graph.mjs +13 -3
- package/dist/esm/graphs/Graph.mjs.map +1 -1
- package/dist/esm/langchain/google-common.mjs +2 -0
- package/dist/esm/langchain/google-common.mjs.map +1 -0
- package/dist/esm/langchain/index.mjs +5 -0
- package/dist/esm/langchain/index.mjs.map +1 -0
- package/dist/esm/langchain/language_models/chat_models.mjs +2 -0
- package/dist/esm/langchain/language_models/chat_models.mjs.map +1 -0
- package/dist/esm/langchain/messages/tool.mjs +2 -0
- package/dist/esm/langchain/messages/tool.mjs.map +1 -0
- package/dist/esm/langchain/messages.mjs +2 -0
- package/dist/esm/langchain/messages.mjs.map +1 -0
- package/dist/esm/langchain/openai.mjs +2 -0
- package/dist/esm/langchain/openai.mjs.map +1 -0
- package/dist/esm/langchain/prompts.mjs +2 -0
- package/dist/esm/langchain/prompts.mjs.map +1 -0
- package/dist/esm/langchain/runnables.mjs +2 -0
- package/dist/esm/langchain/runnables.mjs.map +1 -0
- package/dist/esm/langchain/tools.mjs +2 -0
- package/dist/esm/langchain/tools.mjs.map +1 -0
- package/dist/esm/langchain/utils/env.mjs +2 -0
- package/dist/esm/langchain/utils/env.mjs.map +1 -0
- package/dist/esm/llm/anthropic/index.mjs +146 -54
- package/dist/esm/llm/anthropic/index.mjs.map +1 -1
- package/dist/esm/llm/anthropic/types.mjs.map +1 -1
- package/dist/esm/llm/anthropic/utils/message_inputs.mjs +25 -15
- package/dist/esm/llm/anthropic/utils/message_inputs.mjs.map +1 -1
- package/dist/esm/llm/anthropic/utils/message_outputs.mjs +84 -71
- package/dist/esm/llm/anthropic/utils/message_outputs.mjs.map +1 -1
- package/dist/esm/llm/bedrock/index.mjs +1 -1
- package/dist/esm/llm/bedrock/index.mjs.map +1 -1
- package/dist/esm/llm/bedrock/utils/message_inputs.mjs +214 -4
- package/dist/esm/llm/bedrock/utils/message_inputs.mjs.map +1 -1
- package/dist/esm/llm/bedrock/utils/message_outputs.mjs +2 -1
- package/dist/esm/llm/bedrock/utils/message_outputs.mjs.map +1 -1
- package/dist/esm/llm/google/utils/common.mjs +5 -4
- package/dist/esm/llm/google/utils/common.mjs.map +1 -1
- package/dist/esm/llm/openai/index.mjs +469 -648
- package/dist/esm/llm/openai/index.mjs.map +1 -1
- package/dist/esm/llm/openai/utils/index.mjs +4 -449
- package/dist/esm/llm/openai/utils/index.mjs.map +1 -1
- package/dist/esm/llm/openrouter/index.mjs +57 -175
- package/dist/esm/llm/openrouter/index.mjs.map +1 -1
- package/dist/esm/llm/vertexai/index.mjs +5 -3
- package/dist/esm/llm/vertexai/index.mjs.map +1 -1
- package/dist/esm/main.mjs +4 -0
- package/dist/esm/main.mjs.map +1 -1
- package/dist/esm/messages/cache.mjs +39 -4
- package/dist/esm/messages/cache.mjs.map +1 -1
- package/dist/esm/messages/core.mjs +7 -6
- package/dist/esm/messages/core.mjs.map +1 -1
- package/dist/esm/messages/format.mjs +7 -6
- package/dist/esm/messages/format.mjs.map +1 -1
- package/dist/esm/messages/langchain.mjs +23 -0
- package/dist/esm/messages/langchain.mjs.map +1 -0
- package/dist/esm/messages/prune.mjs +7 -6
- package/dist/esm/messages/prune.mjs.map +1 -1
- package/dist/esm/tools/ToolNode.mjs +5 -1
- package/dist/esm/tools/ToolNode.mjs.map +1 -1
- package/dist/types/agents/AgentContext.d.ts +14 -4
- package/dist/types/agents/__tests__/promptCacheLiveHelpers.d.ts +46 -0
- package/dist/types/index.d.ts +1 -0
- package/dist/types/langchain/google-common.d.ts +1 -0
- package/dist/types/langchain/index.d.ts +8 -0
- package/dist/types/langchain/language_models/chat_models.d.ts +1 -0
- package/dist/types/langchain/messages/tool.d.ts +1 -0
- package/dist/types/langchain/messages.d.ts +2 -0
- package/dist/types/langchain/openai.d.ts +1 -0
- package/dist/types/langchain/prompts.d.ts +1 -0
- package/dist/types/langchain/runnables.d.ts +2 -0
- package/dist/types/langchain/tools.d.ts +2 -0
- package/dist/types/langchain/utils/env.d.ts +1 -0
- package/dist/types/llm/anthropic/index.d.ts +22 -9
- package/dist/types/llm/anthropic/types.d.ts +5 -1
- package/dist/types/llm/anthropic/utils/message_outputs.d.ts +13 -6
- package/dist/types/llm/anthropic/utils/output_parsers.d.ts +1 -1
- package/dist/types/llm/openai/index.d.ts +21 -24
- package/dist/types/llm/openrouter/index.d.ts +11 -9
- package/dist/types/llm/vertexai/index.d.ts +1 -0
- package/dist/types/messages/cache.d.ts +4 -1
- package/dist/types/messages/langchain.d.ts +27 -0
- package/dist/types/types/graph.d.ts +26 -38
- package/dist/types/types/llm.d.ts +3 -3
- package/dist/types/types/run.d.ts +2 -0
- package/dist/types/types/stream.d.ts +1 -1
- package/package.json +80 -17
- package/src/agents/AgentContext.ts +123 -44
- package/src/agents/__tests__/AgentContext.anthropic.live.test.ts +116 -0
- package/src/agents/__tests__/AgentContext.bedrock.live.test.ts +149 -0
- package/src/agents/__tests__/AgentContext.test.ts +155 -2
- package/src/agents/__tests__/promptCacheLiveHelpers.ts +165 -0
- package/src/graphs/Graph.ts +24 -4
- package/src/graphs/__tests__/composition.smoke.test.ts +188 -0
- package/src/index.ts +3 -0
- package/src/langchain/google-common.ts +1 -0
- package/src/langchain/index.ts +8 -0
- package/src/langchain/language_models/chat_models.ts +1 -0
- package/src/langchain/messages/tool.ts +5 -0
- package/src/langchain/messages.ts +21 -0
- package/src/langchain/openai.ts +1 -0
- package/src/langchain/prompts.ts +1 -0
- package/src/langchain/runnables.ts +7 -0
- package/src/langchain/tools.ts +8 -0
- package/src/langchain/utils/env.ts +1 -0
- package/src/llm/anthropic/index.ts +252 -84
- package/src/llm/anthropic/llm.spec.ts +751 -102
- package/src/llm/anthropic/types.ts +9 -1
- package/src/llm/anthropic/utils/message_inputs.ts +43 -20
- package/src/llm/anthropic/utils/message_outputs.ts +119 -101
- package/src/llm/anthropic/utils/server-tool-inputs.test.ts +77 -0
- package/src/llm/bedrock/index.ts +2 -2
- package/src/llm/bedrock/llm.spec.ts +341 -0
- package/src/llm/bedrock/utils/message_inputs.ts +303 -4
- package/src/llm/bedrock/utils/message_outputs.ts +2 -1
- package/src/llm/custom-chat-models.smoke.test.ts +662 -0
- package/src/llm/google/llm.spec.ts +339 -57
- package/src/llm/google/utils/common.ts +53 -48
- package/src/llm/openai/contentBlocks.test.ts +346 -0
- package/src/llm/openai/index.ts +736 -837
- package/src/llm/openai/utils/index.ts +84 -64
- package/src/llm/openrouter/index.ts +124 -247
- package/src/llm/openrouter/reasoning.test.ts +8 -1
- package/src/llm/vertexai/index.ts +11 -5
- package/src/llm/vertexai/llm.spec.ts +28 -1
- package/src/messages/cache.test.ts +106 -4
- package/src/messages/cache.ts +57 -5
- package/src/messages/core.ts +16 -9
- package/src/messages/format.ts +9 -6
- package/src/messages/langchain.ts +39 -0
- package/src/messages/prune.ts +12 -8
- package/src/scripts/caching.ts +2 -3
- package/src/specs/anthropic.simple.test.ts +61 -0
- package/src/specs/summarization.test.ts +58 -61
- package/src/tools/ToolNode.ts +5 -1
- package/src/types/graph.ts +35 -88
- package/src/types/llm.ts +3 -3
- package/src/types/run.ts +2 -0
- package/src/types/stream.ts +1 -1
- package/src/utils/llmConfig.ts +1 -6
|
@@ -1,7 +1,7 @@
|
|
|
1
1
|
'use strict';
|
|
2
2
|
|
|
3
3
|
var messages = require('@langchain/core/messages');
|
|
4
|
-
|
|
4
|
+
require('@langchain/core/outputs');
|
|
5
5
|
var openai_tools = require('@langchain/core/output_parsers/openai_tools');
|
|
6
6
|
|
|
7
7
|
function extractGenericMessageCustomRole(message) {
|
|
@@ -308,458 +308,11 @@ function _convertMessagesToOpenAIParams(messages$1, model, options) {
|
|
|
308
308
|
return completionParam;
|
|
309
309
|
});
|
|
310
310
|
}
|
|
311
|
-
const _FUNCTION_CALL_IDS_MAP_KEY = '__openai_function_call_ids__';
|
|
312
|
-
function _convertReasoningSummaryToOpenAIResponsesParams(reasoning) {
|
|
313
|
-
// combine summary parts that have the the same index and then remove the indexes
|
|
314
|
-
const summary = (reasoning.summary.length > 1
|
|
315
|
-
? reasoning.summary.reduce((acc, curr) => {
|
|
316
|
-
const last = acc.at(-1);
|
|
317
|
-
if (last.index === curr.index) {
|
|
318
|
-
last.text += curr.text;
|
|
319
|
-
}
|
|
320
|
-
else {
|
|
321
|
-
acc.push(curr);
|
|
322
|
-
}
|
|
323
|
-
return acc;
|
|
324
|
-
}, [{ ...reasoning.summary[0] }])
|
|
325
|
-
: reasoning.summary).map((s) => Object.fromEntries(Object.entries(s).filter(([k]) => k !== 'index')));
|
|
326
|
-
return {
|
|
327
|
-
...reasoning,
|
|
328
|
-
summary,
|
|
329
|
-
};
|
|
330
|
-
}
|
|
331
|
-
function _convertMessagesToOpenAIResponsesParams(messages$1, model, zdrEnabled) {
|
|
332
|
-
return messages$1.flatMap((lcMsg) => {
|
|
333
|
-
const additional_kwargs = lcMsg.additional_kwargs;
|
|
334
|
-
let role = messageToOpenAIRole(lcMsg);
|
|
335
|
-
if (role === 'system' && isReasoningModel(model))
|
|
336
|
-
role = 'developer';
|
|
337
|
-
if (role === 'function') {
|
|
338
|
-
throw new Error('Function messages are not supported in Responses API');
|
|
339
|
-
}
|
|
340
|
-
if (role === 'tool') {
|
|
341
|
-
const toolMessage = lcMsg;
|
|
342
|
-
// Handle computer call output
|
|
343
|
-
if (additional_kwargs.type === 'computer_call_output') {
|
|
344
|
-
const output = (() => {
|
|
345
|
-
if (typeof toolMessage.content === 'string') {
|
|
346
|
-
return {
|
|
347
|
-
type: 'computer_screenshot',
|
|
348
|
-
image_url: toolMessage.content,
|
|
349
|
-
};
|
|
350
|
-
}
|
|
351
|
-
if (Array.isArray(toolMessage.content)) {
|
|
352
|
-
const oaiScreenshot = toolMessage.content.find((i) => i.type === 'computer_screenshot');
|
|
353
|
-
if (oaiScreenshot)
|
|
354
|
-
return oaiScreenshot;
|
|
355
|
-
const lcImage = toolMessage.content.find((i) => i.type === 'image_url');
|
|
356
|
-
if (lcImage) {
|
|
357
|
-
return {
|
|
358
|
-
type: 'computer_screenshot',
|
|
359
|
-
image_url: typeof lcImage.image_url === 'string'
|
|
360
|
-
? lcImage.image_url
|
|
361
|
-
: lcImage.image_url.url,
|
|
362
|
-
};
|
|
363
|
-
}
|
|
364
|
-
}
|
|
365
|
-
throw new Error('Invalid computer call output');
|
|
366
|
-
})();
|
|
367
|
-
return {
|
|
368
|
-
type: 'computer_call_output',
|
|
369
|
-
output,
|
|
370
|
-
call_id: toolMessage.tool_call_id,
|
|
371
|
-
};
|
|
372
|
-
}
|
|
373
|
-
return {
|
|
374
|
-
type: 'function_call_output',
|
|
375
|
-
call_id: toolMessage.tool_call_id,
|
|
376
|
-
id: toolMessage.id?.startsWith('fc_') ? toolMessage.id : undefined,
|
|
377
|
-
output: typeof toolMessage.content !== 'string'
|
|
378
|
-
? JSON.stringify(toolMessage.content)
|
|
379
|
-
: toolMessage.content,
|
|
380
|
-
};
|
|
381
|
-
}
|
|
382
|
-
if (role === 'assistant') {
|
|
383
|
-
// if we have the original response items, just reuse them
|
|
384
|
-
if (!zdrEnabled &&
|
|
385
|
-
lcMsg.response_metadata.output != null &&
|
|
386
|
-
Array.isArray(lcMsg.response_metadata.output) &&
|
|
387
|
-
lcMsg.response_metadata.output.length > 0 &&
|
|
388
|
-
lcMsg.response_metadata.output.every((item) => 'type' in item)) {
|
|
389
|
-
return lcMsg.response_metadata.output;
|
|
390
|
-
}
|
|
391
|
-
// otherwise, try to reconstruct the response from what we have
|
|
392
|
-
const input = [];
|
|
393
|
-
// reasoning items
|
|
394
|
-
if (additional_kwargs.reasoning && !zdrEnabled) {
|
|
395
|
-
const reasoningItem = _convertReasoningSummaryToOpenAIResponsesParams(additional_kwargs.reasoning);
|
|
396
|
-
input.push(reasoningItem);
|
|
397
|
-
}
|
|
398
|
-
// ai content
|
|
399
|
-
let { content } = lcMsg;
|
|
400
|
-
if (additional_kwargs.refusal) {
|
|
401
|
-
if (typeof content === 'string') {
|
|
402
|
-
content = [{ type: 'output_text', text: content, annotations: [] }];
|
|
403
|
-
}
|
|
404
|
-
content = [
|
|
405
|
-
...content,
|
|
406
|
-
{ type: 'refusal', refusal: additional_kwargs.refusal },
|
|
407
|
-
];
|
|
408
|
-
}
|
|
409
|
-
input.push({
|
|
410
|
-
type: 'message',
|
|
411
|
-
role: 'assistant',
|
|
412
|
-
...(lcMsg.id && !zdrEnabled && lcMsg.id.startsWith('msg_')
|
|
413
|
-
? { id: lcMsg.id }
|
|
414
|
-
: {}),
|
|
415
|
-
content: typeof content === 'string'
|
|
416
|
-
? content
|
|
417
|
-
: content.flatMap((item) => {
|
|
418
|
-
if (item.type === 'text') {
|
|
419
|
-
return {
|
|
420
|
-
type: 'output_text',
|
|
421
|
-
text: item.text,
|
|
422
|
-
// @ts-expect-error TODO: add types for `annotations`
|
|
423
|
-
annotations: item.annotations ?? [],
|
|
424
|
-
};
|
|
425
|
-
}
|
|
426
|
-
if (item.type === 'output_text' || item.type === 'refusal') {
|
|
427
|
-
return item;
|
|
428
|
-
}
|
|
429
|
-
return [];
|
|
430
|
-
}),
|
|
431
|
-
});
|
|
432
|
-
const functionCallIds = additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY];
|
|
433
|
-
if (messages.isAIMessage(lcMsg) && !!lcMsg.tool_calls?.length) {
|
|
434
|
-
input.push(...lcMsg.tool_calls.map((toolCall) => ({
|
|
435
|
-
type: 'function_call',
|
|
436
|
-
name: toolCall.name,
|
|
437
|
-
arguments: JSON.stringify(toolCall.args),
|
|
438
|
-
call_id: toolCall.id,
|
|
439
|
-
...(zdrEnabled ? { id: functionCallIds?.[toolCall.id] } : {}),
|
|
440
|
-
})));
|
|
441
|
-
}
|
|
442
|
-
else if (additional_kwargs.tool_calls) {
|
|
443
|
-
input.push(...additional_kwargs.tool_calls.map((toolCall) => ({
|
|
444
|
-
type: 'function_call',
|
|
445
|
-
name: toolCall.function.name,
|
|
446
|
-
call_id: toolCall.id,
|
|
447
|
-
arguments: toolCall.function.arguments,
|
|
448
|
-
...(zdrEnabled ? { id: functionCallIds?.[toolCall.id] } : {}),
|
|
449
|
-
})));
|
|
450
|
-
}
|
|
451
|
-
const toolOutputs = (lcMsg.response_metadata.output?.length ?? 0) > 0
|
|
452
|
-
? lcMsg.response_metadata.output
|
|
453
|
-
: additional_kwargs.tool_outputs;
|
|
454
|
-
const fallthroughCallTypes = [
|
|
455
|
-
'computer_call',
|
|
456
|
-
/** @ts-ignore */
|
|
457
|
-
'mcp_call',
|
|
458
|
-
/** @ts-ignore */
|
|
459
|
-
'code_interpreter_call',
|
|
460
|
-
/** @ts-ignore */
|
|
461
|
-
'image_generation_call',
|
|
462
|
-
];
|
|
463
|
-
if (toolOutputs != null) {
|
|
464
|
-
const castToolOutputs = toolOutputs;
|
|
465
|
-
const fallthroughCalls = castToolOutputs.filter((item) => fallthroughCallTypes.includes(item.type));
|
|
466
|
-
if (fallthroughCalls.length > 0)
|
|
467
|
-
input.push(...fallthroughCalls);
|
|
468
|
-
}
|
|
469
|
-
return input;
|
|
470
|
-
}
|
|
471
|
-
if (role === 'user' || role === 'system' || role === 'developer') {
|
|
472
|
-
if (typeof lcMsg.content === 'string') {
|
|
473
|
-
return { type: 'message', role, content: lcMsg.content };
|
|
474
|
-
}
|
|
475
|
-
const messages$1 = [];
|
|
476
|
-
const content = lcMsg.content.flatMap((item) => {
|
|
477
|
-
if (item.type === 'mcp_approval_response') {
|
|
478
|
-
messages$1.push({
|
|
479
|
-
// @ts-ignore
|
|
480
|
-
type: 'mcp_approval_response',
|
|
481
|
-
approval_request_id: item.approval_request_id,
|
|
482
|
-
approve: item.approve,
|
|
483
|
-
});
|
|
484
|
-
}
|
|
485
|
-
if (messages.isDataContentBlock(item)) {
|
|
486
|
-
return messages.convertToProviderContentBlock(item, completionsApiContentBlockConverter);
|
|
487
|
-
}
|
|
488
|
-
if (item.type === 'text') {
|
|
489
|
-
return {
|
|
490
|
-
type: 'input_text',
|
|
491
|
-
text: item.text,
|
|
492
|
-
};
|
|
493
|
-
}
|
|
494
|
-
if (item.type === 'image_url') {
|
|
495
|
-
return {
|
|
496
|
-
type: 'input_image',
|
|
497
|
-
image_url: typeof item.image_url === 'string'
|
|
498
|
-
? item.image_url
|
|
499
|
-
: item.image_url.url,
|
|
500
|
-
detail: typeof item.image_url === 'string'
|
|
501
|
-
? 'auto'
|
|
502
|
-
: item.image_url.detail,
|
|
503
|
-
};
|
|
504
|
-
}
|
|
505
|
-
if (item.type === 'input_text' ||
|
|
506
|
-
item.type === 'input_image' ||
|
|
507
|
-
item.type === 'input_file') {
|
|
508
|
-
return item;
|
|
509
|
-
}
|
|
510
|
-
return [];
|
|
511
|
-
});
|
|
512
|
-
if (content.length > 0) {
|
|
513
|
-
messages$1.push({ type: 'message', role, content });
|
|
514
|
-
}
|
|
515
|
-
return messages$1;
|
|
516
|
-
}
|
|
517
|
-
console.warn(`Unsupported role found when converting to OpenAI Responses API: ${role}`);
|
|
518
|
-
return [];
|
|
519
|
-
});
|
|
520
|
-
}
|
|
521
311
|
function isReasoningModel(model) {
|
|
522
312
|
return model != null && model !== '' && /\b(o\d|gpt-[5-9])\b/i.test(model);
|
|
523
313
|
}
|
|
524
|
-
function _convertOpenAIResponsesMessageToBaseMessage(response) {
|
|
525
|
-
if (response.error) {
|
|
526
|
-
// TODO: add support for `addLangChainErrorFields`
|
|
527
|
-
const error = new Error(response.error.message);
|
|
528
|
-
error.name = response.error.code;
|
|
529
|
-
throw error;
|
|
530
|
-
}
|
|
531
|
-
let messageId;
|
|
532
|
-
const content = [];
|
|
533
|
-
const tool_calls = [];
|
|
534
|
-
const invalid_tool_calls = [];
|
|
535
|
-
const response_metadata = {
|
|
536
|
-
model: response.model,
|
|
537
|
-
created_at: response.created_at,
|
|
538
|
-
id: response.id,
|
|
539
|
-
incomplete_details: response.incomplete_details,
|
|
540
|
-
metadata: response.metadata,
|
|
541
|
-
object: response.object,
|
|
542
|
-
status: response.status,
|
|
543
|
-
user: response.user,
|
|
544
|
-
service_tier: response.service_tier,
|
|
545
|
-
// for compatibility with chat completion calls.
|
|
546
|
-
model_name: response.model,
|
|
547
|
-
};
|
|
548
|
-
const additional_kwargs = {};
|
|
549
|
-
for (const item of response.output) {
|
|
550
|
-
if (item.type === 'message') {
|
|
551
|
-
messageId = item.id;
|
|
552
|
-
content.push(...item.content.flatMap((part) => {
|
|
553
|
-
if (part.type === 'output_text') {
|
|
554
|
-
if ('parsed' in part && part.parsed != null) {
|
|
555
|
-
additional_kwargs.parsed = part.parsed;
|
|
556
|
-
}
|
|
557
|
-
return {
|
|
558
|
-
type: 'text',
|
|
559
|
-
text: part.text,
|
|
560
|
-
annotations: part.annotations,
|
|
561
|
-
};
|
|
562
|
-
}
|
|
563
|
-
if (part.type === 'refusal') {
|
|
564
|
-
additional_kwargs.refusal = part.refusal;
|
|
565
|
-
return [];
|
|
566
|
-
}
|
|
567
|
-
return part;
|
|
568
|
-
}));
|
|
569
|
-
}
|
|
570
|
-
else if (item.type === 'function_call') {
|
|
571
|
-
const fnAdapter = {
|
|
572
|
-
function: { name: item.name, arguments: item.arguments },
|
|
573
|
-
id: item.call_id,
|
|
574
|
-
};
|
|
575
|
-
try {
|
|
576
|
-
tool_calls.push(openai_tools.parseToolCall(fnAdapter, { returnId: true }));
|
|
577
|
-
}
|
|
578
|
-
catch (e) {
|
|
579
|
-
let errMessage;
|
|
580
|
-
if (typeof e === 'object' &&
|
|
581
|
-
e != null &&
|
|
582
|
-
'message' in e &&
|
|
583
|
-
typeof e.message === 'string') {
|
|
584
|
-
errMessage = e.message;
|
|
585
|
-
}
|
|
586
|
-
invalid_tool_calls.push(openai_tools.makeInvalidToolCall(fnAdapter, errMessage));
|
|
587
|
-
}
|
|
588
|
-
additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY] ??= {};
|
|
589
|
-
if (item.id) {
|
|
590
|
-
additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY][item.call_id] = item.id;
|
|
591
|
-
}
|
|
592
|
-
}
|
|
593
|
-
else if (item.type === 'reasoning') {
|
|
594
|
-
additional_kwargs.reasoning = item;
|
|
595
|
-
}
|
|
596
|
-
else {
|
|
597
|
-
additional_kwargs.tool_outputs ??= [];
|
|
598
|
-
additional_kwargs.tool_outputs.push(item);
|
|
599
|
-
}
|
|
600
|
-
}
|
|
601
|
-
return new messages.AIMessage({
|
|
602
|
-
id: messageId,
|
|
603
|
-
content,
|
|
604
|
-
tool_calls,
|
|
605
|
-
invalid_tool_calls,
|
|
606
|
-
usage_metadata: response.usage,
|
|
607
|
-
additional_kwargs,
|
|
608
|
-
response_metadata,
|
|
609
|
-
});
|
|
610
|
-
}
|
|
611
|
-
function _convertOpenAIResponsesDeltaToBaseMessageChunk(chunk) {
|
|
612
|
-
const content = [];
|
|
613
|
-
let generationInfo = {};
|
|
614
|
-
let usage_metadata;
|
|
615
|
-
const tool_call_chunks = [];
|
|
616
|
-
const response_metadata = {};
|
|
617
|
-
const additional_kwargs = {};
|
|
618
|
-
let id;
|
|
619
|
-
if (chunk.type === 'response.output_text.delta') {
|
|
620
|
-
content.push({
|
|
621
|
-
type: 'text',
|
|
622
|
-
text: chunk.delta,
|
|
623
|
-
index: chunk.content_index,
|
|
624
|
-
});
|
|
625
|
-
/** @ts-ignore */
|
|
626
|
-
}
|
|
627
|
-
else if (chunk.type === 'response.output_text_annotation.added') {
|
|
628
|
-
content.push({
|
|
629
|
-
type: 'text',
|
|
630
|
-
text: '',
|
|
631
|
-
/** @ts-ignore */
|
|
632
|
-
annotations: [chunk.annotation],
|
|
633
|
-
/** @ts-ignore */
|
|
634
|
-
index: chunk.content_index,
|
|
635
|
-
});
|
|
636
|
-
}
|
|
637
|
-
else if (chunk.type === 'response.output_item.added' &&
|
|
638
|
-
chunk.item.type === 'message') {
|
|
639
|
-
id = chunk.item.id;
|
|
640
|
-
}
|
|
641
|
-
else if (chunk.type === 'response.output_item.added' &&
|
|
642
|
-
chunk.item.type === 'function_call') {
|
|
643
|
-
tool_call_chunks.push({
|
|
644
|
-
type: 'tool_call_chunk',
|
|
645
|
-
name: chunk.item.name,
|
|
646
|
-
args: chunk.item.arguments,
|
|
647
|
-
id: chunk.item.call_id,
|
|
648
|
-
index: chunk.output_index,
|
|
649
|
-
});
|
|
650
|
-
additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY] = {
|
|
651
|
-
[chunk.item.call_id]: chunk.item.id,
|
|
652
|
-
};
|
|
653
|
-
}
|
|
654
|
-
else if (chunk.type === 'response.output_item.done' &&
|
|
655
|
-
[
|
|
656
|
-
'web_search_call',
|
|
657
|
-
'file_search_call',
|
|
658
|
-
'computer_call',
|
|
659
|
-
'code_interpreter_call',
|
|
660
|
-
'mcp_call',
|
|
661
|
-
'mcp_list_tools',
|
|
662
|
-
'mcp_approval_request',
|
|
663
|
-
'image_generation_call',
|
|
664
|
-
].includes(chunk.item.type)) {
|
|
665
|
-
additional_kwargs.tool_outputs = [chunk.item];
|
|
666
|
-
}
|
|
667
|
-
else if (chunk.type === 'response.created') {
|
|
668
|
-
response_metadata.id = chunk.response.id;
|
|
669
|
-
response_metadata.model_name = chunk.response.model;
|
|
670
|
-
response_metadata.model = chunk.response.model;
|
|
671
|
-
}
|
|
672
|
-
else if (chunk.type === 'response.completed') {
|
|
673
|
-
const msg = _convertOpenAIResponsesMessageToBaseMessage(chunk.response);
|
|
674
|
-
usage_metadata = chunk.response.usage;
|
|
675
|
-
if (chunk.response.text?.format?.type === 'json_schema') {
|
|
676
|
-
additional_kwargs.parsed ??= JSON.parse(msg.text);
|
|
677
|
-
}
|
|
678
|
-
for (const [key, value] of Object.entries(chunk.response)) {
|
|
679
|
-
if (key !== 'id')
|
|
680
|
-
response_metadata[key] = value;
|
|
681
|
-
}
|
|
682
|
-
}
|
|
683
|
-
else if (chunk.type === 'response.function_call_arguments.delta') {
|
|
684
|
-
tool_call_chunks.push({
|
|
685
|
-
type: 'tool_call_chunk',
|
|
686
|
-
args: chunk.delta,
|
|
687
|
-
index: chunk.output_index,
|
|
688
|
-
});
|
|
689
|
-
}
|
|
690
|
-
else if (chunk.type === 'response.web_search_call.completed' ||
|
|
691
|
-
chunk.type === 'response.file_search_call.completed') {
|
|
692
|
-
generationInfo = {
|
|
693
|
-
tool_outputs: {
|
|
694
|
-
id: chunk.item_id,
|
|
695
|
-
type: chunk.type.replace('response.', '').replace('.completed', ''),
|
|
696
|
-
status: 'completed',
|
|
697
|
-
},
|
|
698
|
-
};
|
|
699
|
-
}
|
|
700
|
-
else if (chunk.type === 'response.refusal.done') {
|
|
701
|
-
additional_kwargs.refusal = chunk.refusal;
|
|
702
|
-
}
|
|
703
|
-
else if (chunk.type === 'response.output_item.added' &&
|
|
704
|
-
'item' in chunk &&
|
|
705
|
-
chunk.item.type === 'reasoning') {
|
|
706
|
-
const summary = chunk
|
|
707
|
-
.item.summary
|
|
708
|
-
? chunk.item.summary.map((s, index) => ({
|
|
709
|
-
...s,
|
|
710
|
-
index,
|
|
711
|
-
}))
|
|
712
|
-
: undefined;
|
|
713
|
-
additional_kwargs.reasoning = {
|
|
714
|
-
// We only capture ID in the first chunk or else the concatenated result of all chunks will
|
|
715
|
-
// have an ID field that is repeated once per chunk. There is special handling for the `type`
|
|
716
|
-
// field that prevents this, however.
|
|
717
|
-
id: chunk.item.id,
|
|
718
|
-
type: chunk.item.type,
|
|
719
|
-
...(summary ? { summary } : {}),
|
|
720
|
-
};
|
|
721
|
-
}
|
|
722
|
-
else if (chunk.type === 'response.reasoning_summary_part.added') {
|
|
723
|
-
additional_kwargs.reasoning = {
|
|
724
|
-
type: 'reasoning',
|
|
725
|
-
summary: [{ ...chunk.part, index: chunk.summary_index }],
|
|
726
|
-
};
|
|
727
|
-
}
|
|
728
|
-
else if (chunk.type === 'response.reasoning_summary_text.delta') {
|
|
729
|
-
additional_kwargs.reasoning = {
|
|
730
|
-
type: 'reasoning',
|
|
731
|
-
summary: [
|
|
732
|
-
{ text: chunk.delta, type: 'summary_text', index: chunk.summary_index },
|
|
733
|
-
],
|
|
734
|
-
};
|
|
735
|
-
/** @ts-ignore */
|
|
736
|
-
}
|
|
737
|
-
else if (chunk.type === 'response.image_generation_call.partial_image') {
|
|
738
|
-
// noop/fixme: retaining partial images in a message chunk means that _all_
|
|
739
|
-
// partial images get kept in history, so we don't do anything here.
|
|
740
|
-
return null;
|
|
741
|
-
}
|
|
742
|
-
else {
|
|
743
|
-
return null;
|
|
744
|
-
}
|
|
745
|
-
return new outputs.ChatGenerationChunk({
|
|
746
|
-
// Legacy reasons, `onLLMNewToken` should pulls this out
|
|
747
|
-
text: content.map((part) => part.text).join(''),
|
|
748
|
-
message: new messages.AIMessageChunk({
|
|
749
|
-
id,
|
|
750
|
-
content,
|
|
751
|
-
tool_call_chunks,
|
|
752
|
-
usage_metadata,
|
|
753
|
-
additional_kwargs,
|
|
754
|
-
response_metadata,
|
|
755
|
-
}),
|
|
756
|
-
generationInfo,
|
|
757
|
-
});
|
|
758
|
-
}
|
|
759
314
|
|
|
760
315
|
exports._convertMessagesToOpenAIParams = _convertMessagesToOpenAIParams;
|
|
761
|
-
exports._convertMessagesToOpenAIResponsesParams = _convertMessagesToOpenAIResponsesParams;
|
|
762
|
-
exports._convertOpenAIResponsesDeltaToBaseMessageChunk = _convertOpenAIResponsesDeltaToBaseMessageChunk;
|
|
763
316
|
exports.isReasoningModel = isReasoningModel;
|
|
764
317
|
exports.messageToOpenAIRole = messageToOpenAIRole;
|
|
765
318
|
//# sourceMappingURL=index.cjs.map
|