@librechat/agents 3.1.73 → 3.1.75-dev.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (159) hide show
  1. package/README.md +66 -0
  2. package/dist/cjs/agents/AgentContext.cjs +146 -57
  3. package/dist/cjs/agents/AgentContext.cjs.map +1 -1
  4. package/dist/cjs/graphs/Graph.cjs +13 -3
  5. package/dist/cjs/graphs/Graph.cjs.map +1 -1
  6. package/dist/cjs/llm/anthropic/index.cjs +145 -52
  7. package/dist/cjs/llm/anthropic/index.cjs.map +1 -1
  8. package/dist/cjs/llm/anthropic/types.cjs.map +1 -1
  9. package/dist/cjs/llm/anthropic/utils/message_inputs.cjs +25 -15
  10. package/dist/cjs/llm/anthropic/utils/message_inputs.cjs.map +1 -1
  11. package/dist/cjs/llm/anthropic/utils/message_outputs.cjs +84 -70
  12. package/dist/cjs/llm/anthropic/utils/message_outputs.cjs.map +1 -1
  13. package/dist/cjs/llm/bedrock/index.cjs +1 -1
  14. package/dist/cjs/llm/bedrock/index.cjs.map +1 -1
  15. package/dist/cjs/llm/bedrock/utils/message_inputs.cjs +213 -3
  16. package/dist/cjs/llm/bedrock/utils/message_inputs.cjs.map +1 -1
  17. package/dist/cjs/llm/bedrock/utils/message_outputs.cjs +2 -1
  18. package/dist/cjs/llm/bedrock/utils/message_outputs.cjs.map +1 -1
  19. package/dist/cjs/llm/google/utils/common.cjs +5 -4
  20. package/dist/cjs/llm/google/utils/common.cjs.map +1 -1
  21. package/dist/cjs/llm/openai/index.cjs +468 -647
  22. package/dist/cjs/llm/openai/index.cjs.map +1 -1
  23. package/dist/cjs/llm/openai/utils/index.cjs +1 -448
  24. package/dist/cjs/llm/openai/utils/index.cjs.map +1 -1
  25. package/dist/cjs/llm/openrouter/index.cjs +57 -175
  26. package/dist/cjs/llm/openrouter/index.cjs.map +1 -1
  27. package/dist/cjs/llm/vertexai/index.cjs +5 -3
  28. package/dist/cjs/llm/vertexai/index.cjs.map +1 -1
  29. package/dist/cjs/main.cjs +1 -0
  30. package/dist/cjs/main.cjs.map +1 -1
  31. package/dist/cjs/messages/cache.cjs +39 -4
  32. package/dist/cjs/messages/cache.cjs.map +1 -1
  33. package/dist/cjs/messages/core.cjs +7 -6
  34. package/dist/cjs/messages/core.cjs.map +1 -1
  35. package/dist/cjs/messages/format.cjs +7 -6
  36. package/dist/cjs/messages/format.cjs.map +1 -1
  37. package/dist/cjs/messages/langchain.cjs +26 -0
  38. package/dist/cjs/messages/langchain.cjs.map +1 -0
  39. package/dist/cjs/messages/prune.cjs +7 -6
  40. package/dist/cjs/messages/prune.cjs.map +1 -1
  41. package/dist/cjs/tools/BashExecutor.cjs +21 -11
  42. package/dist/cjs/tools/BashExecutor.cjs.map +1 -1
  43. package/dist/cjs/tools/CodeExecutor.cjs +37 -10
  44. package/dist/cjs/tools/CodeExecutor.cjs.map +1 -1
  45. package/dist/cjs/tools/ProgrammaticToolCalling.cjs +16 -11
  46. package/dist/cjs/tools/ProgrammaticToolCalling.cjs.map +1 -1
  47. package/dist/cjs/tools/ToolNode.cjs +5 -1
  48. package/dist/cjs/tools/ToolNode.cjs.map +1 -1
  49. package/dist/esm/agents/AgentContext.mjs +147 -58
  50. package/dist/esm/agents/AgentContext.mjs.map +1 -1
  51. package/dist/esm/graphs/Graph.mjs +13 -3
  52. package/dist/esm/graphs/Graph.mjs.map +1 -1
  53. package/dist/esm/llm/anthropic/index.mjs +146 -54
  54. package/dist/esm/llm/anthropic/index.mjs.map +1 -1
  55. package/dist/esm/llm/anthropic/types.mjs.map +1 -1
  56. package/dist/esm/llm/anthropic/utils/message_inputs.mjs +25 -15
  57. package/dist/esm/llm/anthropic/utils/message_inputs.mjs.map +1 -1
  58. package/dist/esm/llm/anthropic/utils/message_outputs.mjs +84 -71
  59. package/dist/esm/llm/anthropic/utils/message_outputs.mjs.map +1 -1
  60. package/dist/esm/llm/bedrock/index.mjs +1 -1
  61. package/dist/esm/llm/bedrock/index.mjs.map +1 -1
  62. package/dist/esm/llm/bedrock/utils/message_inputs.mjs +214 -4
  63. package/dist/esm/llm/bedrock/utils/message_inputs.mjs.map +1 -1
  64. package/dist/esm/llm/bedrock/utils/message_outputs.mjs +2 -1
  65. package/dist/esm/llm/bedrock/utils/message_outputs.mjs.map +1 -1
  66. package/dist/esm/llm/google/utils/common.mjs +5 -4
  67. package/dist/esm/llm/google/utils/common.mjs.map +1 -1
  68. package/dist/esm/llm/openai/index.mjs +469 -648
  69. package/dist/esm/llm/openai/index.mjs.map +1 -1
  70. package/dist/esm/llm/openai/utils/index.mjs +4 -449
  71. package/dist/esm/llm/openai/utils/index.mjs.map +1 -1
  72. package/dist/esm/llm/openrouter/index.mjs +57 -175
  73. package/dist/esm/llm/openrouter/index.mjs.map +1 -1
  74. package/dist/esm/llm/vertexai/index.mjs +5 -3
  75. package/dist/esm/llm/vertexai/index.mjs.map +1 -1
  76. package/dist/esm/main.mjs +1 -1
  77. package/dist/esm/messages/cache.mjs +39 -4
  78. package/dist/esm/messages/cache.mjs.map +1 -1
  79. package/dist/esm/messages/core.mjs +7 -6
  80. package/dist/esm/messages/core.mjs.map +1 -1
  81. package/dist/esm/messages/format.mjs +7 -6
  82. package/dist/esm/messages/format.mjs.map +1 -1
  83. package/dist/esm/messages/langchain.mjs +23 -0
  84. package/dist/esm/messages/langchain.mjs.map +1 -0
  85. package/dist/esm/messages/prune.mjs +7 -6
  86. package/dist/esm/messages/prune.mjs.map +1 -1
  87. package/dist/esm/tools/BashExecutor.mjs +22 -12
  88. package/dist/esm/tools/BashExecutor.mjs.map +1 -1
  89. package/dist/esm/tools/CodeExecutor.mjs +37 -11
  90. package/dist/esm/tools/CodeExecutor.mjs.map +1 -1
  91. package/dist/esm/tools/ProgrammaticToolCalling.mjs +17 -12
  92. package/dist/esm/tools/ProgrammaticToolCalling.mjs.map +1 -1
  93. package/dist/esm/tools/ToolNode.mjs +5 -1
  94. package/dist/esm/tools/ToolNode.mjs.map +1 -1
  95. package/dist/types/agents/AgentContext.d.ts +29 -4
  96. package/dist/types/agents/__tests__/promptCacheLiveHelpers.d.ts +46 -0
  97. package/dist/types/llm/anthropic/index.d.ts +22 -9
  98. package/dist/types/llm/anthropic/types.d.ts +5 -1
  99. package/dist/types/llm/anthropic/utils/message_outputs.d.ts +13 -6
  100. package/dist/types/llm/anthropic/utils/output_parsers.d.ts +1 -1
  101. package/dist/types/llm/openai/index.d.ts +21 -24
  102. package/dist/types/llm/openrouter/index.d.ts +11 -9
  103. package/dist/types/llm/vertexai/index.d.ts +1 -0
  104. package/dist/types/messages/cache.d.ts +4 -1
  105. package/dist/types/messages/langchain.d.ts +27 -0
  106. package/dist/types/tools/CodeExecutor.d.ts +6 -0
  107. package/dist/types/types/graph.d.ts +26 -38
  108. package/dist/types/types/llm.d.ts +3 -3
  109. package/dist/types/types/run.d.ts +2 -0
  110. package/dist/types/types/stream.d.ts +1 -1
  111. package/dist/types/types/tools.d.ts +9 -0
  112. package/package.json +17 -16
  113. package/src/agents/AgentContext.ts +189 -71
  114. package/src/agents/__tests__/AgentContext.anthropic.live.test.ts +116 -0
  115. package/src/agents/__tests__/AgentContext.bedrock.live.test.ts +149 -0
  116. package/src/agents/__tests__/AgentContext.test.ts +333 -2
  117. package/src/agents/__tests__/promptCacheLiveHelpers.ts +165 -0
  118. package/src/graphs/Graph.ts +24 -4
  119. package/src/graphs/__tests__/composition.smoke.test.ts +188 -0
  120. package/src/llm/anthropic/index.ts +252 -84
  121. package/src/llm/anthropic/llm.spec.ts +751 -102
  122. package/src/llm/anthropic/types.ts +9 -1
  123. package/src/llm/anthropic/utils/message_inputs.ts +43 -20
  124. package/src/llm/anthropic/utils/message_outputs.ts +119 -101
  125. package/src/llm/anthropic/utils/server-tool-inputs.test.ts +77 -0
  126. package/src/llm/bedrock/index.ts +2 -2
  127. package/src/llm/bedrock/llm.spec.ts +341 -0
  128. package/src/llm/bedrock/utils/message_inputs.ts +303 -4
  129. package/src/llm/bedrock/utils/message_outputs.ts +2 -1
  130. package/src/llm/custom-chat-models.smoke.test.ts +662 -0
  131. package/src/llm/google/llm.spec.ts +339 -57
  132. package/src/llm/google/utils/common.ts +53 -48
  133. package/src/llm/openai/contentBlocks.test.ts +346 -0
  134. package/src/llm/openai/index.ts +736 -837
  135. package/src/llm/openai/utils/index.ts +84 -64
  136. package/src/llm/openrouter/index.ts +124 -247
  137. package/src/llm/openrouter/reasoning.test.ts +8 -1
  138. package/src/llm/vertexai/index.ts +11 -5
  139. package/src/llm/vertexai/llm.spec.ts +28 -1
  140. package/src/messages/cache.test.ts +106 -4
  141. package/src/messages/cache.ts +57 -5
  142. package/src/messages/core.ts +16 -9
  143. package/src/messages/format.ts +9 -6
  144. package/src/messages/langchain.ts +39 -0
  145. package/src/messages/prune.ts +12 -8
  146. package/src/scripts/caching.ts +2 -3
  147. package/src/specs/anthropic.simple.test.ts +61 -0
  148. package/src/specs/summarization.test.ts +58 -61
  149. package/src/tools/BashExecutor.ts +37 -13
  150. package/src/tools/CodeExecutor.ts +55 -11
  151. package/src/tools/ProgrammaticToolCalling.ts +29 -14
  152. package/src/tools/ToolNode.ts +5 -1
  153. package/src/tools/__tests__/ProgrammaticToolCalling.test.ts +60 -0
  154. package/src/types/graph.ts +35 -88
  155. package/src/types/llm.ts +3 -3
  156. package/src/types/run.ts +2 -0
  157. package/src/types/stream.ts +1 -1
  158. package/src/types/tools.ts +9 -0
  159. package/src/utils/llmConfig.ts +1 -6
@@ -1,6 +1,6 @@
1
- import { isDataContentBlock, convertToProviderContentBlock, isAIMessage, AIMessageChunk, ChatMessage, parseBase64DataUrl, parseMimeType, AIMessage } from '@langchain/core/messages';
2
- import { ChatGenerationChunk } from '@langchain/core/outputs';
3
- import { convertLangChainToolCallToOpenAI, parseToolCall, makeInvalidToolCall } from '@langchain/core/output_parsers/openai_tools';
1
+ import { isDataContentBlock, convertToProviderContentBlock, isAIMessage, ChatMessage, parseBase64DataUrl, parseMimeType } from '@langchain/core/messages';
2
+ import '@langchain/core/outputs';
3
+ import { convertLangChainToolCallToOpenAI } from '@langchain/core/output_parsers/openai_tools';
4
4
 
5
5
  function extractGenericMessageCustomRole(message) {
6
6
  if (message.role !== 'system' &&
@@ -306,454 +306,9 @@ function _convertMessagesToOpenAIParams(messages, model, options) {
306
306
  return completionParam;
307
307
  });
308
308
  }
309
- const _FUNCTION_CALL_IDS_MAP_KEY = '__openai_function_call_ids__';
310
- function _convertReasoningSummaryToOpenAIResponsesParams(reasoning) {
311
- // combine summary parts that have the the same index and then remove the indexes
312
- const summary = (reasoning.summary.length > 1
313
- ? reasoning.summary.reduce((acc, curr) => {
314
- const last = acc.at(-1);
315
- if (last.index === curr.index) {
316
- last.text += curr.text;
317
- }
318
- else {
319
- acc.push(curr);
320
- }
321
- return acc;
322
- }, [{ ...reasoning.summary[0] }])
323
- : reasoning.summary).map((s) => Object.fromEntries(Object.entries(s).filter(([k]) => k !== 'index')));
324
- return {
325
- ...reasoning,
326
- summary,
327
- };
328
- }
329
- function _convertMessagesToOpenAIResponsesParams(messages, model, zdrEnabled) {
330
- return messages.flatMap((lcMsg) => {
331
- const additional_kwargs = lcMsg.additional_kwargs;
332
- let role = messageToOpenAIRole(lcMsg);
333
- if (role === 'system' && isReasoningModel(model))
334
- role = 'developer';
335
- if (role === 'function') {
336
- throw new Error('Function messages are not supported in Responses API');
337
- }
338
- if (role === 'tool') {
339
- const toolMessage = lcMsg;
340
- // Handle computer call output
341
- if (additional_kwargs.type === 'computer_call_output') {
342
- const output = (() => {
343
- if (typeof toolMessage.content === 'string') {
344
- return {
345
- type: 'computer_screenshot',
346
- image_url: toolMessage.content,
347
- };
348
- }
349
- if (Array.isArray(toolMessage.content)) {
350
- const oaiScreenshot = toolMessage.content.find((i) => i.type === 'computer_screenshot');
351
- if (oaiScreenshot)
352
- return oaiScreenshot;
353
- const lcImage = toolMessage.content.find((i) => i.type === 'image_url');
354
- if (lcImage) {
355
- return {
356
- type: 'computer_screenshot',
357
- image_url: typeof lcImage.image_url === 'string'
358
- ? lcImage.image_url
359
- : lcImage.image_url.url,
360
- };
361
- }
362
- }
363
- throw new Error('Invalid computer call output');
364
- })();
365
- return {
366
- type: 'computer_call_output',
367
- output,
368
- call_id: toolMessage.tool_call_id,
369
- };
370
- }
371
- return {
372
- type: 'function_call_output',
373
- call_id: toolMessage.tool_call_id,
374
- id: toolMessage.id?.startsWith('fc_') ? toolMessage.id : undefined,
375
- output: typeof toolMessage.content !== 'string'
376
- ? JSON.stringify(toolMessage.content)
377
- : toolMessage.content,
378
- };
379
- }
380
- if (role === 'assistant') {
381
- // if we have the original response items, just reuse them
382
- if (!zdrEnabled &&
383
- lcMsg.response_metadata.output != null &&
384
- Array.isArray(lcMsg.response_metadata.output) &&
385
- lcMsg.response_metadata.output.length > 0 &&
386
- lcMsg.response_metadata.output.every((item) => 'type' in item)) {
387
- return lcMsg.response_metadata.output;
388
- }
389
- // otherwise, try to reconstruct the response from what we have
390
- const input = [];
391
- // reasoning items
392
- if (additional_kwargs.reasoning && !zdrEnabled) {
393
- const reasoningItem = _convertReasoningSummaryToOpenAIResponsesParams(additional_kwargs.reasoning);
394
- input.push(reasoningItem);
395
- }
396
- // ai content
397
- let { content } = lcMsg;
398
- if (additional_kwargs.refusal) {
399
- if (typeof content === 'string') {
400
- content = [{ type: 'output_text', text: content, annotations: [] }];
401
- }
402
- content = [
403
- ...content,
404
- { type: 'refusal', refusal: additional_kwargs.refusal },
405
- ];
406
- }
407
- input.push({
408
- type: 'message',
409
- role: 'assistant',
410
- ...(lcMsg.id && !zdrEnabled && lcMsg.id.startsWith('msg_')
411
- ? { id: lcMsg.id }
412
- : {}),
413
- content: typeof content === 'string'
414
- ? content
415
- : content.flatMap((item) => {
416
- if (item.type === 'text') {
417
- return {
418
- type: 'output_text',
419
- text: item.text,
420
- // @ts-expect-error TODO: add types for `annotations`
421
- annotations: item.annotations ?? [],
422
- };
423
- }
424
- if (item.type === 'output_text' || item.type === 'refusal') {
425
- return item;
426
- }
427
- return [];
428
- }),
429
- });
430
- const functionCallIds = additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY];
431
- if (isAIMessage(lcMsg) && !!lcMsg.tool_calls?.length) {
432
- input.push(...lcMsg.tool_calls.map((toolCall) => ({
433
- type: 'function_call',
434
- name: toolCall.name,
435
- arguments: JSON.stringify(toolCall.args),
436
- call_id: toolCall.id,
437
- ...(zdrEnabled ? { id: functionCallIds?.[toolCall.id] } : {}),
438
- })));
439
- }
440
- else if (additional_kwargs.tool_calls) {
441
- input.push(...additional_kwargs.tool_calls.map((toolCall) => ({
442
- type: 'function_call',
443
- name: toolCall.function.name,
444
- call_id: toolCall.id,
445
- arguments: toolCall.function.arguments,
446
- ...(zdrEnabled ? { id: functionCallIds?.[toolCall.id] } : {}),
447
- })));
448
- }
449
- const toolOutputs = (lcMsg.response_metadata.output?.length ?? 0) > 0
450
- ? lcMsg.response_metadata.output
451
- : additional_kwargs.tool_outputs;
452
- const fallthroughCallTypes = [
453
- 'computer_call',
454
- /** @ts-ignore */
455
- 'mcp_call',
456
- /** @ts-ignore */
457
- 'code_interpreter_call',
458
- /** @ts-ignore */
459
- 'image_generation_call',
460
- ];
461
- if (toolOutputs != null) {
462
- const castToolOutputs = toolOutputs;
463
- const fallthroughCalls = castToolOutputs.filter((item) => fallthroughCallTypes.includes(item.type));
464
- if (fallthroughCalls.length > 0)
465
- input.push(...fallthroughCalls);
466
- }
467
- return input;
468
- }
469
- if (role === 'user' || role === 'system' || role === 'developer') {
470
- if (typeof lcMsg.content === 'string') {
471
- return { type: 'message', role, content: lcMsg.content };
472
- }
473
- const messages = [];
474
- const content = lcMsg.content.flatMap((item) => {
475
- if (item.type === 'mcp_approval_response') {
476
- messages.push({
477
- // @ts-ignore
478
- type: 'mcp_approval_response',
479
- approval_request_id: item.approval_request_id,
480
- approve: item.approve,
481
- });
482
- }
483
- if (isDataContentBlock(item)) {
484
- return convertToProviderContentBlock(item, completionsApiContentBlockConverter);
485
- }
486
- if (item.type === 'text') {
487
- return {
488
- type: 'input_text',
489
- text: item.text,
490
- };
491
- }
492
- if (item.type === 'image_url') {
493
- return {
494
- type: 'input_image',
495
- image_url: typeof item.image_url === 'string'
496
- ? item.image_url
497
- : item.image_url.url,
498
- detail: typeof item.image_url === 'string'
499
- ? 'auto'
500
- : item.image_url.detail,
501
- };
502
- }
503
- if (item.type === 'input_text' ||
504
- item.type === 'input_image' ||
505
- item.type === 'input_file') {
506
- return item;
507
- }
508
- return [];
509
- });
510
- if (content.length > 0) {
511
- messages.push({ type: 'message', role, content });
512
- }
513
- return messages;
514
- }
515
- console.warn(`Unsupported role found when converting to OpenAI Responses API: ${role}`);
516
- return [];
517
- });
518
- }
519
309
  function isReasoningModel(model) {
520
310
  return model != null && model !== '' && /\b(o\d|gpt-[5-9])\b/i.test(model);
521
311
  }
522
- function _convertOpenAIResponsesMessageToBaseMessage(response) {
523
- if (response.error) {
524
- // TODO: add support for `addLangChainErrorFields`
525
- const error = new Error(response.error.message);
526
- error.name = response.error.code;
527
- throw error;
528
- }
529
- let messageId;
530
- const content = [];
531
- const tool_calls = [];
532
- const invalid_tool_calls = [];
533
- const response_metadata = {
534
- model: response.model,
535
- created_at: response.created_at,
536
- id: response.id,
537
- incomplete_details: response.incomplete_details,
538
- metadata: response.metadata,
539
- object: response.object,
540
- status: response.status,
541
- user: response.user,
542
- service_tier: response.service_tier,
543
- // for compatibility with chat completion calls.
544
- model_name: response.model,
545
- };
546
- const additional_kwargs = {};
547
- for (const item of response.output) {
548
- if (item.type === 'message') {
549
- messageId = item.id;
550
- content.push(...item.content.flatMap((part) => {
551
- if (part.type === 'output_text') {
552
- if ('parsed' in part && part.parsed != null) {
553
- additional_kwargs.parsed = part.parsed;
554
- }
555
- return {
556
- type: 'text',
557
- text: part.text,
558
- annotations: part.annotations,
559
- };
560
- }
561
- if (part.type === 'refusal') {
562
- additional_kwargs.refusal = part.refusal;
563
- return [];
564
- }
565
- return part;
566
- }));
567
- }
568
- else if (item.type === 'function_call') {
569
- const fnAdapter = {
570
- function: { name: item.name, arguments: item.arguments },
571
- id: item.call_id,
572
- };
573
- try {
574
- tool_calls.push(parseToolCall(fnAdapter, { returnId: true }));
575
- }
576
- catch (e) {
577
- let errMessage;
578
- if (typeof e === 'object' &&
579
- e != null &&
580
- 'message' in e &&
581
- typeof e.message === 'string') {
582
- errMessage = e.message;
583
- }
584
- invalid_tool_calls.push(makeInvalidToolCall(fnAdapter, errMessage));
585
- }
586
- additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY] ??= {};
587
- if (item.id) {
588
- additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY][item.call_id] = item.id;
589
- }
590
- }
591
- else if (item.type === 'reasoning') {
592
- additional_kwargs.reasoning = item;
593
- }
594
- else {
595
- additional_kwargs.tool_outputs ??= [];
596
- additional_kwargs.tool_outputs.push(item);
597
- }
598
- }
599
- return new AIMessage({
600
- id: messageId,
601
- content,
602
- tool_calls,
603
- invalid_tool_calls,
604
- usage_metadata: response.usage,
605
- additional_kwargs,
606
- response_metadata,
607
- });
608
- }
609
- function _convertOpenAIResponsesDeltaToBaseMessageChunk(chunk) {
610
- const content = [];
611
- let generationInfo = {};
612
- let usage_metadata;
613
- const tool_call_chunks = [];
614
- const response_metadata = {};
615
- const additional_kwargs = {};
616
- let id;
617
- if (chunk.type === 'response.output_text.delta') {
618
- content.push({
619
- type: 'text',
620
- text: chunk.delta,
621
- index: chunk.content_index,
622
- });
623
- /** @ts-ignore */
624
- }
625
- else if (chunk.type === 'response.output_text_annotation.added') {
626
- content.push({
627
- type: 'text',
628
- text: '',
629
- /** @ts-ignore */
630
- annotations: [chunk.annotation],
631
- /** @ts-ignore */
632
- index: chunk.content_index,
633
- });
634
- }
635
- else if (chunk.type === 'response.output_item.added' &&
636
- chunk.item.type === 'message') {
637
- id = chunk.item.id;
638
- }
639
- else if (chunk.type === 'response.output_item.added' &&
640
- chunk.item.type === 'function_call') {
641
- tool_call_chunks.push({
642
- type: 'tool_call_chunk',
643
- name: chunk.item.name,
644
- args: chunk.item.arguments,
645
- id: chunk.item.call_id,
646
- index: chunk.output_index,
647
- });
648
- additional_kwargs[_FUNCTION_CALL_IDS_MAP_KEY] = {
649
- [chunk.item.call_id]: chunk.item.id,
650
- };
651
- }
652
- else if (chunk.type === 'response.output_item.done' &&
653
- [
654
- 'web_search_call',
655
- 'file_search_call',
656
- 'computer_call',
657
- 'code_interpreter_call',
658
- 'mcp_call',
659
- 'mcp_list_tools',
660
- 'mcp_approval_request',
661
- 'image_generation_call',
662
- ].includes(chunk.item.type)) {
663
- additional_kwargs.tool_outputs = [chunk.item];
664
- }
665
- else if (chunk.type === 'response.created') {
666
- response_metadata.id = chunk.response.id;
667
- response_metadata.model_name = chunk.response.model;
668
- response_metadata.model = chunk.response.model;
669
- }
670
- else if (chunk.type === 'response.completed') {
671
- const msg = _convertOpenAIResponsesMessageToBaseMessage(chunk.response);
672
- usage_metadata = chunk.response.usage;
673
- if (chunk.response.text?.format?.type === 'json_schema') {
674
- additional_kwargs.parsed ??= JSON.parse(msg.text);
675
- }
676
- for (const [key, value] of Object.entries(chunk.response)) {
677
- if (key !== 'id')
678
- response_metadata[key] = value;
679
- }
680
- }
681
- else if (chunk.type === 'response.function_call_arguments.delta') {
682
- tool_call_chunks.push({
683
- type: 'tool_call_chunk',
684
- args: chunk.delta,
685
- index: chunk.output_index,
686
- });
687
- }
688
- else if (chunk.type === 'response.web_search_call.completed' ||
689
- chunk.type === 'response.file_search_call.completed') {
690
- generationInfo = {
691
- tool_outputs: {
692
- id: chunk.item_id,
693
- type: chunk.type.replace('response.', '').replace('.completed', ''),
694
- status: 'completed',
695
- },
696
- };
697
- }
698
- else if (chunk.type === 'response.refusal.done') {
699
- additional_kwargs.refusal = chunk.refusal;
700
- }
701
- else if (chunk.type === 'response.output_item.added' &&
702
- 'item' in chunk &&
703
- chunk.item.type === 'reasoning') {
704
- const summary = chunk
705
- .item.summary
706
- ? chunk.item.summary.map((s, index) => ({
707
- ...s,
708
- index,
709
- }))
710
- : undefined;
711
- additional_kwargs.reasoning = {
712
- // We only capture ID in the first chunk or else the concatenated result of all chunks will
713
- // have an ID field that is repeated once per chunk. There is special handling for the `type`
714
- // field that prevents this, however.
715
- id: chunk.item.id,
716
- type: chunk.item.type,
717
- ...(summary ? { summary } : {}),
718
- };
719
- }
720
- else if (chunk.type === 'response.reasoning_summary_part.added') {
721
- additional_kwargs.reasoning = {
722
- type: 'reasoning',
723
- summary: [{ ...chunk.part, index: chunk.summary_index }],
724
- };
725
- }
726
- else if (chunk.type === 'response.reasoning_summary_text.delta') {
727
- additional_kwargs.reasoning = {
728
- type: 'reasoning',
729
- summary: [
730
- { text: chunk.delta, type: 'summary_text', index: chunk.summary_index },
731
- ],
732
- };
733
- /** @ts-ignore */
734
- }
735
- else if (chunk.type === 'response.image_generation_call.partial_image') {
736
- // noop/fixme: retaining partial images in a message chunk means that _all_
737
- // partial images get kept in history, so we don't do anything here.
738
- return null;
739
- }
740
- else {
741
- return null;
742
- }
743
- return new ChatGenerationChunk({
744
- // Legacy reasons, `onLLMNewToken` should pulls this out
745
- text: content.map((part) => part.text).join(''),
746
- message: new AIMessageChunk({
747
- id,
748
- content,
749
- tool_call_chunks,
750
- usage_metadata,
751
- additional_kwargs,
752
- response_metadata,
753
- }),
754
- generationInfo,
755
- });
756
- }
757
312
 
758
- export { _convertMessagesToOpenAIParams, _convertMessagesToOpenAIResponsesParams, _convertOpenAIResponsesDeltaToBaseMessageChunk, isReasoningModel, messageToOpenAIRole };
313
+ export { _convertMessagesToOpenAIParams, isReasoningModel, messageToOpenAIRole };
759
314
  //# sourceMappingURL=index.mjs.map