@eko-ai/eko 2.1.0-alpha.5 → 2.1.0-alpha.7

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -5,6 +5,7 @@ type GlobalConfig = {
5
5
  maxTokens: number;
6
6
  compressThreshold: number;
7
7
  largeTextLength: number;
8
+ maxDialogueImgFileNum: number;
8
9
  };
9
10
  declare const config: GlobalConfig;
10
11
  export default config;
@@ -1 +1 @@
1
- {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/config/index.ts"],"names":[],"mappings":"AAAA,KAAK,YAAY,GAAG;IAClB,IAAI,EAAE,MAAM,CAAC;IACb,QAAQ,EAAE,SAAS,GAAG,KAAK,GAAG,OAAO,CAAC;IACtC,WAAW,EAAE,MAAM,CAAC;IACpB,SAAS,EAAE,MAAM,CAAC;IAClB,iBAAiB,EAAE,MAAM,CAAC;IAC1B,eAAe,EAAE,MAAM,CAAC;CACzB,CAAA;AAED,QAAA,MAAM,MAAM,EAAE,YAOb,CAAC;AAEF,eAAe,MAAM,CAAC"}
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/config/index.ts"],"names":[],"mappings":"AAAA,KAAK,YAAY,GAAG;IAClB,IAAI,EAAE,MAAM,CAAC;IACb,QAAQ,EAAE,SAAS,GAAG,KAAK,GAAG,OAAO,CAAC;IACtC,WAAW,EAAE,MAAM,CAAC;IACpB,SAAS,EAAE,MAAM,CAAC;IAClB,iBAAiB,EAAE,MAAM,CAAC;IAC1B,eAAe,EAAE,MAAM,CAAC;IACxB,qBAAqB,EAAE,MAAM,CAAC;CAC/B,CAAA;AAED,QAAA,MAAM,MAAM,EAAE,YAQb,CAAC;AAEF,eAAe,MAAM,CAAC"}
package/dist/index.cjs.js CHANGED
@@ -11,6 +11,7 @@ const config = {
11
11
  maxTokens: 16000,
12
12
  compressThreshold: 80,
13
13
  largeTextLength: 5000,
14
+ maxDialogueImgFileNum: 2,
14
15
  };
15
16
 
16
17
  var LogLevel;
@@ -17682,7 +17683,7 @@ async function compressAgentMessages(agentContext, rlm, messages, tools) {
17682
17683
  toolName: toolCall.toolName,
17683
17684
  params: args,
17684
17685
  toolResult: toolResult,
17685
- });
17686
+ }, agentContext);
17686
17687
  }
17687
17688
  // handle original messages
17688
17689
  let firstToolIndex = 3;
@@ -17701,6 +17702,7 @@ async function compressAgentMessages(agentContext, rlm, messages, tools) {
17701
17702
  function handleLargeContextMessages(messages) {
17702
17703
  let imageNum = 0;
17703
17704
  let fileNum = 0;
17705
+ let maxNum = config.maxDialogueImgFileNum;
17704
17706
  let longTextTools = {};
17705
17707
  for (let i = messages.length - 1; i >= 0; i--) {
17706
17708
  let message = messages[i];
@@ -17708,7 +17710,7 @@ function handleLargeContextMessages(messages) {
17708
17710
  for (let j = 0; j < message.content.length; j++) {
17709
17711
  let content = message.content[j];
17710
17712
  if (content.type == "image") {
17711
- if (++imageNum == 1) {
17713
+ if (++imageNum <= maxNum) {
17712
17714
  break;
17713
17715
  }
17714
17716
  content = {
@@ -17718,7 +17720,7 @@ function handleLargeContextMessages(messages) {
17718
17720
  message.content[j] = content;
17719
17721
  }
17720
17722
  else if (content.type == "file") {
17721
- if (++fileNum == 1) {
17723
+ if (++fileNum <= maxNum) {
17722
17724
  break;
17723
17725
  }
17724
17726
  content = {
@@ -17739,7 +17741,7 @@ function handleLargeContextMessages(messages) {
17739
17741
  for (let r = 0; r < toolContent.length; r++) {
17740
17742
  let _content = toolContent[r];
17741
17743
  if (_content.type == "image") {
17742
- if (++imageNum == 1) {
17744
+ if (++imageNum <= maxNum) {
17743
17745
  break;
17744
17746
  }
17745
17747
  _content = {
@@ -18385,7 +18387,7 @@ class Agent {
18385
18387
  toolName: result.toolName,
18386
18388
  params: result.args || {},
18387
18389
  toolResult: toolResult,
18388
- });
18390
+ }, agentContext);
18389
18391
  }
18390
18392
  let llmToolResult = this.convertToolResult(result, toolResult, user_messages);
18391
18393
  toolResults.push(llmToolResult);
@@ -18641,7 +18643,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18641
18643
  streamId,
18642
18644
  streamDone: false,
18643
18645
  text: streamText,
18644
- });
18646
+ }, agentContext);
18645
18647
  break;
18646
18648
  }
18647
18649
  case "reasoning": {
@@ -18654,7 +18656,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18654
18656
  streamId,
18655
18657
  streamDone: false,
18656
18658
  text: thinkText,
18657
- });
18659
+ }, agentContext);
18658
18660
  break;
18659
18661
  }
18660
18662
  case "tool-call-delta": {
@@ -18668,7 +18670,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18668
18670
  streamId,
18669
18671
  streamDone: true,
18670
18672
  text: streamText,
18671
- });
18673
+ }, agentContext);
18672
18674
  }
18673
18675
  toolArgsText += chunk.argsTextDelta || "";
18674
18676
  await streamCallback.onMessage({
@@ -18679,7 +18681,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18679
18681
  toolId: chunk.toolCallId,
18680
18682
  toolName: chunk.toolName,
18681
18683
  paramsText: toolArgsText,
18682
- });
18684
+ }, agentContext);
18683
18685
  break;
18684
18686
  }
18685
18687
  case "tool-call": {
@@ -18694,7 +18696,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18694
18696
  toolName: chunk.toolName,
18695
18697
  params: args,
18696
18698
  };
18697
- await streamCallback.onMessage(message);
18699
+ await streamCallback.onMessage(message, agentContext);
18698
18700
  toolParts.push({
18699
18701
  type: "tool-call",
18700
18702
  toolCallId: chunk.toolCallId,
@@ -18711,7 +18713,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18711
18713
  type: "file",
18712
18714
  mimeType: chunk.mimeType,
18713
18715
  data: chunk.data,
18714
- });
18716
+ }, agentContext);
18715
18717
  break;
18716
18718
  }
18717
18719
  case "error": {
@@ -18722,7 +18724,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18722
18724
  nodeId: agentNode.id,
18723
18725
  type: "error",
18724
18726
  error: chunk.error,
18725
- });
18727
+ }, agentContext);
18726
18728
  throw new Error("Plan Error");
18727
18729
  }
18728
18730
  case "finish": {
@@ -18736,7 +18738,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18736
18738
  streamId,
18737
18739
  streamDone: true,
18738
18740
  text: streamText,
18739
- });
18741
+ }, agentContext);
18740
18742
  }
18741
18743
  await streamCallback.onMessage({
18742
18744
  taskId: context.taskId,
@@ -18745,7 +18747,7 @@ async function callLLM(agentContext, rlm, messages, tools, noCompress, toolChoic
18745
18747
  type: "finish",
18746
18748
  finishReason: chunk.finishReason,
18747
18749
  usage: chunk.usage,
18748
- });
18750
+ }, agentContext);
18749
18751
  if (chunk.finishReason === "length" &&
18750
18752
  messages.length >= 10 &&
18751
18753
  !noCompress &&