@agentica/core 0.28.0 → 0.29.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -89,7 +89,7 @@ export async function call<Model extends ILlmSchema.Model>(
89
89
  }) as OpenAI.ChatCompletionTool,
90
90
  ),
91
91
  tool_choice: "auto",
92
- parallel_tool_calls: false,
92
+ // parallel_tool_calls: false,
93
93
  });
94
94
 
95
95
  // ----
@@ -493,7 +493,7 @@ async function correct<Model extends ILlmSchema.Model>(
493
493
  name: call.operation.name,
494
494
  },
495
495
  },
496
- parallel_tool_calls: false,
496
+ // parallel_tool_calls: false,
497
497
  });
498
498
 
499
499
  const chunks = await StreamUtil.readAll(completionStream);
@@ -172,7 +172,7 @@ async function step<Model extends ILlmSchema.Model>(
172
172
  name: CONTAINER.functions[0]!.name,
173
173
  },
174
174
  },
175
- parallel_tool_calls: true,
175
+ // parallel_tool_calls: true,
176
176
  });
177
177
 
178
178
  const chunks = await StreamUtil.readAll(completionStream);
@@ -62,7 +62,7 @@ export async function initialize<Model extends ILlmSchema.Model>(ctx: AgenticaCo
62
62
  },
63
63
  ],
64
64
  tool_choice: "auto",
65
- parallel_tool_calls: false,
65
+ // parallel_tool_calls: false,
66
66
  });
67
67
 
68
68
  const textContext: ({
@@ -171,7 +171,7 @@ async function step<Model extends ILlmSchema.Model>(
171
171
  name: CONTAINER.functions[0]!.name,
172
172
  },
173
173
  },
174
- parallel_tool_calls: false,
174
+ // parallel_tool_calls: false,
175
175
  });
176
176
 
177
177
  const chunks = await StreamUtil.readAll(completionStream);
@@ -5,8 +5,7 @@ import type {
5
5
  ChatCompletionMessageToolCall,
6
6
  } from "openai/resources";
7
7
 
8
- import typia from "typia";
9
-
8
+ // import typia from "typia";
10
9
  import { ByteArrayUtil } from "./ByteArrayUtil";
11
10
  import { ChatGptTokenUsageAggregator } from "./ChatGptTokenUsageAggregator";
12
11
 
@@ -14,10 +13,10 @@ function transformCompletionChunk(source: string | Uint8Array): ChatCompletionCh
14
13
  const str
15
14
  = source instanceof Uint8Array ? ByteArrayUtil.toUtf8(source) : source;
16
15
  const result: ChatCompletionChunk = JSON.parse(str) as ChatCompletionChunk;
17
- const valid = typia.validate<ChatCompletionChunk>(result);
18
- if (valid.success === false) {
19
- console.error("Invalid ChatCompletionChunk", valid.errors);
20
- }
16
+ // const valid = typia.validate<ChatCompletionChunk>(result);
17
+ // if (valid.success === false) {
18
+ // console.error("Invalid ChatCompletionChunk", valid.errors);
19
+ // }
21
20
  return result;
22
21
  }
23
22