@mariozechner/pi-agent-core 0.57.1 → 0.58.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/README.md +34 -0
- package/dist/agent-loop.d.ts +3 -0
- package/dist/agent-loop.d.ts.map +1 -1
- package/dist/agent-loop.js +238 -99
- package/dist/agent-loop.js.map +1 -1
- package/dist/agent.d.ts +15 -1
- package/dist/agent.d.ts.map +1 -1
- package/dist/agent.js +62 -58
- package/dist/agent.js.map +1 -1
- package/dist/types.d.ts +111 -3
- package/dist/types.d.ts.map +1 -1
- package/dist/types.js.map +1 -1
- package/package.json +2 -2
package/README.md
CHANGED
|
@@ -99,6 +99,15 @@ prompt("Read config.json")
|
|
|
99
99
|
└─ agent_end
|
|
100
100
|
```
|
|
101
101
|
|
|
102
|
+
Tool execution mode is configurable:
|
|
103
|
+
|
|
104
|
+
- `parallel` (default): preflight tool calls sequentially, execute allowed tools concurrently, emit final `tool_execution_end` and `toolResult` messages in assistant source order
|
|
105
|
+
- `sequential`: execute tool calls one by one, matching the historical behavior
|
|
106
|
+
|
|
107
|
+
The `beforeToolCall` hook runs after `tool_execution_start` and validated argument parsing. It can block execution. The `afterToolCall` hook runs after tool execution finishes and before `tool_execution_end` and final tool result message events are emitted.
|
|
108
|
+
|
|
109
|
+
When you use the `Agent` class, assistant `message_end` processing is treated as a barrier before tool preflight begins. That means `beforeToolCall` sees agent state that already includes the assistant message that requested the tool call.
|
|
110
|
+
|
|
102
111
|
### continue() Event Sequence
|
|
103
112
|
|
|
104
113
|
`continue()` resumes from existing context without adding a new message. Use it for retries after errors.
|
|
@@ -159,6 +168,23 @@ const agent = new Agent({
|
|
|
159
168
|
// Dynamic API key resolution (for expiring OAuth tokens)
|
|
160
169
|
getApiKey: async (provider) => refreshToken(),
|
|
161
170
|
|
|
171
|
+
// Tool execution mode: "parallel" (default) or "sequential"
|
|
172
|
+
toolExecution: "parallel",
|
|
173
|
+
|
|
174
|
+
// Preflight each tool call after args are validated. Can block execution.
|
|
175
|
+
beforeToolCall: async ({ toolCall, args, context }) => {
|
|
176
|
+
if (toolCall.name === "bash") {
|
|
177
|
+
return { block: true, reason: "bash is disabled" };
|
|
178
|
+
}
|
|
179
|
+
},
|
|
180
|
+
|
|
181
|
+
// Postprocess each tool result before final tool events are emitted.
|
|
182
|
+
afterToolCall: async ({ toolCall, result, isError, context }) => {
|
|
183
|
+
if (!isError) {
|
|
184
|
+
return { details: { ...result.details, audited: true } };
|
|
185
|
+
}
|
|
186
|
+
},
|
|
187
|
+
|
|
162
188
|
// Custom thinking budgets for token-based providers
|
|
163
189
|
thinkingBudgets: {
|
|
164
190
|
minimal: 128,
|
|
@@ -214,6 +240,9 @@ agent.setSystemPrompt("New prompt");
|
|
|
214
240
|
agent.setModel(getModel("openai", "gpt-4o"));
|
|
215
241
|
agent.setThinkingLevel("medium");
|
|
216
242
|
agent.setTools([myTool]);
|
|
243
|
+
agent.setToolExecution("sequential");
|
|
244
|
+
agent.setBeforeToolCall(async ({ toolCall }) => undefined);
|
|
245
|
+
agent.setAfterToolCall(async ({ toolCall, result }) => undefined);
|
|
217
246
|
agent.replaceMessages(newMessages);
|
|
218
247
|
agent.appendMessage(message);
|
|
219
248
|
agent.clearMessages();
|
|
@@ -393,6 +422,9 @@ const context: AgentContext = {
|
|
|
393
422
|
const config: AgentLoopConfig = {
|
|
394
423
|
model: getModel("openai", "gpt-4o"),
|
|
395
424
|
convertToLlm: (msgs) => msgs.filter(m => ["user", "assistant", "toolResult"].includes(m.role)),
|
|
425
|
+
toolExecution: "parallel",
|
|
426
|
+
beforeToolCall: async ({ toolCall, args, context }) => undefined,
|
|
427
|
+
afterToolCall: async ({ toolCall, result, isError, context }) => undefined,
|
|
396
428
|
};
|
|
397
429
|
|
|
398
430
|
const userMessage = { role: "user", content: "Hello", timestamp: Date.now() };
|
|
@@ -407,6 +439,8 @@ for await (const event of agentLoopContinue(context, config)) {
|
|
|
407
439
|
}
|
|
408
440
|
```
|
|
409
441
|
|
|
442
|
+
These low-level streams are observational. They preserve event order, but they do not wait for your async event handling to settle before later producer phases continue. If you need message processing to act as a barrier before tool preflight, use the `Agent` class instead of raw `agentLoop()` or `agentLoopContinue()`.
|
|
443
|
+
|
|
410
444
|
## License
|
|
411
445
|
|
|
412
446
|
MIT
|
package/dist/agent-loop.d.ts
CHANGED
|
@@ -4,6 +4,7 @@
|
|
|
4
4
|
*/
|
|
5
5
|
import { EventStream } from "@mariozechner/pi-ai";
|
|
6
6
|
import type { AgentContext, AgentEvent, AgentLoopConfig, AgentMessage, StreamFn } from "./types.js";
|
|
7
|
+
export type AgentEventSink = (event: AgentEvent) => Promise<void> | void;
|
|
7
8
|
/**
|
|
8
9
|
* Start an agent loop with a new prompt message.
|
|
9
10
|
* The prompt is added to the context and events are emitted for it.
|
|
@@ -18,4 +19,6 @@ export declare function agentLoop(prompts: AgentMessage[], context: AgentContext
|
|
|
18
19
|
* This cannot be validated here since `convertToLlm` is only called once per turn.
|
|
19
20
|
*/
|
|
20
21
|
export declare function agentLoopContinue(context: AgentContext, config: AgentLoopConfig, signal?: AbortSignal, streamFn?: StreamFn): EventStream<AgentEvent, AgentMessage[]>;
|
|
22
|
+
export declare function runAgentLoop(prompts: AgentMessage[], context: AgentContext, config: AgentLoopConfig, emit: AgentEventSink, signal?: AbortSignal, streamFn?: StreamFn): Promise<AgentMessage[]>;
|
|
23
|
+
export declare function runAgentLoopContinue(context: AgentContext, config: AgentLoopConfig, emit: AgentEventSink, signal?: AbortSignal, streamFn?: StreamFn): Promise<AgentMessage[]>;
|
|
21
24
|
//# sourceMappingURL=agent-loop.d.ts.map
|
package/dist/agent-loop.d.ts.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"agent-loop.d.ts","sourceRoot":"","sources":["../src/agent-loop.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,EAGN,WAAW,EAIX,MAAM,qBAAqB,CAAC;AAC7B,OAAO,KAAK,EACX,YAAY,EACZ,UAAU,EACV,eAAe,EACf,YAAY,EAGZ,QAAQ,EACR,MAAM,YAAY,CAAC;AAEpB;;;GAGG;AACH,wBAAgB,SAAS,CACxB,OAAO,EAAE,YAAY,EAAE,EACvB,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAqBzC;AAED;;;;;;;GAOG;AACH,wBAAgB,iBAAiB,CAChC,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAsBzC","sourcesContent":["/**\n * Agent loop that works with AgentMessage throughout.\n * Transforms to Message[] only at the LLM call boundary.\n */\n\nimport {\n\ttype AssistantMessage,\n\ttype Context,\n\tEventStream,\n\tstreamSimple,\n\ttype ToolResultMessage,\n\tvalidateToolArguments,\n} from \"@mariozechner/pi-ai\";\nimport type {\n\tAgentContext,\n\tAgentEvent,\n\tAgentLoopConfig,\n\tAgentMessage,\n\tAgentTool,\n\tAgentToolResult,\n\tStreamFn,\n} from \"./types.js\";\n\n/**\n * Start an agent loop with a new prompt message.\n * The prompt is added to the context and events are emitted for it.\n */\nexport function agentLoop(\n\tprompts: AgentMessage[],\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tconst stream = createAgentStream();\n\n\t(async () => {\n\t\tconst newMessages: AgentMessage[] = [...prompts];\n\t\tconst currentContext: AgentContext = {\n\t\t\t...context,\n\t\t\tmessages: [...context.messages, ...prompts],\n\t\t};\n\n\t\tstream.push({ type: \"agent_start\" });\n\t\tstream.push({ type: \"turn_start\" });\n\t\tfor (const prompt of prompts) {\n\t\t\tstream.push({ type: \"message_start\", message: prompt });\n\t\t\tstream.push({ type: \"message_end\", message: prompt });\n\t\t}\n\n\t\tawait runLoop(currentContext, newMessages, config, signal, stream, streamFn);\n\t})();\n\n\treturn stream;\n}\n\n/**\n * Continue an agent loop from the current context without adding a new message.\n * Used for retries - context already has user message or tool results.\n *\n * **Important:** The last message in context must convert to a `user` or `toolResult` message\n * via `convertToLlm`. If it doesn't, the LLM provider will reject the request.\n * This cannot be validated here since `convertToLlm` is only called once per turn.\n */\nexport function agentLoopContinue(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tif (context.messages.length === 0) {\n\t\tthrow new Error(\"Cannot continue: no messages in context\");\n\t}\n\n\tif (context.messages[context.messages.length - 1].role === \"assistant\") {\n\t\tthrow new Error(\"Cannot continue from message role: assistant\");\n\t}\n\n\tconst stream = createAgentStream();\n\n\t(async () => {\n\t\tconst newMessages: AgentMessage[] = [];\n\t\tconst currentContext: AgentContext = { ...context };\n\n\t\tstream.push({ type: \"agent_start\" });\n\t\tstream.push({ type: \"turn_start\" });\n\n\t\tawait runLoop(currentContext, newMessages, config, signal, stream, streamFn);\n\t})();\n\n\treturn stream;\n}\n\nfunction createAgentStream(): EventStream<AgentEvent, AgentMessage[]> {\n\treturn new EventStream<AgentEvent, AgentMessage[]>(\n\t\t(event: AgentEvent) => event.type === \"agent_end\",\n\t\t(event: AgentEvent) => (event.type === \"agent_end\" ? event.messages : []),\n\t);\n}\n\n/**\n * Main loop logic shared by agentLoop and agentLoopContinue.\n */\nasync function runLoop(\n\tcurrentContext: AgentContext,\n\tnewMessages: AgentMessage[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n\tstreamFn?: StreamFn,\n): Promise<void> {\n\tlet firstTurn = true;\n\t// Check for steering messages at start (user may have typed while waiting)\n\tlet pendingMessages: AgentMessage[] = (await config.getSteeringMessages?.()) || [];\n\n\t// Outer loop: continues when queued follow-up messages arrive after agent would stop\n\twhile (true) {\n\t\tlet hasMoreToolCalls = true;\n\t\tlet steeringAfterTools: AgentMessage[] | null = null;\n\n\t\t// Inner loop: process tool calls and steering messages\n\t\twhile (hasMoreToolCalls || pendingMessages.length > 0) {\n\t\t\tif (!firstTurn) {\n\t\t\t\tstream.push({ type: \"turn_start\" });\n\t\t\t} else {\n\t\t\t\tfirstTurn = false;\n\t\t\t}\n\n\t\t\t// Process pending messages (inject before next assistant response)\n\t\t\tif (pendingMessages.length > 0) {\n\t\t\t\tfor (const message of pendingMessages) {\n\t\t\t\t\tstream.push({ type: \"message_start\", message });\n\t\t\t\t\tstream.push({ type: \"message_end\", message });\n\t\t\t\t\tcurrentContext.messages.push(message);\n\t\t\t\t\tnewMessages.push(message);\n\t\t\t\t}\n\t\t\t\tpendingMessages = [];\n\t\t\t}\n\n\t\t\t// Stream assistant response\n\t\t\tconst message = await streamAssistantResponse(currentContext, config, signal, stream, streamFn);\n\t\t\tnewMessages.push(message);\n\n\t\t\tif (message.stopReason === \"error\" || message.stopReason === \"aborted\") {\n\t\t\t\tstream.push({ type: \"turn_end\", message, toolResults: [] });\n\t\t\t\tstream.push({ type: \"agent_end\", messages: newMessages });\n\t\t\t\tstream.end(newMessages);\n\t\t\t\treturn;\n\t\t\t}\n\n\t\t\t// Check for tool calls\n\t\t\tconst toolCalls = message.content.filter((c) => c.type === \"toolCall\");\n\t\t\thasMoreToolCalls = toolCalls.length > 0;\n\n\t\t\tconst toolResults: ToolResultMessage[] = [];\n\t\t\tif (hasMoreToolCalls) {\n\t\t\t\tconst toolExecution = await executeToolCalls(\n\t\t\t\t\tcurrentContext.tools,\n\t\t\t\t\tmessage,\n\t\t\t\t\tsignal,\n\t\t\t\t\tstream,\n\t\t\t\t\tconfig.getSteeringMessages,\n\t\t\t\t);\n\t\t\t\ttoolResults.push(...toolExecution.toolResults);\n\t\t\t\tsteeringAfterTools = toolExecution.steeringMessages ?? null;\n\n\t\t\t\tfor (const result of toolResults) {\n\t\t\t\t\tcurrentContext.messages.push(result);\n\t\t\t\t\tnewMessages.push(result);\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tstream.push({ type: \"turn_end\", message, toolResults });\n\n\t\t\t// Get steering messages after turn completes\n\t\t\tif (steeringAfterTools && steeringAfterTools.length > 0) {\n\t\t\t\tpendingMessages = steeringAfterTools;\n\t\t\t\tsteeringAfterTools = null;\n\t\t\t} else {\n\t\t\t\tpendingMessages = (await config.getSteeringMessages?.()) || [];\n\t\t\t}\n\t\t}\n\n\t\t// Agent would stop here. Check for follow-up messages.\n\t\tconst followUpMessages = (await config.getFollowUpMessages?.()) || [];\n\t\tif (followUpMessages.length > 0) {\n\t\t\t// Set as pending so inner loop processes them\n\t\t\tpendingMessages = followUpMessages;\n\t\t\tcontinue;\n\t\t}\n\n\t\t// No more messages, exit\n\t\tbreak;\n\t}\n\n\tstream.push({ type: \"agent_end\", messages: newMessages });\n\tstream.end(newMessages);\n}\n\n/**\n * Stream an assistant response from the LLM.\n * This is where AgentMessage[] gets transformed to Message[] for the LLM.\n */\nasync function streamAssistantResponse(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n\tstreamFn?: StreamFn,\n): Promise<AssistantMessage> {\n\t// Apply context transform if configured (AgentMessage[] → AgentMessage[])\n\tlet messages = context.messages;\n\tif (config.transformContext) {\n\t\tmessages = await config.transformContext(messages, signal);\n\t}\n\n\t// Convert to LLM-compatible messages (AgentMessage[] → Message[])\n\tconst llmMessages = await config.convertToLlm(messages);\n\n\t// Build LLM context\n\tconst llmContext: Context = {\n\t\tsystemPrompt: context.systemPrompt,\n\t\tmessages: llmMessages,\n\t\ttools: context.tools,\n\t};\n\n\tconst streamFunction = streamFn || streamSimple;\n\n\t// Resolve API key (important for expiring tokens)\n\tconst resolvedApiKey =\n\t\t(config.getApiKey ? await config.getApiKey(config.model.provider) : undefined) || config.apiKey;\n\n\tconst response = await streamFunction(config.model, llmContext, {\n\t\t...config,\n\t\tapiKey: resolvedApiKey,\n\t\tsignal,\n\t});\n\n\tlet partialMessage: AssistantMessage | null = null;\n\tlet addedPartial = false;\n\n\tfor await (const event of response) {\n\t\tswitch (event.type) {\n\t\t\tcase \"start\":\n\t\t\t\tpartialMessage = event.partial;\n\t\t\t\tcontext.messages.push(partialMessage);\n\t\t\t\taddedPartial = true;\n\t\t\t\tstream.push({ type: \"message_start\", message: { ...partialMessage } });\n\t\t\t\tbreak;\n\n\t\t\tcase \"text_start\":\n\t\t\tcase \"text_delta\":\n\t\t\tcase \"text_end\":\n\t\t\tcase \"thinking_start\":\n\t\t\tcase \"thinking_delta\":\n\t\t\tcase \"thinking_end\":\n\t\t\tcase \"toolcall_start\":\n\t\t\tcase \"toolcall_delta\":\n\t\t\tcase \"toolcall_end\":\n\t\t\t\tif (partialMessage) {\n\t\t\t\t\tpartialMessage = event.partial;\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = partialMessage;\n\t\t\t\t\tstream.push({\n\t\t\t\t\t\ttype: \"message_update\",\n\t\t\t\t\t\tassistantMessageEvent: event,\n\t\t\t\t\t\tmessage: { ...partialMessage },\n\t\t\t\t\t});\n\t\t\t\t}\n\t\t\t\tbreak;\n\n\t\t\tcase \"done\":\n\t\t\tcase \"error\": {\n\t\t\t\tconst finalMessage = await response.result();\n\t\t\t\tif (addedPartial) {\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = finalMessage;\n\t\t\t\t} else {\n\t\t\t\t\tcontext.messages.push(finalMessage);\n\t\t\t\t}\n\t\t\t\tif (!addedPartial) {\n\t\t\t\t\tstream.push({ type: \"message_start\", message: { ...finalMessage } });\n\t\t\t\t}\n\t\t\t\tstream.push({ type: \"message_end\", message: finalMessage });\n\t\t\t\treturn finalMessage;\n\t\t\t}\n\t\t}\n\t}\n\n\treturn await response.result();\n}\n\n/**\n * Execute tool calls from an assistant message.\n */\nasync function executeToolCalls(\n\ttools: AgentTool<any>[] | undefined,\n\tassistantMessage: AssistantMessage,\n\tsignal: AbortSignal | undefined,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n\tgetSteeringMessages?: AgentLoopConfig[\"getSteeringMessages\"],\n): Promise<{ toolResults: ToolResultMessage[]; steeringMessages?: AgentMessage[] }> {\n\tconst toolCalls = assistantMessage.content.filter((c) => c.type === \"toolCall\");\n\tconst results: ToolResultMessage[] = [];\n\tlet steeringMessages: AgentMessage[] | undefined;\n\n\tfor (let index = 0; index < toolCalls.length; index++) {\n\t\tconst toolCall = toolCalls[index];\n\t\tconst tool = tools?.find((t) => t.name === toolCall.name);\n\n\t\tstream.push({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\n\t\tlet result: AgentToolResult<any>;\n\t\tlet isError = false;\n\n\t\ttry {\n\t\t\tif (!tool) throw new Error(`Tool ${toolCall.name} not found`);\n\n\t\t\tconst validatedArgs = validateToolArguments(tool, toolCall);\n\n\t\t\tresult = await tool.execute(toolCall.id, validatedArgs, signal, (partialResult) => {\n\t\t\t\tstream.push({\n\t\t\t\t\ttype: \"tool_execution_update\",\n\t\t\t\t\ttoolCallId: toolCall.id,\n\t\t\t\t\ttoolName: toolCall.name,\n\t\t\t\t\targs: toolCall.arguments,\n\t\t\t\t\tpartialResult,\n\t\t\t\t});\n\t\t\t});\n\t\t} catch (e) {\n\t\t\tresult = {\n\t\t\t\tcontent: [{ type: \"text\", text: e instanceof Error ? e.message : String(e) }],\n\t\t\t\tdetails: {},\n\t\t\t};\n\t\t\tisError = true;\n\t\t}\n\n\t\tstream.push({\n\t\t\ttype: \"tool_execution_end\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\tresult,\n\t\t\tisError,\n\t\t});\n\n\t\tconst toolResultMessage: ToolResultMessage = {\n\t\t\trole: \"toolResult\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\tcontent: result.content,\n\t\t\tdetails: result.details,\n\t\t\tisError,\n\t\t\ttimestamp: Date.now(),\n\t\t};\n\n\t\tresults.push(toolResultMessage);\n\t\tstream.push({ type: \"message_start\", message: toolResultMessage });\n\t\tstream.push({ type: \"message_end\", message: toolResultMessage });\n\n\t\t// Check for steering messages - skip remaining tools if user interrupted\n\t\tif (getSteeringMessages) {\n\t\t\tconst steering = await getSteeringMessages();\n\t\t\tif (steering.length > 0) {\n\t\t\t\tsteeringMessages = steering;\n\t\t\t\tconst remainingCalls = toolCalls.slice(index + 1);\n\t\t\t\tfor (const skipped of remainingCalls) {\n\t\t\t\t\tresults.push(skipToolCall(skipped, stream));\n\t\t\t\t}\n\t\t\t\tbreak;\n\t\t\t}\n\t\t}\n\t}\n\n\treturn { toolResults: results, steeringMessages };\n}\n\nfunction skipToolCall(\n\ttoolCall: Extract<AssistantMessage[\"content\"][number], { type: \"toolCall\" }>,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n): ToolResultMessage {\n\tconst result: AgentToolResult<any> = {\n\t\tcontent: [{ type: \"text\", text: \"Skipped due to queued user message.\" }],\n\t\tdetails: {},\n\t};\n\n\tstream.push({\n\t\ttype: \"tool_execution_start\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\targs: toolCall.arguments,\n\t});\n\tstream.push({\n\t\ttype: \"tool_execution_end\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tresult,\n\t\tisError: true,\n\t});\n\n\tconst toolResultMessage: ToolResultMessage = {\n\t\trole: \"toolResult\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tcontent: result.content,\n\t\tdetails: {},\n\t\tisError: true,\n\t\ttimestamp: Date.now(),\n\t};\n\n\tstream.push({ type: \"message_start\", message: toolResultMessage });\n\tstream.push({ type: \"message_end\", message: toolResultMessage });\n\n\treturn toolResultMessage;\n}\n"]}
|
|
1
|
+
{"version":3,"file":"agent-loop.d.ts","sourceRoot":"","sources":["../src/agent-loop.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,EAGN,WAAW,EAIX,MAAM,qBAAqB,CAAC;AAC7B,OAAO,KAAK,EACX,YAAY,EACZ,UAAU,EACV,eAAe,EACf,YAAY,EAIZ,QAAQ,EACR,MAAM,YAAY,CAAC;AAEpB,MAAM,MAAM,cAAc,GAAG,CAAC,KAAK,EAAE,UAAU,KAAK,OAAO,CAAC,IAAI,CAAC,GAAG,IAAI,CAAC;AAEzE;;;GAGG;AACH,wBAAgB,SAAS,CACxB,OAAO,EAAE,YAAY,EAAE,EACvB,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAiBzC;AAED;;;;;;;GAOG;AACH,wBAAgB,iBAAiB,CAChC,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAwBzC;AAED,wBAAsB,YAAY,CACjC,OAAO,EAAE,YAAY,EAAE,EACvB,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,IAAI,EAAE,cAAc,EACpB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,OAAO,CAAC,YAAY,EAAE,CAAC,CAgBzB;AAED,wBAAsB,oBAAoB,CACzC,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,IAAI,EAAE,cAAc,EACpB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,OAAO,CAAC,YAAY,EAAE,CAAC,CAiBzB","sourcesContent":["/**\n * Agent loop that works with AgentMessage throughout.\n * Transforms to Message[] only at the LLM call boundary.\n */\n\nimport {\n\ttype AssistantMessage,\n\ttype Context,\n\tEventStream,\n\tstreamSimple,\n\ttype ToolResultMessage,\n\tvalidateToolArguments,\n} from \"@mariozechner/pi-ai\";\nimport type {\n\tAgentContext,\n\tAgentEvent,\n\tAgentLoopConfig,\n\tAgentMessage,\n\tAgentTool,\n\tAgentToolCall,\n\tAgentToolResult,\n\tStreamFn,\n} from \"./types.js\";\n\nexport type AgentEventSink = (event: AgentEvent) => Promise<void> | void;\n\n/**\n * Start an agent loop with a new prompt message.\n * The prompt is added to the context and events are emitted for it.\n */\nexport function agentLoop(\n\tprompts: AgentMessage[],\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tconst stream = createAgentStream();\n\n\tvoid runAgentLoop(\n\t\tprompts,\n\t\tcontext,\n\t\tconfig,\n\t\tasync (event) => {\n\t\t\tstream.push(event);\n\t\t},\n\t\tsignal,\n\t\tstreamFn,\n\t).then((messages) => {\n\t\tstream.end(messages);\n\t});\n\n\treturn stream;\n}\n\n/**\n * Continue an agent loop from the current context without adding a new message.\n * Used for retries - context already has user message or tool results.\n *\n * **Important:** The last message in context must convert to a `user` or `toolResult` message\n * via `convertToLlm`. If it doesn't, the LLM provider will reject the request.\n * This cannot be validated here since `convertToLlm` is only called once per turn.\n */\nexport function agentLoopContinue(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tif (context.messages.length === 0) {\n\t\tthrow new Error(\"Cannot continue: no messages in context\");\n\t}\n\n\tif (context.messages[context.messages.length - 1].role === \"assistant\") {\n\t\tthrow new Error(\"Cannot continue from message role: assistant\");\n\t}\n\n\tconst stream = createAgentStream();\n\n\tvoid runAgentLoopContinue(\n\t\tcontext,\n\t\tconfig,\n\t\tasync (event) => {\n\t\t\tstream.push(event);\n\t\t},\n\t\tsignal,\n\t\tstreamFn,\n\t).then((messages) => {\n\t\tstream.end(messages);\n\t});\n\n\treturn stream;\n}\n\nexport async function runAgentLoop(\n\tprompts: AgentMessage[],\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\temit: AgentEventSink,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): Promise<AgentMessage[]> {\n\tconst newMessages: AgentMessage[] = [...prompts];\n\tconst currentContext: AgentContext = {\n\t\t...context,\n\t\tmessages: [...context.messages, ...prompts],\n\t};\n\n\tawait emit({ type: \"agent_start\" });\n\tawait emit({ type: \"turn_start\" });\n\tfor (const prompt of prompts) {\n\t\tawait emit({ type: \"message_start\", message: prompt });\n\t\tawait emit({ type: \"message_end\", message: prompt });\n\t}\n\n\tawait runLoop(currentContext, newMessages, config, signal, emit, streamFn);\n\treturn newMessages;\n}\n\nexport async function runAgentLoopContinue(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\temit: AgentEventSink,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): Promise<AgentMessage[]> {\n\tif (context.messages.length === 0) {\n\t\tthrow new Error(\"Cannot continue: no messages in context\");\n\t}\n\n\tif (context.messages[context.messages.length - 1].role === \"assistant\") {\n\t\tthrow new Error(\"Cannot continue from message role: assistant\");\n\t}\n\n\tconst newMessages: AgentMessage[] = [];\n\tconst currentContext: AgentContext = { ...context };\n\n\tawait emit({ type: \"agent_start\" });\n\tawait emit({ type: \"turn_start\" });\n\n\tawait runLoop(currentContext, newMessages, config, signal, emit, streamFn);\n\treturn newMessages;\n}\n\nfunction createAgentStream(): EventStream<AgentEvent, AgentMessage[]> {\n\treturn new EventStream<AgentEvent, AgentMessage[]>(\n\t\t(event: AgentEvent) => event.type === \"agent_end\",\n\t\t(event: AgentEvent) => (event.type === \"agent_end\" ? event.messages : []),\n\t);\n}\n\n/**\n * Main loop logic shared by agentLoop and agentLoopContinue.\n */\nasync function runLoop(\n\tcurrentContext: AgentContext,\n\tnewMessages: AgentMessage[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n\tstreamFn?: StreamFn,\n): Promise<void> {\n\tlet firstTurn = true;\n\t// Check for steering messages at start (user may have typed while waiting)\n\tlet pendingMessages: AgentMessage[] = (await config.getSteeringMessages?.()) || [];\n\n\t// Outer loop: continues when queued follow-up messages arrive after agent would stop\n\twhile (true) {\n\t\tlet hasMoreToolCalls = true;\n\t\tlet steeringAfterTools: AgentMessage[] | null = null;\n\n\t\t// Inner loop: process tool calls and steering messages\n\t\twhile (hasMoreToolCalls || pendingMessages.length > 0) {\n\t\t\tif (!firstTurn) {\n\t\t\t\tawait emit({ type: \"turn_start\" });\n\t\t\t} else {\n\t\t\t\tfirstTurn = false;\n\t\t\t}\n\n\t\t\t// Process pending messages (inject before next assistant response)\n\t\t\tif (pendingMessages.length > 0) {\n\t\t\t\tfor (const message of pendingMessages) {\n\t\t\t\t\tawait emit({ type: \"message_start\", message });\n\t\t\t\t\tawait emit({ type: \"message_end\", message });\n\t\t\t\t\tcurrentContext.messages.push(message);\n\t\t\t\t\tnewMessages.push(message);\n\t\t\t\t}\n\t\t\t\tpendingMessages = [];\n\t\t\t}\n\n\t\t\t// Stream assistant response\n\t\t\tconst message = await streamAssistantResponse(currentContext, config, signal, emit, streamFn);\n\t\t\tnewMessages.push(message);\n\n\t\t\tif (message.stopReason === \"error\" || message.stopReason === \"aborted\") {\n\t\t\t\tawait emit({ type: \"turn_end\", message, toolResults: [] });\n\t\t\t\tawait emit({ type: \"agent_end\", messages: newMessages });\n\t\t\t\treturn;\n\t\t\t}\n\n\t\t\t// Check for tool calls\n\t\t\tconst toolCalls = message.content.filter((c) => c.type === \"toolCall\");\n\t\t\thasMoreToolCalls = toolCalls.length > 0;\n\n\t\t\tconst toolResults: ToolResultMessage[] = [];\n\t\t\tif (hasMoreToolCalls) {\n\t\t\t\tconst toolExecution = await executeToolCalls(currentContext, message, config, signal, emit);\n\t\t\t\ttoolResults.push(...toolExecution.toolResults);\n\t\t\t\tsteeringAfterTools = toolExecution.steeringMessages ?? null;\n\n\t\t\t\tfor (const result of toolResults) {\n\t\t\t\t\tcurrentContext.messages.push(result);\n\t\t\t\t\tnewMessages.push(result);\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tawait emit({ type: \"turn_end\", message, toolResults });\n\n\t\t\t// Get steering messages after turn completes\n\t\t\tif (steeringAfterTools && steeringAfterTools.length > 0) {\n\t\t\t\tpendingMessages = steeringAfterTools;\n\t\t\t\tsteeringAfterTools = null;\n\t\t\t} else {\n\t\t\t\tpendingMessages = (await config.getSteeringMessages?.()) || [];\n\t\t\t}\n\t\t}\n\n\t\t// Agent would stop here. Check for follow-up messages.\n\t\tconst followUpMessages = (await config.getFollowUpMessages?.()) || [];\n\t\tif (followUpMessages.length > 0) {\n\t\t\t// Set as pending so inner loop processes them\n\t\t\tpendingMessages = followUpMessages;\n\t\t\tcontinue;\n\t\t}\n\n\t\t// No more messages, exit\n\t\tbreak;\n\t}\n\n\tawait emit({ type: \"agent_end\", messages: newMessages });\n}\n\n/**\n * Stream an assistant response from the LLM.\n * This is where AgentMessage[] gets transformed to Message[] for the LLM.\n */\nasync function streamAssistantResponse(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n\tstreamFn?: StreamFn,\n): Promise<AssistantMessage> {\n\t// Apply context transform if configured (AgentMessage[] → AgentMessage[])\n\tlet messages = context.messages;\n\tif (config.transformContext) {\n\t\tmessages = await config.transformContext(messages, signal);\n\t}\n\n\t// Convert to LLM-compatible messages (AgentMessage[] → Message[])\n\tconst llmMessages = await config.convertToLlm(messages);\n\n\t// Build LLM context\n\tconst llmContext: Context = {\n\t\tsystemPrompt: context.systemPrompt,\n\t\tmessages: llmMessages,\n\t\ttools: context.tools,\n\t};\n\n\tconst streamFunction = streamFn || streamSimple;\n\n\t// Resolve API key (important for expiring tokens)\n\tconst resolvedApiKey =\n\t\t(config.getApiKey ? await config.getApiKey(config.model.provider) : undefined) || config.apiKey;\n\n\tconst response = await streamFunction(config.model, llmContext, {\n\t\t...config,\n\t\tapiKey: resolvedApiKey,\n\t\tsignal,\n\t});\n\n\tlet partialMessage: AssistantMessage | null = null;\n\tlet addedPartial = false;\n\n\tfor await (const event of response) {\n\t\tswitch (event.type) {\n\t\t\tcase \"start\":\n\t\t\t\tpartialMessage = event.partial;\n\t\t\t\tcontext.messages.push(partialMessage);\n\t\t\t\taddedPartial = true;\n\t\t\t\tawait emit({ type: \"message_start\", message: { ...partialMessage } });\n\t\t\t\tbreak;\n\n\t\t\tcase \"text_start\":\n\t\t\tcase \"text_delta\":\n\t\t\tcase \"text_end\":\n\t\t\tcase \"thinking_start\":\n\t\t\tcase \"thinking_delta\":\n\t\t\tcase \"thinking_end\":\n\t\t\tcase \"toolcall_start\":\n\t\t\tcase \"toolcall_delta\":\n\t\t\tcase \"toolcall_end\":\n\t\t\t\tif (partialMessage) {\n\t\t\t\t\tpartialMessage = event.partial;\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = partialMessage;\n\t\t\t\t\tawait emit({\n\t\t\t\t\t\ttype: \"message_update\",\n\t\t\t\t\t\tassistantMessageEvent: event,\n\t\t\t\t\t\tmessage: { ...partialMessage },\n\t\t\t\t\t});\n\t\t\t\t}\n\t\t\t\tbreak;\n\n\t\t\tcase \"done\":\n\t\t\tcase \"error\": {\n\t\t\t\tconst finalMessage = await response.result();\n\t\t\t\tif (addedPartial) {\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = finalMessage;\n\t\t\t\t} else {\n\t\t\t\t\tcontext.messages.push(finalMessage);\n\t\t\t\t}\n\t\t\t\tif (!addedPartial) {\n\t\t\t\t\tawait emit({ type: \"message_start\", message: { ...finalMessage } });\n\t\t\t\t}\n\t\t\t\tawait emit({ type: \"message_end\", message: finalMessage });\n\t\t\t\treturn finalMessage;\n\t\t\t}\n\t\t}\n\t}\n\n\tconst finalMessage = await response.result();\n\tif (addedPartial) {\n\t\tcontext.messages[context.messages.length - 1] = finalMessage;\n\t} else {\n\t\tcontext.messages.push(finalMessage);\n\t\tawait emit({ type: \"message_start\", message: { ...finalMessage } });\n\t}\n\tawait emit({ type: \"message_end\", message: finalMessage });\n\treturn finalMessage;\n}\n\n/**\n * Execute tool calls from an assistant message.\n */\nasync function executeToolCalls(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<{ toolResults: ToolResultMessage[]; steeringMessages?: AgentMessage[] }> {\n\tconst toolCalls = assistantMessage.content.filter((c) => c.type === \"toolCall\");\n\tif (config.toolExecution === \"sequential\") {\n\t\treturn executeToolCallsSequential(currentContext, assistantMessage, toolCalls, config, signal, emit);\n\t}\n\treturn executeToolCallsParallel(currentContext, assistantMessage, toolCalls, config, signal, emit);\n}\n\nasync function executeToolCallsSequential(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\ttoolCalls: AgentToolCall[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<{ toolResults: ToolResultMessage[]; steeringMessages?: AgentMessage[] }> {\n\tconst results: ToolResultMessage[] = [];\n\tlet steeringMessages: AgentMessage[] | undefined;\n\n\tfor (let index = 0; index < toolCalls.length; index++) {\n\t\tconst toolCall = toolCalls[index];\n\t\tawait emit({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\n\t\tconst preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);\n\t\tif (preparation.kind === \"immediate\") {\n\t\t\tresults.push(await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit));\n\t\t} else {\n\t\t\tconst executed = await executePreparedToolCall(preparation, signal, emit);\n\t\t\tresults.push(\n\t\t\t\tawait finalizeExecutedToolCall(\n\t\t\t\t\tcurrentContext,\n\t\t\t\t\tassistantMessage,\n\t\t\t\t\tpreparation,\n\t\t\t\t\texecuted,\n\t\t\t\t\tconfig,\n\t\t\t\t\tsignal,\n\t\t\t\t\temit,\n\t\t\t\t),\n\t\t\t);\n\t\t}\n\n\t\tif (config.getSteeringMessages) {\n\t\t\tconst steering = await config.getSteeringMessages();\n\t\t\tif (steering.length > 0) {\n\t\t\t\tsteeringMessages = steering;\n\t\t\t\tconst remainingCalls = toolCalls.slice(index + 1);\n\t\t\t\tfor (const skipped of remainingCalls) {\n\t\t\t\t\tresults.push(await skipToolCall(skipped, emit));\n\t\t\t\t}\n\t\t\t\tbreak;\n\t\t\t}\n\t\t}\n\t}\n\n\treturn { toolResults: results, steeringMessages };\n}\n\nasync function executeToolCallsParallel(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\ttoolCalls: AgentToolCall[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<{ toolResults: ToolResultMessage[]; steeringMessages?: AgentMessage[] }> {\n\tconst results: ToolResultMessage[] = [];\n\tconst runnableCalls: PreparedToolCall[] = [];\n\tlet steeringMessages: AgentMessage[] | undefined;\n\n\tfor (let index = 0; index < toolCalls.length; index++) {\n\t\tconst toolCall = toolCalls[index];\n\t\tawait emit({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\n\t\tconst preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);\n\t\tif (preparation.kind === \"immediate\") {\n\t\t\tresults.push(await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit));\n\t\t} else {\n\t\t\trunnableCalls.push(preparation);\n\t\t}\n\n\t\tif (config.getSteeringMessages) {\n\t\t\tconst steering = await config.getSteeringMessages();\n\t\t\tif (steering.length > 0) {\n\t\t\t\tsteeringMessages = steering;\n\t\t\t\tfor (const runnable of runnableCalls) {\n\t\t\t\t\tresults.push(await skipToolCall(runnable.toolCall, emit, { emitStart: false }));\n\t\t\t\t}\n\t\t\t\tconst remainingCalls = toolCalls.slice(index + 1);\n\t\t\t\tfor (const skipped of remainingCalls) {\n\t\t\t\t\tresults.push(await skipToolCall(skipped, emit));\n\t\t\t\t}\n\t\t\t\treturn { toolResults: results, steeringMessages };\n\t\t\t}\n\t\t}\n\t}\n\n\tconst runningCalls = runnableCalls.map((prepared) => ({\n\t\tprepared,\n\t\texecution: executePreparedToolCall(prepared, signal, emit),\n\t}));\n\n\tfor (const running of runningCalls) {\n\t\tconst executed = await running.execution;\n\t\tresults.push(\n\t\t\tawait finalizeExecutedToolCall(\n\t\t\t\tcurrentContext,\n\t\t\t\tassistantMessage,\n\t\t\t\trunning.prepared,\n\t\t\t\texecuted,\n\t\t\t\tconfig,\n\t\t\t\tsignal,\n\t\t\t\temit,\n\t\t\t),\n\t\t);\n\t}\n\n\tif (!steeringMessages && config.getSteeringMessages) {\n\t\tconst steering = await config.getSteeringMessages();\n\t\tif (steering.length > 0) {\n\t\t\tsteeringMessages = steering;\n\t\t}\n\t}\n\n\treturn { toolResults: results, steeringMessages };\n}\n\ntype PreparedToolCall = {\n\tkind: \"prepared\";\n\ttoolCall: AgentToolCall;\n\ttool: AgentTool<any>;\n\targs: unknown;\n};\n\ntype ImmediateToolCallOutcome = {\n\tkind: \"immediate\";\n\tresult: AgentToolResult<any>;\n\tisError: boolean;\n};\n\ntype ExecutedToolCallOutcome = {\n\tresult: AgentToolResult<any>;\n\tisError: boolean;\n};\n\nasync function prepareToolCall(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\ttoolCall: AgentToolCall,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n): Promise<PreparedToolCall | ImmediateToolCallOutcome> {\n\tconst tool = currentContext.tools?.find((t) => t.name === toolCall.name);\n\tif (!tool) {\n\t\treturn {\n\t\t\tkind: \"immediate\",\n\t\t\tresult: createErrorToolResult(`Tool ${toolCall.name} not found`),\n\t\t\tisError: true,\n\t\t};\n\t}\n\n\ttry {\n\t\tconst validatedArgs = validateToolArguments(tool, toolCall);\n\t\tif (config.beforeToolCall) {\n\t\t\tconst beforeResult = await config.beforeToolCall(\n\t\t\t\t{\n\t\t\t\t\tassistantMessage,\n\t\t\t\t\ttoolCall,\n\t\t\t\t\targs: validatedArgs,\n\t\t\t\t\tcontext: currentContext,\n\t\t\t\t},\n\t\t\t\tsignal,\n\t\t\t);\n\t\t\tif (beforeResult?.block) {\n\t\t\t\treturn {\n\t\t\t\t\tkind: \"immediate\",\n\t\t\t\t\tresult: createErrorToolResult(beforeResult.reason || \"Tool execution was blocked\"),\n\t\t\t\t\tisError: true,\n\t\t\t\t};\n\t\t\t}\n\t\t}\n\t\treturn {\n\t\t\tkind: \"prepared\",\n\t\t\ttoolCall,\n\t\t\ttool,\n\t\t\targs: validatedArgs,\n\t\t};\n\t} catch (error) {\n\t\treturn {\n\t\t\tkind: \"immediate\",\n\t\t\tresult: createErrorToolResult(error instanceof Error ? error.message : String(error)),\n\t\t\tisError: true,\n\t\t};\n\t}\n}\n\nasync function executePreparedToolCall(\n\tprepared: PreparedToolCall,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ExecutedToolCallOutcome> {\n\tconst updateEvents: Promise<void>[] = [];\n\n\ttry {\n\t\tconst result = await prepared.tool.execute(\n\t\t\tprepared.toolCall.id,\n\t\t\tprepared.args as never,\n\t\t\tsignal,\n\t\t\t(partialResult) => {\n\t\t\t\tupdateEvents.push(\n\t\t\t\t\tPromise.resolve(\n\t\t\t\t\t\temit({\n\t\t\t\t\t\t\ttype: \"tool_execution_update\",\n\t\t\t\t\t\t\ttoolCallId: prepared.toolCall.id,\n\t\t\t\t\t\t\ttoolName: prepared.toolCall.name,\n\t\t\t\t\t\t\targs: prepared.toolCall.arguments,\n\t\t\t\t\t\t\tpartialResult,\n\t\t\t\t\t\t}),\n\t\t\t\t\t),\n\t\t\t\t);\n\t\t\t},\n\t\t);\n\t\tawait Promise.all(updateEvents);\n\t\treturn { result, isError: false };\n\t} catch (error) {\n\t\tawait Promise.all(updateEvents);\n\t\treturn {\n\t\t\tresult: createErrorToolResult(error instanceof Error ? error.message : String(error)),\n\t\t\tisError: true,\n\t\t};\n\t}\n}\n\nasync function finalizeExecutedToolCall(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\tprepared: PreparedToolCall,\n\texecuted: ExecutedToolCallOutcome,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage> {\n\tlet result = executed.result;\n\tlet isError = executed.isError;\n\n\tif (config.afterToolCall) {\n\t\tconst afterResult = await config.afterToolCall(\n\t\t\t{\n\t\t\t\tassistantMessage,\n\t\t\t\ttoolCall: prepared.toolCall,\n\t\t\t\targs: prepared.args,\n\t\t\t\tresult,\n\t\t\t\tisError,\n\t\t\t\tcontext: currentContext,\n\t\t\t},\n\t\t\tsignal,\n\t\t);\n\t\tif (afterResult) {\n\t\t\tresult = {\n\t\t\t\tcontent: afterResult.content ?? result.content,\n\t\t\t\tdetails: afterResult.details ?? result.details,\n\t\t\t};\n\t\t\tisError = afterResult.isError ?? isError;\n\t\t}\n\t}\n\n\treturn await emitToolCallOutcome(prepared.toolCall, result, isError, emit);\n}\n\nfunction createErrorToolResult(message: string): AgentToolResult<any> {\n\treturn {\n\t\tcontent: [{ type: \"text\", text: message }],\n\t\tdetails: {},\n\t};\n}\n\nasync function emitToolCallOutcome(\n\ttoolCall: AgentToolCall,\n\tresult: AgentToolResult<any>,\n\tisError: boolean,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage> {\n\tawait emit({\n\t\ttype: \"tool_execution_end\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tresult,\n\t\tisError,\n\t});\n\n\tconst toolResultMessage: ToolResultMessage = {\n\t\trole: \"toolResult\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tcontent: result.content,\n\t\tdetails: result.details,\n\t\tisError,\n\t\ttimestamp: Date.now(),\n\t};\n\n\tawait emit({ type: \"message_start\", message: toolResultMessage });\n\tawait emit({ type: \"message_end\", message: toolResultMessage });\n\treturn toolResultMessage;\n}\n\nasync function skipToolCall(\n\ttoolCall: AgentToolCall,\n\temit: AgentEventSink,\n\toptions?: { emitStart?: boolean },\n): Promise<ToolResultMessage> {\n\tconst result = createErrorToolResult(\"Skipped due to queued user message.\");\n\n\tif (options?.emitStart !== false) {\n\t\tawait emit({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\t}\n\n\treturn await emitToolCallOutcome(toolCall, result, true, emit);\n}\n"]}
|
package/dist/agent-loop.js
CHANGED
|
@@ -9,20 +9,11 @@ import { EventStream, streamSimple, validateToolArguments, } from "@mariozechner
|
|
|
9
9
|
*/
|
|
10
10
|
export function agentLoop(prompts, context, config, signal, streamFn) {
|
|
11
11
|
const stream = createAgentStream();
|
|
12
|
-
(async () => {
|
|
13
|
-
|
|
14
|
-
|
|
15
|
-
|
|
16
|
-
|
|
17
|
-
};
|
|
18
|
-
stream.push({ type: "agent_start" });
|
|
19
|
-
stream.push({ type: "turn_start" });
|
|
20
|
-
for (const prompt of prompts) {
|
|
21
|
-
stream.push({ type: "message_start", message: prompt });
|
|
22
|
-
stream.push({ type: "message_end", message: prompt });
|
|
23
|
-
}
|
|
24
|
-
await runLoop(currentContext, newMessages, config, signal, stream, streamFn);
|
|
25
|
-
})();
|
|
12
|
+
void runAgentLoop(prompts, context, config, async (event) => {
|
|
13
|
+
stream.push(event);
|
|
14
|
+
}, signal, streamFn).then((messages) => {
|
|
15
|
+
stream.end(messages);
|
|
16
|
+
});
|
|
26
17
|
return stream;
|
|
27
18
|
}
|
|
28
19
|
/**
|
|
@@ -41,22 +32,49 @@ export function agentLoopContinue(context, config, signal, streamFn) {
|
|
|
41
32
|
throw new Error("Cannot continue from message role: assistant");
|
|
42
33
|
}
|
|
43
34
|
const stream = createAgentStream();
|
|
44
|
-
(async () => {
|
|
45
|
-
|
|
46
|
-
|
|
47
|
-
stream.
|
|
48
|
-
|
|
49
|
-
await runLoop(currentContext, newMessages, config, signal, stream, streamFn);
|
|
50
|
-
})();
|
|
35
|
+
void runAgentLoopContinue(context, config, async (event) => {
|
|
36
|
+
stream.push(event);
|
|
37
|
+
}, signal, streamFn).then((messages) => {
|
|
38
|
+
stream.end(messages);
|
|
39
|
+
});
|
|
51
40
|
return stream;
|
|
52
41
|
}
|
|
42
|
+
export async function runAgentLoop(prompts, context, config, emit, signal, streamFn) {
|
|
43
|
+
const newMessages = [...prompts];
|
|
44
|
+
const currentContext = {
|
|
45
|
+
...context,
|
|
46
|
+
messages: [...context.messages, ...prompts],
|
|
47
|
+
};
|
|
48
|
+
await emit({ type: "agent_start" });
|
|
49
|
+
await emit({ type: "turn_start" });
|
|
50
|
+
for (const prompt of prompts) {
|
|
51
|
+
await emit({ type: "message_start", message: prompt });
|
|
52
|
+
await emit({ type: "message_end", message: prompt });
|
|
53
|
+
}
|
|
54
|
+
await runLoop(currentContext, newMessages, config, signal, emit, streamFn);
|
|
55
|
+
return newMessages;
|
|
56
|
+
}
|
|
57
|
+
export async function runAgentLoopContinue(context, config, emit, signal, streamFn) {
|
|
58
|
+
if (context.messages.length === 0) {
|
|
59
|
+
throw new Error("Cannot continue: no messages in context");
|
|
60
|
+
}
|
|
61
|
+
if (context.messages[context.messages.length - 1].role === "assistant") {
|
|
62
|
+
throw new Error("Cannot continue from message role: assistant");
|
|
63
|
+
}
|
|
64
|
+
const newMessages = [];
|
|
65
|
+
const currentContext = { ...context };
|
|
66
|
+
await emit({ type: "agent_start" });
|
|
67
|
+
await emit({ type: "turn_start" });
|
|
68
|
+
await runLoop(currentContext, newMessages, config, signal, emit, streamFn);
|
|
69
|
+
return newMessages;
|
|
70
|
+
}
|
|
53
71
|
function createAgentStream() {
|
|
54
72
|
return new EventStream((event) => event.type === "agent_end", (event) => (event.type === "agent_end" ? event.messages : []));
|
|
55
73
|
}
|
|
56
74
|
/**
|
|
57
75
|
* Main loop logic shared by agentLoop and agentLoopContinue.
|
|
58
76
|
*/
|
|
59
|
-
async function runLoop(currentContext, newMessages, config, signal,
|
|
77
|
+
async function runLoop(currentContext, newMessages, config, signal, emit, streamFn) {
|
|
60
78
|
let firstTurn = true;
|
|
61
79
|
// Check for steering messages at start (user may have typed while waiting)
|
|
62
80
|
let pendingMessages = (await config.getSteeringMessages?.()) || [];
|
|
@@ -67,7 +85,7 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
|
|
|
67
85
|
// Inner loop: process tool calls and steering messages
|
|
68
86
|
while (hasMoreToolCalls || pendingMessages.length > 0) {
|
|
69
87
|
if (!firstTurn) {
|
|
70
|
-
|
|
88
|
+
await emit({ type: "turn_start" });
|
|
71
89
|
}
|
|
72
90
|
else {
|
|
73
91
|
firstTurn = false;
|
|
@@ -75,20 +93,19 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
|
|
|
75
93
|
// Process pending messages (inject before next assistant response)
|
|
76
94
|
if (pendingMessages.length > 0) {
|
|
77
95
|
for (const message of pendingMessages) {
|
|
78
|
-
|
|
79
|
-
|
|
96
|
+
await emit({ type: "message_start", message });
|
|
97
|
+
await emit({ type: "message_end", message });
|
|
80
98
|
currentContext.messages.push(message);
|
|
81
99
|
newMessages.push(message);
|
|
82
100
|
}
|
|
83
101
|
pendingMessages = [];
|
|
84
102
|
}
|
|
85
103
|
// Stream assistant response
|
|
86
|
-
const message = await streamAssistantResponse(currentContext, config, signal,
|
|
104
|
+
const message = await streamAssistantResponse(currentContext, config, signal, emit, streamFn);
|
|
87
105
|
newMessages.push(message);
|
|
88
106
|
if (message.stopReason === "error" || message.stopReason === "aborted") {
|
|
89
|
-
|
|
90
|
-
|
|
91
|
-
stream.end(newMessages);
|
|
107
|
+
await emit({ type: "turn_end", message, toolResults: [] });
|
|
108
|
+
await emit({ type: "agent_end", messages: newMessages });
|
|
92
109
|
return;
|
|
93
110
|
}
|
|
94
111
|
// Check for tool calls
|
|
@@ -96,7 +113,7 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
|
|
|
96
113
|
hasMoreToolCalls = toolCalls.length > 0;
|
|
97
114
|
const toolResults = [];
|
|
98
115
|
if (hasMoreToolCalls) {
|
|
99
|
-
const toolExecution = await executeToolCalls(currentContext
|
|
116
|
+
const toolExecution = await executeToolCalls(currentContext, message, config, signal, emit);
|
|
100
117
|
toolResults.push(...toolExecution.toolResults);
|
|
101
118
|
steeringAfterTools = toolExecution.steeringMessages ?? null;
|
|
102
119
|
for (const result of toolResults) {
|
|
@@ -104,7 +121,7 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
|
|
|
104
121
|
newMessages.push(result);
|
|
105
122
|
}
|
|
106
123
|
}
|
|
107
|
-
|
|
124
|
+
await emit({ type: "turn_end", message, toolResults });
|
|
108
125
|
// Get steering messages after turn completes
|
|
109
126
|
if (steeringAfterTools && steeringAfterTools.length > 0) {
|
|
110
127
|
pendingMessages = steeringAfterTools;
|
|
@@ -124,14 +141,13 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
|
|
|
124
141
|
// No more messages, exit
|
|
125
142
|
break;
|
|
126
143
|
}
|
|
127
|
-
|
|
128
|
-
stream.end(newMessages);
|
|
144
|
+
await emit({ type: "agent_end", messages: newMessages });
|
|
129
145
|
}
|
|
130
146
|
/**
|
|
131
147
|
* Stream an assistant response from the LLM.
|
|
132
148
|
* This is where AgentMessage[] gets transformed to Message[] for the LLM.
|
|
133
149
|
*/
|
|
134
|
-
async function streamAssistantResponse(context, config, signal,
|
|
150
|
+
async function streamAssistantResponse(context, config, signal, emit, streamFn) {
|
|
135
151
|
// Apply context transform if configured (AgentMessage[] → AgentMessage[])
|
|
136
152
|
let messages = context.messages;
|
|
137
153
|
if (config.transformContext) {
|
|
@@ -161,7 +177,7 @@ async function streamAssistantResponse(context, config, signal, stream, streamFn
|
|
|
161
177
|
partialMessage = event.partial;
|
|
162
178
|
context.messages.push(partialMessage);
|
|
163
179
|
addedPartial = true;
|
|
164
|
-
|
|
180
|
+
await emit({ type: "message_start", message: { ...partialMessage } });
|
|
165
181
|
break;
|
|
166
182
|
case "text_start":
|
|
167
183
|
case "text_delta":
|
|
@@ -175,7 +191,7 @@ async function streamAssistantResponse(context, config, signal, stream, streamFn
|
|
|
175
191
|
if (partialMessage) {
|
|
176
192
|
partialMessage = event.partial;
|
|
177
193
|
context.messages[context.messages.length - 1] = partialMessage;
|
|
178
|
-
|
|
194
|
+
await emit({
|
|
179
195
|
type: "message_update",
|
|
180
196
|
assistantMessageEvent: event,
|
|
181
197
|
message: { ...partialMessage },
|
|
@@ -192,117 +208,240 @@ async function streamAssistantResponse(context, config, signal, stream, streamFn
|
|
|
192
208
|
context.messages.push(finalMessage);
|
|
193
209
|
}
|
|
194
210
|
if (!addedPartial) {
|
|
195
|
-
|
|
211
|
+
await emit({ type: "message_start", message: { ...finalMessage } });
|
|
196
212
|
}
|
|
197
|
-
|
|
213
|
+
await emit({ type: "message_end", message: finalMessage });
|
|
198
214
|
return finalMessage;
|
|
199
215
|
}
|
|
200
216
|
}
|
|
201
217
|
}
|
|
202
|
-
|
|
218
|
+
const finalMessage = await response.result();
|
|
219
|
+
if (addedPartial) {
|
|
220
|
+
context.messages[context.messages.length - 1] = finalMessage;
|
|
221
|
+
}
|
|
222
|
+
else {
|
|
223
|
+
context.messages.push(finalMessage);
|
|
224
|
+
await emit({ type: "message_start", message: { ...finalMessage } });
|
|
225
|
+
}
|
|
226
|
+
await emit({ type: "message_end", message: finalMessage });
|
|
227
|
+
return finalMessage;
|
|
203
228
|
}
|
|
204
229
|
/**
|
|
205
230
|
* Execute tool calls from an assistant message.
|
|
206
231
|
*/
|
|
207
|
-
async function executeToolCalls(
|
|
232
|
+
async function executeToolCalls(currentContext, assistantMessage, config, signal, emit) {
|
|
208
233
|
const toolCalls = assistantMessage.content.filter((c) => c.type === "toolCall");
|
|
234
|
+
if (config.toolExecution === "sequential") {
|
|
235
|
+
return executeToolCallsSequential(currentContext, assistantMessage, toolCalls, config, signal, emit);
|
|
236
|
+
}
|
|
237
|
+
return executeToolCallsParallel(currentContext, assistantMessage, toolCalls, config, signal, emit);
|
|
238
|
+
}
|
|
239
|
+
async function executeToolCallsSequential(currentContext, assistantMessage, toolCalls, config, signal, emit) {
|
|
209
240
|
const results = [];
|
|
210
241
|
let steeringMessages;
|
|
211
242
|
for (let index = 0; index < toolCalls.length; index++) {
|
|
212
243
|
const toolCall = toolCalls[index];
|
|
213
|
-
|
|
214
|
-
stream.push({
|
|
244
|
+
await emit({
|
|
215
245
|
type: "tool_execution_start",
|
|
216
246
|
toolCallId: toolCall.id,
|
|
217
247
|
toolName: toolCall.name,
|
|
218
248
|
args: toolCall.arguments,
|
|
219
249
|
});
|
|
220
|
-
|
|
221
|
-
|
|
222
|
-
|
|
223
|
-
if (!tool)
|
|
224
|
-
throw new Error(`Tool ${toolCall.name} not found`);
|
|
225
|
-
const validatedArgs = validateToolArguments(tool, toolCall);
|
|
226
|
-
result = await tool.execute(toolCall.id, validatedArgs, signal, (partialResult) => {
|
|
227
|
-
stream.push({
|
|
228
|
-
type: "tool_execution_update",
|
|
229
|
-
toolCallId: toolCall.id,
|
|
230
|
-
toolName: toolCall.name,
|
|
231
|
-
args: toolCall.arguments,
|
|
232
|
-
partialResult,
|
|
233
|
-
});
|
|
234
|
-
});
|
|
250
|
+
const preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);
|
|
251
|
+
if (preparation.kind === "immediate") {
|
|
252
|
+
results.push(await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit));
|
|
235
253
|
}
|
|
236
|
-
|
|
237
|
-
|
|
238
|
-
|
|
239
|
-
|
|
240
|
-
|
|
241
|
-
|
|
254
|
+
else {
|
|
255
|
+
const executed = await executePreparedToolCall(preparation, signal, emit);
|
|
256
|
+
results.push(await finalizeExecutedToolCall(currentContext, assistantMessage, preparation, executed, config, signal, emit));
|
|
257
|
+
}
|
|
258
|
+
if (config.getSteeringMessages) {
|
|
259
|
+
const steering = await config.getSteeringMessages();
|
|
260
|
+
if (steering.length > 0) {
|
|
261
|
+
steeringMessages = steering;
|
|
262
|
+
const remainingCalls = toolCalls.slice(index + 1);
|
|
263
|
+
for (const skipped of remainingCalls) {
|
|
264
|
+
results.push(await skipToolCall(skipped, emit));
|
|
265
|
+
}
|
|
266
|
+
break;
|
|
267
|
+
}
|
|
242
268
|
}
|
|
243
|
-
|
|
244
|
-
|
|
269
|
+
}
|
|
270
|
+
return { toolResults: results, steeringMessages };
|
|
271
|
+
}
|
|
272
|
+
async function executeToolCallsParallel(currentContext, assistantMessage, toolCalls, config, signal, emit) {
|
|
273
|
+
const results = [];
|
|
274
|
+
const runnableCalls = [];
|
|
275
|
+
let steeringMessages;
|
|
276
|
+
for (let index = 0; index < toolCalls.length; index++) {
|
|
277
|
+
const toolCall = toolCalls[index];
|
|
278
|
+
await emit({
|
|
279
|
+
type: "tool_execution_start",
|
|
245
280
|
toolCallId: toolCall.id,
|
|
246
281
|
toolName: toolCall.name,
|
|
247
|
-
|
|
248
|
-
isError,
|
|
282
|
+
args: toolCall.arguments,
|
|
249
283
|
});
|
|
250
|
-
const
|
|
251
|
-
|
|
252
|
-
|
|
253
|
-
|
|
254
|
-
|
|
255
|
-
|
|
256
|
-
|
|
257
|
-
|
|
258
|
-
|
|
259
|
-
results.push(toolResultMessage);
|
|
260
|
-
stream.push({ type: "message_start", message: toolResultMessage });
|
|
261
|
-
stream.push({ type: "message_end", message: toolResultMessage });
|
|
262
|
-
// Check for steering messages - skip remaining tools if user interrupted
|
|
263
|
-
if (getSteeringMessages) {
|
|
264
|
-
const steering = await getSteeringMessages();
|
|
284
|
+
const preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);
|
|
285
|
+
if (preparation.kind === "immediate") {
|
|
286
|
+
results.push(await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit));
|
|
287
|
+
}
|
|
288
|
+
else {
|
|
289
|
+
runnableCalls.push(preparation);
|
|
290
|
+
}
|
|
291
|
+
if (config.getSteeringMessages) {
|
|
292
|
+
const steering = await config.getSteeringMessages();
|
|
265
293
|
if (steering.length > 0) {
|
|
266
294
|
steeringMessages = steering;
|
|
295
|
+
for (const runnable of runnableCalls) {
|
|
296
|
+
results.push(await skipToolCall(runnable.toolCall, emit, { emitStart: false }));
|
|
297
|
+
}
|
|
267
298
|
const remainingCalls = toolCalls.slice(index + 1);
|
|
268
299
|
for (const skipped of remainingCalls) {
|
|
269
|
-
results.push(skipToolCall(skipped,
|
|
300
|
+
results.push(await skipToolCall(skipped, emit));
|
|
270
301
|
}
|
|
271
|
-
|
|
302
|
+
return { toolResults: results, steeringMessages };
|
|
272
303
|
}
|
|
273
304
|
}
|
|
274
305
|
}
|
|
306
|
+
const runningCalls = runnableCalls.map((prepared) => ({
|
|
307
|
+
prepared,
|
|
308
|
+
execution: executePreparedToolCall(prepared, signal, emit),
|
|
309
|
+
}));
|
|
310
|
+
for (const running of runningCalls) {
|
|
311
|
+
const executed = await running.execution;
|
|
312
|
+
results.push(await finalizeExecutedToolCall(currentContext, assistantMessage, running.prepared, executed, config, signal, emit));
|
|
313
|
+
}
|
|
314
|
+
if (!steeringMessages && config.getSteeringMessages) {
|
|
315
|
+
const steering = await config.getSteeringMessages();
|
|
316
|
+
if (steering.length > 0) {
|
|
317
|
+
steeringMessages = steering;
|
|
318
|
+
}
|
|
319
|
+
}
|
|
275
320
|
return { toolResults: results, steeringMessages };
|
|
276
321
|
}
|
|
277
|
-
function
|
|
278
|
-
const
|
|
279
|
-
|
|
322
|
+
async function prepareToolCall(currentContext, assistantMessage, toolCall, config, signal) {
|
|
323
|
+
const tool = currentContext.tools?.find((t) => t.name === toolCall.name);
|
|
324
|
+
if (!tool) {
|
|
325
|
+
return {
|
|
326
|
+
kind: "immediate",
|
|
327
|
+
result: createErrorToolResult(`Tool ${toolCall.name} not found`),
|
|
328
|
+
isError: true,
|
|
329
|
+
};
|
|
330
|
+
}
|
|
331
|
+
try {
|
|
332
|
+
const validatedArgs = validateToolArguments(tool, toolCall);
|
|
333
|
+
if (config.beforeToolCall) {
|
|
334
|
+
const beforeResult = await config.beforeToolCall({
|
|
335
|
+
assistantMessage,
|
|
336
|
+
toolCall,
|
|
337
|
+
args: validatedArgs,
|
|
338
|
+
context: currentContext,
|
|
339
|
+
}, signal);
|
|
340
|
+
if (beforeResult?.block) {
|
|
341
|
+
return {
|
|
342
|
+
kind: "immediate",
|
|
343
|
+
result: createErrorToolResult(beforeResult.reason || "Tool execution was blocked"),
|
|
344
|
+
isError: true,
|
|
345
|
+
};
|
|
346
|
+
}
|
|
347
|
+
}
|
|
348
|
+
return {
|
|
349
|
+
kind: "prepared",
|
|
350
|
+
toolCall,
|
|
351
|
+
tool,
|
|
352
|
+
args: validatedArgs,
|
|
353
|
+
};
|
|
354
|
+
}
|
|
355
|
+
catch (error) {
|
|
356
|
+
return {
|
|
357
|
+
kind: "immediate",
|
|
358
|
+
result: createErrorToolResult(error instanceof Error ? error.message : String(error)),
|
|
359
|
+
isError: true,
|
|
360
|
+
};
|
|
361
|
+
}
|
|
362
|
+
}
|
|
363
|
+
async function executePreparedToolCall(prepared, signal, emit) {
|
|
364
|
+
const updateEvents = [];
|
|
365
|
+
try {
|
|
366
|
+
const result = await prepared.tool.execute(prepared.toolCall.id, prepared.args, signal, (partialResult) => {
|
|
367
|
+
updateEvents.push(Promise.resolve(emit({
|
|
368
|
+
type: "tool_execution_update",
|
|
369
|
+
toolCallId: prepared.toolCall.id,
|
|
370
|
+
toolName: prepared.toolCall.name,
|
|
371
|
+
args: prepared.toolCall.arguments,
|
|
372
|
+
partialResult,
|
|
373
|
+
})));
|
|
374
|
+
});
|
|
375
|
+
await Promise.all(updateEvents);
|
|
376
|
+
return { result, isError: false };
|
|
377
|
+
}
|
|
378
|
+
catch (error) {
|
|
379
|
+
await Promise.all(updateEvents);
|
|
380
|
+
return {
|
|
381
|
+
result: createErrorToolResult(error instanceof Error ? error.message : String(error)),
|
|
382
|
+
isError: true,
|
|
383
|
+
};
|
|
384
|
+
}
|
|
385
|
+
}
|
|
386
|
+
async function finalizeExecutedToolCall(currentContext, assistantMessage, prepared, executed, config, signal, emit) {
|
|
387
|
+
let result = executed.result;
|
|
388
|
+
let isError = executed.isError;
|
|
389
|
+
if (config.afterToolCall) {
|
|
390
|
+
const afterResult = await config.afterToolCall({
|
|
391
|
+
assistantMessage,
|
|
392
|
+
toolCall: prepared.toolCall,
|
|
393
|
+
args: prepared.args,
|
|
394
|
+
result,
|
|
395
|
+
isError,
|
|
396
|
+
context: currentContext,
|
|
397
|
+
}, signal);
|
|
398
|
+
if (afterResult) {
|
|
399
|
+
result = {
|
|
400
|
+
content: afterResult.content ?? result.content,
|
|
401
|
+
details: afterResult.details ?? result.details,
|
|
402
|
+
};
|
|
403
|
+
isError = afterResult.isError ?? isError;
|
|
404
|
+
}
|
|
405
|
+
}
|
|
406
|
+
return await emitToolCallOutcome(prepared.toolCall, result, isError, emit);
|
|
407
|
+
}
|
|
408
|
+
function createErrorToolResult(message) {
|
|
409
|
+
return {
|
|
410
|
+
content: [{ type: "text", text: message }],
|
|
280
411
|
details: {},
|
|
281
412
|
};
|
|
282
|
-
|
|
283
|
-
|
|
284
|
-
|
|
285
|
-
toolName: toolCall.name,
|
|
286
|
-
args: toolCall.arguments,
|
|
287
|
-
});
|
|
288
|
-
stream.push({
|
|
413
|
+
}
|
|
414
|
+
async function emitToolCallOutcome(toolCall, result, isError, emit) {
|
|
415
|
+
await emit({
|
|
289
416
|
type: "tool_execution_end",
|
|
290
417
|
toolCallId: toolCall.id,
|
|
291
418
|
toolName: toolCall.name,
|
|
292
419
|
result,
|
|
293
|
-
isError
|
|
420
|
+
isError,
|
|
294
421
|
});
|
|
295
422
|
const toolResultMessage = {
|
|
296
423
|
role: "toolResult",
|
|
297
424
|
toolCallId: toolCall.id,
|
|
298
425
|
toolName: toolCall.name,
|
|
299
426
|
content: result.content,
|
|
300
|
-
details:
|
|
301
|
-
isError
|
|
427
|
+
details: result.details,
|
|
428
|
+
isError,
|
|
302
429
|
timestamp: Date.now(),
|
|
303
430
|
};
|
|
304
|
-
|
|
305
|
-
|
|
431
|
+
await emit({ type: "message_start", message: toolResultMessage });
|
|
432
|
+
await emit({ type: "message_end", message: toolResultMessage });
|
|
306
433
|
return toolResultMessage;
|
|
307
434
|
}
|
|
435
|
+
async function skipToolCall(toolCall, emit, options) {
|
|
436
|
+
const result = createErrorToolResult("Skipped due to queued user message.");
|
|
437
|
+
if (options?.emitStart !== false) {
|
|
438
|
+
await emit({
|
|
439
|
+
type: "tool_execution_start",
|
|
440
|
+
toolCallId: toolCall.id,
|
|
441
|
+
toolName: toolCall.name,
|
|
442
|
+
args: toolCall.arguments,
|
|
443
|
+
});
|
|
444
|
+
}
|
|
445
|
+
return await emitToolCallOutcome(toolCall, result, true, emit);
|
|
446
|
+
}
|
|
308
447
|
//# sourceMappingURL=agent-loop.js.map
|