@draht/agent-core 2026.3.14 → 2026.3.25

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/README.md CHANGED
@@ -99,6 +99,15 @@ prompt("Read config.json")
99
99
  └─ agent_end
100
100
  ```
101
101
 
102
+ Tool execution mode is configurable:
103
+
104
+ - `parallel` (default): preflight tool calls sequentially, execute allowed tools concurrently, emit final `tool_execution_end` and `toolResult` messages in assistant source order
105
+ - `sequential`: execute tool calls one by one, matching the historical behavior
106
+
107
+ The `beforeToolCall` hook runs after `tool_execution_start` and validated argument parsing. It can block execution. The `afterToolCall` hook runs after tool execution finishes and before `tool_execution_end` and final tool result message events are emitted.
108
+
109
+ When you use the `Agent` class, assistant `message_end` processing is treated as a barrier before tool preflight begins. That means `beforeToolCall` sees agent state that already includes the assistant message that requested the tool call.
110
+
102
111
  ### continue() Event Sequence
103
112
 
104
113
  `continue()` resumes from existing context without adding a new message. Use it for retries after errors.
@@ -159,6 +168,23 @@ const agent = new Agent({
159
168
  // Dynamic API key resolution (for expiring OAuth tokens)
160
169
  getApiKey: async (provider) => refreshToken(),
161
170
 
171
+ // Tool execution mode: "parallel" (default) or "sequential"
172
+ toolExecution: "parallel",
173
+
174
+ // Preflight each tool call after args are validated. Can block execution.
175
+ beforeToolCall: async ({ toolCall, args, context }) => {
176
+ if (toolCall.name === "bash") {
177
+ return { block: true, reason: "bash is disabled" };
178
+ }
179
+ },
180
+
181
+ // Postprocess each tool result before final tool events are emitted.
182
+ afterToolCall: async ({ toolCall, result, isError, context }) => {
183
+ if (!isError) {
184
+ return { details: { ...result.details, audited: true } };
185
+ }
186
+ },
187
+
162
188
  // Custom thinking budgets for token-based providers
163
189
  thinkingBudgets: {
164
190
  minimal: 128,
@@ -214,6 +240,9 @@ agent.setSystemPrompt("New prompt");
214
240
  agent.setModel(getModel("openai", "gpt-4o"));
215
241
  agent.setThinkingLevel("medium");
216
242
  agent.setTools([myTool]);
243
+ agent.setToolExecution("sequential");
244
+ agent.setBeforeToolCall(async ({ toolCall }) => undefined);
245
+ agent.setAfterToolCall(async ({ toolCall, result }) => undefined);
217
246
  agent.replaceMessages(newMessages);
218
247
  agent.appendMessage(message);
219
248
  agent.clearMessages();
@@ -281,10 +310,10 @@ agent.clearAllQueues();
281
310
 
282
311
  Use clearSteeringQueue, clearFollowUpQueue, or clearAllQueues to drop queued messages.
283
312
 
284
- When steering messages are detected after a tool completes:
285
- 1. Remaining tools are skipped with error results
313
+ When steering messages are detected after a turn completes:
314
+ 1. All tool calls from the current assistant message have already finished
286
315
  2. Steering messages are injected
287
- 3. LLM responds to the interruption
316
+ 3. The LLM responds on the next turn
288
317
 
289
318
  Follow-up messages are checked only when there are no more tool calls and no steering messages. If any are queued, they are injected and another turn runs.
290
319
 
@@ -393,6 +422,9 @@ const context: AgentContext = {
393
422
  const config: AgentLoopConfig = {
394
423
  model: getModel("openai", "gpt-4o"),
395
424
  convertToLlm: (msgs) => msgs.filter(m => ["user", "assistant", "toolResult"].includes(m.role)),
425
+ toolExecution: "parallel",
426
+ beforeToolCall: async ({ toolCall, args, context }) => undefined,
427
+ afterToolCall: async ({ toolCall, result, isError, context }) => undefined,
396
428
  };
397
429
 
398
430
  const userMessage = { role: "user", content: "Hello", timestamp: Date.now() };
@@ -407,6 +439,8 @@ for await (const event of agentLoopContinue(context, config)) {
407
439
  }
408
440
  ```
409
441
 
442
+ These low-level streams are observational. They preserve event order, but they do not wait for your async event handling to settle before later producer phases continue. If you need message processing to act as a barrier before tool preflight, use the `Agent` class instead of raw `agentLoop()` or `agentLoopContinue()`.
443
+
410
444
  ## License
411
445
 
412
446
  MIT
@@ -4,6 +4,7 @@
4
4
  */
5
5
  import { EventStream } from "@draht/ai";
6
6
  import type { AgentContext, AgentEvent, AgentLoopConfig, AgentMessage, StreamFn } from "./types.js";
7
+ export type AgentEventSink = (event: AgentEvent) => Promise<void> | void;
7
8
  /**
8
9
  * Start an agent loop with a new prompt message.
9
10
  * The prompt is added to the context and events are emitted for it.
@@ -18,4 +19,6 @@ export declare function agentLoop(prompts: AgentMessage[], context: AgentContext
18
19
  * This cannot be validated here since `convertToLlm` is only called once per turn.
19
20
  */
20
21
  export declare function agentLoopContinue(context: AgentContext, config: AgentLoopConfig, signal?: AbortSignal, streamFn?: StreamFn): EventStream<AgentEvent, AgentMessage[]>;
22
+ export declare function runAgentLoop(prompts: AgentMessage[], context: AgentContext, config: AgentLoopConfig, emit: AgentEventSink, signal?: AbortSignal, streamFn?: StreamFn): Promise<AgentMessage[]>;
23
+ export declare function runAgentLoopContinue(context: AgentContext, config: AgentLoopConfig, emit: AgentEventSink, signal?: AbortSignal, streamFn?: StreamFn): Promise<AgentMessage[]>;
21
24
  //# sourceMappingURL=agent-loop.d.ts.map
@@ -1 +1 @@
1
- {"version":3,"file":"agent-loop.d.ts","sourceRoot":"","sources":["../src/agent-loop.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,EAGN,WAAW,EAIX,MAAM,WAAW,CAAC;AACnB,OAAO,KAAK,EACX,YAAY,EACZ,UAAU,EACV,eAAe,EACf,YAAY,EAGZ,QAAQ,EACR,MAAM,YAAY,CAAC;AAEpB;;;GAGG;AACH,wBAAgB,SAAS,CACxB,OAAO,EAAE,YAAY,EAAE,EACvB,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAqBzC;AAED;;;;;;;GAOG;AACH,wBAAgB,iBAAiB,CAChC,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAsBzC","sourcesContent":["/**\n * Agent loop that works with AgentMessage throughout.\n * Transforms to Message[] only at the LLM call boundary.\n */\n\nimport {\n\ttype AssistantMessage,\n\ttype Context,\n\tEventStream,\n\tstreamSimple,\n\ttype ToolResultMessage,\n\tvalidateToolArguments,\n} from \"@draht/ai\";\nimport type {\n\tAgentContext,\n\tAgentEvent,\n\tAgentLoopConfig,\n\tAgentMessage,\n\tAgentTool,\n\tAgentToolResult,\n\tStreamFn,\n} from \"./types.js\";\n\n/**\n * Start an agent loop with a new prompt message.\n * The prompt is added to the context and events are emitted for it.\n */\nexport function agentLoop(\n\tprompts: AgentMessage[],\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tconst stream = createAgentStream();\n\n\t(async () => {\n\t\tconst newMessages: AgentMessage[] = [...prompts];\n\t\tconst currentContext: AgentContext = {\n\t\t\t...context,\n\t\t\tmessages: [...context.messages, ...prompts],\n\t\t};\n\n\t\tstream.push({ type: \"agent_start\" });\n\t\tstream.push({ type: \"turn_start\" });\n\t\tfor (const prompt of prompts) {\n\t\t\tstream.push({ type: \"message_start\", message: prompt });\n\t\t\tstream.push({ type: \"message_end\", message: prompt });\n\t\t}\n\n\t\tawait runLoop(currentContext, newMessages, config, signal, stream, streamFn);\n\t})();\n\n\treturn stream;\n}\n\n/**\n * Continue an agent loop from the current context without adding a new message.\n * Used for retries - context already has user message or tool results.\n *\n * **Important:** The last message in context must convert to a `user` or `toolResult` message\n * via `convertToLlm`. If it doesn't, the LLM provider will reject the request.\n * This cannot be validated here since `convertToLlm` is only called once per turn.\n */\nexport function agentLoopContinue(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tif (context.messages.length === 0) {\n\t\tthrow new Error(\"Cannot continue: no messages in context\");\n\t}\n\n\tif (context.messages[context.messages.length - 1].role === \"assistant\") {\n\t\tthrow new Error(\"Cannot continue from message role: assistant\");\n\t}\n\n\tconst stream = createAgentStream();\n\n\t(async () => {\n\t\tconst newMessages: AgentMessage[] = [];\n\t\tconst currentContext: AgentContext = { ...context };\n\n\t\tstream.push({ type: \"agent_start\" });\n\t\tstream.push({ type: \"turn_start\" });\n\n\t\tawait runLoop(currentContext, newMessages, config, signal, stream, streamFn);\n\t})();\n\n\treturn stream;\n}\n\nfunction createAgentStream(): EventStream<AgentEvent, AgentMessage[]> {\n\treturn new EventStream<AgentEvent, AgentMessage[]>(\n\t\t(event: AgentEvent) => event.type === \"agent_end\",\n\t\t(event: AgentEvent) => (event.type === \"agent_end\" ? event.messages : []),\n\t);\n}\n\n/**\n * Main loop logic shared by agentLoop and agentLoopContinue.\n */\nasync function runLoop(\n\tcurrentContext: AgentContext,\n\tnewMessages: AgentMessage[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n\tstreamFn?: StreamFn,\n): Promise<void> {\n\tlet firstTurn = true;\n\t// Check for steering messages at start (user may have typed while waiting)\n\tlet pendingMessages: AgentMessage[] = (await config.getSteeringMessages?.()) || [];\n\n\t// Outer loop: continues when queued follow-up messages arrive after agent would stop\n\twhile (true) {\n\t\tlet hasMoreToolCalls = true;\n\t\tlet steeringAfterTools: AgentMessage[] | null = null;\n\n\t\t// Inner loop: process tool calls and steering messages\n\t\twhile (hasMoreToolCalls || pendingMessages.length > 0) {\n\t\t\tif (!firstTurn) {\n\t\t\t\tstream.push({ type: \"turn_start\" });\n\t\t\t} else {\n\t\t\t\tfirstTurn = false;\n\t\t\t}\n\n\t\t\t// Process pending messages (inject before next assistant response)\n\t\t\tif (pendingMessages.length > 0) {\n\t\t\t\tfor (const message of pendingMessages) {\n\t\t\t\t\tstream.push({ type: \"message_start\", message });\n\t\t\t\t\tstream.push({ type: \"message_end\", message });\n\t\t\t\t\tcurrentContext.messages.push(message);\n\t\t\t\t\tnewMessages.push(message);\n\t\t\t\t}\n\t\t\t\tpendingMessages = [];\n\t\t\t}\n\n\t\t\t// Stream assistant response\n\t\t\tconst message = await streamAssistantResponse(currentContext, config, signal, stream, streamFn);\n\t\t\tnewMessages.push(message);\n\n\t\t\tif (message.stopReason === \"error\" || message.stopReason === \"aborted\") {\n\t\t\t\tstream.push({ type: \"turn_end\", message, toolResults: [] });\n\t\t\t\tstream.push({ type: \"agent_end\", messages: newMessages });\n\t\t\t\tstream.end(newMessages);\n\t\t\t\treturn;\n\t\t\t}\n\n\t\t\t// Check for tool calls\n\t\t\tconst toolCalls = message.content.filter((c) => c.type === \"toolCall\");\n\t\t\thasMoreToolCalls = toolCalls.length > 0;\n\n\t\t\tconst toolResults: ToolResultMessage[] = [];\n\t\t\tif (hasMoreToolCalls) {\n\t\t\t\tconst toolExecution = await executeToolCalls(\n\t\t\t\t\tcurrentContext.tools,\n\t\t\t\t\tmessage,\n\t\t\t\t\tsignal,\n\t\t\t\t\tstream,\n\t\t\t\t\tconfig.getSteeringMessages,\n\t\t\t\t);\n\t\t\t\ttoolResults.push(...toolExecution.toolResults);\n\t\t\t\tsteeringAfterTools = toolExecution.steeringMessages ?? null;\n\n\t\t\t\tfor (const result of toolResults) {\n\t\t\t\t\tcurrentContext.messages.push(result);\n\t\t\t\t\tnewMessages.push(result);\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tstream.push({ type: \"turn_end\", message, toolResults });\n\n\t\t\t// Get steering messages after turn completes\n\t\t\tif (steeringAfterTools && steeringAfterTools.length > 0) {\n\t\t\t\tpendingMessages = steeringAfterTools;\n\t\t\t\tsteeringAfterTools = null;\n\t\t\t} else {\n\t\t\t\tpendingMessages = (await config.getSteeringMessages?.()) || [];\n\t\t\t}\n\t\t}\n\n\t\t// Agent would stop here. Check for follow-up messages.\n\t\tconst followUpMessages = (await config.getFollowUpMessages?.()) || [];\n\t\tif (followUpMessages.length > 0) {\n\t\t\t// Set as pending so inner loop processes them\n\t\t\tpendingMessages = followUpMessages;\n\t\t\tcontinue;\n\t\t}\n\n\t\t// No more messages, exit\n\t\tbreak;\n\t}\n\n\tstream.push({ type: \"agent_end\", messages: newMessages });\n\tstream.end(newMessages);\n}\n\n/**\n * Stream an assistant response from the LLM.\n * This is where AgentMessage[] gets transformed to Message[] for the LLM.\n */\nasync function streamAssistantResponse(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n\tstreamFn?: StreamFn,\n): Promise<AssistantMessage> {\n\t// Apply context transform if configured (AgentMessage[] → AgentMessage[])\n\tlet messages = context.messages;\n\tif (config.transformContext) {\n\t\tmessages = await config.transformContext(messages, signal);\n\t}\n\n\t// Convert to LLM-compatible messages (AgentMessage[] → Message[])\n\tconst llmMessages = await config.convertToLlm(messages);\n\n\t// Build LLM context\n\tconst llmContext: Context = {\n\t\tsystemPrompt: context.systemPrompt,\n\t\tmessages: llmMessages,\n\t\ttools: context.tools,\n\t};\n\n\tconst streamFunction = streamFn || streamSimple;\n\n\t// Resolve API key (important for expiring tokens)\n\tconst resolvedApiKey =\n\t\t(config.getApiKey ? await config.getApiKey(config.model.provider) : undefined) || config.apiKey;\n\n\tconst response = await streamFunction(config.model, llmContext, {\n\t\t...config,\n\t\tapiKey: resolvedApiKey,\n\t\tsignal,\n\t});\n\n\tlet partialMessage: AssistantMessage | null = null;\n\tlet addedPartial = false;\n\n\tfor await (const event of response) {\n\t\tswitch (event.type) {\n\t\t\tcase \"start\":\n\t\t\t\tpartialMessage = event.partial;\n\t\t\t\tcontext.messages.push(partialMessage);\n\t\t\t\taddedPartial = true;\n\t\t\t\tstream.push({ type: \"message_start\", message: { ...partialMessage } });\n\t\t\t\tbreak;\n\n\t\t\tcase \"text_start\":\n\t\t\tcase \"text_delta\":\n\t\t\tcase \"text_end\":\n\t\t\tcase \"thinking_start\":\n\t\t\tcase \"thinking_delta\":\n\t\t\tcase \"thinking_end\":\n\t\t\tcase \"toolcall_start\":\n\t\t\tcase \"toolcall_delta\":\n\t\t\tcase \"toolcall_end\":\n\t\t\t\tif (partialMessage) {\n\t\t\t\t\tpartialMessage = event.partial;\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = partialMessage;\n\t\t\t\t\tstream.push({\n\t\t\t\t\t\ttype: \"message_update\",\n\t\t\t\t\t\tassistantMessageEvent: event,\n\t\t\t\t\t\tmessage: { ...partialMessage },\n\t\t\t\t\t});\n\t\t\t\t}\n\t\t\t\tbreak;\n\n\t\t\tcase \"done\":\n\t\t\tcase \"error\": {\n\t\t\t\tconst finalMessage = await response.result();\n\t\t\t\tif (addedPartial) {\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = finalMessage;\n\t\t\t\t} else {\n\t\t\t\t\tcontext.messages.push(finalMessage);\n\t\t\t\t}\n\t\t\t\tif (!addedPartial) {\n\t\t\t\t\tstream.push({ type: \"message_start\", message: { ...finalMessage } });\n\t\t\t\t}\n\t\t\t\tstream.push({ type: \"message_end\", message: finalMessage });\n\t\t\t\treturn finalMessage;\n\t\t\t}\n\t\t}\n\t}\n\n\treturn await response.result();\n}\n\n/**\n * Execute tool calls from an assistant message.\n */\nasync function executeToolCalls(\n\ttools: AgentTool<any>[] | undefined,\n\tassistantMessage: AssistantMessage,\n\tsignal: AbortSignal | undefined,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n\tgetSteeringMessages?: AgentLoopConfig[\"getSteeringMessages\"],\n): Promise<{ toolResults: ToolResultMessage[]; steeringMessages?: AgentMessage[] }> {\n\tconst toolCalls = assistantMessage.content.filter((c) => c.type === \"toolCall\");\n\tconst results: ToolResultMessage[] = [];\n\tlet steeringMessages: AgentMessage[] | undefined;\n\n\tfor (let index = 0; index < toolCalls.length; index++) {\n\t\tconst toolCall = toolCalls[index];\n\t\tconst tool = tools?.find((t) => t.name === toolCall.name);\n\n\t\tstream.push({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\n\t\tlet result: AgentToolResult<any>;\n\t\tlet isError = false;\n\n\t\ttry {\n\t\t\tif (!tool) throw new Error(`Tool ${toolCall.name} not found`);\n\n\t\t\tconst validatedArgs = validateToolArguments(tool, toolCall);\n\n\t\t\tresult = await tool.execute(toolCall.id, validatedArgs, signal, (partialResult) => {\n\t\t\t\tstream.push({\n\t\t\t\t\ttype: \"tool_execution_update\",\n\t\t\t\t\ttoolCallId: toolCall.id,\n\t\t\t\t\ttoolName: toolCall.name,\n\t\t\t\t\targs: toolCall.arguments,\n\t\t\t\t\tpartialResult,\n\t\t\t\t});\n\t\t\t});\n\t\t} catch (e) {\n\t\t\tresult = {\n\t\t\t\tcontent: [{ type: \"text\", text: e instanceof Error ? e.message : String(e) }],\n\t\t\t\tdetails: {},\n\t\t\t};\n\t\t\tisError = true;\n\t\t}\n\n\t\tstream.push({\n\t\t\ttype: \"tool_execution_end\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\tresult,\n\t\t\tisError,\n\t\t});\n\n\t\tconst toolResultMessage: ToolResultMessage = {\n\t\t\trole: \"toolResult\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\tcontent: result.content,\n\t\t\tdetails: result.details,\n\t\t\tisError,\n\t\t\ttimestamp: Date.now(),\n\t\t};\n\n\t\tresults.push(toolResultMessage);\n\t\tstream.push({ type: \"message_start\", message: toolResultMessage });\n\t\tstream.push({ type: \"message_end\", message: toolResultMessage });\n\n\t\t// Check for steering messages - skip remaining tools if user interrupted\n\t\tif (getSteeringMessages) {\n\t\t\tconst steering = await getSteeringMessages();\n\t\t\tif (steering.length > 0) {\n\t\t\t\tsteeringMessages = steering;\n\t\t\t\tconst remainingCalls = toolCalls.slice(index + 1);\n\t\t\t\tfor (const skipped of remainingCalls) {\n\t\t\t\t\tresults.push(skipToolCall(skipped, stream));\n\t\t\t\t}\n\t\t\t\tbreak;\n\t\t\t}\n\t\t}\n\t}\n\n\treturn { toolResults: results, steeringMessages };\n}\n\nfunction skipToolCall(\n\ttoolCall: Extract<AssistantMessage[\"content\"][number], { type: \"toolCall\" }>,\n\tstream: EventStream<AgentEvent, AgentMessage[]>,\n): ToolResultMessage {\n\tconst result: AgentToolResult<any> = {\n\t\tcontent: [{ type: \"text\", text: \"Skipped due to queued user message.\" }],\n\t\tdetails: {},\n\t};\n\n\tstream.push({\n\t\ttype: \"tool_execution_start\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\targs: toolCall.arguments,\n\t});\n\tstream.push({\n\t\ttype: \"tool_execution_end\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tresult,\n\t\tisError: true,\n\t});\n\n\tconst toolResultMessage: ToolResultMessage = {\n\t\trole: \"toolResult\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tcontent: result.content,\n\t\tdetails: {},\n\t\tisError: true,\n\t\ttimestamp: Date.now(),\n\t};\n\n\tstream.push({ type: \"message_start\", message: toolResultMessage });\n\tstream.push({ type: \"message_end\", message: toolResultMessage });\n\n\treturn toolResultMessage;\n}\n"]}
1
+ {"version":3,"file":"agent-loop.d.ts","sourceRoot":"","sources":["../src/agent-loop.ts"],"names":[],"mappings":"AAAA;;;GAGG;AAEH,OAAO,EAGN,WAAW,EAIX,MAAM,WAAW,CAAC;AACnB,OAAO,KAAK,EACX,YAAY,EACZ,UAAU,EACV,eAAe,EACf,YAAY,EAIZ,QAAQ,EACR,MAAM,YAAY,CAAC;AAEpB,MAAM,MAAM,cAAc,GAAG,CAAC,KAAK,EAAE,UAAU,KAAK,OAAO,CAAC,IAAI,CAAC,GAAG,IAAI,CAAC;AAEzE;;;GAGG;AACH,wBAAgB,SAAS,CACxB,OAAO,EAAE,YAAY,EAAE,EACvB,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CAqBzC;AAED;;;;;;;GAOG;AACH,wBAAgB,iBAAiB,CAChC,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,WAAW,CAAC,UAAU,EAAE,YAAY,EAAE,CAAC,CA4BzC;AAED,wBAAsB,YAAY,CACjC,OAAO,EAAE,YAAY,EAAE,EACvB,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,IAAI,EAAE,cAAc,EACpB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,OAAO,CAAC,YAAY,EAAE,CAAC,CAgBzB;AAED,wBAAsB,oBAAoB,CACzC,OAAO,EAAE,YAAY,EACrB,MAAM,EAAE,eAAe,EACvB,IAAI,EAAE,cAAc,EACpB,MAAM,CAAC,EAAE,WAAW,EACpB,QAAQ,CAAC,EAAE,QAAQ,GACjB,OAAO,CAAC,YAAY,EAAE,CAAC,CAiBzB","sourcesContent":["/**\n * Agent loop that works with AgentMessage throughout.\n * Transforms to Message[] only at the LLM call boundary.\n */\n\nimport {\n\ttype AssistantMessage,\n\ttype Context,\n\tEventStream,\n\tstreamSimple,\n\ttype ToolResultMessage,\n\tvalidateToolArguments,\n} from \"@draht/ai\";\nimport type {\n\tAgentContext,\n\tAgentEvent,\n\tAgentLoopConfig,\n\tAgentMessage,\n\tAgentTool,\n\tAgentToolCall,\n\tAgentToolResult,\n\tStreamFn,\n} from \"./types.js\";\n\nexport type AgentEventSink = (event: AgentEvent) => Promise<void> | void;\n\n/**\n * Start an agent loop with a new prompt message.\n * The prompt is added to the context and events are emitted for it.\n */\nexport function agentLoop(\n\tprompts: AgentMessage[],\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tconst stream = createAgentStream();\n\n\tvoid runAgentLoop(\n\t\tprompts,\n\t\tcontext,\n\t\tconfig,\n\t\tasync (event) => {\n\t\t\tstream.push(event);\n\t\t},\n\t\tsignal,\n\t\tstreamFn,\n\t)\n\t\t.then((messages) => {\n\t\t\tstream.end(messages);\n\t\t})\n\t\t.catch(() => {\n\t\t\tstream.end([]);\n\t\t});\n\n\treturn stream;\n}\n\n/**\n * Continue an agent loop from the current context without adding a new message.\n * Used for retries - context already has user message or tool results.\n *\n * **Important:** The last message in context must convert to a `user` or `toolResult` message\n * via `convertToLlm`. If it doesn't, the LLM provider will reject the request.\n * This cannot be validated here since `convertToLlm` is only called once per turn.\n */\nexport function agentLoopContinue(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): EventStream<AgentEvent, AgentMessage[]> {\n\tif (context.messages.length === 0) {\n\t\tthrow new Error(\"Cannot continue: no messages in context\");\n\t}\n\n\tif (context.messages[context.messages.length - 1].role === \"assistant\") {\n\t\tthrow new Error(\"Cannot continue from message role: assistant\");\n\t}\n\n\tconst stream = createAgentStream();\n\n\tvoid runAgentLoopContinue(\n\t\tcontext,\n\t\tconfig,\n\t\tasync (event) => {\n\t\t\tstream.push(event);\n\t\t},\n\t\tsignal,\n\t\tstreamFn,\n\t)\n\t\t.then((messages) => {\n\t\t\tstream.end(messages);\n\t\t})\n\t\t.catch(() => {\n\t\t\tstream.end([]);\n\t\t});\n\n\treturn stream;\n}\n\nexport async function runAgentLoop(\n\tprompts: AgentMessage[],\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\temit: AgentEventSink,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): Promise<AgentMessage[]> {\n\tconst newMessages: AgentMessage[] = [...prompts];\n\tconst currentContext: AgentContext = {\n\t\t...context,\n\t\tmessages: [...context.messages, ...prompts],\n\t};\n\n\tawait emit({ type: \"agent_start\" });\n\tawait emit({ type: \"turn_start\" });\n\tfor (const prompt of prompts) {\n\t\tawait emit({ type: \"message_start\", message: prompt });\n\t\tawait emit({ type: \"message_end\", message: prompt });\n\t}\n\n\tawait runLoop(currentContext, newMessages, config, signal, emit, streamFn);\n\treturn newMessages;\n}\n\nexport async function runAgentLoopContinue(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\temit: AgentEventSink,\n\tsignal?: AbortSignal,\n\tstreamFn?: StreamFn,\n): Promise<AgentMessage[]> {\n\tif (context.messages.length === 0) {\n\t\tthrow new Error(\"Cannot continue: no messages in context\");\n\t}\n\n\tif (context.messages[context.messages.length - 1].role === \"assistant\") {\n\t\tthrow new Error(\"Cannot continue from message role: assistant\");\n\t}\n\n\tconst newMessages: AgentMessage[] = [];\n\tconst currentContext: AgentContext = { ...context };\n\n\tawait emit({ type: \"agent_start\" });\n\tawait emit({ type: \"turn_start\" });\n\n\tawait runLoop(currentContext, newMessages, config, signal, emit, streamFn);\n\treturn newMessages;\n}\n\nfunction createAgentStream(): EventStream<AgentEvent, AgentMessage[]> {\n\treturn new EventStream<AgentEvent, AgentMessage[]>(\n\t\t(event: AgentEvent) => event.type === \"agent_end\",\n\t\t(event: AgentEvent) => (event.type === \"agent_end\" ? event.messages : []),\n\t);\n}\n\n/**\n * Main loop logic shared by agentLoop and agentLoopContinue.\n */\nasync function runLoop(\n\tcurrentContext: AgentContext,\n\tnewMessages: AgentMessage[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n\tstreamFn?: StreamFn,\n): Promise<void> {\n\tlet firstTurn = true;\n\t// Check for steering messages at start (user may have typed while waiting)\n\tlet pendingMessages: AgentMessage[] = (await config.getSteeringMessages?.()) || [];\n\n\t// Outer loop: continues when queued follow-up messages arrive after agent would stop\n\twhile (true) {\n\t\tlet hasMoreToolCalls = true;\n\n\t\t// Inner loop: process tool calls and steering messages\n\t\twhile (hasMoreToolCalls || pendingMessages.length > 0) {\n\t\t\tif (!firstTurn) {\n\t\t\t\tawait emit({ type: \"turn_start\" });\n\t\t\t} else {\n\t\t\t\tfirstTurn = false;\n\t\t\t}\n\n\t\t\t// Process pending messages (inject before next assistant response)\n\t\t\tif (pendingMessages.length > 0) {\n\t\t\t\tfor (const message of pendingMessages) {\n\t\t\t\t\tawait emit({ type: \"message_start\", message });\n\t\t\t\t\tawait emit({ type: \"message_end\", message });\n\t\t\t\t\tcurrentContext.messages.push(message);\n\t\t\t\t\tnewMessages.push(message);\n\t\t\t\t}\n\t\t\t\tpendingMessages = [];\n\t\t\t}\n\n\t\t\t// Stream assistant response\n\t\t\tconst message = await streamAssistantResponse(currentContext, config, signal, emit, streamFn);\n\t\t\tnewMessages.push(message);\n\n\t\t\tif (message.stopReason === \"error\" || message.stopReason === \"aborted\") {\n\t\t\t\tawait emit({ type: \"turn_end\", message, toolResults: [] });\n\t\t\t\tawait emit({ type: \"agent_end\", messages: newMessages });\n\t\t\t\treturn;\n\t\t\t}\n\n\t\t\t// Check for tool calls\n\t\t\tconst toolCalls = message.content.filter((c) => c.type === \"toolCall\");\n\t\t\thasMoreToolCalls = toolCalls.length > 0;\n\n\t\t\tconst toolResults: ToolResultMessage[] = [];\n\t\t\tif (hasMoreToolCalls) {\n\t\t\t\ttoolResults.push(...(await executeToolCalls(currentContext, message, config, signal, emit)));\n\n\t\t\t\tfor (const result of toolResults) {\n\t\t\t\t\tcurrentContext.messages.push(result);\n\t\t\t\t\tnewMessages.push(result);\n\t\t\t\t}\n\t\t\t}\n\n\t\t\tawait emit({ type: \"turn_end\", message, toolResults });\n\n\t\t\tpendingMessages = (await config.getSteeringMessages?.()) || [];\n\t\t}\n\n\t\t// Agent would stop here. Check for follow-up messages.\n\t\tconst followUpMessages = (await config.getFollowUpMessages?.()) || [];\n\t\tif (followUpMessages.length > 0) {\n\t\t\t// Set as pending so inner loop processes them\n\t\t\tpendingMessages = followUpMessages;\n\t\t\tcontinue;\n\t\t}\n\n\t\t// No more messages, exit\n\t\tbreak;\n\t}\n\n\tawait emit({ type: \"agent_end\", messages: newMessages });\n}\n\n/**\n * Stream an assistant response from the LLM.\n * This is where AgentMessage[] gets transformed to Message[] for the LLM.\n */\nasync function streamAssistantResponse(\n\tcontext: AgentContext,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n\tstreamFn?: StreamFn,\n): Promise<AssistantMessage> {\n\t// Apply context transform if configured (AgentMessage[] → AgentMessage[])\n\tlet messages = context.messages;\n\tif (config.transformContext) {\n\t\tmessages = await config.transformContext(messages, signal);\n\t}\n\n\t// Convert to LLM-compatible messages (AgentMessage[] → Message[])\n\tconst llmMessages = await config.convertToLlm(messages);\n\n\t// Build LLM context\n\tconst llmContext: Context = {\n\t\tsystemPrompt: context.systemPrompt,\n\t\tmessages: llmMessages,\n\t\ttools: context.tools,\n\t};\n\n\tconst streamFunction = streamFn || streamSimple;\n\n\t// Resolve API key (important for expiring tokens)\n\tconst resolvedApiKey =\n\t\t(config.getApiKey ? await config.getApiKey(config.model.provider) : undefined) || config.apiKey;\n\n\tconst response = await streamFunction(config.model, llmContext, {\n\t\t...config,\n\t\tapiKey: resolvedApiKey,\n\t\tsignal,\n\t});\n\n\tlet partialMessage: AssistantMessage | null = null;\n\tlet addedPartial = false;\n\n\tfor await (const event of response) {\n\t\tswitch (event.type) {\n\t\t\tcase \"start\":\n\t\t\t\tpartialMessage = event.partial;\n\t\t\t\tcontext.messages.push(partialMessage);\n\t\t\t\taddedPartial = true;\n\t\t\t\tawait emit({ type: \"message_start\", message: { ...partialMessage } });\n\t\t\t\tbreak;\n\n\t\t\tcase \"text_start\":\n\t\t\tcase \"text_delta\":\n\t\t\tcase \"text_end\":\n\t\t\tcase \"thinking_start\":\n\t\t\tcase \"thinking_delta\":\n\t\t\tcase \"thinking_end\":\n\t\t\tcase \"toolcall_start\":\n\t\t\tcase \"toolcall_delta\":\n\t\t\tcase \"toolcall_end\":\n\t\t\t\tif (partialMessage) {\n\t\t\t\t\tpartialMessage = event.partial;\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = partialMessage;\n\t\t\t\t\tawait emit({\n\t\t\t\t\t\ttype: \"message_update\",\n\t\t\t\t\t\tassistantMessageEvent: event,\n\t\t\t\t\t\tmessage: { ...partialMessage },\n\t\t\t\t\t});\n\t\t\t\t}\n\t\t\t\tbreak;\n\n\t\t\tcase \"done\":\n\t\t\tcase \"error\": {\n\t\t\t\tconst finalMessage = await response.result();\n\t\t\t\tif (addedPartial) {\n\t\t\t\t\tcontext.messages[context.messages.length - 1] = finalMessage;\n\t\t\t\t} else {\n\t\t\t\t\tcontext.messages.push(finalMessage);\n\t\t\t\t}\n\t\t\t\tif (!addedPartial) {\n\t\t\t\t\tawait emit({ type: \"message_start\", message: { ...finalMessage } });\n\t\t\t\t}\n\t\t\t\tawait emit({ type: \"message_end\", message: finalMessage });\n\t\t\t\treturn finalMessage;\n\t\t\t}\n\t\t}\n\t}\n\n\tconst finalMessage = await response.result();\n\tif (addedPartial) {\n\t\tcontext.messages[context.messages.length - 1] = finalMessage;\n\t} else {\n\t\tcontext.messages.push(finalMessage);\n\t\tawait emit({ type: \"message_start\", message: { ...finalMessage } });\n\t}\n\tawait emit({ type: \"message_end\", message: finalMessage });\n\treturn finalMessage;\n}\n\n/**\n * Execute tool calls from an assistant message.\n */\nasync function executeToolCalls(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage[]> {\n\tconst toolCalls = assistantMessage.content.filter((c) => c.type === \"toolCall\");\n\tif (config.toolExecution === \"sequential\") {\n\t\treturn executeToolCallsSequential(currentContext, assistantMessage, toolCalls, config, signal, emit);\n\t}\n\treturn executeToolCallsParallel(currentContext, assistantMessage, toolCalls, config, signal, emit);\n}\n\nasync function executeToolCallsSequential(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\ttoolCalls: AgentToolCall[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage[]> {\n\tconst results: ToolResultMessage[] = [];\n\n\tfor (const toolCall of toolCalls) {\n\t\tawait emit({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\n\t\tconst preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);\n\t\tif (preparation.kind === \"immediate\") {\n\t\t\tresults.push(await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit));\n\t\t} else {\n\t\t\tconst executed = await executePreparedToolCall(preparation, signal, emit);\n\t\t\tresults.push(\n\t\t\t\tawait finalizeExecutedToolCall(\n\t\t\t\t\tcurrentContext,\n\t\t\t\t\tassistantMessage,\n\t\t\t\t\tpreparation,\n\t\t\t\t\texecuted,\n\t\t\t\t\tconfig,\n\t\t\t\t\tsignal,\n\t\t\t\t\temit,\n\t\t\t\t),\n\t\t\t);\n\t\t}\n\t}\n\n\treturn results;\n}\n\nasync function executeToolCallsParallel(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\ttoolCalls: AgentToolCall[],\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage[]> {\n\tconst results: (ToolResultMessage | null)[] = new Array(toolCalls.length).fill(null);\n\tconst runnableCalls: { index: number; prepared: PreparedToolCall }[] = [];\n\n\tfor (let i = 0; i < toolCalls.length; i++) {\n\t\tconst toolCall = toolCalls[i];\n\t\tawait emit({\n\t\t\ttype: \"tool_execution_start\",\n\t\t\ttoolCallId: toolCall.id,\n\t\t\ttoolName: toolCall.name,\n\t\t\targs: toolCall.arguments,\n\t\t});\n\n\t\tconst preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);\n\t\tif (preparation.kind === \"immediate\") {\n\t\t\tresults[i] = await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit);\n\t\t} else {\n\t\t\trunnableCalls.push({ index: i, prepared: preparation });\n\t\t}\n\t}\n\n\tconst runningCalls = runnableCalls.map((entry) => ({\n\t\tindex: entry.index,\n\t\tprepared: entry.prepared,\n\t\texecution: executePreparedToolCall(entry.prepared, signal, emit),\n\t}));\n\n\tfor (const running of runningCalls) {\n\t\tconst executed = await running.execution;\n\t\tresults[running.index] = await finalizeExecutedToolCall(\n\t\t\tcurrentContext,\n\t\t\tassistantMessage,\n\t\t\trunning.prepared,\n\t\t\texecuted,\n\t\t\tconfig,\n\t\t\tsignal,\n\t\t\temit,\n\t\t);\n\t}\n\n\treturn results.filter((r): r is ToolResultMessage => r !== null);\n}\n\ntype PreparedToolCall = {\n\tkind: \"prepared\";\n\ttoolCall: AgentToolCall;\n\ttool: AgentTool<any>;\n\targs: unknown;\n};\n\ntype ImmediateToolCallOutcome = {\n\tkind: \"immediate\";\n\tresult: AgentToolResult<any>;\n\tisError: boolean;\n};\n\ntype ExecutedToolCallOutcome = {\n\tresult: AgentToolResult<any>;\n\tisError: boolean;\n};\n\nasync function prepareToolCall(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\ttoolCall: AgentToolCall,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n): Promise<PreparedToolCall | ImmediateToolCallOutcome> {\n\tconst tool = currentContext.tools?.find((t) => t.name === toolCall.name);\n\tif (!tool) {\n\t\treturn {\n\t\t\tkind: \"immediate\",\n\t\t\tresult: createErrorToolResult(`Tool ${toolCall.name} not found`),\n\t\t\tisError: true,\n\t\t};\n\t}\n\n\ttry {\n\t\tconst validatedArgs = validateToolArguments(tool, toolCall);\n\t\tif (config.beforeToolCall) {\n\t\t\tconst beforeResult = await config.beforeToolCall(\n\t\t\t\t{\n\t\t\t\t\tassistantMessage,\n\t\t\t\t\ttoolCall,\n\t\t\t\t\targs: validatedArgs,\n\t\t\t\t\tcontext: currentContext,\n\t\t\t\t},\n\t\t\t\tsignal,\n\t\t\t);\n\t\t\tif (beforeResult?.block) {\n\t\t\t\treturn {\n\t\t\t\t\tkind: \"immediate\",\n\t\t\t\t\tresult: createErrorToolResult(beforeResult.reason || \"Tool execution was blocked\"),\n\t\t\t\t\tisError: true,\n\t\t\t\t};\n\t\t\t}\n\t\t}\n\t\treturn {\n\t\t\tkind: \"prepared\",\n\t\t\ttoolCall,\n\t\t\ttool,\n\t\t\targs: validatedArgs,\n\t\t};\n\t} catch (error) {\n\t\treturn {\n\t\t\tkind: \"immediate\",\n\t\t\tresult: createErrorToolResult(error instanceof Error ? error.message : String(error)),\n\t\t\tisError: true,\n\t\t};\n\t}\n}\n\nasync function executePreparedToolCall(\n\tprepared: PreparedToolCall,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ExecutedToolCallOutcome> {\n\tconst updateEvents: Promise<void>[] = [];\n\n\ttry {\n\t\tconst result = await prepared.tool.execute(\n\t\t\tprepared.toolCall.id,\n\t\t\tprepared.args as never,\n\t\t\tsignal,\n\t\t\t(partialResult) => {\n\t\t\t\tupdateEvents.push(\n\t\t\t\t\tPromise.resolve(\n\t\t\t\t\t\temit({\n\t\t\t\t\t\t\ttype: \"tool_execution_update\",\n\t\t\t\t\t\t\ttoolCallId: prepared.toolCall.id,\n\t\t\t\t\t\t\ttoolName: prepared.toolCall.name,\n\t\t\t\t\t\t\targs: prepared.toolCall.arguments,\n\t\t\t\t\t\t\tpartialResult,\n\t\t\t\t\t\t}),\n\t\t\t\t\t),\n\t\t\t\t);\n\t\t\t},\n\t\t);\n\t\tawait Promise.all(updateEvents);\n\t\treturn { result, isError: false };\n\t} catch (error) {\n\t\tawait Promise.all(updateEvents);\n\t\treturn {\n\t\t\tresult: createErrorToolResult(error instanceof Error ? error.message : String(error)),\n\t\t\tisError: true,\n\t\t};\n\t}\n}\n\nasync function finalizeExecutedToolCall(\n\tcurrentContext: AgentContext,\n\tassistantMessage: AssistantMessage,\n\tprepared: PreparedToolCall,\n\texecuted: ExecutedToolCallOutcome,\n\tconfig: AgentLoopConfig,\n\tsignal: AbortSignal | undefined,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage> {\n\tlet result = executed.result;\n\tlet isError = executed.isError;\n\n\tif (config.afterToolCall) {\n\t\tconst afterResult = await config.afterToolCall(\n\t\t\t{\n\t\t\t\tassistantMessage,\n\t\t\t\ttoolCall: prepared.toolCall,\n\t\t\t\targs: prepared.args,\n\t\t\t\tresult,\n\t\t\t\tisError,\n\t\t\t\tcontext: currentContext,\n\t\t\t},\n\t\t\tsignal,\n\t\t);\n\t\tif (afterResult) {\n\t\t\tresult = {\n\t\t\t\tcontent: afterResult.content ?? result.content,\n\t\t\t\tdetails: afterResult.details ?? result.details,\n\t\t\t};\n\t\t\tisError = afterResult.isError ?? isError;\n\t\t}\n\t}\n\n\treturn await emitToolCallOutcome(prepared.toolCall, result, isError, emit);\n}\n\nfunction createErrorToolResult(message: string): AgentToolResult<any> {\n\treturn {\n\t\tcontent: [{ type: \"text\", text: message }],\n\t\tdetails: {},\n\t};\n}\n\nasync function emitToolCallOutcome(\n\ttoolCall: AgentToolCall,\n\tresult: AgentToolResult<any>,\n\tisError: boolean,\n\temit: AgentEventSink,\n): Promise<ToolResultMessage> {\n\tawait emit({\n\t\ttype: \"tool_execution_end\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tresult,\n\t\tisError,\n\t});\n\n\tconst toolResultMessage: ToolResultMessage = {\n\t\trole: \"toolResult\",\n\t\ttoolCallId: toolCall.id,\n\t\ttoolName: toolCall.name,\n\t\tcontent: result.content,\n\t\tdetails: result.details,\n\t\tisError,\n\t\ttimestamp: Date.now(),\n\t};\n\n\tawait emit({ type: \"message_start\", message: toolResultMessage });\n\tawait emit({ type: \"message_end\", message: toolResultMessage });\n\treturn toolResultMessage;\n}\n"]}
@@ -9,20 +9,15 @@ import { EventStream, streamSimple, validateToolArguments, } from "@draht/ai";
9
9
  */
10
10
  export function agentLoop(prompts, context, config, signal, streamFn) {
11
11
  const stream = createAgentStream();
12
- (async () => {
13
- const newMessages = [...prompts];
14
- const currentContext = {
15
- ...context,
16
- messages: [...context.messages, ...prompts],
17
- };
18
- stream.push({ type: "agent_start" });
19
- stream.push({ type: "turn_start" });
20
- for (const prompt of prompts) {
21
- stream.push({ type: "message_start", message: prompt });
22
- stream.push({ type: "message_end", message: prompt });
23
- }
24
- await runLoop(currentContext, newMessages, config, signal, stream, streamFn);
25
- })();
12
+ void runAgentLoop(prompts, context, config, async (event) => {
13
+ stream.push(event);
14
+ }, signal, streamFn)
15
+ .then((messages) => {
16
+ stream.end(messages);
17
+ })
18
+ .catch(() => {
19
+ stream.end([]);
20
+ });
26
21
  return stream;
27
22
  }
28
23
  /**
@@ -41,33 +36,63 @@ export function agentLoopContinue(context, config, signal, streamFn) {
41
36
  throw new Error("Cannot continue from message role: assistant");
42
37
  }
43
38
  const stream = createAgentStream();
44
- (async () => {
45
- const newMessages = [];
46
- const currentContext = { ...context };
47
- stream.push({ type: "agent_start" });
48
- stream.push({ type: "turn_start" });
49
- await runLoop(currentContext, newMessages, config, signal, stream, streamFn);
50
- })();
39
+ void runAgentLoopContinue(context, config, async (event) => {
40
+ stream.push(event);
41
+ }, signal, streamFn)
42
+ .then((messages) => {
43
+ stream.end(messages);
44
+ })
45
+ .catch(() => {
46
+ stream.end([]);
47
+ });
51
48
  return stream;
52
49
  }
50
+ export async function runAgentLoop(prompts, context, config, emit, signal, streamFn) {
51
+ const newMessages = [...prompts];
52
+ const currentContext = {
53
+ ...context,
54
+ messages: [...context.messages, ...prompts],
55
+ };
56
+ await emit({ type: "agent_start" });
57
+ await emit({ type: "turn_start" });
58
+ for (const prompt of prompts) {
59
+ await emit({ type: "message_start", message: prompt });
60
+ await emit({ type: "message_end", message: prompt });
61
+ }
62
+ await runLoop(currentContext, newMessages, config, signal, emit, streamFn);
63
+ return newMessages;
64
+ }
65
+ export async function runAgentLoopContinue(context, config, emit, signal, streamFn) {
66
+ if (context.messages.length === 0) {
67
+ throw new Error("Cannot continue: no messages in context");
68
+ }
69
+ if (context.messages[context.messages.length - 1].role === "assistant") {
70
+ throw new Error("Cannot continue from message role: assistant");
71
+ }
72
+ const newMessages = [];
73
+ const currentContext = { ...context };
74
+ await emit({ type: "agent_start" });
75
+ await emit({ type: "turn_start" });
76
+ await runLoop(currentContext, newMessages, config, signal, emit, streamFn);
77
+ return newMessages;
78
+ }
53
79
  function createAgentStream() {
54
80
  return new EventStream((event) => event.type === "agent_end", (event) => (event.type === "agent_end" ? event.messages : []));
55
81
  }
56
82
  /**
57
83
  * Main loop logic shared by agentLoop and agentLoopContinue.
58
84
  */
59
- async function runLoop(currentContext, newMessages, config, signal, stream, streamFn) {
85
+ async function runLoop(currentContext, newMessages, config, signal, emit, streamFn) {
60
86
  let firstTurn = true;
61
87
  // Check for steering messages at start (user may have typed while waiting)
62
88
  let pendingMessages = (await config.getSteeringMessages?.()) || [];
63
89
  // Outer loop: continues when queued follow-up messages arrive after agent would stop
64
90
  while (true) {
65
91
  let hasMoreToolCalls = true;
66
- let steeringAfterTools = null;
67
92
  // Inner loop: process tool calls and steering messages
68
93
  while (hasMoreToolCalls || pendingMessages.length > 0) {
69
94
  if (!firstTurn) {
70
- stream.push({ type: "turn_start" });
95
+ await emit({ type: "turn_start" });
71
96
  }
72
97
  else {
73
98
  firstTurn = false;
@@ -75,20 +100,19 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
75
100
  // Process pending messages (inject before next assistant response)
76
101
  if (pendingMessages.length > 0) {
77
102
  for (const message of pendingMessages) {
78
- stream.push({ type: "message_start", message });
79
- stream.push({ type: "message_end", message });
103
+ await emit({ type: "message_start", message });
104
+ await emit({ type: "message_end", message });
80
105
  currentContext.messages.push(message);
81
106
  newMessages.push(message);
82
107
  }
83
108
  pendingMessages = [];
84
109
  }
85
110
  // Stream assistant response
86
- const message = await streamAssistantResponse(currentContext, config, signal, stream, streamFn);
111
+ const message = await streamAssistantResponse(currentContext, config, signal, emit, streamFn);
87
112
  newMessages.push(message);
88
113
  if (message.stopReason === "error" || message.stopReason === "aborted") {
89
- stream.push({ type: "turn_end", message, toolResults: [] });
90
- stream.push({ type: "agent_end", messages: newMessages });
91
- stream.end(newMessages);
114
+ await emit({ type: "turn_end", message, toolResults: [] });
115
+ await emit({ type: "agent_end", messages: newMessages });
92
116
  return;
93
117
  }
94
118
  // Check for tool calls
@@ -96,23 +120,14 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
96
120
  hasMoreToolCalls = toolCalls.length > 0;
97
121
  const toolResults = [];
98
122
  if (hasMoreToolCalls) {
99
- const toolExecution = await executeToolCalls(currentContext.tools, message, signal, stream, config.getSteeringMessages);
100
- toolResults.push(...toolExecution.toolResults);
101
- steeringAfterTools = toolExecution.steeringMessages ?? null;
123
+ toolResults.push(...(await executeToolCalls(currentContext, message, config, signal, emit)));
102
124
  for (const result of toolResults) {
103
125
  currentContext.messages.push(result);
104
126
  newMessages.push(result);
105
127
  }
106
128
  }
107
- stream.push({ type: "turn_end", message, toolResults });
108
- // Get steering messages after turn completes
109
- if (steeringAfterTools && steeringAfterTools.length > 0) {
110
- pendingMessages = steeringAfterTools;
111
- steeringAfterTools = null;
112
- }
113
- else {
114
- pendingMessages = (await config.getSteeringMessages?.()) || [];
115
- }
129
+ await emit({ type: "turn_end", message, toolResults });
130
+ pendingMessages = (await config.getSteeringMessages?.()) || [];
116
131
  }
117
132
  // Agent would stop here. Check for follow-up messages.
118
133
  const followUpMessages = (await config.getFollowUpMessages?.()) || [];
@@ -124,14 +139,13 @@ async function runLoop(currentContext, newMessages, config, signal, stream, stre
124
139
  // No more messages, exit
125
140
  break;
126
141
  }
127
- stream.push({ type: "agent_end", messages: newMessages });
128
- stream.end(newMessages);
142
+ await emit({ type: "agent_end", messages: newMessages });
129
143
  }
130
144
  /**
131
145
  * Stream an assistant response from the LLM.
132
146
  * This is where AgentMessage[] gets transformed to Message[] for the LLM.
133
147
  */
134
- async function streamAssistantResponse(context, config, signal, stream, streamFn) {
148
+ async function streamAssistantResponse(context, config, signal, emit, streamFn) {
135
149
  // Apply context transform if configured (AgentMessage[] → AgentMessage[])
136
150
  let messages = context.messages;
137
151
  if (config.transformContext) {
@@ -161,7 +175,7 @@ async function streamAssistantResponse(context, config, signal, stream, streamFn
161
175
  partialMessage = event.partial;
162
176
  context.messages.push(partialMessage);
163
177
  addedPartial = true;
164
- stream.push({ type: "message_start", message: { ...partialMessage } });
178
+ await emit({ type: "message_start", message: { ...partialMessage } });
165
179
  break;
166
180
  case "text_start":
167
181
  case "text_delta":
@@ -175,7 +189,7 @@ async function streamAssistantResponse(context, config, signal, stream, streamFn
175
189
  if (partialMessage) {
176
190
  partialMessage = event.partial;
177
191
  context.messages[context.messages.length - 1] = partialMessage;
178
- stream.push({
192
+ await emit({
179
193
  type: "message_update",
180
194
  assistantMessageEvent: event,
181
195
  message: { ...partialMessage },
@@ -192,117 +206,195 @@ async function streamAssistantResponse(context, config, signal, stream, streamFn
192
206
  context.messages.push(finalMessage);
193
207
  }
194
208
  if (!addedPartial) {
195
- stream.push({ type: "message_start", message: { ...finalMessage } });
209
+ await emit({ type: "message_start", message: { ...finalMessage } });
196
210
  }
197
- stream.push({ type: "message_end", message: finalMessage });
211
+ await emit({ type: "message_end", message: finalMessage });
198
212
  return finalMessage;
199
213
  }
200
214
  }
201
215
  }
202
- return await response.result();
216
+ const finalMessage = await response.result();
217
+ if (addedPartial) {
218
+ context.messages[context.messages.length - 1] = finalMessage;
219
+ }
220
+ else {
221
+ context.messages.push(finalMessage);
222
+ await emit({ type: "message_start", message: { ...finalMessage } });
223
+ }
224
+ await emit({ type: "message_end", message: finalMessage });
225
+ return finalMessage;
203
226
  }
204
227
  /**
205
228
  * Execute tool calls from an assistant message.
206
229
  */
207
- async function executeToolCalls(tools, assistantMessage, signal, stream, getSteeringMessages) {
230
+ async function executeToolCalls(currentContext, assistantMessage, config, signal, emit) {
208
231
  const toolCalls = assistantMessage.content.filter((c) => c.type === "toolCall");
232
+ if (config.toolExecution === "sequential") {
233
+ return executeToolCallsSequential(currentContext, assistantMessage, toolCalls, config, signal, emit);
234
+ }
235
+ return executeToolCallsParallel(currentContext, assistantMessage, toolCalls, config, signal, emit);
236
+ }
237
+ async function executeToolCallsSequential(currentContext, assistantMessage, toolCalls, config, signal, emit) {
209
238
  const results = [];
210
- let steeringMessages;
211
- for (let index = 0; index < toolCalls.length; index++) {
212
- const toolCall = toolCalls[index];
213
- const tool = tools?.find((t) => t.name === toolCall.name);
214
- stream.push({
239
+ for (const toolCall of toolCalls) {
240
+ await emit({
215
241
  type: "tool_execution_start",
216
242
  toolCallId: toolCall.id,
217
243
  toolName: toolCall.name,
218
244
  args: toolCall.arguments,
219
245
  });
220
- let result;
221
- let isError = false;
222
- try {
223
- if (!tool)
224
- throw new Error(`Tool ${toolCall.name} not found`);
225
- const validatedArgs = validateToolArguments(tool, toolCall);
226
- result = await tool.execute(toolCall.id, validatedArgs, signal, (partialResult) => {
227
- stream.push({
228
- type: "tool_execution_update",
229
- toolCallId: toolCall.id,
230
- toolName: toolCall.name,
231
- args: toolCall.arguments,
232
- partialResult,
233
- });
234
- });
246
+ const preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);
247
+ if (preparation.kind === "immediate") {
248
+ results.push(await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit));
235
249
  }
236
- catch (e) {
237
- result = {
238
- content: [{ type: "text", text: e instanceof Error ? e.message : String(e) }],
239
- details: {},
240
- };
241
- isError = true;
250
+ else {
251
+ const executed = await executePreparedToolCall(preparation, signal, emit);
252
+ results.push(await finalizeExecutedToolCall(currentContext, assistantMessage, preparation, executed, config, signal, emit));
242
253
  }
243
- stream.push({
244
- type: "tool_execution_end",
254
+ }
255
+ return results;
256
+ }
257
+ async function executeToolCallsParallel(currentContext, assistantMessage, toolCalls, config, signal, emit) {
258
+ const results = new Array(toolCalls.length).fill(null);
259
+ const runnableCalls = [];
260
+ for (let i = 0; i < toolCalls.length; i++) {
261
+ const toolCall = toolCalls[i];
262
+ await emit({
263
+ type: "tool_execution_start",
245
264
  toolCallId: toolCall.id,
246
265
  toolName: toolCall.name,
247
- result,
248
- isError,
266
+ args: toolCall.arguments,
249
267
  });
250
- const toolResultMessage = {
251
- role: "toolResult",
252
- toolCallId: toolCall.id,
253
- toolName: toolCall.name,
254
- content: result.content,
255
- details: result.details,
256
- isError,
257
- timestamp: Date.now(),
268
+ const preparation = await prepareToolCall(currentContext, assistantMessage, toolCall, config, signal);
269
+ if (preparation.kind === "immediate") {
270
+ results[i] = await emitToolCallOutcome(toolCall, preparation.result, preparation.isError, emit);
271
+ }
272
+ else {
273
+ runnableCalls.push({ index: i, prepared: preparation });
274
+ }
275
+ }
276
+ const runningCalls = runnableCalls.map((entry) => ({
277
+ index: entry.index,
278
+ prepared: entry.prepared,
279
+ execution: executePreparedToolCall(entry.prepared, signal, emit),
280
+ }));
281
+ for (const running of runningCalls) {
282
+ const executed = await running.execution;
283
+ results[running.index] = await finalizeExecutedToolCall(currentContext, assistantMessage, running.prepared, executed, config, signal, emit);
284
+ }
285
+ return results.filter((r) => r !== null);
286
+ }
287
+ async function prepareToolCall(currentContext, assistantMessage, toolCall, config, signal) {
288
+ const tool = currentContext.tools?.find((t) => t.name === toolCall.name);
289
+ if (!tool) {
290
+ return {
291
+ kind: "immediate",
292
+ result: createErrorToolResult(`Tool ${toolCall.name} not found`),
293
+ isError: true,
258
294
  };
259
- results.push(toolResultMessage);
260
- stream.push({ type: "message_start", message: toolResultMessage });
261
- stream.push({ type: "message_end", message: toolResultMessage });
262
- // Check for steering messages - skip remaining tools if user interrupted
263
- if (getSteeringMessages) {
264
- const steering = await getSteeringMessages();
265
- if (steering.length > 0) {
266
- steeringMessages = steering;
267
- const remainingCalls = toolCalls.slice(index + 1);
268
- for (const skipped of remainingCalls) {
269
- results.push(skipToolCall(skipped, stream));
270
- }
271
- break;
295
+ }
296
+ try {
297
+ const validatedArgs = validateToolArguments(tool, toolCall);
298
+ if (config.beforeToolCall) {
299
+ const beforeResult = await config.beforeToolCall({
300
+ assistantMessage,
301
+ toolCall,
302
+ args: validatedArgs,
303
+ context: currentContext,
304
+ }, signal);
305
+ if (beforeResult?.block) {
306
+ return {
307
+ kind: "immediate",
308
+ result: createErrorToolResult(beforeResult.reason || "Tool execution was blocked"),
309
+ isError: true,
310
+ };
272
311
  }
273
312
  }
313
+ return {
314
+ kind: "prepared",
315
+ toolCall,
316
+ tool,
317
+ args: validatedArgs,
318
+ };
319
+ }
320
+ catch (error) {
321
+ return {
322
+ kind: "immediate",
323
+ result: createErrorToolResult(error instanceof Error ? error.message : String(error)),
324
+ isError: true,
325
+ };
326
+ }
327
+ }
328
+ async function executePreparedToolCall(prepared, signal, emit) {
329
+ const updateEvents = [];
330
+ try {
331
+ const result = await prepared.tool.execute(prepared.toolCall.id, prepared.args, signal, (partialResult) => {
332
+ updateEvents.push(Promise.resolve(emit({
333
+ type: "tool_execution_update",
334
+ toolCallId: prepared.toolCall.id,
335
+ toolName: prepared.toolCall.name,
336
+ args: prepared.toolCall.arguments,
337
+ partialResult,
338
+ })));
339
+ });
340
+ await Promise.all(updateEvents);
341
+ return { result, isError: false };
342
+ }
343
+ catch (error) {
344
+ await Promise.all(updateEvents);
345
+ return {
346
+ result: createErrorToolResult(error instanceof Error ? error.message : String(error)),
347
+ isError: true,
348
+ };
349
+ }
350
+ }
351
+ async function finalizeExecutedToolCall(currentContext, assistantMessage, prepared, executed, config, signal, emit) {
352
+ let result = executed.result;
353
+ let isError = executed.isError;
354
+ if (config.afterToolCall) {
355
+ const afterResult = await config.afterToolCall({
356
+ assistantMessage,
357
+ toolCall: prepared.toolCall,
358
+ args: prepared.args,
359
+ result,
360
+ isError,
361
+ context: currentContext,
362
+ }, signal);
363
+ if (afterResult) {
364
+ result = {
365
+ content: afterResult.content ?? result.content,
366
+ details: afterResult.details ?? result.details,
367
+ };
368
+ isError = afterResult.isError ?? isError;
369
+ }
274
370
  }
275
- return { toolResults: results, steeringMessages };
371
+ return await emitToolCallOutcome(prepared.toolCall, result, isError, emit);
276
372
  }
277
- function skipToolCall(toolCall, stream) {
278
- const result = {
279
- content: [{ type: "text", text: "Skipped due to queued user message." }],
373
+ function createErrorToolResult(message) {
374
+ return {
375
+ content: [{ type: "text", text: message }],
280
376
  details: {},
281
377
  };
282
- stream.push({
283
- type: "tool_execution_start",
284
- toolCallId: toolCall.id,
285
- toolName: toolCall.name,
286
- args: toolCall.arguments,
287
- });
288
- stream.push({
378
+ }
379
+ async function emitToolCallOutcome(toolCall, result, isError, emit) {
380
+ await emit({
289
381
  type: "tool_execution_end",
290
382
  toolCallId: toolCall.id,
291
383
  toolName: toolCall.name,
292
384
  result,
293
- isError: true,
385
+ isError,
294
386
  });
295
387
  const toolResultMessage = {
296
388
  role: "toolResult",
297
389
  toolCallId: toolCall.id,
298
390
  toolName: toolCall.name,
299
391
  content: result.content,
300
- details: {},
301
- isError: true,
392
+ details: result.details,
393
+ isError,
302
394
  timestamp: Date.now(),
303
395
  };
304
- stream.push({ type: "message_start", message: toolResultMessage });
305
- stream.push({ type: "message_end", message: toolResultMessage });
396
+ await emit({ type: "message_start", message: toolResultMessage });
397
+ await emit({ type: "message_end", message: toolResultMessage });
306
398
  return toolResultMessage;
307
399
  }
308
400
  //# sourceMappingURL=agent-loop.js.map