@kenkaiiii/gg-agent 4.2.2 → 4.2.3

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs ADDED
@@ -0,0 +1,370 @@
1
+ "use strict";
2
+ var __defProp = Object.defineProperty;
3
+ var __getOwnPropDesc = Object.getOwnPropertyDescriptor;
4
+ var __getOwnPropNames = Object.getOwnPropertyNames;
5
+ var __hasOwnProp = Object.prototype.hasOwnProperty;
6
+ var __export = (target, all) => {
7
+ for (var name in all)
8
+ __defProp(target, name, { get: all[name], enumerable: true });
9
+ };
10
+ var __copyProps = (to, from, except, desc) => {
11
+ if (from && typeof from === "object" || typeof from === "function") {
12
+ for (let key of __getOwnPropNames(from))
13
+ if (!__hasOwnProp.call(to, key) && key !== except)
14
+ __defProp(to, key, { get: () => from[key], enumerable: !(desc = __getOwnPropDesc(from, key)) || desc.enumerable });
15
+ }
16
+ return to;
17
+ };
18
+ var __toCommonJS = (mod) => __copyProps(__defProp({}, "__esModule", { value: true }), mod);
19
+
20
+ // src/index.ts
21
+ var index_exports = {};
22
+ __export(index_exports, {
23
+ Agent: () => Agent,
24
+ AgentStream: () => AgentStream,
25
+ agentLoop: () => agentLoop,
26
+ isContextOverflow: () => isContextOverflow
27
+ });
28
+ module.exports = __toCommonJS(index_exports);
29
+
30
+ // src/agent.ts
31
+ var import_gg_ai2 = require("@kenkaiiii/gg-ai");
32
+
33
+ // src/agent-loop.ts
34
+ var import_gg_ai = require("@kenkaiiii/gg-ai");
35
+ var DEFAULT_MAX_TURNS = 100;
36
+ function isContextOverflow(err) {
37
+ if (!(err instanceof Error)) return false;
38
+ const msg = err.message.toLowerCase();
39
+ return msg.includes("prompt is too long") || msg.includes("context_length_exceeded") || msg.includes("maximum context length") || msg.includes("token") && msg.includes("exceed");
40
+ }
41
+ function isOverloaded(err) {
42
+ if (!(err instanceof Error)) return false;
43
+ const msg = err.message.toLowerCase();
44
+ return msg.includes("overloaded") || msg.includes("rate limit") || msg.includes("too many requests") || msg.includes("429") || msg.includes("529");
45
+ }
46
+ async function* agentLoop(messages, options) {
47
+ const maxTurns = options.maxTurns ?? DEFAULT_MAX_TURNS;
48
+ const maxContinuations = options.maxContinuations ?? 5;
49
+ const toolMap = new Map((options.tools ?? []).map((t) => [t.name, t]));
50
+ const totalUsage = { inputTokens: 0, outputTokens: 0 };
51
+ let turn = 0;
52
+ let consecutivePauses = 0;
53
+ let overflowRetries = 0;
54
+ let overloadRetries = 0;
55
+ const MAX_OVERFLOW_RETRIES = 3;
56
+ const MAX_OVERLOAD_RETRIES = 3;
57
+ const OVERLOAD_RETRY_DELAY_MS = 3e3;
58
+ while (turn < maxTurns) {
59
+ options.signal?.throwIfAborted();
60
+ turn++;
61
+ if (options.transformContext) {
62
+ const transformed = await options.transformContext(messages);
63
+ if (transformed !== messages) {
64
+ messages.length = 0;
65
+ messages.push(...transformed);
66
+ }
67
+ }
68
+ let response;
69
+ try {
70
+ const result = (0, import_gg_ai.stream)({
71
+ provider: options.provider,
72
+ model: options.model,
73
+ messages,
74
+ tools: options.tools,
75
+ serverTools: options.serverTools,
76
+ webSearch: options.webSearch,
77
+ maxTokens: options.maxTokens,
78
+ temperature: options.temperature,
79
+ thinking: options.thinking,
80
+ apiKey: options.apiKey,
81
+ baseUrl: options.baseUrl,
82
+ signal: options.signal,
83
+ accountId: options.accountId,
84
+ cacheRetention: options.cacheRetention,
85
+ compaction: options.compaction
86
+ });
87
+ result.response.catch(() => {
88
+ });
89
+ for await (const event of result) {
90
+ if (event.type === "text_delta") {
91
+ yield { type: "text_delta", text: event.text };
92
+ } else if (event.type === "thinking_delta") {
93
+ yield { type: "thinking_delta", text: event.text };
94
+ } else if (event.type === "server_toolcall") {
95
+ yield {
96
+ type: "server_tool_call",
97
+ id: event.id,
98
+ name: event.name,
99
+ input: event.input
100
+ };
101
+ } else if (event.type === "server_toolresult") {
102
+ yield {
103
+ type: "server_tool_result",
104
+ toolUseId: event.toolUseId,
105
+ resultType: event.resultType,
106
+ data: event.data
107
+ };
108
+ }
109
+ }
110
+ response = await result.response;
111
+ } catch (err) {
112
+ if (overflowRetries < MAX_OVERFLOW_RETRIES && isContextOverflow(err) && options.transformContext) {
113
+ overflowRetries++;
114
+ const transformed = await options.transformContext(messages, { force: true });
115
+ if (transformed !== messages) {
116
+ messages.length = 0;
117
+ messages.push(...transformed);
118
+ }
119
+ turn--;
120
+ continue;
121
+ }
122
+ if (overloadRetries < MAX_OVERLOAD_RETRIES && isOverloaded(err)) {
123
+ overloadRetries++;
124
+ await new Promise((r) => setTimeout(r, OVERLOAD_RETRY_DELAY_MS));
125
+ turn--;
126
+ continue;
127
+ }
128
+ throw err;
129
+ }
130
+ overflowRetries = 0;
131
+ overloadRetries = 0;
132
+ totalUsage.inputTokens += response.usage.inputTokens;
133
+ totalUsage.outputTokens += response.usage.outputTokens;
134
+ if (response.usage.cacheRead) {
135
+ totalUsage.cacheRead = (totalUsage.cacheRead ?? 0) + response.usage.cacheRead;
136
+ }
137
+ if (response.usage.cacheWrite) {
138
+ totalUsage.cacheWrite = (totalUsage.cacheWrite ?? 0) + response.usage.cacheWrite;
139
+ }
140
+ messages.push(response.message);
141
+ yield {
142
+ type: "turn_end",
143
+ turn,
144
+ stopReason: response.stopReason,
145
+ usage: response.usage
146
+ };
147
+ if (response.stopReason === "pause_turn") {
148
+ consecutivePauses++;
149
+ if (consecutivePauses >= maxContinuations) {
150
+ break;
151
+ }
152
+ continue;
153
+ }
154
+ consecutivePauses = 0;
155
+ if (response.stopReason !== "tool_use") {
156
+ yield {
157
+ type: "agent_done",
158
+ totalTurns: turn,
159
+ totalUsage: { ...totalUsage }
160
+ };
161
+ return {
162
+ message: response.message,
163
+ totalTurns: turn,
164
+ totalUsage: { ...totalUsage }
165
+ };
166
+ }
167
+ const allToolCalls = extractToolCalls(response.message.content);
168
+ const toolCalls = [];
169
+ const toolResults = [];
170
+ for (const tc of allToolCalls) {
171
+ if (tc.name.startsWith("$")) {
172
+ toolResults.push({
173
+ type: "tool_result",
174
+ toolCallId: tc.id,
175
+ content: JSON.stringify(tc.args)
176
+ });
177
+ } else {
178
+ toolCalls.push(tc);
179
+ }
180
+ }
181
+ const eventStream = new import_gg_ai.EventStream();
182
+ const executions = toolCalls.map(async (toolCall) => {
183
+ const startTime = Date.now();
184
+ eventStream.push({
185
+ type: "tool_call_start",
186
+ toolCallId: toolCall.id,
187
+ name: toolCall.name,
188
+ args: toolCall.args
189
+ });
190
+ let resultContent;
191
+ let details;
192
+ let isError = false;
193
+ const tool = toolMap.get(toolCall.name);
194
+ if (!tool) {
195
+ resultContent = `Unknown tool: ${toolCall.name}`;
196
+ isError = true;
197
+ } else {
198
+ try {
199
+ const parsed = tool.parameters.parse(toolCall.args);
200
+ const ctx = {
201
+ signal: options.signal ?? AbortSignal.timeout(3e5),
202
+ toolCallId: toolCall.id,
203
+ onUpdate: (update) => {
204
+ eventStream.push({
205
+ type: "tool_call_update",
206
+ toolCallId: toolCall.id,
207
+ update
208
+ });
209
+ }
210
+ };
211
+ const raw = await tool.execute(parsed, ctx);
212
+ const normalized = normalizeToolResult(raw);
213
+ resultContent = normalized.content;
214
+ details = normalized.details;
215
+ } catch (err) {
216
+ isError = true;
217
+ resultContent = err instanceof Error ? err.message : String(err);
218
+ }
219
+ }
220
+ const durationMs = Date.now() - startTime;
221
+ eventStream.push({
222
+ type: "tool_call_end",
223
+ toolCallId: toolCall.id,
224
+ result: resultContent,
225
+ details,
226
+ isError,
227
+ durationMs
228
+ });
229
+ return { toolCallId: toolCall.id, content: resultContent, isError };
230
+ });
231
+ const abortHandler = () => eventStream.abort(new Error("aborted"));
232
+ options.signal?.addEventListener("abort", abortHandler, { once: true });
233
+ Promise.all(executions).then((results) => {
234
+ for (const tc of toolCalls) {
235
+ const r = results.find((x) => x.toolCallId === tc.id);
236
+ toolResults.push({
237
+ type: "tool_result",
238
+ toolCallId: tc.id,
239
+ content: r.content,
240
+ isError: r.isError || void 0
241
+ });
242
+ }
243
+ eventStream.close();
244
+ }).catch((err) => eventStream.abort(err instanceof Error ? err : new Error(String(err))));
245
+ try {
246
+ for await (const event of eventStream) {
247
+ yield event;
248
+ }
249
+ } finally {
250
+ options.signal?.removeEventListener("abort", abortHandler);
251
+ for (const tc of toolCalls) {
252
+ if (!toolResults.some((r) => r.toolCallId === tc.id)) {
253
+ toolResults.push({
254
+ type: "tool_result",
255
+ toolCallId: tc.id,
256
+ content: "Tool execution was aborted.",
257
+ isError: true
258
+ });
259
+ }
260
+ }
261
+ messages.push({ role: "tool", content: toolResults });
262
+ }
263
+ }
264
+ let lastAssistant;
265
+ for (let i = messages.length - 1; i >= 0; i--) {
266
+ if (messages[i].role === "assistant") {
267
+ lastAssistant = messages[i];
268
+ break;
269
+ }
270
+ }
271
+ yield {
272
+ type: "agent_done",
273
+ totalTurns: turn,
274
+ totalUsage: { ...totalUsage }
275
+ };
276
+ return {
277
+ message: lastAssistant ?? { role: "assistant", content: [] },
278
+ totalTurns: turn,
279
+ totalUsage: { ...totalUsage }
280
+ };
281
+ }
282
+ function normalizeToolResult(raw) {
283
+ return typeof raw === "string" ? { content: raw } : raw;
284
+ }
285
+ function extractToolCalls(content) {
286
+ if (typeof content === "string") return [];
287
+ return content.filter((part) => part.type === "tool_call");
288
+ }
289
+
290
+ // src/agent.ts
291
+ var AgentStream = class {
292
+ events;
293
+ resultPromise;
294
+ resolveResult;
295
+ rejectResult;
296
+ hasConsumer = false;
297
+ constructor(generator, onDone) {
298
+ this.events = new import_gg_ai2.EventStream();
299
+ this.resultPromise = new Promise((resolve, reject) => {
300
+ this.resolveResult = resolve;
301
+ this.rejectResult = reject;
302
+ });
303
+ this.pump(generator, onDone);
304
+ }
305
+ async pump(generator, onDone) {
306
+ try {
307
+ let next = await generator.next();
308
+ while (!next.done) {
309
+ this.events.push(next.value);
310
+ next = await generator.next();
311
+ }
312
+ this.events.close();
313
+ this.resolveResult(next.value);
314
+ } catch (err) {
315
+ const error = err instanceof Error ? err : new Error(String(err));
316
+ this.events.abort(error);
317
+ this.rejectResult(error);
318
+ } finally {
319
+ onDone();
320
+ }
321
+ }
322
+ [Symbol.asyncIterator]() {
323
+ this.hasConsumer = true;
324
+ return this.events[Symbol.asyncIterator]();
325
+ }
326
+ then(onfulfilled, onrejected) {
327
+ this.drainEvents().catch(() => {
328
+ });
329
+ return this.resultPromise.then(onfulfilled, onrejected);
330
+ }
331
+ async drainEvents() {
332
+ if (this.hasConsumer) return;
333
+ this.hasConsumer = true;
334
+ for await (const _ of this.events) {
335
+ }
336
+ }
337
+ };
338
+ var Agent = class {
339
+ messages = [];
340
+ _running = false;
341
+ options;
342
+ constructor(options) {
343
+ this.options = options;
344
+ if (options.system) {
345
+ this.messages.push({ role: "system", content: options.system });
346
+ }
347
+ }
348
+ get running() {
349
+ return this._running;
350
+ }
351
+ prompt(content) {
352
+ if (this._running) {
353
+ throw new Error("Agent is already running");
354
+ }
355
+ this._running = true;
356
+ this.messages.push({ role: "user", content });
357
+ const generator = agentLoop(this.messages, this.options);
358
+ return new AgentStream(generator, () => {
359
+ this._running = false;
360
+ });
361
+ }
362
+ };
363
+ // Annotate the CommonJS export names for ESM import in node:
364
+ 0 && (module.exports = {
365
+ Agent,
366
+ AgentStream,
367
+ agentLoop,
368
+ isContextOverflow
369
+ });
370
+ //# sourceMappingURL=index.cjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/index.ts","../src/agent.ts","../src/agent-loop.ts"],"sourcesContent":["// Core\nexport { Agent, AgentStream } from \"./agent.js\";\nexport { agentLoop, isContextOverflow } from \"./agent-loop.js\";\n\n// Types\nexport type {\n StructuredToolResult,\n ToolExecuteResult,\n ToolContext,\n AgentTool,\n AgentTextDeltaEvent,\n AgentThinkingDeltaEvent,\n AgentToolCallStartEvent,\n AgentToolCallUpdateEvent,\n AgentToolCallEndEvent,\n AgentServerToolCallEvent,\n AgentServerToolResultEvent,\n AgentTurnEndEvent,\n AgentDoneEvent,\n AgentErrorEvent,\n AgentEvent,\n AgentOptions,\n AgentResult,\n} from \"./types.js\";\n","import { EventStream, type Message } from \"@kenkaiiii/gg-ai\";\nimport { agentLoop } from \"./agent-loop.js\";\nimport type { AgentEvent, AgentOptions, AgentResult } from \"./types.js\";\n\n// ── AgentStream ─────────────────────────────────────────────\n\n/**\n * Dual-nature result: async iterable for streaming events,\n * thenable for awaiting the final AgentResult.\n *\n * ```ts\n * // Stream events\n * for await (const event of agent.prompt(\"hello\")) { ... }\n *\n * // Or just await the result\n * const result = await agent.prompt(\"hello\");\n * ```\n */\nexport class AgentStream implements AsyncIterable<AgentEvent> {\n private events: EventStream<AgentEvent>;\n private resultPromise: Promise<AgentResult>;\n private resolveResult!: (r: AgentResult) => void;\n private rejectResult!: (e: Error) => void;\n private hasConsumer = false;\n\n constructor(generator: AsyncGenerator<AgentEvent, AgentResult>, onDone: () => void) {\n this.events = new EventStream<AgentEvent>();\n this.resultPromise = new Promise<AgentResult>((resolve, reject) => {\n this.resolveResult = resolve;\n this.rejectResult = reject;\n });\n this.pump(generator, onDone);\n }\n\n private async pump(\n generator: AsyncGenerator<AgentEvent, AgentResult>,\n onDone: () => void,\n ): Promise<void> {\n try {\n let next = await generator.next();\n while (!next.done) {\n this.events.push(next.value);\n next = await generator.next();\n }\n this.events.close();\n this.resolveResult(next.value);\n } catch (err) {\n const error = err instanceof Error ? err : new Error(String(err));\n this.events.abort(error);\n this.rejectResult(error);\n } finally {\n onDone();\n }\n }\n\n [Symbol.asyncIterator](): AsyncIterator<AgentEvent> {\n this.hasConsumer = true;\n return this.events[Symbol.asyncIterator]();\n }\n\n then<TResult1 = AgentResult, TResult2 = never>(\n onfulfilled?: ((value: AgentResult) => TResult1 | PromiseLike<TResult1>) | null,\n onrejected?: ((reason: unknown) => TResult2 | PromiseLike<TResult2>) | null,\n ): Promise<TResult1 | TResult2> {\n this.drainEvents().catch(() => {});\n return this.resultPromise.then(onfulfilled, onrejected);\n }\n\n private async drainEvents(): Promise<void> {\n if (this.hasConsumer) return;\n this.hasConsumer = true;\n for await (const _ of this.events) {\n // consume silently\n }\n }\n}\n\n// ── Agent ───────────────────────────────────────────────────\n\nexport class Agent {\n private messages: Message[] = [];\n private _running = false;\n private options: AgentOptions;\n\n constructor(options: AgentOptions) {\n this.options = options;\n if (options.system) {\n this.messages.push({ role: \"system\", content: options.system });\n }\n }\n\n get running(): boolean {\n return this._running;\n }\n\n prompt(content: string): AgentStream {\n if (this._running) {\n throw new Error(\"Agent is already running\");\n }\n this._running = true;\n\n this.messages.push({ role: \"user\", content });\n\n const generator = agentLoop(this.messages, this.options);\n return new AgentStream(generator, () => {\n this._running = false;\n });\n }\n}\n","import {\n stream,\n EventStream,\n type Message,\n type ToolCall,\n type ToolResult,\n type Usage,\n type ContentPart,\n type AssistantMessage,\n} from \"@kenkaiiii/gg-ai\";\nimport type {\n AgentEvent,\n AgentOptions,\n AgentResult,\n AgentTool,\n ToolContext,\n ToolExecuteResult,\n StructuredToolResult,\n} from \"./types.js\";\n\nconst DEFAULT_MAX_TURNS = 100;\n\n/**\n * Detect context window overflow errors from LLM providers.\n * Anthropic: \"prompt is too long: N tokens > M maximum\"\n * OpenAI: \"context_length_exceeded\" / \"maximum context length\"\n */\nexport function isContextOverflow(err: unknown): boolean {\n if (!(err instanceof Error)) return false;\n const msg = err.message.toLowerCase();\n return (\n msg.includes(\"prompt is too long\") ||\n msg.includes(\"context_length_exceeded\") ||\n msg.includes(\"maximum context length\") ||\n (msg.includes(\"token\") && msg.includes(\"exceed\"))\n );\n}\n\n/**\n * Detect overloaded/rate-limit errors from LLM providers.\n * HTTP 429 (rate limit) or 529/503 (overloaded).\n */\nexport function isOverloaded(err: unknown): boolean {\n if (!(err instanceof Error)) return false;\n const msg = err.message.toLowerCase();\n return (\n msg.includes(\"overloaded\") ||\n msg.includes(\"rate limit\") ||\n msg.includes(\"too many requests\") ||\n msg.includes(\"429\") ||\n msg.includes(\"529\")\n );\n}\n\nexport async function* agentLoop(\n messages: Message[],\n options: AgentOptions,\n): AsyncGenerator<AgentEvent, AgentResult> {\n const maxTurns = options.maxTurns ?? DEFAULT_MAX_TURNS;\n const maxContinuations = options.maxContinuations ?? 5;\n const toolMap = new Map<string, AgentTool>((options.tools ?? []).map((t) => [t.name, t]));\n\n const totalUsage: Usage = { inputTokens: 0, outputTokens: 0 };\n let turn = 0;\n let consecutivePauses = 0;\n let overflowRetries = 0;\n let overloadRetries = 0;\n const MAX_OVERFLOW_RETRIES = 3;\n const MAX_OVERLOAD_RETRIES = 3;\n const OVERLOAD_RETRY_DELAY_MS = 3_000;\n\n while (turn < maxTurns) {\n options.signal?.throwIfAborted();\n turn++;\n\n // ── Mid-loop context transform (compaction / truncation) ──\n if (options.transformContext) {\n const transformed = await options.transformContext(messages);\n if (transformed !== messages) {\n messages.length = 0;\n messages.push(...transformed);\n }\n }\n\n // ── Call LLM with overflow recovery ──\n let response;\n try {\n const result = stream({\n provider: options.provider,\n model: options.model,\n messages,\n tools: options.tools,\n serverTools: options.serverTools,\n webSearch: options.webSearch,\n maxTokens: options.maxTokens,\n temperature: options.temperature,\n thinking: options.thinking,\n apiKey: options.apiKey,\n baseUrl: options.baseUrl,\n signal: options.signal,\n accountId: options.accountId,\n cacheRetention: options.cacheRetention,\n compaction: options.compaction,\n });\n\n // Suppress unhandled rejection if the iterator path throws first\n result.response.catch(() => {});\n\n // Forward streaming deltas\n for await (const event of result) {\n if (event.type === \"text_delta\") {\n yield { type: \"text_delta\" as const, text: event.text };\n } else if (event.type === \"thinking_delta\") {\n yield { type: \"thinking_delta\" as const, text: event.text };\n } else if (event.type === \"server_toolcall\") {\n yield {\n type: \"server_tool_call\" as const,\n id: event.id,\n name: event.name,\n input: event.input,\n };\n } else if (event.type === \"server_toolresult\") {\n yield {\n type: \"server_tool_result\" as const,\n toolUseId: event.toolUseId,\n resultType: event.resultType,\n data: event.data,\n };\n }\n }\n\n response = await result.response;\n } catch (err) {\n // Context overflow: force-compact via transformContext and retry (up to 3 times)\n if (\n overflowRetries < MAX_OVERFLOW_RETRIES &&\n isContextOverflow(err) &&\n options.transformContext\n ) {\n overflowRetries++;\n const transformed = await options.transformContext(messages, { force: true });\n if (transformed !== messages) {\n messages.length = 0;\n messages.push(...transformed);\n }\n turn--; // Don't count the failed turn\n continue;\n }\n // Overloaded / rate-limited: wait 3s and retry (up to 3 times)\n if (overloadRetries < MAX_OVERLOAD_RETRIES && isOverloaded(err)) {\n overloadRetries++;\n await new Promise((r) => setTimeout(r, OVERLOAD_RETRY_DELAY_MS));\n turn--; // Don't count the failed turn\n continue;\n }\n throw err;\n }\n\n // Reset retry counters after successful call\n overflowRetries = 0;\n overloadRetries = 0;\n\n // Accumulate usage\n totalUsage.inputTokens += response.usage.inputTokens;\n totalUsage.outputTokens += response.usage.outputTokens;\n if (response.usage.cacheRead) {\n totalUsage.cacheRead = (totalUsage.cacheRead ?? 0) + response.usage.cacheRead;\n }\n if (response.usage.cacheWrite) {\n totalUsage.cacheWrite = (totalUsage.cacheWrite ?? 0) + response.usage.cacheWrite;\n }\n\n // Append assistant message to conversation\n messages.push(response.message);\n\n yield {\n type: \"turn_end\" as const,\n turn,\n stopReason: response.stopReason,\n usage: response.usage,\n };\n\n // Server-side tool hit iteration limit — re-send to continue.\n // Do NOT add an extra user message; the API detects the trailing\n // server_tool_use block and resumes automatically.\n if (response.stopReason === \"pause_turn\") {\n consecutivePauses++;\n if (consecutivePauses >= maxContinuations) {\n break; // Safety limit — fall through to agent_done below\n }\n continue;\n }\n consecutivePauses = 0;\n\n // If not tool_use, we're done\n if (response.stopReason !== \"tool_use\") {\n yield {\n type: \"agent_done\" as const,\n totalTurns: turn,\n totalUsage: { ...totalUsage },\n };\n return {\n message: response.message,\n totalTurns: turn,\n totalUsage: { ...totalUsage },\n };\n }\n\n // Extract tool calls — separate client-executed from provider built-in (e.g. Moonshot $web_search)\n const allToolCalls = extractToolCalls(response.message.content);\n const toolCalls: ToolCall[] = [];\n const toolResults: ToolResult[] = [];\n\n for (const tc of allToolCalls) {\n if (tc.name.startsWith(\"$\")) {\n // Provider built-in tool (e.g. Moonshot $web_search) — not locally executed.\n // Still needs a tool_result for the message history round-trip.\n toolResults.push({\n type: \"tool_result\",\n toolCallId: tc.id,\n content: JSON.stringify(tc.args),\n });\n } else {\n toolCalls.push(tc);\n }\n }\n const eventStream = new EventStream<AgentEvent>();\n\n // Launch all tool calls in parallel\n const executions = toolCalls.map(async (toolCall) => {\n const startTime = Date.now();\n\n eventStream.push({\n type: \"tool_call_start\" as const,\n toolCallId: toolCall.id,\n name: toolCall.name,\n args: toolCall.args,\n });\n\n let resultContent: string;\n let details: unknown;\n let isError = false;\n\n const tool = toolMap.get(toolCall.name);\n if (!tool) {\n resultContent = `Unknown tool: ${toolCall.name}`;\n isError = true;\n } else {\n try {\n const parsed = tool.parameters.parse(toolCall.args);\n const ctx: ToolContext = {\n signal: options.signal ?? AbortSignal.timeout(300_000),\n toolCallId: toolCall.id,\n onUpdate: (update: unknown) => {\n eventStream.push({\n type: \"tool_call_update\" as const,\n toolCallId: toolCall.id,\n update,\n });\n },\n };\n const raw = await tool.execute(parsed, ctx);\n const normalized = normalizeToolResult(raw);\n resultContent = normalized.content;\n details = normalized.details;\n } catch (err) {\n isError = true;\n resultContent = err instanceof Error ? err.message : String(err);\n }\n }\n\n const durationMs = Date.now() - startTime;\n\n eventStream.push({\n type: \"tool_call_end\" as const,\n toolCallId: toolCall.id,\n result: resultContent,\n details,\n isError,\n durationMs,\n });\n\n return { toolCallId: toolCall.id, content: resultContent, isError };\n });\n\n // Abort the tool event stream when the signal fires so Ctrl+C\n // doesn't hang waiting for long-running tools to finish.\n const abortHandler = () => eventStream.abort(new Error(\"aborted\"));\n options.signal?.addEventListener(\"abort\", abortHandler, { once: true });\n\n // Close event stream when all tools complete\n Promise.all(executions)\n .then((results) => {\n for (const tc of toolCalls) {\n const r = results.find((x) => x.toolCallId === tc.id)!;\n toolResults.push({\n type: \"tool_result\",\n toolCallId: tc.id,\n content: r.content,\n isError: r.isError || undefined,\n });\n }\n eventStream.close();\n })\n .catch((err) => eventStream.abort(err instanceof Error ? err : new Error(String(err))));\n\n // Yield events as they arrive from parallel tools\n try {\n for await (const event of eventStream) {\n yield event;\n }\n } finally {\n options.signal?.removeEventListener(\"abort\", abortHandler);\n\n // Ensure every tool_use has a matching tool_result, even on abort.\n // Without this, an aborted turn leaves an orphaned tool_use in the\n // message history which causes Anthropic API 400 errors on the next\n // request.\n for (const tc of toolCalls) {\n if (!toolResults.some((r) => r.toolCallId === tc.id)) {\n toolResults.push({\n type: \"tool_result\",\n toolCallId: tc.id,\n content: \"Tool execution was aborted.\",\n isError: true,\n });\n }\n }\n messages.push({ role: \"tool\", content: toolResults });\n }\n }\n\n // Exceeded max turns — return last assistant message\n let lastAssistant: AssistantMessage | undefined;\n for (let i = messages.length - 1; i >= 0; i--) {\n if (messages[i]!.role === \"assistant\") {\n lastAssistant = messages[i] as AssistantMessage;\n break;\n }\n }\n\n yield {\n type: \"agent_done\" as const,\n totalTurns: turn,\n totalUsage: { ...totalUsage },\n };\n\n return {\n message: lastAssistant ?? { role: \"assistant\" as const, content: [] },\n totalTurns: turn,\n totalUsage: { ...totalUsage },\n };\n}\n\nfunction normalizeToolResult(raw: ToolExecuteResult): StructuredToolResult {\n return typeof raw === \"string\" ? { content: raw } : raw;\n}\n\nfunction extractToolCalls(content: string | ContentPart[]): ToolCall[] {\n if (typeof content === \"string\") return [];\n return content.filter((part): part is ToolCall => part.type === \"tool_call\");\n}\n"],"mappings":";;;;;;;;;;;;;;;;;;;;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;AAAA;;;ACAA,IAAAA,gBAA0C;;;ACA1C,mBASO;AAWP,IAAM,oBAAoB;AAOnB,SAAS,kBAAkB,KAAuB;AACvD,MAAI,EAAE,eAAe,OAAQ,QAAO;AACpC,QAAM,MAAM,IAAI,QAAQ,YAAY;AACpC,SACE,IAAI,SAAS,oBAAoB,KACjC,IAAI,SAAS,yBAAyB,KACtC,IAAI,SAAS,wBAAwB,KACpC,IAAI,SAAS,OAAO,KAAK,IAAI,SAAS,QAAQ;AAEnD;AAMO,SAAS,aAAa,KAAuB;AAClD,MAAI,EAAE,eAAe,OAAQ,QAAO;AACpC,QAAM,MAAM,IAAI,QAAQ,YAAY;AACpC,SACE,IAAI,SAAS,YAAY,KACzB,IAAI,SAAS,YAAY,KACzB,IAAI,SAAS,mBAAmB,KAChC,IAAI,SAAS,KAAK,KAClB,IAAI,SAAS,KAAK;AAEtB;AAEA,gBAAuB,UACrB,UACA,SACyC;AACzC,QAAM,WAAW,QAAQ,YAAY;AACrC,QAAM,mBAAmB,QAAQ,oBAAoB;AACrD,QAAM,UAAU,IAAI,KAAwB,QAAQ,SAAS,CAAC,GAAG,IAAI,CAAC,MAAM,CAAC,EAAE,MAAM,CAAC,CAAC,CAAC;AAExF,QAAM,aAAoB,EAAE,aAAa,GAAG,cAAc,EAAE;AAC5D,MAAI,OAAO;AACX,MAAI,oBAAoB;AACxB,MAAI,kBAAkB;AACtB,MAAI,kBAAkB;AACtB,QAAM,uBAAuB;AAC7B,QAAM,uBAAuB;AAC7B,QAAM,0BAA0B;AAEhC,SAAO,OAAO,UAAU;AACtB,YAAQ,QAAQ,eAAe;AAC/B;AAGA,QAAI,QAAQ,kBAAkB;AAC5B,YAAM,cAAc,MAAM,QAAQ,iBAAiB,QAAQ;AAC3D,UAAI,gBAAgB,UAAU;AAC5B,iBAAS,SAAS;AAClB,iBAAS,KAAK,GAAG,WAAW;AAAA,MAC9B;AAAA,IACF;AAGA,QAAI;AACJ,QAAI;AACF,YAAM,aAAS,qBAAO;AAAA,QACpB,UAAU,QAAQ;AAAA,QAClB,OAAO,QAAQ;AAAA,QACf;AAAA,QACA,OAAO,QAAQ;AAAA,QACf,aAAa,QAAQ;AAAA,QACrB,WAAW,QAAQ;AAAA,QACnB,WAAW,QAAQ;AAAA,QACnB,aAAa,QAAQ;AAAA,QACrB,UAAU,QAAQ;AAAA,QAClB,QAAQ,QAAQ;AAAA,QAChB,SAAS,QAAQ;AAAA,QACjB,QAAQ,QAAQ;AAAA,QAChB,WAAW,QAAQ;AAAA,QACnB,gBAAgB,QAAQ;AAAA,QACxB,YAAY,QAAQ;AAAA,MACtB,CAAC;AAGD,aAAO,SAAS,MAAM,MAAM;AAAA,MAAC,CAAC;AAG9B,uBAAiB,SAAS,QAAQ;AAChC,YAAI,MAAM,SAAS,cAAc;AAC/B,gBAAM,EAAE,MAAM,cAAuB,MAAM,MAAM,KAAK;AAAA,QACxD,WAAW,MAAM,SAAS,kBAAkB;AAC1C,gBAAM,EAAE,MAAM,kBAA2B,MAAM,MAAM,KAAK;AAAA,QAC5D,WAAW,MAAM,SAAS,mBAAmB;AAC3C,gBAAM;AAAA,YACJ,MAAM;AAAA,YACN,IAAI,MAAM;AAAA,YACV,MAAM,MAAM;AAAA,YACZ,OAAO,MAAM;AAAA,UACf;AAAA,QACF,WAAW,MAAM,SAAS,qBAAqB;AAC7C,gBAAM;AAAA,YACJ,MAAM;AAAA,YACN,WAAW,MAAM;AAAA,YACjB,YAAY,MAAM;AAAA,YAClB,MAAM,MAAM;AAAA,UACd;AAAA,QACF;AAAA,MACF;AAEA,iBAAW,MAAM,OAAO;AAAA,IAC1B,SAAS,KAAK;AAEZ,UACE,kBAAkB,wBAClB,kBAAkB,GAAG,KACrB,QAAQ,kBACR;AACA;AACA,cAAM,cAAc,MAAM,QAAQ,iBAAiB,UAAU,EAAE,OAAO,KAAK,CAAC;AAC5E,YAAI,gBAAgB,UAAU;AAC5B,mBAAS,SAAS;AAClB,mBAAS,KAAK,GAAG,WAAW;AAAA,QAC9B;AACA;AACA;AAAA,MACF;AAEA,UAAI,kBAAkB,wBAAwB,aAAa,GAAG,GAAG;AAC/D;AACA,cAAM,IAAI,QAAQ,CAAC,MAAM,WAAW,GAAG,uBAAuB,CAAC;AAC/D;AACA;AAAA,MACF;AACA,YAAM;AAAA,IACR;AAGA,sBAAkB;AAClB,sBAAkB;AAGlB,eAAW,eAAe,SAAS,MAAM;AACzC,eAAW,gBAAgB,SAAS,MAAM;AAC1C,QAAI,SAAS,MAAM,WAAW;AAC5B,iBAAW,aAAa,WAAW,aAAa,KAAK,SAAS,MAAM;AAAA,IACtE;AACA,QAAI,SAAS,MAAM,YAAY;AAC7B,iBAAW,cAAc,WAAW,cAAc,KAAK,SAAS,MAAM;AAAA,IACxE;AAGA,aAAS,KAAK,SAAS,OAAO;AAE9B,UAAM;AAAA,MACJ,MAAM;AAAA,MACN;AAAA,MACA,YAAY,SAAS;AAAA,MACrB,OAAO,SAAS;AAAA,IAClB;AAKA,QAAI,SAAS,eAAe,cAAc;AACxC;AACA,UAAI,qBAAqB,kBAAkB;AACzC;AAAA,MACF;AACA;AAAA,IACF;AACA,wBAAoB;AAGpB,QAAI,SAAS,eAAe,YAAY;AACtC,YAAM;AAAA,QACJ,MAAM;AAAA,QACN,YAAY;AAAA,QACZ,YAAY,EAAE,GAAG,WAAW;AAAA,MAC9B;AACA,aAAO;AAAA,QACL,SAAS,SAAS;AAAA,QAClB,YAAY;AAAA,QACZ,YAAY,EAAE,GAAG,WAAW;AAAA,MAC9B;AAAA,IACF;AAGA,UAAM,eAAe,iBAAiB,SAAS,QAAQ,OAAO;AAC9D,UAAM,YAAwB,CAAC;AAC/B,UAAM,cAA4B,CAAC;AAEnC,eAAW,MAAM,cAAc;AAC7B,UAAI,GAAG,KAAK,WAAW,GAAG,GAAG;AAG3B,oBAAY,KAAK;AAAA,UACf,MAAM;AAAA,UACN,YAAY,GAAG;AAAA,UACf,SAAS,KAAK,UAAU,GAAG,IAAI;AAAA,QACjC,CAAC;AAAA,MACH,OAAO;AACL,kBAAU,KAAK,EAAE;AAAA,MACnB;AAAA,IACF;AACA,UAAM,cAAc,IAAI,yBAAwB;AAGhD,UAAM,aAAa,UAAU,IAAI,OAAO,aAAa;AACnD,YAAM,YAAY,KAAK,IAAI;AAE3B,kBAAY,KAAK;AAAA,QACf,MAAM;AAAA,QACN,YAAY,SAAS;AAAA,QACrB,MAAM,SAAS;AAAA,QACf,MAAM,SAAS;AAAA,MACjB,CAAC;AAED,UAAI;AACJ,UAAI;AACJ,UAAI,UAAU;AAEd,YAAM,OAAO,QAAQ,IAAI,SAAS,IAAI;AACtC,UAAI,CAAC,MAAM;AACT,wBAAgB,iBAAiB,SAAS,IAAI;AAC9C,kBAAU;AAAA,MACZ,OAAO;AACL,YAAI;AACF,gBAAM,SAAS,KAAK,WAAW,MAAM,SAAS,IAAI;AAClD,gBAAM,MAAmB;AAAA,YACvB,QAAQ,QAAQ,UAAU,YAAY,QAAQ,GAAO;AAAA,YACrD,YAAY,SAAS;AAAA,YACrB,UAAU,CAAC,WAAoB;AAC7B,0BAAY,KAAK;AAAA,gBACf,MAAM;AAAA,gBACN,YAAY,SAAS;AAAA,gBACrB;AAAA,cACF,CAAC;AAAA,YACH;AAAA,UACF;AACA,gBAAM,MAAM,MAAM,KAAK,QAAQ,QAAQ,GAAG;AAC1C,gBAAM,aAAa,oBAAoB,GAAG;AAC1C,0BAAgB,WAAW;AAC3B,oBAAU,WAAW;AAAA,QACvB,SAAS,KAAK;AACZ,oBAAU;AACV,0BAAgB,eAAe,QAAQ,IAAI,UAAU,OAAO,GAAG;AAAA,QACjE;AAAA,MACF;AAEA,YAAM,aAAa,KAAK,IAAI,IAAI;AAEhC,kBAAY,KAAK;AAAA,QACf,MAAM;AAAA,QACN,YAAY,SAAS;AAAA,QACrB,QAAQ;AAAA,QACR;AAAA,QACA;AAAA,QACA;AAAA,MACF,CAAC;AAED,aAAO,EAAE,YAAY,SAAS,IAAI,SAAS,eAAe,QAAQ;AAAA,IACpE,CAAC;AAID,UAAM,eAAe,MAAM,YAAY,MAAM,IAAI,MAAM,SAAS,CAAC;AACjE,YAAQ,QAAQ,iBAAiB,SAAS,cAAc,EAAE,MAAM,KAAK,CAAC;AAGtE,YAAQ,IAAI,UAAU,EACnB,KAAK,CAAC,YAAY;AACjB,iBAAW,MAAM,WAAW;AAC1B,cAAM,IAAI,QAAQ,KAAK,CAAC,MAAM,EAAE,eAAe,GAAG,EAAE;AACpD,oBAAY,KAAK;AAAA,UACf,MAAM;AAAA,UACN,YAAY,GAAG;AAAA,UACf,SAAS,EAAE;AAAA,UACX,SAAS,EAAE,WAAW;AAAA,QACxB,CAAC;AAAA,MACH;AACA,kBAAY,MAAM;AAAA,IACpB,CAAC,EACA,MAAM,CAAC,QAAQ,YAAY,MAAM,eAAe,QAAQ,MAAM,IAAI,MAAM,OAAO,GAAG,CAAC,CAAC,CAAC;AAGxF,QAAI;AACF,uBAAiB,SAAS,aAAa;AACrC,cAAM;AAAA,MACR;AAAA,IACF,UAAE;AACA,cAAQ,QAAQ,oBAAoB,SAAS,YAAY;AAMzD,iBAAW,MAAM,WAAW;AAC1B,YAAI,CAAC,YAAY,KAAK,CAAC,MAAM,EAAE,eAAe,GAAG,EAAE,GAAG;AACpD,sBAAY,KAAK;AAAA,YACf,MAAM;AAAA,YACN,YAAY,GAAG;AAAA,YACf,SAAS;AAAA,YACT,SAAS;AAAA,UACX,CAAC;AAAA,QACH;AAAA,MACF;AACA,eAAS,KAAK,EAAE,MAAM,QAAQ,SAAS,YAAY,CAAC;AAAA,IACtD;AAAA,EACF;AAGA,MAAI;AACJ,WAAS,IAAI,SAAS,SAAS,GAAG,KAAK,GAAG,KAAK;AAC7C,QAAI,SAAS,CAAC,EAAG,SAAS,aAAa;AACrC,sBAAgB,SAAS,CAAC;AAC1B;AAAA,IACF;AAAA,EACF;AAEA,QAAM;AAAA,IACJ,MAAM;AAAA,IACN,YAAY;AAAA,IACZ,YAAY,EAAE,GAAG,WAAW;AAAA,EAC9B;AAEA,SAAO;AAAA,IACL,SAAS,iBAAiB,EAAE,MAAM,aAAsB,SAAS,CAAC,EAAE;AAAA,IACpE,YAAY;AAAA,IACZ,YAAY,EAAE,GAAG,WAAW;AAAA,EAC9B;AACF;AAEA,SAAS,oBAAoB,KAA8C;AACzE,SAAO,OAAO,QAAQ,WAAW,EAAE,SAAS,IAAI,IAAI;AACtD;AAEA,SAAS,iBAAiB,SAA6C;AACrE,MAAI,OAAO,YAAY,SAAU,QAAO,CAAC;AACzC,SAAO,QAAQ,OAAO,CAAC,SAA2B,KAAK,SAAS,WAAW;AAC7E;;;ADvVO,IAAM,cAAN,MAAuD;AAAA,EACpD;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA,cAAc;AAAA,EAEtB,YAAY,WAAoD,QAAoB;AAClF,SAAK,SAAS,IAAI,0BAAwB;AAC1C,SAAK,gBAAgB,IAAI,QAAqB,CAAC,SAAS,WAAW;AACjE,WAAK,gBAAgB;AACrB,WAAK,eAAe;AAAA,IACtB,CAAC;AACD,SAAK,KAAK,WAAW,MAAM;AAAA,EAC7B;AAAA,EAEA,MAAc,KACZ,WACA,QACe;AACf,QAAI;AACF,UAAI,OAAO,MAAM,UAAU,KAAK;AAChC,aAAO,CAAC,KAAK,MAAM;AACjB,aAAK,OAAO,KAAK,KAAK,KAAK;AAC3B,eAAO,MAAM,UAAU,KAAK;AAAA,MAC9B;AACA,WAAK,OAAO,MAAM;AAClB,WAAK,cAAc,KAAK,KAAK;AAAA,IAC/B,SAAS,KAAK;AACZ,YAAM,QAAQ,eAAe,QAAQ,MAAM,IAAI,MAAM,OAAO,GAAG,CAAC;AAChE,WAAK,OAAO,MAAM,KAAK;AACvB,WAAK,aAAa,KAAK;AAAA,IACzB,UAAE;AACA,aAAO;AAAA,IACT;AAAA,EACF;AAAA,EAEA,CAAC,OAAO,aAAa,IAA+B;AAClD,SAAK,cAAc;AACnB,WAAO,KAAK,OAAO,OAAO,aAAa,EAAE;AAAA,EAC3C;AAAA,EAEA,KACE,aACA,YAC8B;AAC9B,SAAK,YAAY,EAAE,MAAM,MAAM;AAAA,IAAC,CAAC;AACjC,WAAO,KAAK,cAAc,KAAK,aAAa,UAAU;AAAA,EACxD;AAAA,EAEA,MAAc,cAA6B;AACzC,QAAI,KAAK,YAAa;AACtB,SAAK,cAAc;AACnB,qBAAiB,KAAK,KAAK,QAAQ;AAAA,IAEnC;AAAA,EACF;AACF;AAIO,IAAM,QAAN,MAAY;AAAA,EACT,WAAsB,CAAC;AAAA,EACvB,WAAW;AAAA,EACX;AAAA,EAER,YAAY,SAAuB;AACjC,SAAK,UAAU;AACf,QAAI,QAAQ,QAAQ;AAClB,WAAK,SAAS,KAAK,EAAE,MAAM,UAAU,SAAS,QAAQ,OAAO,CAAC;AAAA,IAChE;AAAA,EACF;AAAA,EAEA,IAAI,UAAmB;AACrB,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,OAAO,SAA8B;AACnC,QAAI,KAAK,UAAU;AACjB,YAAM,IAAI,MAAM,0BAA0B;AAAA,IAC5C;AACA,SAAK,WAAW;AAEhB,SAAK,SAAS,KAAK,EAAE,MAAM,QAAQ,QAAQ,CAAC;AAE5C,UAAM,YAAY,UAAU,KAAK,UAAU,KAAK,OAAO;AACvD,WAAO,IAAI,YAAY,WAAW,MAAM;AACtC,WAAK,WAAW;AAAA,IAClB,CAAC;AAAA,EACH;AACF;","names":["import_gg_ai"]}
@@ -0,0 +1,154 @@
1
+ import { z } from 'zod';
2
+ import { StreamOptions, Tool, ServerToolDefinition, Message, StopReason, Usage, AssistantMessage } from '@kenkaiiii/gg-ai';
3
+
4
+ interface StructuredToolResult {
5
+ content: string;
6
+ details?: unknown;
7
+ }
8
+ type ToolExecuteResult = string | StructuredToolResult;
9
+ interface ToolContext {
10
+ signal: AbortSignal;
11
+ toolCallId: string;
12
+ onUpdate?: (update: unknown) => void;
13
+ }
14
+ interface AgentTool<T extends z.ZodType = z.ZodType> extends Tool {
15
+ parameters: T;
16
+ execute: (args: z.infer<T>, context: ToolContext) => ToolExecuteResult | Promise<ToolExecuteResult>;
17
+ }
18
+ interface AgentTextDeltaEvent {
19
+ type: "text_delta";
20
+ text: string;
21
+ }
22
+ interface AgentThinkingDeltaEvent {
23
+ type: "thinking_delta";
24
+ text: string;
25
+ }
26
+ interface AgentToolCallStartEvent {
27
+ type: "tool_call_start";
28
+ toolCallId: string;
29
+ name: string;
30
+ args: Record<string, unknown>;
31
+ }
32
+ interface AgentToolCallUpdateEvent {
33
+ type: "tool_call_update";
34
+ toolCallId: string;
35
+ update: unknown;
36
+ }
37
+ interface AgentToolCallEndEvent {
38
+ type: "tool_call_end";
39
+ toolCallId: string;
40
+ result: string;
41
+ details?: unknown;
42
+ isError: boolean;
43
+ durationMs: number;
44
+ }
45
+ interface AgentTurnEndEvent {
46
+ type: "turn_end";
47
+ turn: number;
48
+ stopReason: StopReason;
49
+ usage: Usage;
50
+ }
51
+ interface AgentDoneEvent {
52
+ type: "agent_done";
53
+ totalTurns: number;
54
+ totalUsage: Usage;
55
+ }
56
+ interface AgentErrorEvent {
57
+ type: "error";
58
+ error: Error;
59
+ }
60
+ interface AgentServerToolCallEvent {
61
+ type: "server_tool_call";
62
+ id: string;
63
+ name: string;
64
+ input: unknown;
65
+ }
66
+ interface AgentServerToolResultEvent {
67
+ type: "server_tool_result";
68
+ toolUseId: string;
69
+ resultType: string;
70
+ data: unknown;
71
+ }
72
+ type AgentEvent = AgentTextDeltaEvent | AgentThinkingDeltaEvent | AgentToolCallStartEvent | AgentToolCallUpdateEvent | AgentToolCallEndEvent | AgentServerToolCallEvent | AgentServerToolResultEvent | AgentTurnEndEvent | AgentDoneEvent | AgentErrorEvent;
73
+ interface AgentOptions {
74
+ provider: StreamOptions["provider"];
75
+ model: string;
76
+ system?: string;
77
+ tools?: AgentTool[];
78
+ serverTools?: ServerToolDefinition[];
79
+ maxTurns?: number;
80
+ maxTokens?: number;
81
+ temperature?: number;
82
+ thinking?: StreamOptions["thinking"];
83
+ apiKey?: string;
84
+ baseUrl?: string;
85
+ signal?: AbortSignal;
86
+ accountId?: string;
87
+ cacheRetention?: StreamOptions["cacheRetention"];
88
+ /** Enable provider-native web search. */
89
+ webSearch?: boolean;
90
+ /** Enable server-side compaction (Anthropic only, beta). */
91
+ compaction?: boolean;
92
+ /** Max consecutive pause_turn continuations before stopping (default: 5).
93
+ * Prevents infinite loops when server-side tools keep pausing. */
94
+ maxContinuations?: number;
95
+ /**
96
+ * Called before each LLM call. Allows the caller to inspect and transform
97
+ * the messages array (e.g. compaction, truncation). Return the same array
98
+ * for no-op, or a new array to replace the conversation context.
99
+ *
100
+ * When `options.force` is true, the caller should compact unconditionally
101
+ * (e.g. after a context overflow error from the API).
102
+ */
103
+ transformContext?: (messages: Message[], options?: {
104
+ force?: boolean;
105
+ }) => Message[] | Promise<Message[]>;
106
+ }
107
+ interface AgentResult {
108
+ message: AssistantMessage;
109
+ totalTurns: number;
110
+ totalUsage: Usage;
111
+ }
112
+
113
+ /**
114
+ * Dual-nature result: async iterable for streaming events,
115
+ * thenable for awaiting the final AgentResult.
116
+ *
117
+ * ```ts
118
+ * // Stream events
119
+ * for await (const event of agent.prompt("hello")) { ... }
120
+ *
121
+ * // Or just await the result
122
+ * const result = await agent.prompt("hello");
123
+ * ```
124
+ */
125
+ declare class AgentStream implements AsyncIterable<AgentEvent> {
126
+ private events;
127
+ private resultPromise;
128
+ private resolveResult;
129
+ private rejectResult;
130
+ private hasConsumer;
131
+ constructor(generator: AsyncGenerator<AgentEvent, AgentResult>, onDone: () => void);
132
+ private pump;
133
+ [Symbol.asyncIterator](): AsyncIterator<AgentEvent>;
134
+ then<TResult1 = AgentResult, TResult2 = never>(onfulfilled?: ((value: AgentResult) => TResult1 | PromiseLike<TResult1>) | null, onrejected?: ((reason: unknown) => TResult2 | PromiseLike<TResult2>) | null): Promise<TResult1 | TResult2>;
135
+ private drainEvents;
136
+ }
137
+ declare class Agent {
138
+ private messages;
139
+ private _running;
140
+ private options;
141
+ constructor(options: AgentOptions);
142
+ get running(): boolean;
143
+ prompt(content: string): AgentStream;
144
+ }
145
+
146
+ /**
147
+ * Detect context window overflow errors from LLM providers.
148
+ * Anthropic: "prompt is too long: N tokens > M maximum"
149
+ * OpenAI: "context_length_exceeded" / "maximum context length"
150
+ */
151
+ declare function isContextOverflow(err: unknown): boolean;
152
+ declare function agentLoop(messages: Message[], options: AgentOptions): AsyncGenerator<AgentEvent, AgentResult>;
153
+
154
+ export { Agent, type AgentDoneEvent, type AgentErrorEvent, type AgentEvent, type AgentOptions, type AgentResult, type AgentServerToolCallEvent, type AgentServerToolResultEvent, AgentStream, type AgentTextDeltaEvent, type AgentThinkingDeltaEvent, type AgentTool, type AgentToolCallEndEvent, type AgentToolCallStartEvent, type AgentToolCallUpdateEvent, type AgentTurnEndEvent, type StructuredToolResult, type ToolContext, type ToolExecuteResult, agentLoop, isContextOverflow };