@easynet/agent-llm 1.0.11 → 1.0.13

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (48) hide show
  1. package/dist/{createAgentLlM.d.ts → api/create-agent-llm.d.ts} +2 -2
  2. package/dist/api/create-agent-llm.d.ts.map +1 -0
  3. package/dist/{chunk-HKET766S.js → chunk-MYGWM3YH.js} +205 -134
  4. package/dist/chunk-MYGWM3YH.js.map +1 -0
  5. package/dist/{cli.d.ts → cli/index.d.ts} +1 -1
  6. package/dist/cli/index.d.ts.map +1 -0
  7. package/dist/{cli.js → cli/index.js} +3 -3
  8. package/dist/cli/index.js.map +1 -0
  9. package/dist/config/index.d.ts +7 -0
  10. package/dist/config/index.d.ts.map +1 -0
  11. package/dist/{loadLlmConfig.d.ts → config/loader.d.ts} +1 -1
  12. package/dist/config/loader.d.ts.map +1 -0
  13. package/dist/{config.d.ts → config/parser.d.ts} +2 -2
  14. package/dist/config/parser.d.ts.map +1 -0
  15. package/dist/extensions/index.d.ts +8 -0
  16. package/dist/extensions/index.d.ts.map +1 -0
  17. package/dist/{loadLLMExtensions.d.ts → extensions/loader.d.ts} +1 -1
  18. package/dist/extensions/loader.d.ts.map +1 -0
  19. package/dist/{npmProviderProtocol.d.ts → extensions/npm-protocol.d.ts} +2 -2
  20. package/dist/{npmProviderProtocol.d.ts.map → extensions/npm-protocol.d.ts.map} +1 -1
  21. package/dist/index.d.ts +14 -14
  22. package/dist/index.d.ts.map +1 -1
  23. package/dist/index.js +2 -2
  24. package/dist/index.js.map +1 -1
  25. package/dist/{llmAdapter.d.ts → langchain/adapter.d.ts} +1 -1
  26. package/dist/langchain/adapter.d.ts.map +1 -0
  27. package/dist/langchain/index.d.ts +6 -0
  28. package/dist/langchain/index.d.ts.map +1 -0
  29. package/dist/providers/openai.d.ts.map +1 -1
  30. package/dist/{chatModelRegistry.d.ts → registry/chat-model.d.ts} +2 -2
  31. package/dist/registry/chat-model.d.ts.map +1 -0
  32. package/dist/{factory.d.ts → registry/client.d.ts} +2 -2
  33. package/dist/registry/client.d.ts.map +1 -0
  34. package/dist/registry/index.d.ts +8 -0
  35. package/dist/registry/index.d.ts.map +1 -0
  36. package/dist/types.d.ts +42 -0
  37. package/dist/types.d.ts.map +1 -1
  38. package/package.json +3 -3
  39. package/dist/chatModelRegistry.d.ts.map +0 -1
  40. package/dist/chunk-HKET766S.js.map +0 -1
  41. package/dist/cli.d.ts.map +0 -1
  42. package/dist/cli.js.map +0 -1
  43. package/dist/config.d.ts.map +0 -1
  44. package/dist/createAgentLlM.d.ts.map +0 -1
  45. package/dist/factory.d.ts.map +0 -1
  46. package/dist/llmAdapter.d.ts.map +0 -1
  47. package/dist/loadLLMExtensions.d.ts.map +0 -1
  48. package/dist/loadLlmConfig.d.ts.map +0 -1
@@ -1,4 +1,4 @@
1
- import { createChatModelFromLlmConfig } from "./llmAdapter.js";
1
+ import { createChatModelFromLlmConfig } from "../langchain/adapter.js";
2
2
  export interface CreateAgentLlMOptions {
3
3
  /** Path to YAML config file. If omitted, uses llm.yaml in the current directory (cwd). */
4
4
  configPath?: string;
@@ -12,4 +12,4 @@ export interface CreateAgentLlMOptions {
12
12
  * - createAgentLlM({ configPath: "...", installNpmIfMissing: true }) — options object
13
13
  */
14
14
  export declare function createAgentLlM(configPathOrOptions?: string | CreateAgentLlMOptions): Promise<ReturnType<typeof createChatModelFromLlmConfig>>;
15
- //# sourceMappingURL=createAgentLlM.d.ts.map
15
+ //# sourceMappingURL=create-agent-llm.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"create-agent-llm.d.ts","sourceRoot":"","sources":["../../src/api/create-agent-llm.ts"],"names":[],"mappings":"AAMA,OAAO,EAAE,4BAA4B,EAAE,MAAM,yBAAyB,CAAC;AAIvE,MAAM,WAAW,qBAAqB;IACpC,0FAA0F;IAC1F,UAAU,CAAC,EAAE,MAAM,CAAC;IACpB,kGAAkG;IAClG,mBAAmB,CAAC,EAAE,OAAO,CAAC;CAC/B;AAgDD;;;;;GAKG;AACH,wBAAsB,cAAc,CAClC,mBAAmB,CAAC,EAAE,MAAM,GAAG,qBAAqB,GACnD,OAAO,CAAC,UAAU,CAAC,OAAO,4BAA4B,CAAC,CAAC,CAiB1D"}
@@ -1,4 +1,4 @@
1
- // src/config.ts
1
+ // src/config/parser.ts
2
2
  var DEFAULT_LLM_ID = "default";
3
3
  var RESERVED_KEYS = /* @__PURE__ */ new Set([
4
4
  "default",
@@ -108,81 +108,48 @@ function normalizeLlmConfig(o) {
108
108
  return config;
109
109
  }
110
110
 
111
- // src/chatModelRegistry.ts
112
- var CHAT_MODEL_FACTORIES = /* @__PURE__ */ new Map();
113
- function registerChatModelProvider(providerName, factory) {
114
- CHAT_MODEL_FACTORIES.set(providerName.toLowerCase(), factory);
115
- }
116
- function getChatModelFactory(providerName) {
117
- return CHAT_MODEL_FACTORIES.get(providerName.toLowerCase());
111
+ // src/config/loader.ts
112
+ import { readFileSync, existsSync } from "fs";
113
+ import { parse as parseYaml } from "yaml";
114
+ function substituteEnv(obj) {
115
+ if (obj === null || obj === void 0) return obj;
116
+ if (typeof obj === "string") {
117
+ const m = obj.match(/^\$\{(\w+)\}$/);
118
+ return m ? process.env[m[1]] ?? obj : obj;
119
+ }
120
+ if (Array.isArray(obj)) return obj.map(substituteEnv);
121
+ if (typeof obj === "object") {
122
+ const out = {};
123
+ for (const [k, v] of Object.entries(obj)) out[k] = substituteEnv(v);
124
+ return out;
125
+ }
126
+ return obj;
118
127
  }
119
-
120
- // src/llmAdapter.ts
121
- import { ChatOpenAI } from "@langchain/openai";
122
- var DEFAULT_MODEL = "gpt-4o-mini";
123
- function normalizeAgentLlMError(e, context) {
124
- if (e instanceof Error) return new Error(`${context}: ${e.message}`, { cause: e });
125
- return new Error(`${context}: ${String(e)}`);
128
+ function parseLlmYaml(content, options = {}) {
129
+ const { substituteEnv: doSub = true } = options;
130
+ const parsed = parseYaml(content);
131
+ const llm = parsed?.llm;
132
+ if (llm == null) return void 0;
133
+ return doSub ? substituteEnv(llm) : llm;
126
134
  }
127
- function createChatModelFromLlmConfig(options) {
128
- const { llmSection, modelEnv, apiKeyEnv } = options;
129
- let defaultId;
130
- let configs;
135
+ function loadLlmConfig(filePath, options = {}) {
136
+ if (typeof filePath !== "string" || filePath.trim().length === 0) {
137
+ throw new Error("agent-llm: loadLlmConfig requires a non-empty file path");
138
+ }
139
+ if (!existsSync(filePath)) return null;
140
+ let raw;
131
141
  try {
132
- const parsed = parseLlmSection(llmSection ?? null);
133
- defaultId = parsed.defaultId;
134
- configs = parsed.configs;
142
+ raw = readFileSync(filePath, "utf8");
135
143
  } catch (e) {
136
- throw normalizeAgentLlMError(e, "agent-llm: failed to parse llm section");
137
- }
138
- const defaultConfig = configs.find((c) => c.id === defaultId) ?? configs[0];
139
- if (!defaultConfig) {
140
- const model2 = modelEnv ?? process.env.OPENAI_MODEL ?? DEFAULT_MODEL;
141
- const apiKey2 = apiKeyEnv ?? process.env.OPENAI_API_KEY;
142
- try {
143
- return new ChatOpenAI({
144
- model: model2,
145
- temperature: 0,
146
- ...apiKey2 ? { apiKey: apiKey2 } : {}
147
- });
148
- } catch (e) {
149
- throw normalizeAgentLlMError(e, "agent-llm: failed to create default ChatOpenAI");
150
- }
151
- }
152
- const provider = defaultConfig.provider ?? "openai";
153
- const chatModelFactory = getChatModelFactory(provider);
154
- if (chatModelFactory) {
155
- const config = {
156
- ...defaultConfig,
157
- model: modelEnv ?? defaultConfig.model,
158
- temperature: typeof defaultConfig.temperature === "number" ? defaultConfig.temperature : 0
159
- };
160
- try {
161
- return chatModelFactory(config);
162
- } catch (e) {
163
- throw normalizeAgentLlMError(e, `agent-llm: failed to create ChatModel for provider "${provider}"`);
164
- }
165
- }
166
- const model = modelEnv ?? defaultConfig?.model ?? process.env.OPENAI_MODEL ?? DEFAULT_MODEL;
167
- let apiKey = apiKeyEnv ?? defaultConfig?.apiKey ?? process.env.OPENAI_API_KEY;
168
- let baseURL = defaultConfig?.baseURL;
169
- if (baseURL && !baseURL.replace(/\/$/, "").endsWith("/v1")) {
170
- baseURL = baseURL.replace(/\/$/, "") + "/v1";
171
- }
172
- if (baseURL && apiKey === void 0) {
173
- apiKey = "ollama";
144
+ const msg = e instanceof Error ? e.message : String(e);
145
+ throw new Error(`agent-llm: failed to read config file ${filePath}: ${msg}`, { cause: e });
174
146
  }
175
- const temperature = typeof defaultConfig?.temperature === "number" ? defaultConfig.temperature : 0;
176
- const constructorOptions = {
177
- model,
178
- temperature,
179
- ...apiKey ? { apiKey } : {},
180
- ...baseURL ? { configuration: { baseURL } } : {}
181
- };
182
147
  try {
183
- return new ChatOpenAI(constructorOptions);
148
+ const llm = parseLlmYaml(raw, options);
149
+ return llm ?? null;
184
150
  } catch (e) {
185
- throw normalizeAgentLlMError(e, "agent-llm: failed to create ChatOpenAI from config");
151
+ const msg = e instanceof Error ? e.message : String(e);
152
+ throw new Error(`agent-llm: failed to parse config file ${filePath}: ${msg}`, { cause: e });
186
153
  }
187
154
  }
188
155
 
@@ -231,6 +198,31 @@ function createOpenAIChatClient(config) {
231
198
  const usage = resp.usage ? { promptTokens: resp.usage.prompt_tokens, completionTokens: resp.usage.completion_tokens } : void 0;
232
199
  return { content, usage };
233
200
  },
201
+ async *chatStream(messages) {
202
+ const stream = await client.chat.completions.create({
203
+ model,
204
+ temperature,
205
+ messages: messages.map((m) => ({ role: m.role, content: m.content })),
206
+ stream: true
207
+ });
208
+ let usage;
209
+ for await (const chunk of stream) {
210
+ const delta = chunk.choices[0]?.delta;
211
+ const content = delta?.content ?? "";
212
+ const finishReason = chunk.choices[0]?.finish_reason;
213
+ if (chunk.usage) {
214
+ usage = {
215
+ promptTokens: chunk.usage.prompt_tokens,
216
+ completionTokens: chunk.usage.completion_tokens
217
+ };
218
+ }
219
+ yield {
220
+ content,
221
+ done: finishReason !== null && finishReason !== void 0,
222
+ usage
223
+ };
224
+ }
225
+ },
234
226
  async chatWithTools(messages, tools, _options) {
235
227
  const resp = await client.chat.completions.create({
236
228
  model,
@@ -262,6 +254,52 @@ function createOpenAIChatClient(config) {
262
254
  },
263
255
  usage
264
256
  };
257
+ },
258
+ async *chatWithToolsStream(messages, tools, _options) {
259
+ const stream = await client.chat.completions.create({
260
+ model,
261
+ temperature,
262
+ messages: messages.map(serializeMessage),
263
+ tools: tools.map((t) => ({
264
+ type: "function",
265
+ function: {
266
+ name: t.function.name,
267
+ description: t.function.description,
268
+ parameters: t.function.parameters ?? void 0
269
+ }
270
+ })),
271
+ stream: true,
272
+ stream_options: { include_usage: true }
273
+ });
274
+ let usage;
275
+ for await (const chunk of stream) {
276
+ const delta = chunk.choices[0]?.delta;
277
+ const finishReason = chunk.choices[0]?.finish_reason;
278
+ if (chunk.usage) {
279
+ usage = {
280
+ promptTokens: chunk.usage.prompt_tokens,
281
+ completionTokens: chunk.usage.completion_tokens
282
+ };
283
+ }
284
+ const toolCalls = delta?.tool_calls?.map((tc) => ({
285
+ index: tc.index,
286
+ id: tc.id,
287
+ type: tc.type,
288
+ function: tc.function ? {
289
+ name: tc.function.name,
290
+ arguments: tc.function.arguments
291
+ } : void 0
292
+ }));
293
+ yield {
294
+ delta: {
295
+ role: delta?.role,
296
+ content: delta?.content ?? null,
297
+ tool_calls: toolCalls
298
+ },
299
+ done: finishReason !== null && finishReason !== void 0,
300
+ usage
301
+ };
302
+ }
265
303
  }
266
304
  };
267
305
  }
@@ -316,8 +354,86 @@ function registerProvider(name, factory) {
316
354
  PROVIDERS[name.toLowerCase()] = factory;
317
355
  }
318
356
 
319
- // src/loadLLMExtensions.ts
320
- import { readdirSync, readFileSync, existsSync } from "fs";
357
+ // src/registry/chat-model.ts
358
+ var CHAT_MODEL_FACTORIES = /* @__PURE__ */ new Map();
359
+ function registerChatModelProvider(providerName, factory) {
360
+ CHAT_MODEL_FACTORIES.set(providerName.toLowerCase(), factory);
361
+ }
362
+ function getChatModelFactory(providerName) {
363
+ return CHAT_MODEL_FACTORIES.get(providerName.toLowerCase());
364
+ }
365
+
366
+ // src/langchain/adapter.ts
367
+ import { ChatOpenAI } from "@langchain/openai";
368
+ var DEFAULT_MODEL = "gpt-4o-mini";
369
+ function normalizeAgentLlMError(e, context) {
370
+ if (e instanceof Error) return new Error(`${context}: ${e.message}`, { cause: e });
371
+ return new Error(`${context}: ${String(e)}`);
372
+ }
373
+ function createChatModelFromLlmConfig(options) {
374
+ const { llmSection, modelEnv, apiKeyEnv } = options;
375
+ let defaultId;
376
+ let configs;
377
+ try {
378
+ const parsed = parseLlmSection(llmSection ?? null);
379
+ defaultId = parsed.defaultId;
380
+ configs = parsed.configs;
381
+ } catch (e) {
382
+ throw normalizeAgentLlMError(e, "agent-llm: failed to parse llm section");
383
+ }
384
+ const defaultConfig = configs.find((c) => c.id === defaultId) ?? configs[0];
385
+ if (!defaultConfig) {
386
+ const model2 = modelEnv ?? process.env.OPENAI_MODEL ?? DEFAULT_MODEL;
387
+ const apiKey2 = apiKeyEnv ?? process.env.OPENAI_API_KEY;
388
+ try {
389
+ return new ChatOpenAI({
390
+ model: model2,
391
+ temperature: 0,
392
+ ...apiKey2 ? { apiKey: apiKey2 } : {}
393
+ });
394
+ } catch (e) {
395
+ throw normalizeAgentLlMError(e, "agent-llm: failed to create default ChatOpenAI");
396
+ }
397
+ }
398
+ const provider = defaultConfig.provider ?? "openai";
399
+ const chatModelFactory = getChatModelFactory(provider);
400
+ if (chatModelFactory) {
401
+ const config = {
402
+ ...defaultConfig,
403
+ model: modelEnv ?? defaultConfig.model,
404
+ temperature: typeof defaultConfig.temperature === "number" ? defaultConfig.temperature : 0
405
+ };
406
+ try {
407
+ return chatModelFactory(config);
408
+ } catch (e) {
409
+ throw normalizeAgentLlMError(e, `agent-llm: failed to create ChatModel for provider "${provider}"`);
410
+ }
411
+ }
412
+ const model = modelEnv ?? defaultConfig?.model ?? process.env.OPENAI_MODEL ?? DEFAULT_MODEL;
413
+ let apiKey = apiKeyEnv ?? defaultConfig?.apiKey ?? process.env.OPENAI_API_KEY;
414
+ let baseURL = defaultConfig?.baseURL;
415
+ if (baseURL && !baseURL.replace(/\/$/, "").endsWith("/v1")) {
416
+ baseURL = baseURL.replace(/\/$/, "") + "/v1";
417
+ }
418
+ if (baseURL && apiKey === void 0) {
419
+ apiKey = "ollama";
420
+ }
421
+ const temperature = typeof defaultConfig?.temperature === "number" ? defaultConfig.temperature : 0;
422
+ const constructorOptions = {
423
+ model,
424
+ temperature,
425
+ ...apiKey ? { apiKey } : {},
426
+ ...baseURL ? { configuration: { baseURL } } : {}
427
+ };
428
+ try {
429
+ return new ChatOpenAI(constructorOptions);
430
+ } catch (e) {
431
+ throw normalizeAgentLlMError(e, "agent-llm: failed to create ChatOpenAI from config");
432
+ }
433
+ }
434
+
435
+ // src/extensions/loader.ts
436
+ import { readdirSync, readFileSync as readFileSync2, existsSync as existsSync2 } from "fs";
321
437
  import { join } from "path";
322
438
  import { pathToFileURL } from "url";
323
439
  var loadedPackages = /* @__PURE__ */ new Set();
@@ -356,9 +472,9 @@ function resolveLLMExtensionPackages(types) {
356
472
  );
357
473
  }
358
474
  function readPackageProviderName(pkgPath) {
359
- if (!existsSync(pkgPath)) return null;
475
+ if (!existsSync2(pkgPath)) return null;
360
476
  try {
361
- const raw = readFileSync(pkgPath, "utf-8");
477
+ const raw = readFileSync2(pkgPath, "utf-8");
362
478
  const pkg = JSON.parse(raw);
363
479
  const declared = pkg.agentLlmProvider === true || Array.isArray(pkg.keywords) && pkg.keywords.includes("agent-llm-provider");
364
480
  return declared && typeof pkg.name === "string" ? pkg.name : null;
@@ -369,7 +485,7 @@ function readPackageProviderName(pkgPath) {
369
485
  function discoverLLMExtensions(cwd = process.cwd()) {
370
486
  const dir = typeof cwd === "string" && cwd.trim().length > 0 ? cwd : process.cwd();
371
487
  const nodeModules = join(dir, "node_modules");
372
- if (!existsSync(nodeModules)) return [];
488
+ if (!existsSync2(nodeModules)) return [];
373
489
  const names = [];
374
490
  const seen = /* @__PURE__ */ new Set();
375
491
  try {
@@ -401,13 +517,13 @@ function discoverLLMExtensions(cwd = process.cwd()) {
401
517
  }
402
518
  async function loadModuleFromPath(pkgDir) {
403
519
  const pkgJsonPath = join(pkgDir, "package.json");
404
- if (!existsSync(pkgJsonPath)) {
520
+ if (!existsSync2(pkgJsonPath)) {
405
521
  throw new Error(`package.json not found in ${pkgDir}`);
406
522
  }
407
- const pkgJson = JSON.parse(readFileSync(pkgJsonPath, "utf-8"));
523
+ const pkgJson = JSON.parse(readFileSync2(pkgJsonPath, "utf-8"));
408
524
  const main = pkgJson?.main ?? "index.js";
409
525
  const entryPath = join(pkgDir, main);
410
- if (!existsSync(entryPath)) {
526
+ if (!existsSync2(entryPath)) {
411
527
  throw new Error(`Entry ${main} not found in ${pkgDir}`);
412
528
  }
413
529
  const entryUrl = pathToFileURL(entryPath).href;
@@ -424,9 +540,9 @@ async function loadLLMExtensions(extensionPackages, options) {
424
540
  loadedPackages.add(pkg);
425
541
  let loaded = false;
426
542
  const cwdPkgDir = join(cwd, "node_modules", pkg);
427
- const cwdIsProject = existsSync(join(cwd, "package.json")) && (() => {
543
+ const cwdIsProject = existsSync2(join(cwd, "package.json")) && (() => {
428
544
  try {
429
- const name = JSON.parse(readFileSync(join(cwd, "package.json"), "utf-8")).name;
545
+ const name = JSON.parse(readFileSync2(join(cwd, "package.json"), "utf-8")).name;
430
546
  return name === pkg;
431
547
  } catch {
432
548
  return false;
@@ -434,7 +550,7 @@ async function loadLLMExtensions(extensionPackages, options) {
434
550
  })();
435
551
  const dirsToTry = [cwdPkgDir, ...cwdIsProject ? [cwd] : []];
436
552
  for (const pkgDir of dirsToTry) {
437
- if (!existsSync(join(pkgDir, "package.json"))) continue;
553
+ if (!existsSync2(join(pkgDir, "package.json"))) continue;
438
554
  try {
439
555
  const m = await loadModuleFromPath(pkgDir);
440
556
  const ext = getExtensionFromModule(m);
@@ -480,7 +596,7 @@ async function loadDiscoveredExtensions(cwd = process.cwd()) {
480
596
  return names;
481
597
  }
482
598
 
483
- // src/npmProviderProtocol.ts
599
+ // src/extensions/npm-protocol.ts
484
600
  import { execSync } from "child_process";
485
601
  var NPM_PROTOCOL_PREFIX = "npm:";
486
602
  function parseNpmProviderSpec(spec) {
@@ -636,52 +752,7 @@ async function createChatModelFromLlmConfigWithNpm(options) {
636
752
  }
637
753
  }
638
754
 
639
- // src/loadLlmConfig.ts
640
- import { readFileSync as readFileSync2, existsSync as existsSync2 } from "fs";
641
- import { parse as parseYaml } from "yaml";
642
- function substituteEnv(obj) {
643
- if (obj === null || obj === void 0) return obj;
644
- if (typeof obj === "string") {
645
- const m = obj.match(/^\$\{(\w+)\}$/);
646
- return m ? process.env[m[1]] ?? obj : obj;
647
- }
648
- if (Array.isArray(obj)) return obj.map(substituteEnv);
649
- if (typeof obj === "object") {
650
- const out = {};
651
- for (const [k, v] of Object.entries(obj)) out[k] = substituteEnv(v);
652
- return out;
653
- }
654
- return obj;
655
- }
656
- function parseLlmYaml(content, options = {}) {
657
- const { substituteEnv: doSub = true } = options;
658
- const parsed = parseYaml(content);
659
- const llm = parsed?.llm;
660
- if (llm == null) return void 0;
661
- return doSub ? substituteEnv(llm) : llm;
662
- }
663
- function loadLlmConfig(filePath, options = {}) {
664
- if (typeof filePath !== "string" || filePath.trim().length === 0) {
665
- throw new Error("agent-llm: loadLlmConfig requires a non-empty file path");
666
- }
667
- if (!existsSync2(filePath)) return null;
668
- let raw;
669
- try {
670
- raw = readFileSync2(filePath, "utf8");
671
- } catch (e) {
672
- const msg = e instanceof Error ? e.message : String(e);
673
- throw new Error(`agent-llm: failed to read config file ${filePath}: ${msg}`, { cause: e });
674
- }
675
- try {
676
- const llm = parseLlmYaml(raw, options);
677
- return llm ?? null;
678
- } catch (e) {
679
- const msg = e instanceof Error ? e.message : String(e);
680
- throw new Error(`agent-llm: failed to parse config file ${filePath}: ${msg}`, { cause: e });
681
- }
682
- }
683
-
684
- // src/createAgentLlM.ts
755
+ // src/api/create-agent-llm.ts
685
756
  import { join as join2 } from "path";
686
757
  function resolveDefaultConfigPath() {
687
758
  const cwd = process.cwd();
@@ -717,14 +788,17 @@ async function createAgentLlM(configPathOrOptions) {
717
788
 
718
789
  export {
719
790
  parseLlmSection,
720
- registerChatModelProvider,
721
- getChatModelFactory,
722
- createChatModelFromLlmConfig,
791
+ substituteEnv,
792
+ parseLlmYaml,
793
+ loadLlmConfig,
723
794
  createOpenAIChatClient,
724
795
  createOpenAIImageClient,
725
796
  createOpenAIClient,
726
797
  createClient,
727
798
  registerProvider,
799
+ registerChatModelProvider,
800
+ getChatModelFactory,
801
+ createChatModelFromLlmConfig,
728
802
  resolveLLMExtensionPackages,
729
803
  discoverLLMExtensions,
730
804
  loadLLMExtensions,
@@ -736,9 +810,6 @@ export {
736
810
  resolveNpmProvider,
737
811
  resolveLlmSectionWithNpm,
738
812
  createChatModelFromLlmConfigWithNpm,
739
- substituteEnv,
740
- parseLlmYaml,
741
- loadLlmConfig,
742
813
  createAgentLlM
743
814
  };
744
- //# sourceMappingURL=chunk-HKET766S.js.map
815
+ //# sourceMappingURL=chunk-MYGWM3YH.js.map
@@ -0,0 +1 @@
1
+ {"version":3,"sources":["../src/config/parser.ts","../src/config/loader.ts","../src/providers/openai.ts","../src/providers/index.ts","../src/registry/chat-model.ts","../src/langchain/adapter.ts","../src/extensions/loader.ts","../src/extensions/npm-protocol.ts","../src/api/create-agent-llm.ts"],"sourcesContent":["/**\n * Parse agent.yaml llm section into normalized LLMConfig[] and default id.\n * Supports: flat (each model keyed by name), instances[], or single object.\n */\n\nimport type { LLMConfig } from \"../types.js\";\n\nconst DEFAULT_LLM_ID = \"default\";\n\nconst RESERVED_KEYS = new Set([\n \"default\",\n \"instances\",\n \"catalog\",\n \"provider\",\n \"model\",\n \"temperature\",\n \"apiKey\",\n \"baseURL\",\n \"base_url\",\n \"type\",\n \"id\",\n]);\n\n/**\n * Parse llm section: flat (each model keyed by name), default+instances, or single object.\n */\nexport function parseLlmSection(section: unknown): { defaultId: string; configs: LLMConfig[] } {\n if (section == null || typeof section !== \"object\") {\n return { defaultId: DEFAULT_LLM_ID, configs: [] };\n }\n\n if (Array.isArray(section)) {\n const configs = section\n .filter((i): i is Record<string, unknown> => i != null && typeof i === \"object\")\n .map((item, i) => normalizeLlmConfig({ ...item, id: item.id ?? item.name ?? String(i) }))\n .filter((c): c is LLMConfig => c != null);\n const defaultId = configs.length > 0 ? configs[0]!.id : DEFAULT_LLM_ID;\n return { defaultId, configs };\n }\n\n const s = section as Record<string, unknown>;\n\n const flatEntries = Object.entries(s).filter(\n ([k, v]) => !RESERVED_KEYS.has(k) && v != null && typeof v === \"object\" && !Array.isArray(v)\n );\n if (flatEntries.length > 0) {\n const configs: LLMConfig[] = [];\n for (const [id, entry] of flatEntries) {\n const c = entryToLlmConfig(id, entry as Record<string, unknown>);\n if (c) configs.push(c);\n }\n const defaultId =\n typeof s.default === \"string\" && s.default && flatEntries.some(([k]) => k === s.default)\n ? s.default\n : configs.length > 0\n ? configs[0]!.id\n : DEFAULT_LLM_ID;\n return { defaultId, configs };\n }\n\n if (Array.isArray(s.instances)) {\n const configs = (s.instances as unknown[])\n .filter((i): i is Record<string, unknown> => i != null && typeof i === \"object\")\n .map((i) => normalizeLlmConfig(i))\n .filter((c): c is LLMConfig => c != null);\n const defaultId =\n typeof s.default === \"string\" && s.default\n ? s.default\n : configs.length > 0\n ? configs[0]!.id\n : DEFAULT_LLM_ID;\n return { defaultId, configs };\n }\n\n if (typeof s.provider === \"string\" || typeof s.model === \"string\" || typeof (s as { name?: string }).name === \"string\") {\n const one = singleObjectToLlmConfig(s);\n return { defaultId: one.id, configs: [one] };\n }\n\n return { defaultId: DEFAULT_LLM_ID, configs: [] };\n}\n\nconst EXTENSION_OPTION_KEYS = [\"featureKey\", \"tenant\", \"authToken\", \"verifySSL\", \"bypassAuth\", \"host\", \"resolveHost\", \"timeoutMs\", \"options\"];\n\nfunction entryToLlmConfig(id: string, entry: Record<string, unknown>): LLMConfig | null {\n const opts = entry.options as Record<string, unknown> | undefined;\n const baseURL =\n typeof entry.base_url === \"string\"\n ? entry.base_url\n : typeof entry.baseURL === \"string\"\n ? entry.baseURL\n : undefined;\n const model = typeof entry.name === \"string\" ? entry.name : typeof entry.model === \"string\" ? entry.model : undefined;\n const provider = typeof entry.provider === \"string\" && entry.provider ? entry.provider : \"openai\";\n const config: LLMConfig = {\n id,\n type: \"chat\",\n provider,\n model,\n temperature: typeof opts?.temperature === \"number\" ? opts.temperature : typeof entry.temperature === \"number\" ? entry.temperature : undefined,\n apiKey: typeof opts?.apiKey === \"string\" ? opts.apiKey : typeof entry.apiKey === \"string\" ? entry.apiKey : undefined,\n baseURL,\n };\n if (typeof entry.type === \"string\" && entry.type === \"image\") config.type = \"image\";\n if (opts && typeof opts === \"object\") (config as Record<string, unknown>).options = opts;\n for (const k of EXTENSION_OPTION_KEYS) {\n if (entry[k] !== undefined) (config as Record<string, unknown>)[k] = entry[k];\n else if (opts && opts[k] !== undefined) (config as Record<string, unknown>)[k] = opts[k];\n }\n return config;\n}\n\nfunction singleObjectToLlmConfig(s: Record<string, unknown>): LLMConfig {\n const one: LLMConfig = {\n id: DEFAULT_LLM_ID,\n type: \"chat\",\n provider: typeof s.provider === \"string\" ? s.provider : \"openai\",\n model: typeof s.model === \"string\" ? s.model : (typeof (s as { name?: string }).name === \"string\" ? (s as { name: string }).name : undefined),\n temperature: typeof s.temperature === \"number\" ? s.temperature : undefined,\n apiKey: typeof s.apiKey === \"string\" ? s.apiKey : undefined,\n baseURL:\n typeof s.baseURL === \"string\" ? s.baseURL : typeof s.base_url === \"string\" ? s.base_url : undefined,\n };\n Object.keys(s).forEach((k) => {\n if (![\"id\", \"type\", \"provider\", \"model\", \"name\", \"temperature\", \"apiKey\", \"baseURL\", \"base_url\", \"default\", \"instances\"].includes(k)) {\n (one as Record<string, unknown>)[k] = s[k];\n }\n });\n return one;\n}\n\nfunction normalizeLlmConfig(o: Record<string, unknown>): LLMConfig | null {\n const id = typeof o.id === \"string\" && o.id ? o.id : DEFAULT_LLM_ID;\n const type = o.type === \"image\" ? \"image\" : \"chat\";\n const provider = typeof o.provider === \"string\" && o.provider ? o.provider : \"openai\";\n const opts = o.options as Record<string, unknown> | undefined;\n const config: LLMConfig = {\n id,\n type,\n provider,\n model: typeof o.model === \"string\" ? o.model : (typeof o.name === \"string\" ? o.name : undefined),\n temperature:\n typeof o.temperature === \"number\"\n ? o.temperature\n : typeof opts?.temperature === \"number\"\n ? opts.temperature\n : undefined,\n apiKey:\n typeof o.apiKey === \"string\"\n ? o.apiKey\n : typeof opts?.apiKey === \"string\"\n ? opts.apiKey\n : undefined,\n baseURL: typeof o.baseURL === \"string\" ? o.baseURL : (typeof o.base_url === \"string\" ? o.base_url : undefined),\n };\n Object.keys(o).forEach((k) => {\n if (![\"id\", \"type\", \"provider\", \"model\", \"name\", \"temperature\", \"apiKey\", \"baseURL\", \"base_url\"].includes(k)) {\n (config as Record<string, unknown>)[k] = o[k];\n }\n });\n return config;\n}\n","/**\n * Load and parse LLM config from YAML (e.g. llm.yaml).\n * Supports ${VAR} substitution from process.env.\n */\n\nimport { readFileSync, existsSync } from \"node:fs\";\nimport { parse as parseYaml } from \"yaml\";\n\nexport interface LoadLlmConfigOptions {\n /** Replace ${VAR} with process.env.VAR. Default true. */\n substituteEnv?: boolean;\n}\n\n/**\n * Recursively replace ${VAR} in strings with process.env.VAR.\n */\nexport function substituteEnv(obj: unknown): unknown {\n if (obj === null || obj === undefined) return obj;\n if (typeof obj === \"string\") {\n const m = obj.match(/^\\$\\{(\\w+)\\}$/);\n return m ? (process.env[m[1]] ?? obj) : obj;\n }\n if (Array.isArray(obj)) return obj.map(substituteEnv);\n if (typeof obj === \"object\") {\n const out: Record<string, unknown> = {};\n for (const [k, v] of Object.entries(obj)) out[k] = substituteEnv(v);\n return out;\n }\n return obj;\n}\n\n/**\n * Parse YAML string and return the llm section (top-level key \"llm\").\n * Returns undefined if content has no llm key.\n */\nexport function parseLlmYaml(\n content: string,\n options: LoadLlmConfigOptions = {}\n): unknown {\n const { substituteEnv: doSub = true } = options;\n const parsed = parseYaml(content) as { llm?: unknown };\n const llm = parsed?.llm;\n if (llm == null) return undefined;\n return doSub ? substituteEnv(llm) : llm;\n}\n\n/**\n * Load LLM config from a YAML file (e.g. llm.yaml in current directory).\n * Returns the llm section for use with createChatModelFromLlmConfig or parseLlmSection.\n * Returns null if file does not exist or has no llm key.\n * Throws with clear message if file exists but read or parse fails.\n */\nexport function loadLlmConfig(\n filePath: string,\n options: LoadLlmConfigOptions = {}\n): unknown | null {\n if (typeof filePath !== \"string\" || filePath.trim().length === 0) {\n throw new Error(\"agent-llm: loadLlmConfig requires a non-empty file path\");\n }\n if (!existsSync(filePath)) return null;\n let raw: string;\n try {\n raw = readFileSync(filePath, \"utf8\");\n } catch (e) {\n const msg = e instanceof Error ? e.message : String(e);\n throw new Error(`agent-llm: failed to read config file ${filePath}: ${msg}`, { cause: e });\n }\n try {\n const llm = parseLlmYaml(raw, options);\n return llm ?? null;\n } catch (e) {\n const msg = e instanceof Error ? e.message : String(e);\n throw new Error(`agent-llm: failed to parse config file ${filePath}: ${msg}`, { cause: e });\n }\n}\n","/**\n * OpenAI-compatible format: chat (/v1/chat/completions) and image.\n * Supports baseURL for Azure, local proxy, and other compatible endpoints.\n */\n\nimport OpenAI from \"openai\";\nimport type {\n LLMConfig,\n ChatMessage,\n ChatResult,\n ImageResult,\n ILLMClient,\n ChatWithToolsMessage,\n ChatWithToolsResult,\n ToolDefinition,\n ChatStreamChunk,\n ChatWithToolsStreamChunk,\n} from \"../types.js\";\n\nfunction getApiKey(config: LLMConfig): string {\n const key = config.apiKey ?? process.env.OPENAI_API_KEY ?? \"\";\n if (!key) throw new Error(\"OpenAI-compatible apiKey required (config.apiKey or OPENAI_API_KEY)\");\n return key;\n}\n\nfunction createOpenAIClientOptions(config: LLMConfig): { apiKey: string; baseURL?: string } {\n const opts: { apiKey: string; baseURL?: string } = { apiKey: getApiKey(config) };\n if (typeof config.baseURL === \"string\" && config.baseURL) opts.baseURL = config.baseURL;\n return opts;\n}\n\nfunction serializeMessage(\n m: ChatWithToolsMessage\n): OpenAI.Chat.Completions.ChatCompletionMessageParam {\n if (m.role === \"tool\")\n return { role: \"tool\", content: m.content, tool_call_id: m.tool_call_id };\n if (m.role === \"assistant\" && \"tool_calls\" in m && m.tool_calls?.length) {\n return {\n role: \"assistant\",\n content: m.content ?? null,\n tool_calls: m.tool_calls.map((tc) => ({\n id: tc.id,\n type: \"function\" as const,\n function: { name: tc.function.name, arguments: tc.function.arguments },\n })),\n };\n }\n return { role: m.role, content: (m as ChatMessage).content };\n}\n\nexport function createOpenAIChatClient(config: LLMConfig): ILLMClient {\n const client = new OpenAI(createOpenAIClientOptions(config));\n const model = config.model ?? process.env.OPENAI_MODEL ?? \"gpt-4o-mini\";\n const temperature = config.temperature ?? 0;\n\n return {\n id: config.id,\n type: \"chat\",\n async chat(messages: ChatMessage[]): Promise<ChatResult> {\n const resp = await client.chat.completions.create({\n model,\n temperature,\n messages: messages.map((m) => ({ role: m.role, content: m.content })),\n });\n const content = resp.choices[0]?.message?.content ?? \"\";\n const usage = resp.usage\n ? { promptTokens: resp.usage.prompt_tokens, completionTokens: resp.usage.completion_tokens }\n : undefined;\n return { content, usage };\n },\n async *chatStream(messages: ChatMessage[]): AsyncGenerator<ChatStreamChunk, void, unknown> {\n const stream = await client.chat.completions.create({\n model,\n temperature,\n messages: messages.map((m) => ({ role: m.role, content: m.content })),\n stream: true,\n });\n\n let usage: { promptTokens?: number; completionTokens?: number } | undefined;\n\n for await (const chunk of stream) {\n const delta = chunk.choices[0]?.delta;\n const content = delta?.content ?? \"\";\n const finishReason = chunk.choices[0]?.finish_reason;\n\n // Capture usage from the final chunk if available\n if (chunk.usage) {\n usage = {\n promptTokens: chunk.usage.prompt_tokens,\n completionTokens: chunk.usage.completion_tokens,\n };\n }\n\n yield {\n content,\n done: finishReason !== null && finishReason !== undefined,\n usage,\n };\n }\n },\n async chatWithTools(\n messages: ChatWithToolsMessage[],\n tools: ToolDefinition[],\n _options?: { timeoutMs?: number }\n ): Promise<ChatWithToolsResult> {\n const resp = await client.chat.completions.create({\n model,\n temperature,\n messages: messages.map(serializeMessage),\n tools: tools.map((t) => ({\n type: \"function\" as const,\n function: {\n name: t.function.name,\n description: t.function.description,\n parameters: (t.function.parameters ?? undefined) as Record<string, unknown> | undefined,\n },\n })),\n });\n const msg = resp.choices[0]?.message;\n const usage = resp.usage\n ? { promptTokens: resp.usage.prompt_tokens, completionTokens: resp.usage.completion_tokens }\n : undefined;\n return {\n message: {\n role: \"assistant\",\n content: msg?.content ?? null,\n tool_calls: msg?.tool_calls?.map((tc) => ({\n id: tc.id,\n type: \"function\" as const,\n function: {\n name: tc.function?.name ?? \"\",\n arguments: tc.function?.arguments ?? \"\",\n },\n })),\n },\n usage,\n };\n },\n async *chatWithToolsStream(\n messages: ChatWithToolsMessage[],\n tools: ToolDefinition[],\n _options?: { timeoutMs?: number }\n ): AsyncGenerator<ChatWithToolsStreamChunk, void, unknown> {\n const stream = await client.chat.completions.create({\n model,\n temperature,\n messages: messages.map(serializeMessage),\n tools: tools.map((t) => ({\n type: \"function\" as const,\n function: {\n name: t.function.name,\n description: t.function.description,\n parameters: (t.function.parameters ?? undefined) as Record<string, unknown> | undefined,\n },\n })),\n stream: true,\n stream_options: { include_usage: true },\n });\n\n let usage: { promptTokens?: number; completionTokens?: number } | undefined;\n\n for await (const chunk of stream) {\n const delta = chunk.choices[0]?.delta;\n const finishReason = chunk.choices[0]?.finish_reason;\n\n // Capture usage from the final chunk if available\n if (chunk.usage) {\n usage = {\n promptTokens: chunk.usage.prompt_tokens,\n completionTokens: chunk.usage.completion_tokens,\n };\n }\n\n const toolCalls = delta?.tool_calls?.map((tc) => ({\n index: tc.index,\n id: tc.id,\n type: tc.type as \"function\" | undefined,\n function: tc.function\n ? {\n name: tc.function.name,\n arguments: tc.function.arguments,\n }\n : undefined,\n }));\n\n yield {\n delta: {\n role: delta?.role as \"assistant\" | undefined,\n content: delta?.content ?? null,\n tool_calls: toolCalls,\n },\n done: finishReason !== null && finishReason !== undefined,\n usage,\n };\n }\n },\n };\n}\n\nexport function createOpenAIImageClient(config: LLMConfig): ILLMClient {\n const client = new OpenAI(createOpenAIClientOptions(config));\n const model = (config.model as string) ?? \"dall-e-3\";\n\n return {\n id: config.id,\n type: \"image\",\n async chat(): Promise<ChatResult> {\n throw new Error(\"OpenAI image model does not support chat; use generateImage()\");\n },\n async generateImage(options: { prompt: string; size?: string; n?: number }): Promise<ImageResult> {\n const resp = await client.images.generate({\n model,\n prompt: options.prompt,\n size: (options.size as \"1024x1024\" | \"1792x1024\" | \"1024x1792\") ?? \"1024x1024\",\n n: options.n ?? 1,\n response_format: \"url\",\n });\n const url = resp.data?.[0]?.url ?? undefined;\n return { url };\n },\n };\n}\n\nexport function createOpenAIClient(config: LLMConfig): ILLMClient {\n if (config.type === \"image\") return createOpenAIImageClient(config);\n return createOpenAIChatClient(config);\n}\n","/**\n * Supports OpenAI-compatible and extension providers.\n */\n\nimport type { LLMConfig, ILLMClient } from \"../types.js\";\nimport { createOpenAIClient } from \"./openai.js\";\n\nconst OPENAI_COMPATIBLE = \"openai-compatible\";\n\nfunction createOpenAICompat(config: LLMConfig): ILLMClient {\n return createOpenAIClient(config);\n}\n\nconst PROVIDERS: Record<string, (config: LLMConfig) => ILLMClient> = {\n openai: createOpenAICompat,\n [OPENAI_COMPATIBLE]: createOpenAICompat,\n};\n\nexport function createClient(config: LLMConfig): ILLMClient {\n const p = (config.provider ?? \"\").toLowerCase();\n const fn = PROVIDERS[p];\n if (!fn) {\n const supported = [...new Set([...Object.keys(PROVIDERS), \"extension providers\"])].sort().join(\", \");\n throw new Error(\n `Unsupported LLM provider: ${config.provider}. Supported: ${supported}.`\n );\n }\n return fn(config);\n}\n\nexport function registerProvider(name: string, factory: (config: LLMConfig) => ILLMClient): void {\n PROVIDERS[name.toLowerCase()] = factory;\n}\n","/**\n * Registry for LangChain ChatModel by provider name.\n * Extensions register via registerChatModelProvider; llmAdapter uses getChatModelFactory.\n */\n\nimport type { BaseChatModel } from \"@langchain/core/language_models/chat_models\";\nimport type { LLMConfig } from \"../types.js\";\n\nexport type ChatModelFactory = (config: LLMConfig) => BaseChatModel;\n\nconst CHAT_MODEL_FACTORIES = new Map<string, ChatModelFactory>();\n\n/**\n * Register a ChatModel factory for a provider name.\n * Called by extensions (e.g. wallee-llm) on load.\n */\nexport function registerChatModelProvider(providerName: string, factory: ChatModelFactory): void {\n CHAT_MODEL_FACTORIES.set(providerName.toLowerCase(), factory);\n}\n\n/**\n * Get the ChatModel factory for a provider name, if registered.\n */\nexport function getChatModelFactory(providerName: string): ChatModelFactory | undefined {\n return CHAT_MODEL_FACTORIES.get(providerName.toLowerCase());\n}\n","/**\n * Build LangChain ChatModel from agent.yaml llm section.\n * Supports single object, default + instances, and flat keyed configs.\n * When provider is registered by an extension, uses that extension's ChatModel;\n * otherwise uses ChatOpenAI.\n */\n\nimport { ChatOpenAI } from \"@langchain/openai\";\nimport type { BaseChatModel } from \"@langchain/core/language_models/chat_models\";\nimport { parseLlmSection } from \"../config/parser.js\";\nimport { getChatModelFactory } from \"../registry/chat-model.js\";\n\nconst DEFAULT_MODEL = \"gpt-4o-mini\";\n\nexport interface CreateChatModelFromLlmConfigOptions {\n /** agent.yaml llm section (raw or parsed); compatible with AgentConfigLlmSection / AgentConfigLlm */\n llmSection?: unknown;\n /** Override model from env */\n modelEnv?: string;\n /** Override API key from env */\n apiKeyEnv?: string;\n}\n\nfunction normalizeAgentLlMError(e: unknown, context: string): Error {\n if (e instanceof Error) return new Error(`${context}: ${e.message}`, { cause: e });\n return new Error(`${context}: ${String(e)}`);\n}\n\n/**\n * Create a LangChain ChatModel from agent config llm section.\n * Uses extension-registered ChatModel when available; otherwise ChatOpenAI.\n */\nexport function createChatModelFromLlmConfig(\n options: CreateChatModelFromLlmConfigOptions\n): BaseChatModel {\n const { llmSection, modelEnv, apiKeyEnv } = options;\n let defaultId: string;\n let configs: ReturnType<typeof parseLlmSection>[\"configs\"];\n try {\n const parsed = parseLlmSection(llmSection ?? null);\n defaultId = parsed.defaultId;\n configs = parsed.configs;\n } catch (e) {\n throw normalizeAgentLlMError(e, \"agent-llm: failed to parse llm section\");\n }\n const defaultConfig = configs.find((c) => c.id === defaultId) ?? configs[0];\n\n if (!defaultConfig) {\n const model =\n modelEnv ?? process.env.OPENAI_MODEL ?? DEFAULT_MODEL;\n const apiKey = apiKeyEnv ?? process.env.OPENAI_API_KEY;\n try {\n return new ChatOpenAI({\n model,\n temperature: 0,\n ...(apiKey ? { apiKey } : {}),\n });\n } catch (e) {\n throw normalizeAgentLlMError(e, \"agent-llm: failed to create default ChatOpenAI\");\n }\n }\n\n const provider = (defaultConfig as { provider?: string }).provider ?? \"openai\";\n const chatModelFactory = getChatModelFactory(provider);\n if (chatModelFactory) {\n const config = {\n ...defaultConfig,\n model: modelEnv ?? defaultConfig.model,\n temperature:\n typeof defaultConfig.temperature === \"number\"\n ? defaultConfig.temperature\n : 0,\n };\n try {\n return chatModelFactory(config);\n } catch (e) {\n throw normalizeAgentLlMError(e, `agent-llm: failed to create ChatModel for provider \"${provider}\"`);\n }\n }\n\n const model =\n modelEnv ??\n defaultConfig?.model ??\n process.env.OPENAI_MODEL ??\n DEFAULT_MODEL;\n\n let apiKey =\n apiKeyEnv ?? defaultConfig?.apiKey ?? process.env.OPENAI_API_KEY;\n let baseURL = defaultConfig?.baseURL;\n // OpenAI client appends path (e.g. /chat/completions) to baseURL; Ollama and OpenAI-compatible APIs expect /v1/chat/completions.\n if (baseURL && !baseURL.replace(/\\/$/, \"\").endsWith(\"/v1\")) {\n baseURL = baseURL.replace(/\\/$/, \"\") + \"/v1\";\n }\n // OpenAI client throws if apiKey is undefined; Ollama and many compatible endpoints accept a dummy.\n if (baseURL && apiKey === undefined) {\n apiKey = \"ollama\";\n }\n\n const temperature =\n typeof defaultConfig?.temperature === \"number\" ? defaultConfig.temperature : 0;\n\n const constructorOptions: ConstructorParameters<typeof ChatOpenAI>[0] = {\n model,\n temperature,\n ...(apiKey ? { apiKey } : {}),\n ...(baseURL ? { configuration: { baseURL } } : {}),\n };\n\n try {\n return new ChatOpenAI(constructorOptions);\n } catch (e) {\n throw normalizeAgentLlMError(e, \"agent-llm: failed to create ChatOpenAI from config\");\n }\n}\n","/**\n * Load LLM extensions by npm package name or by dynamic discovery.\n * Extensions implement ILLMExtension (getLLMExtension()). Framework does registration so extension need not call register.\n */\n\nimport { readdirSync, readFileSync, existsSync } from \"node:fs\";\nimport { join } from \"node:path\";\nimport { pathToFileURL } from \"node:url\";\nimport type { ILLMExtension } from \"../types.js\";\nimport { AGENT_LLM_PROVIDER_FIELD } from \"../types.js\";\nimport { registerProvider } from \"../providers/index.js\";\nimport { registerChatModelProvider, type ChatModelFactory } from \"../registry/chat-model.js\";\n\nconst loadedPackages = new Set<string>();\n\nfunction isLLMExtension(m: unknown): m is ILLMExtension {\n if (m == null || typeof m !== \"object\") return false;\n const e = m as ILLMExtension;\n if (typeof e.providerName !== \"string\") return false;\n const hasRegister = typeof e.register === \"function\";\n const hasFactories = typeof e.createClient === \"function\" && typeof e.createChatModel === \"function\";\n return hasRegister || hasFactories;\n}\n\nfunction registerExtension(ext: ILLMExtension): void {\n if (typeof ext.createClient === \"function\" && typeof ext.createChatModel === \"function\") {\n registerProvider(ext.providerName, ext.createClient);\n registerChatModelProvider(ext.providerName, ext.createChatModel as ChatModelFactory);\n return;\n }\n if (typeof ext.register === \"function\") {\n ext.register();\n }\n}\n\nfunction getExtensionFromModule(m: Record<string, unknown>): ILLMExtension | null {\n if (isLLMExtension(m)) return m;\n if (typeof m.getLLMExtension === \"function\") {\n const ext = m.getLLMExtension();\n return isLLMExtension(ext) ? ext : null;\n }\n if (isLLMExtension(m.default)) return m.default;\n return null;\n}\n\n/**\n * Resolve llm.type to a list of npm package names to load.\n */\nexport function resolveLLMExtensionPackages(types?: string | string[]): string[] {\n if (types == null) return [];\n const typeList = Array.isArray(types) ? types : [types];\n return typeList.filter(\n (t): t is string => typeof t === \"string\" && String(t).trim().length > 0\n );\n}\n\n/**\n * Scan node_modules (optionally under cwd) for packages that declare agent-llm provider.\n * A package is included if package.json has \"agentLlmProvider\": true or keywords includes \"agent-llm-provider\".\n * @param cwd Directory containing node_modules; default process.cwd()\n * @returns List of package names that can be passed to loadLLMExtensions()\n */\nfunction readPackageProviderName(pkgPath: string): string | null {\n if (!existsSync(pkgPath)) return null;\n try {\n const raw = readFileSync(pkgPath, \"utf-8\");\n const pkg = JSON.parse(raw) as { agentLlmProvider?: boolean; keywords?: string[]; name?: string };\n const declared =\n pkg.agentLlmProvider === true ||\n (Array.isArray(pkg.keywords) && pkg.keywords.includes(\"agent-llm-provider\"));\n return declared && typeof pkg.name === \"string\" ? pkg.name : null;\n } catch {\n return null;\n }\n}\n\nexport function discoverLLMExtensions(cwd: string = process.cwd()): string[] {\n const dir = typeof cwd === \"string\" && cwd.trim().length > 0 ? cwd : process.cwd();\n const nodeModules = join(dir, \"node_modules\");\n if (!existsSync(nodeModules)) return [];\n const names: string[] = [];\n const seen = new Set<string>();\n try {\n const entries = readdirSync(nodeModules, { withFileTypes: true });\n for (const e of entries) {\n if (e.name.startsWith(\".\") || e.name === \"node\") continue;\n if (e.name.startsWith(\"@\") && e.isDirectory()) {\n const scopePath = join(nodeModules, e.name);\n const scopeEntries = readdirSync(scopePath, { withFileTypes: true });\n for (const se of scopeEntries) {\n if (!se.isDirectory()) continue;\n const name = readPackageProviderName(join(scopePath, se.name, \"package.json\"));\n if (name && !seen.has(name)) {\n seen.add(name);\n names.push(name);\n }\n }\n continue;\n }\n const name = readPackageProviderName(join(nodeModules, e.name, \"package.json\"));\n if (name && !seen.has(name)) {\n seen.add(name);\n names.push(name);\n }\n }\n } catch {\n // no node_modules or not readable\n }\n return names;\n}\n\n/**\n * Load a module from a package directory (e.g. cwd/node_modules/packageName).\n * Reads package.json \"main\" and imports that file.\n */\nasync function loadModuleFromPath(pkgDir: string): Promise<Record<string, unknown>> {\n const pkgJsonPath = join(pkgDir, \"package.json\");\n if (!existsSync(pkgJsonPath)) {\n throw new Error(`package.json not found in ${pkgDir}`);\n }\n const pkgJson = JSON.parse(readFileSync(pkgJsonPath, \"utf-8\")) as { main?: string };\n const main = pkgJson?.main ?? \"index.js\";\n const entryPath = join(pkgDir, main);\n if (!existsSync(entryPath)) {\n throw new Error(`Entry ${main} not found in ${pkgDir}`);\n }\n const entryUrl = pathToFileURL(entryPath).href;\n return (await import(/* @vite-ignore */ entryUrl)) as Record<string, unknown>;\n}\n\nexport interface LoadLLMExtensionsOptions {\n /** If set, try loading each package from cwd/node_modules/<package> first (so npx @easynet/agent-llm from a project dir finds local provider). */\n cwd?: string;\n}\n\n/**\n * Load LLM extensions by npm package name.\n * Prefers ILLMExtension (getLLMExtension() or default export); falls back to registerLLMExtension().\n * When options.cwd is set, tries loading from cwd/node_modules/<package> first so running from a provider project (e.g. wallee-llm) uses the local package.\n * Safe to call multiple times; each package is loaded at most once.\n * @param extensionPackages npm package names; when omitted, loads none\n */\nexport async function loadLLMExtensions(\n extensionPackages?: string[],\n options?: LoadLLMExtensionsOptions\n): Promise<void> {\n const packages = Array.isArray(extensionPackages)\n ? extensionPackages.filter((p): p is string => typeof p === \"string\" && String(p).trim().length > 0)\n : [];\n const cwd = options?.cwd ?? process.cwd();\n for (const pkg of packages) {\n if (loadedPackages.has(pkg)) continue;\n loadedPackages.add(pkg);\n let loaded = false;\n // Try loading from cwd/node_modules first, or from cwd itself when cwd is the provider project (e.g. running from wallee-llm repo)\n const cwdPkgDir = join(cwd, \"node_modules\", pkg);\n const cwdIsProject =\n existsSync(join(cwd, \"package.json\")) &&\n (() => {\n try {\n const name = (JSON.parse(readFileSync(join(cwd, \"package.json\"), \"utf-8\")) as { name?: string }).name;\n return name === pkg;\n } catch {\n return false;\n }\n })();\n const dirsToTry = [cwdPkgDir, ...(cwdIsProject ? [cwd] : [])];\n for (const pkgDir of dirsToTry) {\n if (!existsSync(join(pkgDir, \"package.json\"))) continue;\n try {\n const m = await loadModuleFromPath(pkgDir);\n const ext = getExtensionFromModule(m);\n if (ext) {\n registerExtension(ext);\n loaded = true;\n break;\n }\n if (typeof (m as { registerLLMExtension?: () => void }).registerLLMExtension === \"function\") {\n (m as { registerLLMExtension: () => void }).registerLLMExtension();\n loaded = true;\n break;\n }\n } catch {\n // try next path\n }\n }\n if (loaded) continue;\n try {\n const m = await import(/* @vite-ignore */ pkg) as Record<string, unknown>;\n const ext = getExtensionFromModule(m);\n if (ext) {\n registerExtension(ext);\n continue;\n }\n if (typeof (m as { registerLLMExtension?: () => void }).registerLLMExtension === \"function\") {\n (m as { registerLLMExtension: () => void }).registerLLMExtension();\n }\n } catch (e) {\n // extension not installed or load failed — continue with other packages (fault tolerance)\n const msg = e instanceof Error ? e.message : String(e);\n if (typeof process !== \"undefined\" && process.emitWarning) {\n process.emitWarning(`[agent-llm] Failed to load extension \"${pkg}\": ${msg}`, { code: \"AGENT_LLM_EXTENSION_LOAD\" });\n }\n }\n }\n}\n\n/**\n * Discover provider packages in node_modules and load them.\n * Equivalent to loadLLMExtensions(discoverLLMExtensions(cwd)).\n * @param cwd Directory containing node_modules; default process.cwd()\n */\nexport async function loadDiscoveredExtensions(cwd: string = process.cwd()): Promise<string[]> {\n const dir = typeof cwd === \"string\" && cwd.trim().length > 0 ? cwd : process.cwd();\n const names = discoverLLMExtensions(dir);\n await loadLLMExtensions(names);\n return names;\n}\n","/**\n * npm: protocol in provider — specify an npm package name (and optional version) in config's model provider.\n * Format: npm:<package-name> or npm:<package-name>@<version> or npm:<package-name>#<provider-name> or npm:<package-name>@<version>#<provider-name>\n * When installNpmIfMissing is true, the framework will run npm install <package>[@version] if the package is not found.\n */\n\nimport { execSync } from \"node:child_process\";\nimport { loadLLMExtensions } from \"./loader.js\";\nimport { getChatModelFactory } from \"../registry/chat-model.js\";\nimport { createChatModelFromLlmConfig } from \"../langchain/adapter.js\";\nimport type { CreateChatModelFromLlmConfigOptions } from \"../langchain/adapter.js\";\nimport type { BaseChatModel } from \"@langchain/core/language_models/chat_models\";\n\nexport const NPM_PROTOCOL_PREFIX = \"npm:\";\n\n/** Parse npm: spec into package name, optional version, and optional #provider fragment. */\nexport function parseNpmProviderSpec(spec: string): { packageName: string; version?: string; provider?: string } | null {\n if (typeof spec !== \"string\" || !spec.startsWith(NPM_PROTOCOL_PREFIX)) return null;\n const rest = spec.slice(NPM_PROTOCOL_PREFIX.length);\n const hashIdx = rest.indexOf(\"#\");\n const beforeHash = hashIdx >= 0 ? rest.slice(0, hashIdx).trim() : rest.trim();\n const fragmentProvider = hashIdx >= 0 ? rest.slice(hashIdx + 1).trim() : null;\n // Extract version: last @ not part of scoped package (e.g. pkg@0.1.0 or @scope/pkg@1.0.0)\n const versionMatch = beforeHash.match(/@([^/]+)$/);\n const packageName = versionMatch ? beforeHash.slice(0, -versionMatch[0].length).trim() : beforeHash;\n const version = versionMatch ? versionMatch[1] : undefined;\n if (!packageName) return null;\n return { packageName, version, provider: fragmentProvider ?? undefined };\n}\n\nexport interface EnsureNpmPackageInstalledOptions {\n /** Version to install (e.g. 0.1.0, latest). If set, runs npm install <package>@<version>. */\n version?: string;\n /** Working directory for npm install. Default: process.cwd() */\n cwd?: string;\n}\n\n/**\n * Check if a provider spec uses the npm: protocol.\n */\nexport function isNpmProviderSpec(spec: unknown): spec is string {\n return typeof spec === \"string\" && spec.startsWith(NPM_PROTOCOL_PREFIX);\n}\n\nfunction isModuleNotFoundError(err: unknown): boolean {\n const msg = err instanceof Error ? err.message : String(err);\n return (\n msg.includes(\"Cannot find module\") ||\n msg.includes(\"Cannot find package\") ||\n msg.includes(\"MODULE_NOT_FOUND\") ||\n msg.includes(\"ERR_MODULE_NOT_FOUND\")\n );\n}\n\n/**\n * Ensure an npm package is installed: if it cannot be resolved, run npm install <packageName>[@version] from cwd.\n * Call this before loading a provider package when the config uses npm:<package-name>[@version].\n */\nexport async function ensureNpmPackageInstalled(\n packageName: string,\n options: EnsureNpmPackageInstalledOptions = {}\n): Promise<void> {\n if (typeof packageName !== \"string\" || packageName.trim().length === 0) {\n throw new Error(\"agent-llm: ensureNpmPackageInstalled requires a non-empty package name\");\n }\n const cwd = options.cwd ?? process.cwd();\n const version = options.version;\n try {\n await import(/* @vite-ignore */ packageName);\n return;\n } catch (err) {\n if (!isModuleNotFoundError(err)) throw err;\n }\n const installSpec = version ? `${packageName}@${version}` : packageName;\n try {\n execSync(`npm install ${installSpec}`, {\n cwd,\n stdio: \"inherit\",\n encoding: \"utf-8\",\n });\n } catch (e) {\n const msg = e instanceof Error ? e.message : String(e);\n throw new Error(`agent-llm: npm install failed for ${installSpec}: ${msg}`, { cause: e });\n }\n}\n\nexport interface ResolveNpmProviderOptions {\n /** If true (default), run npm install <package> when the package is not found. */\n installNpmIfMissing?: boolean;\n /** Working directory for npm install. Default: process.cwd() */\n cwd?: string;\n}\n\n/**\n * Resolve an npm provider spec to the concrete provider name.\n * - npm:<package> → load package, use its default provider\n * - npm:<package>@<version> → load package@version, use its default provider\n * - npm:<package>#<provider> → load package, use provider from fragment\n * - npm:<package>@<version>#<provider> → load package@version, use provider from fragment\n * When installNpmIfMissing is true, installs the package (with optional version) if not found.\n * Returns the provider name to use, or null if spec is not npm: protocol.\n */\nfunction normalizeAgentLlMError(e: unknown, context: string): Error {\n if (e instanceof Error) {\n return new Error(`${context}: ${e.message}`, { cause: e });\n }\n return new Error(`${context}: ${String(e)}`);\n}\n\nexport async function resolveNpmProvider(\n spec: string,\n options: ResolveNpmProviderOptions = {}\n): Promise<string | null> {\n if (typeof spec !== \"string\" || spec.trim().length === 0) return null;\n const parsed = parseNpmProviderSpec(spec);\n if (!parsed) return null;\n const { packageName, version, provider: fragmentProvider } = parsed;\n\n const installNpmIfMissing = options.installNpmIfMissing !== false;\n const cwd = options.cwd ?? process.cwd();\n\n const load = async (): Promise<void> => {\n await loadLLMExtensions([packageName], { cwd });\n };\n\n try {\n await load();\n } catch (err) {\n if (installNpmIfMissing && isModuleNotFoundError(err)) {\n try {\n await ensureNpmPackageInstalled(packageName, { version, cwd });\n await load();\n } catch (installErr) {\n throw normalizeAgentLlMError(installErr, `agent-llm: failed to install or load npm provider \"${packageName}\"`);\n }\n } else {\n throw normalizeAgentLlMError(err, `agent-llm: failed to load npm provider \"${packageName}\"`);\n }\n }\n\n // If we need a specific provider (fragment) but loadLLMExtensions didn't register it (e.g. package was missing and didn't throw), try installing and load again\n if (fragmentProvider && installNpmIfMissing && !getChatModelFactory(fragmentProvider)) {\n try {\n await ensureNpmPackageInstalled(packageName, { version, cwd });\n await load();\n } catch (installErr) {\n throw normalizeAgentLlMError(installErr, `agent-llm: failed to install or load npm provider \"${packageName}\" (provider \"${fragmentProvider}\" not registered)`);\n }\n if (!getChatModelFactory(fragmentProvider)) {\n throw new Error(`agent-llm: package \"${packageName}\" was installed but did not register provider \"${fragmentProvider}\". Check that the package exports getLLMExtension() or registerLLMExtension().`);\n }\n }\n\n if (fragmentProvider) return fragmentProvider;\n\n try {\n const m = await import(/* @vite-ignore */ packageName);\n if (\n typeof (m as { getDefaultProviderName?: () => string }).getDefaultProviderName === \"function\"\n ) {\n return (m as { getDefaultProviderName: () => string }).getDefaultProviderName();\n }\n } catch {\n // ignore\n }\n throw new Error(\n `Provider spec ${spec} has no #provider fragment and the package does not export getDefaultProviderName(). Use e.g. npm:${packageName}#<provider-name>.`\n );\n}\n\nexport interface ResolveLlmSectionWithNpmOptions extends ResolveNpmProviderOptions {}\n\n/**\n * Recursively resolve all provider values that use the npm: protocol in a clone of the llm section.\n * When installNpmIfMissing is true, installs any npm: package that is not found.\n */\nexport async function resolveLlmSectionWithNpm(\n llmSection: unknown,\n options: ResolveLlmSectionWithNpmOptions = {}\n): Promise<unknown> {\n if (llmSection == null) return llmSection;\n if (Array.isArray(llmSection)) {\n const out: unknown[] = [];\n for (let i = 0; i < llmSection.length; i++) {\n try {\n out.push(await resolveLlmSectionWithNpm(llmSection[i], options));\n } catch (e) {\n throw normalizeAgentLlMError(e, `agent-llm: failed to resolve llm section at index ${i}`);\n }\n }\n return out;\n }\n if (typeof llmSection === \"object\") {\n const out: Record<string, unknown> = {};\n for (const [k, v] of Object.entries(llmSection)) {\n if (k === \"provider\" && isNpmProviderSpec(v)) {\n try {\n const resolved = await resolveNpmProvider(v as string, options);\n out[k] = resolved ?? v;\n } catch (e) {\n throw normalizeAgentLlMError(e, `agent-llm: failed to resolve provider \"${String(v)}\"`);\n }\n continue;\n }\n try {\n out[k] = await resolveLlmSectionWithNpm(v, options);\n } catch (e) {\n throw normalizeAgentLlMError(e, `agent-llm: failed to resolve llm section key \"${k}\"`);\n }\n }\n return out;\n }\n return llmSection;\n}\n\nexport interface CreateChatModelFromLlmConfigWithNpmOptions\n extends CreateChatModelFromLlmConfigOptions,\n ResolveNpmProviderOptions {}\n\n/**\n * Create a LangChain ChatModel from llm section, resolving any provider values that use the npm: protocol.\n * Use when config has provider: \"npm:<package>\" or \"npm:<package>#<provider>\".\n * When installNpmIfMissing is true (default), the framework will run npm install <package> if the package is not found.\n */\nexport async function createChatModelFromLlmConfigWithNpm(\n options: CreateChatModelFromLlmConfigWithNpmOptions\n): Promise<BaseChatModel> {\n try {\n const { installNpmIfMissing, cwd, ...rest } = options;\n const resolvedSection = await resolveLlmSectionWithNpm(options.llmSection ?? null, {\n installNpmIfMissing,\n cwd,\n });\n return createChatModelFromLlmConfig({\n ...rest,\n llmSection: resolvedSection,\n });\n } catch (e) {\n throw normalizeAgentLlMError(e, \"agent-llm: createChatModelFromLlmConfigWithNpm failed\");\n }\n}\n","/**\n * Return a LangChain-formatted LLM from config (llm.yaml in current directory).\n * Use this LLM with LangChain's createAgent (e.g. createToolCallingAgent + AgentExecutor).\n */\nimport { join } from \"node:path\";\nimport { existsSync } from \"node:fs\";\nimport { createChatModelFromLlmConfig } from \"../langchain/adapter.js\";\nimport { createChatModelFromLlmConfigWithNpm } from \"../extensions/npm-protocol.js\";\nimport { loadLlmConfig } from \"../config/loader.js\";\n\nexport interface CreateAgentLlMOptions {\n /** Path to YAML config file. If omitted, uses llm.yaml in the current directory (cwd). */\n configPath?: string;\n /** If true (default), run npm install when provider is npm:<package> and package is not found. */\n installNpmIfMissing?: boolean;\n}\n\nfunction resolveDefaultConfigPath(): string {\n const cwd = process.cwd();\n return join(cwd, \"llm.yaml\");\n}\n\nfunction normalizeOptions(\n configPathOrOptions?: string | CreateAgentLlMOptions\n): CreateAgentLlMOptions {\n if (configPathOrOptions == null) return {};\n if (typeof configPathOrOptions === \"string\") return { configPath: configPathOrOptions };\n return configPathOrOptions;\n}\n\nfunction normalizeCreateOptions(\n configPathOrOptions?: string | CreateAgentLlMOptions\n): CreateAgentLlMOptions {\n if (configPathOrOptions == null) return {};\n if (typeof configPathOrOptions === \"string\") return { configPath: configPathOrOptions };\n return configPathOrOptions;\n}\n\nfunction normalizeAgentLlMError(e: unknown, context: string): Error {\n if (e instanceof Error) return new Error(`${context}: ${e.message}`, { cause: e });\n return new Error(`${context}: ${String(e)}`);\n}\n\n/**\n * Internal: create LLM from config without resolving npm: providers (sync).\n */\nfunction createAgentLlMInternal(\n configPathOrOptions?: string | CreateAgentLlMOptions\n): ReturnType<typeof createChatModelFromLlmConfig> {\n try {\n const options = normalizeOptions(configPathOrOptions);\n const configPath = options.configPath ?? resolveDefaultConfigPath();\n const llmSection = loadLlmConfig(configPath);\n if (llmSection == null) {\n throw new Error(`No LLM config at ${configPath}. Add llm.yaml in the current directory, or pass configPath.`);\n }\n return createChatModelFromLlmConfig({ llmSection });\n } catch (e) {\n if (e instanceof Error && e.message.includes(\"No LLM config\")) throw e;\n throw normalizeAgentLlMError(e, \"agent-llm: createAgentLlM failed\");\n }\n}\n\n/**\n * Create a LangChain-formatted LLM from config. Resolves npm: providers and optionally installs packages.\n * - createAgentLlM() — uses llm.yaml in the current directory\n * - createAgentLlM(\"path/to/llm.yaml\") — use specific config file\n * - createAgentLlM({ configPath: \"...\", installNpmIfMissing: true }) — options object\n */\nexport async function createAgentLlM(\n configPathOrOptions?: string | CreateAgentLlMOptions\n): Promise<ReturnType<typeof createChatModelFromLlmConfig>> {\n try {\n const options = normalizeCreateOptions(configPathOrOptions);\n const configPath = options.configPath ?? resolveDefaultConfigPath();\n const llmSection = loadLlmConfig(configPath);\n if (llmSection == null) {\n throw new Error(`No LLM config at ${configPath}. Add llm.yaml in the current directory, or pass configPath.`);\n }\n return await createChatModelFromLlmConfigWithNpm({\n llmSection,\n installNpmIfMissing: options.installNpmIfMissing !== false,\n cwd: process.cwd(),\n });\n } catch (e) {\n if (e instanceof Error && e.message.includes(\"No LLM config\")) throw e;\n throw normalizeAgentLlMError(e, \"agent-llm: createAgentLlM failed\");\n }\n}\n"],"mappings":";AAOA,IAAM,iBAAiB;AAEvB,IAAM,gBAAgB,oBAAI,IAAI;AAAA,EAC5B;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AACF,CAAC;AAKM,SAAS,gBAAgB,SAA+D;AAC7F,MAAI,WAAW,QAAQ,OAAO,YAAY,UAAU;AAClD,WAAO,EAAE,WAAW,gBAAgB,SAAS,CAAC,EAAE;AAAA,EAClD;AAEA,MAAI,MAAM,QAAQ,OAAO,GAAG;AAC1B,UAAM,UAAU,QACb,OAAO,CAAC,MAAoC,KAAK,QAAQ,OAAO,MAAM,QAAQ,EAC9E,IAAI,CAAC,MAAM,MAAM,mBAAmB,EAAE,GAAG,MAAM,IAAI,KAAK,MAAM,KAAK,QAAQ,OAAO,CAAC,EAAE,CAAC,CAAC,EACvF,OAAO,CAAC,MAAsB,KAAK,IAAI;AAC1C,UAAM,YAAY,QAAQ,SAAS,IAAI,QAAQ,CAAC,EAAG,KAAK;AACxD,WAAO,EAAE,WAAW,QAAQ;AAAA,EAC9B;AAEA,QAAM,IAAI;AAEV,QAAM,cAAc,OAAO,QAAQ,CAAC,EAAE;AAAA,IACpC,CAAC,CAAC,GAAG,CAAC,MAAM,CAAC,cAAc,IAAI,CAAC,KAAK,KAAK,QAAQ,OAAO,MAAM,YAAY,CAAC,MAAM,QAAQ,CAAC;AAAA,EAC7F;AACA,MAAI,YAAY,SAAS,GAAG;AAC1B,UAAM,UAAuB,CAAC;AAC9B,eAAW,CAAC,IAAI,KAAK,KAAK,aAAa;AACrC,YAAM,IAAI,iBAAiB,IAAI,KAAgC;AAC/D,UAAI,EAAG,SAAQ,KAAK,CAAC;AAAA,IACvB;AACA,UAAM,YACJ,OAAO,EAAE,YAAY,YAAY,EAAE,WAAW,YAAY,KAAK,CAAC,CAAC,CAAC,MAAM,MAAM,EAAE,OAAO,IACnF,EAAE,UACF,QAAQ,SAAS,IACf,QAAQ,CAAC,EAAG,KACZ;AACR,WAAO,EAAE,WAAW,QAAQ;AAAA,EAC9B;AAEA,MAAI,MAAM,QAAQ,EAAE,SAAS,GAAG;AAC9B,UAAM,UAAW,EAAE,UAChB,OAAO,CAAC,MAAoC,KAAK,QAAQ,OAAO,MAAM,QAAQ,EAC9E,IAAI,CAAC,MAAM,mBAAmB,CAAC,CAAC,EAChC,OAAO,CAAC,MAAsB,KAAK,IAAI;AAC1C,UAAM,YACJ,OAAO,EAAE,YAAY,YAAY,EAAE,UAC/B,EAAE,UACF,QAAQ,SAAS,IACf,QAAQ,CAAC,EAAG,KACZ;AACR,WAAO,EAAE,WAAW,QAAQ;AAAA,EAC9B;AAEA,MAAI,OAAO,EAAE,aAAa,YAAY,OAAO,EAAE,UAAU,YAAY,OAAQ,EAAwB,SAAS,UAAU;AACtH,UAAM,MAAM,wBAAwB,CAAC;AACrC,WAAO,EAAE,WAAW,IAAI,IAAI,SAAS,CAAC,GAAG,EAAE;AAAA,EAC7C;AAEA,SAAO,EAAE,WAAW,gBAAgB,SAAS,CAAC,EAAE;AAClD;AAEA,IAAM,wBAAwB,CAAC,cAAc,UAAU,aAAa,aAAa,cAAc,QAAQ,eAAe,aAAa,SAAS;AAE5I,SAAS,iBAAiB,IAAY,OAAkD;AACtF,QAAM,OAAO,MAAM;AACnB,QAAM,UACJ,OAAO,MAAM,aAAa,WACtB,MAAM,WACN,OAAO,MAAM,YAAY,WACvB,MAAM,UACN;AACR,QAAM,QAAQ,OAAO,MAAM,SAAS,WAAW,MAAM,OAAO,OAAO,MAAM,UAAU,WAAW,MAAM,QAAQ;AAC5G,QAAM,WAAW,OAAO,MAAM,aAAa,YAAY,MAAM,WAAW,MAAM,WAAW;AACzF,QAAM,SAAoB;AAAA,IACxB;AAAA,IACA,MAAM;AAAA,IACN;AAAA,IACA;AAAA,IACA,aAAa,OAAO,MAAM,gBAAgB,WAAW,KAAK,cAAc,OAAO,MAAM,gBAAgB,WAAW,MAAM,cAAc;AAAA,IACpI,QAAQ,OAAO,MAAM,WAAW,WAAW,KAAK,SAAS,OAAO,MAAM,WAAW,WAAW,MAAM,SAAS;AAAA,IAC3G;AAAA,EACF;AACA,MAAI,OAAO,MAAM,SAAS,YAAY,MAAM,SAAS,QAAS,QAAO,OAAO;AAC5E,MAAI,QAAQ,OAAO,SAAS,SAAU,CAAC,OAAmC,UAAU;AACpF,aAAW,KAAK,uBAAuB;AACrC,QAAI,MAAM,CAAC,MAAM,OAAW,CAAC,OAAmC,CAAC,IAAI,MAAM,CAAC;AAAA,aACnE,QAAQ,KAAK,CAAC,MAAM,OAAW,CAAC,OAAmC,CAAC,IAAI,KAAK,CAAC;AAAA,EACzF;AACA,SAAO;AACT;AAEA,SAAS,wBAAwB,GAAuC;AACtE,QAAM,MAAiB;AAAA,IACrB,IAAI;AAAA,IACJ,MAAM;AAAA,IACN,UAAU,OAAO,EAAE,aAAa,WAAW,EAAE,WAAW;AAAA,IACxD,OAAO,OAAO,EAAE,UAAU,WAAW,EAAE,QAAS,OAAQ,EAAwB,SAAS,WAAY,EAAuB,OAAO;AAAA,IACnI,aAAa,OAAO,EAAE,gBAAgB,WAAW,EAAE,cAAc;AAAA,IACjE,QAAQ,OAAO,EAAE,WAAW,WAAW,EAAE,SAAS;AAAA,IAClD,SACE,OAAO,EAAE,YAAY,WAAW,EAAE,UAAU,OAAO,EAAE,aAAa,WAAW,EAAE,WAAW;AAAA,EAC9F;AACA,SAAO,KAAK,CAAC,EAAE,QAAQ,CAAC,MAAM;AAC5B,QAAI,CAAC,CAAC,MAAM,QAAQ,YAAY,SAAS,QAAQ,eAAe,UAAU,WAAW,YAAY,WAAW,WAAW,EAAE,SAAS,CAAC,GAAG;AACpI,MAAC,IAAgC,CAAC,IAAI,EAAE,CAAC;AAAA,IAC3C;AAAA,EACF,CAAC;AACD,SAAO;AACT;AAEA,SAAS,mBAAmB,GAA8C;AACxE,QAAM,KAAK,OAAO,EAAE,OAAO,YAAY,EAAE,KAAK,EAAE,KAAK;AACrD,QAAM,OAAO,EAAE,SAAS,UAAU,UAAU;AAC5C,QAAM,WAAW,OAAO,EAAE,aAAa,YAAY,EAAE,WAAW,EAAE,WAAW;AAC7E,QAAM,OAAO,EAAE;AACf,QAAM,SAAoB;AAAA,IACxB;AAAA,IACA;AAAA,IACA;AAAA,IACA,OAAO,OAAO,EAAE,UAAU,WAAW,EAAE,QAAS,OAAO,EAAE,SAAS,WAAW,EAAE,OAAO;AAAA,IACtF,aACE,OAAO,EAAE,gBAAgB,WACrB,EAAE,cACF,OAAO,MAAM,gBAAgB,WAC3B,KAAK,cACL;AAAA,IACR,QACE,OAAO,EAAE,WAAW,WAChB,EAAE,SACF,OAAO,MAAM,WAAW,WACtB,KAAK,SACL;AAAA,IACR,SAAS,OAAO,EAAE,YAAY,WAAW,EAAE,UAAW,OAAO,EAAE,aAAa,WAAW,EAAE,WAAW;AAAA,EACtG;AACA,SAAO,KAAK,CAAC,EAAE,QAAQ,CAAC,MAAM;AAC5B,QAAI,CAAC,CAAC,MAAM,QAAQ,YAAY,SAAS,QAAQ,eAAe,UAAU,WAAW,UAAU,EAAE,SAAS,CAAC,GAAG;AAC5G,MAAC,OAAmC,CAAC,IAAI,EAAE,CAAC;AAAA,IAC9C;AAAA,EACF,CAAC;AACD,SAAO;AACT;;;AC5JA,SAAS,cAAc,kBAAkB;AACzC,SAAS,SAAS,iBAAiB;AAU5B,SAAS,cAAc,KAAuB;AACnD,MAAI,QAAQ,QAAQ,QAAQ,OAAW,QAAO;AAC9C,MAAI,OAAO,QAAQ,UAAU;AAC3B,UAAM,IAAI,IAAI,MAAM,eAAe;AACnC,WAAO,IAAK,QAAQ,IAAI,EAAE,CAAC,CAAC,KAAK,MAAO;AAAA,EAC1C;AACA,MAAI,MAAM,QAAQ,GAAG,EAAG,QAAO,IAAI,IAAI,aAAa;AACpD,MAAI,OAAO,QAAQ,UAAU;AAC3B,UAAM,MAA+B,CAAC;AACtC,eAAW,CAAC,GAAG,CAAC,KAAK,OAAO,QAAQ,GAAG,EAAG,KAAI,CAAC,IAAI,cAAc,CAAC;AAClE,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAMO,SAAS,aACd,SACA,UAAgC,CAAC,GACxB;AACT,QAAM,EAAE,eAAe,QAAQ,KAAK,IAAI;AACxC,QAAM,SAAS,UAAU,OAAO;AAChC,QAAM,MAAM,QAAQ;AACpB,MAAI,OAAO,KAAM,QAAO;AACxB,SAAO,QAAQ,cAAc,GAAG,IAAI;AACtC;AAQO,SAAS,cACd,UACA,UAAgC,CAAC,GACjB;AAChB,MAAI,OAAO,aAAa,YAAY,SAAS,KAAK,EAAE,WAAW,GAAG;AAChE,UAAM,IAAI,MAAM,yDAAyD;AAAA,EAC3E;AACA,MAAI,CAAC,WAAW,QAAQ,EAAG,QAAO;AAClC,MAAI;AACJ,MAAI;AACF,UAAM,aAAa,UAAU,MAAM;AAAA,EACrC,SAAS,GAAG;AACV,UAAM,MAAM,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AACrD,UAAM,IAAI,MAAM,yCAAyC,QAAQ,KAAK,GAAG,IAAI,EAAE,OAAO,EAAE,CAAC;AAAA,EAC3F;AACA,MAAI;AACF,UAAM,MAAM,aAAa,KAAK,OAAO;AACrC,WAAO,OAAO;AAAA,EAChB,SAAS,GAAG;AACV,UAAM,MAAM,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AACrD,UAAM,IAAI,MAAM,0CAA0C,QAAQ,KAAK,GAAG,IAAI,EAAE,OAAO,EAAE,CAAC;AAAA,EAC5F;AACF;;;ACrEA,OAAO,YAAY;AAcnB,SAAS,UAAU,QAA2B;AAC5C,QAAM,MAAM,OAAO,UAAU,QAAQ,IAAI,kBAAkB;AAC3D,MAAI,CAAC,IAAK,OAAM,IAAI,MAAM,qEAAqE;AAC/F,SAAO;AACT;AAEA,SAAS,0BAA0B,QAAyD;AAC1F,QAAM,OAA6C,EAAE,QAAQ,UAAU,MAAM,EAAE;AAC/E,MAAI,OAAO,OAAO,YAAY,YAAY,OAAO,QAAS,MAAK,UAAU,OAAO;AAChF,SAAO;AACT;AAEA,SAAS,iBACP,GACoD;AACpD,MAAI,EAAE,SAAS;AACb,WAAO,EAAE,MAAM,QAAQ,SAAS,EAAE,SAAS,cAAc,EAAE,aAAa;AAC1E,MAAI,EAAE,SAAS,eAAe,gBAAgB,KAAK,EAAE,YAAY,QAAQ;AACvE,WAAO;AAAA,MACL,MAAM;AAAA,MACN,SAAS,EAAE,WAAW;AAAA,MACtB,YAAY,EAAE,WAAW,IAAI,CAAC,QAAQ;AAAA,QACpC,IAAI,GAAG;AAAA,QACP,MAAM;AAAA,QACN,UAAU,EAAE,MAAM,GAAG,SAAS,MAAM,WAAW,GAAG,SAAS,UAAU;AAAA,MACvE,EAAE;AAAA,IACJ;AAAA,EACF;AACA,SAAO,EAAE,MAAM,EAAE,MAAM,SAAU,EAAkB,QAAQ;AAC7D;AAEO,SAAS,uBAAuB,QAA+B;AACpE,QAAM,SAAS,IAAI,OAAO,0BAA0B,MAAM,CAAC;AAC3D,QAAM,QAAQ,OAAO,SAAS,QAAQ,IAAI,gBAAgB;AAC1D,QAAM,cAAc,OAAO,eAAe;AAE1C,SAAO;AAAA,IACL,IAAI,OAAO;AAAA,IACX,MAAM;AAAA,IACN,MAAM,KAAK,UAA8C;AACvD,YAAM,OAAO,MAAM,OAAO,KAAK,YAAY,OAAO;AAAA,QAChD;AAAA,QACA;AAAA,QACA,UAAU,SAAS,IAAI,CAAC,OAAO,EAAE,MAAM,EAAE,MAAM,SAAS,EAAE,QAAQ,EAAE;AAAA,MACtE,CAAC;AACD,YAAM,UAAU,KAAK,QAAQ,CAAC,GAAG,SAAS,WAAW;AACrD,YAAM,QAAQ,KAAK,QACf,EAAE,cAAc,KAAK,MAAM,eAAe,kBAAkB,KAAK,MAAM,kBAAkB,IACzF;AACJ,aAAO,EAAE,SAAS,MAAM;AAAA,IAC1B;AAAA,IACA,OAAO,WAAW,UAAyE;AACzF,YAAM,SAAS,MAAM,OAAO,KAAK,YAAY,OAAO;AAAA,QAClD;AAAA,QACA;AAAA,QACA,UAAU,SAAS,IAAI,CAAC,OAAO,EAAE,MAAM,EAAE,MAAM,SAAS,EAAE,QAAQ,EAAE;AAAA,QACpE,QAAQ;AAAA,MACV,CAAC;AAED,UAAI;AAEJ,uBAAiB,SAAS,QAAQ;AAChC,cAAM,QAAQ,MAAM,QAAQ,CAAC,GAAG;AAChC,cAAM,UAAU,OAAO,WAAW;AAClC,cAAM,eAAe,MAAM,QAAQ,CAAC,GAAG;AAGvC,YAAI,MAAM,OAAO;AACf,kBAAQ;AAAA,YACN,cAAc,MAAM,MAAM;AAAA,YAC1B,kBAAkB,MAAM,MAAM;AAAA,UAChC;AAAA,QACF;AAEA,cAAM;AAAA,UACJ;AAAA,UACA,MAAM,iBAAiB,QAAQ,iBAAiB;AAAA,UAChD;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,IACA,MAAM,cACJ,UACA,OACA,UAC8B;AAC9B,YAAM,OAAO,MAAM,OAAO,KAAK,YAAY,OAAO;AAAA,QAChD;AAAA,QACA;AAAA,QACA,UAAU,SAAS,IAAI,gBAAgB;AAAA,QACvC,OAAO,MAAM,IAAI,CAAC,OAAO;AAAA,UACvB,MAAM;AAAA,UACN,UAAU;AAAA,YACR,MAAM,EAAE,SAAS;AAAA,YACjB,aAAa,EAAE,SAAS;AAAA,YACxB,YAAa,EAAE,SAAS,cAAc;AAAA,UACxC;AAAA,QACF,EAAE;AAAA,MACJ,CAAC;AACD,YAAM,MAAM,KAAK,QAAQ,CAAC,GAAG;AAC7B,YAAM,QAAQ,KAAK,QACf,EAAE,cAAc,KAAK,MAAM,eAAe,kBAAkB,KAAK,MAAM,kBAAkB,IACzF;AACJ,aAAO;AAAA,QACL,SAAS;AAAA,UACP,MAAM;AAAA,UACN,SAAS,KAAK,WAAW;AAAA,UACzB,YAAY,KAAK,YAAY,IAAI,CAAC,QAAQ;AAAA,YACxC,IAAI,GAAG;AAAA,YACP,MAAM;AAAA,YACN,UAAU;AAAA,cACR,MAAM,GAAG,UAAU,QAAQ;AAAA,cAC3B,WAAW,GAAG,UAAU,aAAa;AAAA,YACvC;AAAA,UACF,EAAE;AAAA,QACJ;AAAA,QACA;AAAA,MACF;AAAA,IACF;AAAA,IACA,OAAO,oBACL,UACA,OACA,UACyD;AACzD,YAAM,SAAS,MAAM,OAAO,KAAK,YAAY,OAAO;AAAA,QAClD;AAAA,QACA;AAAA,QACA,UAAU,SAAS,IAAI,gBAAgB;AAAA,QACvC,OAAO,MAAM,IAAI,CAAC,OAAO;AAAA,UACvB,MAAM;AAAA,UACN,UAAU;AAAA,YACR,MAAM,EAAE,SAAS;AAAA,YACjB,aAAa,EAAE,SAAS;AAAA,YACxB,YAAa,EAAE,SAAS,cAAc;AAAA,UACxC;AAAA,QACF,EAAE;AAAA,QACF,QAAQ;AAAA,QACR,gBAAgB,EAAE,eAAe,KAAK;AAAA,MACxC,CAAC;AAED,UAAI;AAEJ,uBAAiB,SAAS,QAAQ;AAChC,cAAM,QAAQ,MAAM,QAAQ,CAAC,GAAG;AAChC,cAAM,eAAe,MAAM,QAAQ,CAAC,GAAG;AAGvC,YAAI,MAAM,OAAO;AACf,kBAAQ;AAAA,YACN,cAAc,MAAM,MAAM;AAAA,YAC1B,kBAAkB,MAAM,MAAM;AAAA,UAChC;AAAA,QACF;AAEA,cAAM,YAAY,OAAO,YAAY,IAAI,CAAC,QAAQ;AAAA,UAChD,OAAO,GAAG;AAAA,UACV,IAAI,GAAG;AAAA,UACP,MAAM,GAAG;AAAA,UACT,UAAU,GAAG,WACT;AAAA,YACE,MAAM,GAAG,SAAS;AAAA,YAClB,WAAW,GAAG,SAAS;AAAA,UACzB,IACA;AAAA,QACN,EAAE;AAEF,cAAM;AAAA,UACJ,OAAO;AAAA,YACL,MAAM,OAAO;AAAA,YACb,SAAS,OAAO,WAAW;AAAA,YAC3B,YAAY;AAAA,UACd;AAAA,UACA,MAAM,iBAAiB,QAAQ,iBAAiB;AAAA,UAChD;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAEO,SAAS,wBAAwB,QAA+B;AACrE,QAAM,SAAS,IAAI,OAAO,0BAA0B,MAAM,CAAC;AAC3D,QAAM,QAAS,OAAO,SAAoB;AAE1C,SAAO;AAAA,IACL,IAAI,OAAO;AAAA,IACX,MAAM;AAAA,IACN,MAAM,OAA4B;AAChC,YAAM,IAAI,MAAM,+DAA+D;AAAA,IACjF;AAAA,IACA,MAAM,cAAc,SAA8E;AAChG,YAAM,OAAO,MAAM,OAAO,OAAO,SAAS;AAAA,QACxC;AAAA,QACA,QAAQ,QAAQ;AAAA,QAChB,MAAO,QAAQ,QAAoD;AAAA,QACnE,GAAG,QAAQ,KAAK;AAAA,QAChB,iBAAiB;AAAA,MACnB,CAAC;AACD,YAAM,MAAM,KAAK,OAAO,CAAC,GAAG,OAAO;AACnC,aAAO,EAAE,IAAI;AAAA,IACf;AAAA,EACF;AACF;AAEO,SAAS,mBAAmB,QAA+B;AAChE,MAAI,OAAO,SAAS,QAAS,QAAO,wBAAwB,MAAM;AAClE,SAAO,uBAAuB,MAAM;AACtC;;;AC3NA,IAAM,oBAAoB;AAE1B,SAAS,mBAAmB,QAA+B;AACzD,SAAO,mBAAmB,MAAM;AAClC;AAEA,IAAM,YAA+D;AAAA,EACnE,QAAQ;AAAA,EACR,CAAC,iBAAiB,GAAG;AACvB;AAEO,SAAS,aAAa,QAA+B;AAC1D,QAAM,KAAK,OAAO,YAAY,IAAI,YAAY;AAC9C,QAAM,KAAK,UAAU,CAAC;AACtB,MAAI,CAAC,IAAI;AACP,UAAM,YAAY,CAAC,GAAG,oBAAI,IAAI,CAAC,GAAG,OAAO,KAAK,SAAS,GAAG,qBAAqB,CAAC,CAAC,EAAE,KAAK,EAAE,KAAK,IAAI;AACnG,UAAM,IAAI;AAAA,MACR,6BAA6B,OAAO,QAAQ,gBAAgB,SAAS;AAAA,IACvE;AAAA,EACF;AACA,SAAO,GAAG,MAAM;AAClB;AAEO,SAAS,iBAAiB,MAAc,SAAkD;AAC/F,YAAU,KAAK,YAAY,CAAC,IAAI;AAClC;;;ACtBA,IAAM,uBAAuB,oBAAI,IAA8B;AAMxD,SAAS,0BAA0B,cAAsB,SAAiC;AAC/F,uBAAqB,IAAI,aAAa,YAAY,GAAG,OAAO;AAC9D;AAKO,SAAS,oBAAoB,cAAoD;AACtF,SAAO,qBAAqB,IAAI,aAAa,YAAY,CAAC;AAC5D;;;AClBA,SAAS,kBAAkB;AAK3B,IAAM,gBAAgB;AAWtB,SAAS,uBAAuB,GAAY,SAAwB;AAClE,MAAI,aAAa,MAAO,QAAO,IAAI,MAAM,GAAG,OAAO,KAAK,EAAE,OAAO,IAAI,EAAE,OAAO,EAAE,CAAC;AACjF,SAAO,IAAI,MAAM,GAAG,OAAO,KAAK,OAAO,CAAC,CAAC,EAAE;AAC7C;AAMO,SAAS,6BACd,SACe;AACf,QAAM,EAAE,YAAY,UAAU,UAAU,IAAI;AAC5C,MAAI;AACJ,MAAI;AACJ,MAAI;AACF,UAAM,SAAS,gBAAgB,cAAc,IAAI;AACjD,gBAAY,OAAO;AACnB,cAAU,OAAO;AAAA,EACnB,SAAS,GAAG;AACV,UAAM,uBAAuB,GAAG,wCAAwC;AAAA,EAC1E;AACA,QAAM,gBAAgB,QAAQ,KAAK,CAAC,MAAM,EAAE,OAAO,SAAS,KAAK,QAAQ,CAAC;AAE1E,MAAI,CAAC,eAAe;AAClB,UAAMA,SACJ,YAAY,QAAQ,IAAI,gBAAgB;AAC1C,UAAMC,UAAS,aAAa,QAAQ,IAAI;AACxC,QAAI;AACF,aAAO,IAAI,WAAW;AAAA,QACpB,OAAAD;AAAA,QACA,aAAa;AAAA,QACb,GAAIC,UAAS,EAAE,QAAAA,QAAO,IAAI,CAAC;AAAA,MAC7B,CAAC;AAAA,IACH,SAAS,GAAG;AACV,YAAM,uBAAuB,GAAG,gDAAgD;AAAA,IAClF;AAAA,EACF;AAEA,QAAM,WAAY,cAAwC,YAAY;AACtE,QAAM,mBAAmB,oBAAoB,QAAQ;AACrD,MAAI,kBAAkB;AACpB,UAAM,SAAS;AAAA,MACb,GAAG;AAAA,MACH,OAAO,YAAY,cAAc;AAAA,MACjC,aACE,OAAO,cAAc,gBAAgB,WACjC,cAAc,cACd;AAAA,IACR;AACA,QAAI;AACF,aAAO,iBAAiB,MAAM;AAAA,IAChC,SAAS,GAAG;AACV,YAAM,uBAAuB,GAAG,uDAAuD,QAAQ,GAAG;AAAA,IACpG;AAAA,EACF;AAEA,QAAM,QACJ,YACA,eAAe,SACf,QAAQ,IAAI,gBACZ;AAEF,MAAI,SACF,aAAa,eAAe,UAAU,QAAQ,IAAI;AACpD,MAAI,UAAU,eAAe;AAE7B,MAAI,WAAW,CAAC,QAAQ,QAAQ,OAAO,EAAE,EAAE,SAAS,KAAK,GAAG;AAC1D,cAAU,QAAQ,QAAQ,OAAO,EAAE,IAAI;AAAA,EACzC;AAEA,MAAI,WAAW,WAAW,QAAW;AACnC,aAAS;AAAA,EACX;AAEA,QAAM,cACJ,OAAO,eAAe,gBAAgB,WAAW,cAAc,cAAc;AAE/E,QAAM,qBAAkE;AAAA,IACtE;AAAA,IACA;AAAA,IACA,GAAI,SAAS,EAAE,OAAO,IAAI,CAAC;AAAA,IAC3B,GAAI,UAAU,EAAE,eAAe,EAAE,QAAQ,EAAE,IAAI,CAAC;AAAA,EAClD;AAEA,MAAI;AACF,WAAO,IAAI,WAAW,kBAAkB;AAAA,EAC1C,SAAS,GAAG;AACV,UAAM,uBAAuB,GAAG,oDAAoD;AAAA,EACtF;AACF;;;AC5GA,SAAS,aAAa,gBAAAC,eAAc,cAAAC,mBAAkB;AACtD,SAAS,YAAY;AACrB,SAAS,qBAAqB;AAM9B,IAAM,iBAAiB,oBAAI,IAAY;AAEvC,SAAS,eAAe,GAAgC;AACtD,MAAI,KAAK,QAAQ,OAAO,MAAM,SAAU,QAAO;AAC/C,QAAM,IAAI;AACV,MAAI,OAAO,EAAE,iBAAiB,SAAU,QAAO;AAC/C,QAAM,cAAc,OAAO,EAAE,aAAa;AAC1C,QAAM,eAAe,OAAO,EAAE,iBAAiB,cAAc,OAAO,EAAE,oBAAoB;AAC1F,SAAO,eAAe;AACxB;AAEA,SAAS,kBAAkB,KAA0B;AACnD,MAAI,OAAO,IAAI,iBAAiB,cAAc,OAAO,IAAI,oBAAoB,YAAY;AACvF,qBAAiB,IAAI,cAAc,IAAI,YAAY;AACnD,8BAA0B,IAAI,cAAc,IAAI,eAAmC;AACnF;AAAA,EACF;AACA,MAAI,OAAO,IAAI,aAAa,YAAY;AACtC,QAAI,SAAS;AAAA,EACf;AACF;AAEA,SAAS,uBAAuB,GAAkD;AAChF,MAAI,eAAe,CAAC,EAAG,QAAO;AAC9B,MAAI,OAAO,EAAE,oBAAoB,YAAY;AAC3C,UAAM,MAAM,EAAE,gBAAgB;AAC9B,WAAO,eAAe,GAAG,IAAI,MAAM;AAAA,EACrC;AACA,MAAI,eAAe,EAAE,OAAO,EAAG,QAAO,EAAE;AACxC,SAAO;AACT;AAKO,SAAS,4BAA4B,OAAqC;AAC/E,MAAI,SAAS,KAAM,QAAO,CAAC;AAC3B,QAAM,WAAW,MAAM,QAAQ,KAAK,IAAI,QAAQ,CAAC,KAAK;AACtD,SAAO,SAAS;AAAA,IACd,CAAC,MAAmB,OAAO,MAAM,YAAY,OAAO,CAAC,EAAE,KAAK,EAAE,SAAS;AAAA,EACzE;AACF;AAQA,SAAS,wBAAwB,SAAgC;AAC/D,MAAI,CAACC,YAAW,OAAO,EAAG,QAAO;AACjC,MAAI;AACF,UAAM,MAAMC,cAAa,SAAS,OAAO;AACzC,UAAM,MAAM,KAAK,MAAM,GAAG;AAC1B,UAAM,WACJ,IAAI,qBAAqB,QACxB,MAAM,QAAQ,IAAI,QAAQ,KAAK,IAAI,SAAS,SAAS,oBAAoB;AAC5E,WAAO,YAAY,OAAO,IAAI,SAAS,WAAW,IAAI,OAAO;AAAA,EAC/D,QAAQ;AACN,WAAO;AAAA,EACT;AACF;AAEO,SAAS,sBAAsB,MAAc,QAAQ,IAAI,GAAa;AAC3E,QAAM,MAAM,OAAO,QAAQ,YAAY,IAAI,KAAK,EAAE,SAAS,IAAI,MAAM,QAAQ,IAAI;AACjF,QAAM,cAAc,KAAK,KAAK,cAAc;AAC5C,MAAI,CAACD,YAAW,WAAW,EAAG,QAAO,CAAC;AACtC,QAAM,QAAkB,CAAC;AACzB,QAAM,OAAO,oBAAI,IAAY;AAC7B,MAAI;AACF,UAAM,UAAU,YAAY,aAAa,EAAE,eAAe,KAAK,CAAC;AAChE,eAAW,KAAK,SAAS;AACvB,UAAI,EAAE,KAAK,WAAW,GAAG,KAAK,EAAE,SAAS,OAAQ;AACjD,UAAI,EAAE,KAAK,WAAW,GAAG,KAAK,EAAE,YAAY,GAAG;AAC7C,cAAM,YAAY,KAAK,aAAa,EAAE,IAAI;AAC1C,cAAM,eAAe,YAAY,WAAW,EAAE,eAAe,KAAK,CAAC;AACnE,mBAAW,MAAM,cAAc;AAC7B,cAAI,CAAC,GAAG,YAAY,EAAG;AACvB,gBAAME,QAAO,wBAAwB,KAAK,WAAW,GAAG,MAAM,cAAc,CAAC;AAC7E,cAAIA,SAAQ,CAAC,KAAK,IAAIA,KAAI,GAAG;AAC3B,iBAAK,IAAIA,KAAI;AACb,kBAAM,KAAKA,KAAI;AAAA,UACjB;AAAA,QACF;AACA;AAAA,MACF;AACA,YAAM,OAAO,wBAAwB,KAAK,aAAa,EAAE,MAAM,cAAc,CAAC;AAC9E,UAAI,QAAQ,CAAC,KAAK,IAAI,IAAI,GAAG;AAC3B,aAAK,IAAI,IAAI;AACb,cAAM,KAAK,IAAI;AAAA,MACjB;AAAA,IACF;AAAA,EACF,QAAQ;AAAA,EAER;AACA,SAAO;AACT;AAMA,eAAe,mBAAmB,QAAkD;AAClF,QAAM,cAAc,KAAK,QAAQ,cAAc;AAC/C,MAAI,CAACF,YAAW,WAAW,GAAG;AAC5B,UAAM,IAAI,MAAM,6BAA6B,MAAM,EAAE;AAAA,EACvD;AACA,QAAM,UAAU,KAAK,MAAMC,cAAa,aAAa,OAAO,CAAC;AAC7D,QAAM,OAAO,SAAS,QAAQ;AAC9B,QAAM,YAAY,KAAK,QAAQ,IAAI;AACnC,MAAI,CAACD,YAAW,SAAS,GAAG;AAC1B,UAAM,IAAI,MAAM,SAAS,IAAI,iBAAiB,MAAM,EAAE;AAAA,EACxD;AACA,QAAM,WAAW,cAAc,SAAS,EAAE;AAC1C,SAAQ,MAAM;AAAA;AAAA,IAA0B;AAAA;AAC1C;AAcA,eAAsB,kBACpB,mBACA,SACe;AACf,QAAM,WAAW,MAAM,QAAQ,iBAAiB,IAC5C,kBAAkB,OAAO,CAAC,MAAmB,OAAO,MAAM,YAAY,OAAO,CAAC,EAAE,KAAK,EAAE,SAAS,CAAC,IACjG,CAAC;AACL,QAAM,MAAM,SAAS,OAAO,QAAQ,IAAI;AACxC,aAAW,OAAO,UAAU;AAC1B,QAAI,eAAe,IAAI,GAAG,EAAG;AAC7B,mBAAe,IAAI,GAAG;AACtB,QAAI,SAAS;AAEb,UAAM,YAAY,KAAK,KAAK,gBAAgB,GAAG;AAC/C,UAAM,eACJA,YAAW,KAAK,KAAK,cAAc,CAAC,MACnC,MAAM;AACL,UAAI;AACF,cAAM,OAAQ,KAAK,MAAMC,cAAa,KAAK,KAAK,cAAc,GAAG,OAAO,CAAC,EAAwB;AACjG,eAAO,SAAS;AAAA,MAClB,QAAQ;AACN,eAAO;AAAA,MACT;AAAA,IACF,GAAG;AACL,UAAM,YAAY,CAAC,WAAW,GAAI,eAAe,CAAC,GAAG,IAAI,CAAC,CAAE;AAC5D,eAAW,UAAU,WAAW;AAC9B,UAAI,CAACD,YAAW,KAAK,QAAQ,cAAc,CAAC,EAAG;AAC/C,UAAI;AACF,cAAM,IAAI,MAAM,mBAAmB,MAAM;AACzC,cAAM,MAAM,uBAAuB,CAAC;AACpC,YAAI,KAAK;AACP,4BAAkB,GAAG;AACrB,mBAAS;AACT;AAAA,QACF;AACA,YAAI,OAAQ,EAA4C,yBAAyB,YAAY;AAC3F,UAAC,EAA2C,qBAAqB;AACjE,mBAAS;AACT;AAAA,QACF;AAAA,MACF,QAAQ;AAAA,MAER;AAAA,IACF;AACA,QAAI,OAAQ;AACZ,QAAI;AACF,YAAM,IAAI,MAAM;AAAA;AAAA,QAA0B;AAAA;AAC1C,YAAM,MAAM,uBAAuB,CAAC;AACpC,UAAI,KAAK;AACP,0BAAkB,GAAG;AACrB;AAAA,MACF;AACA,UAAI,OAAQ,EAA4C,yBAAyB,YAAY;AAC3F,QAAC,EAA2C,qBAAqB;AAAA,MACnE;AAAA,IACF,SAAS,GAAG;AAEV,YAAM,MAAM,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AACrD,UAAI,OAAO,YAAY,eAAe,QAAQ,aAAa;AACzD,gBAAQ,YAAY,yCAAyC,GAAG,MAAM,GAAG,IAAI,EAAE,MAAM,2BAA2B,CAAC;AAAA,MACnH;AAAA,IACF;AAAA,EACF;AACF;AAOA,eAAsB,yBAAyB,MAAc,QAAQ,IAAI,GAAsB;AAC7F,QAAM,MAAM,OAAO,QAAQ,YAAY,IAAI,KAAK,EAAE,SAAS,IAAI,MAAM,QAAQ,IAAI;AACjF,QAAM,QAAQ,sBAAsB,GAAG;AACvC,QAAM,kBAAkB,KAAK;AAC7B,SAAO;AACT;;;ACnNA,SAAS,gBAAgB;AAOlB,IAAM,sBAAsB;AAG5B,SAAS,qBAAqB,MAAmF;AACtH,MAAI,OAAO,SAAS,YAAY,CAAC,KAAK,WAAW,mBAAmB,EAAG,QAAO;AAC9E,QAAM,OAAO,KAAK,MAAM,oBAAoB,MAAM;AAClD,QAAM,UAAU,KAAK,QAAQ,GAAG;AAChC,QAAM,aAAa,WAAW,IAAI,KAAK,MAAM,GAAG,OAAO,EAAE,KAAK,IAAI,KAAK,KAAK;AAC5E,QAAM,mBAAmB,WAAW,IAAI,KAAK,MAAM,UAAU,CAAC,EAAE,KAAK,IAAI;AAEzE,QAAM,eAAe,WAAW,MAAM,WAAW;AACjD,QAAM,cAAc,eAAe,WAAW,MAAM,GAAG,CAAC,aAAa,CAAC,EAAE,MAAM,EAAE,KAAK,IAAI;AACzF,QAAM,UAAU,eAAe,aAAa,CAAC,IAAI;AACjD,MAAI,CAAC,YAAa,QAAO;AACzB,SAAO,EAAE,aAAa,SAAS,UAAU,oBAAoB,OAAU;AACzE;AAYO,SAAS,kBAAkB,MAA+B;AAC/D,SAAO,OAAO,SAAS,YAAY,KAAK,WAAW,mBAAmB;AACxE;AAEA,SAAS,sBAAsB,KAAuB;AACpD,QAAM,MAAM,eAAe,QAAQ,IAAI,UAAU,OAAO,GAAG;AAC3D,SACE,IAAI,SAAS,oBAAoB,KACjC,IAAI,SAAS,qBAAqB,KAClC,IAAI,SAAS,kBAAkB,KAC/B,IAAI,SAAS,sBAAsB;AAEvC;AAMA,eAAsB,0BACpB,aACA,UAA4C,CAAC,GAC9B;AACf,MAAI,OAAO,gBAAgB,YAAY,YAAY,KAAK,EAAE,WAAW,GAAG;AACtE,UAAM,IAAI,MAAM,wEAAwE;AAAA,EAC1F;AACA,QAAM,MAAM,QAAQ,OAAO,QAAQ,IAAI;AACvC,QAAM,UAAU,QAAQ;AACxB,MAAI;AACF,UAAM;AAAA;AAAA,MAA0B;AAAA;AAChC;AAAA,EACF,SAAS,KAAK;AACZ,QAAI,CAAC,sBAAsB,GAAG,EAAG,OAAM;AAAA,EACzC;AACA,QAAM,cAAc,UAAU,GAAG,WAAW,IAAI,OAAO,KAAK;AAC5D,MAAI;AACF,aAAS,eAAe,WAAW,IAAI;AAAA,MACrC;AAAA,MACA,OAAO;AAAA,MACP,UAAU;AAAA,IACZ,CAAC;AAAA,EACH,SAAS,GAAG;AACV,UAAM,MAAM,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AACrD,UAAM,IAAI,MAAM,qCAAqC,WAAW,KAAK,GAAG,IAAI,EAAE,OAAO,EAAE,CAAC;AAAA,EAC1F;AACF;AAkBA,SAASG,wBAAuB,GAAY,SAAwB;AAClE,MAAI,aAAa,OAAO;AACtB,WAAO,IAAI,MAAM,GAAG,OAAO,KAAK,EAAE,OAAO,IAAI,EAAE,OAAO,EAAE,CAAC;AAAA,EAC3D;AACA,SAAO,IAAI,MAAM,GAAG,OAAO,KAAK,OAAO,CAAC,CAAC,EAAE;AAC7C;AAEA,eAAsB,mBACpB,MACA,UAAqC,CAAC,GACd;AACxB,MAAI,OAAO,SAAS,YAAY,KAAK,KAAK,EAAE,WAAW,EAAG,QAAO;AACjE,QAAM,SAAS,qBAAqB,IAAI;AACxC,MAAI,CAAC,OAAQ,QAAO;AACpB,QAAM,EAAE,aAAa,SAAS,UAAU,iBAAiB,IAAI;AAE7D,QAAM,sBAAsB,QAAQ,wBAAwB;AAC5D,QAAM,MAAM,QAAQ,OAAO,QAAQ,IAAI;AAEvC,QAAM,OAAO,YAA2B;AACtC,UAAM,kBAAkB,CAAC,WAAW,GAAG,EAAE,IAAI,CAAC;AAAA,EAChD;AAEA,MAAI;AACF,UAAM,KAAK;AAAA,EACb,SAAS,KAAK;AACZ,QAAI,uBAAuB,sBAAsB,GAAG,GAAG;AACrD,UAAI;AACF,cAAM,0BAA0B,aAAa,EAAE,SAAS,IAAI,CAAC;AAC7D,cAAM,KAAK;AAAA,MACb,SAAS,YAAY;AACnB,cAAMA,wBAAuB,YAAY,sDAAsD,WAAW,GAAG;AAAA,MAC/G;AAAA,IACF,OAAO;AACL,YAAMA,wBAAuB,KAAK,2CAA2C,WAAW,GAAG;AAAA,IAC7F;AAAA,EACF;AAGA,MAAI,oBAAoB,uBAAuB,CAAC,oBAAoB,gBAAgB,GAAG;AACrF,QAAI;AACF,YAAM,0BAA0B,aAAa,EAAE,SAAS,IAAI,CAAC;AAC7D,YAAM,KAAK;AAAA,IACb,SAAS,YAAY;AACnB,YAAMA,wBAAuB,YAAY,sDAAsD,WAAW,gBAAgB,gBAAgB,mBAAmB;AAAA,IAC/J;AACA,QAAI,CAAC,oBAAoB,gBAAgB,GAAG;AAC1C,YAAM,IAAI,MAAM,uBAAuB,WAAW,kDAAkD,gBAAgB,gFAAgF;AAAA,IACtM;AAAA,EACF;AAEA,MAAI,iBAAkB,QAAO;AAE7B,MAAI;AACF,UAAM,IAAI,MAAM;AAAA;AAAA,MAA0B;AAAA;AAC1C,QACE,OAAQ,EAAgD,2BAA2B,YACnF;AACA,aAAQ,EAA+C,uBAAuB;AAAA,IAChF;AAAA,EACF,QAAQ;AAAA,EAER;AACA,QAAM,IAAI;AAAA,IACR,iBAAiB,IAAI,qGAAqG,WAAW;AAAA,EACvI;AACF;AAQA,eAAsB,yBACpB,YACA,UAA2C,CAAC,GAC1B;AAClB,MAAI,cAAc,KAAM,QAAO;AAC/B,MAAI,MAAM,QAAQ,UAAU,GAAG;AAC7B,UAAM,MAAiB,CAAC;AACxB,aAAS,IAAI,GAAG,IAAI,WAAW,QAAQ,KAAK;AAC1C,UAAI;AACF,YAAI,KAAK,MAAM,yBAAyB,WAAW,CAAC,GAAG,OAAO,CAAC;AAAA,MACjE,SAAS,GAAG;AACV,cAAMA,wBAAuB,GAAG,qDAAqD,CAAC,EAAE;AAAA,MAC1F;AAAA,IACF;AACA,WAAO;AAAA,EACT;AACA,MAAI,OAAO,eAAe,UAAU;AAClC,UAAM,MAA+B,CAAC;AACtC,eAAW,CAAC,GAAG,CAAC,KAAK,OAAO,QAAQ,UAAU,GAAG;AAC/C,UAAI,MAAM,cAAc,kBAAkB,CAAC,GAAG;AAC5C,YAAI;AACF,gBAAM,WAAW,MAAM,mBAAmB,GAAa,OAAO;AAC9D,cAAI,CAAC,IAAI,YAAY;AAAA,QACvB,SAAS,GAAG;AACV,gBAAMA,wBAAuB,GAAG,0CAA0C,OAAO,CAAC,CAAC,GAAG;AAAA,QACxF;AACA;AAAA,MACF;AACA,UAAI;AACF,YAAI,CAAC,IAAI,MAAM,yBAAyB,GAAG,OAAO;AAAA,MACpD,SAAS,GAAG;AACV,cAAMA,wBAAuB,GAAG,iDAAiD,CAAC,GAAG;AAAA,MACvF;AAAA,IACF;AACA,WAAO;AAAA,EACT;AACA,SAAO;AACT;AAWA,eAAsB,oCACpB,SACwB;AACxB,MAAI;AACF,UAAM,EAAE,qBAAqB,KAAK,GAAG,KAAK,IAAI;AAC9C,UAAM,kBAAkB,MAAM,yBAAyB,QAAQ,cAAc,MAAM;AAAA,MACjF;AAAA,MACA;AAAA,IACF,CAAC;AACD,WAAO,6BAA6B;AAAA,MAClC,GAAG;AAAA,MACH,YAAY;AAAA,IACd,CAAC;AAAA,EACH,SAAS,GAAG;AACV,UAAMA,wBAAuB,GAAG,uDAAuD;AAAA,EACzF;AACF;;;AC5OA,SAAS,QAAAC,aAAY;AAarB,SAAS,2BAAmC;AAC1C,QAAM,MAAM,QAAQ,IAAI;AACxB,SAAOC,MAAK,KAAK,UAAU;AAC7B;AAUA,SAAS,uBACP,qBACuB;AACvB,MAAI,uBAAuB,KAAM,QAAO,CAAC;AACzC,MAAI,OAAO,wBAAwB,SAAU,QAAO,EAAE,YAAY,oBAAoB;AACtF,SAAO;AACT;AAEA,SAASC,wBAAuB,GAAY,SAAwB;AAClE,MAAI,aAAa,MAAO,QAAO,IAAI,MAAM,GAAG,OAAO,KAAK,EAAE,OAAO,IAAI,EAAE,OAAO,EAAE,CAAC;AACjF,SAAO,IAAI,MAAM,GAAG,OAAO,KAAK,OAAO,CAAC,CAAC,EAAE;AAC7C;AA4BA,eAAsB,eACpB,qBAC0D;AAC1D,MAAI;AACF,UAAM,UAAU,uBAAuB,mBAAmB;AAC1D,UAAM,aAAa,QAAQ,cAAc,yBAAyB;AAClE,UAAM,aAAa,cAAc,UAAU;AAC3C,QAAI,cAAc,MAAM;AACtB,YAAM,IAAI,MAAM,oBAAoB,UAAU,8DAA8D;AAAA,IAC9G;AACA,WAAO,MAAM,oCAAoC;AAAA,MAC/C;AAAA,MACA,qBAAqB,QAAQ,wBAAwB;AAAA,MACrD,KAAK,QAAQ,IAAI;AAAA,IACnB,CAAC;AAAA,EACH,SAAS,GAAG;AACV,QAAI,aAAa,SAAS,EAAE,QAAQ,SAAS,eAAe,EAAG,OAAM;AACrE,UAAMC,wBAAuB,GAAG,kCAAkC;AAAA,EACpE;AACF;","names":["model","apiKey","readFileSync","existsSync","existsSync","readFileSync","name","normalizeAgentLlMError","join","join","normalizeAgentLlMError","normalizeAgentLlMError"]}
@@ -7,4 +7,4 @@
7
7
  * Use --tools to enable the agent loop with get_weather/add_numbers tools.
8
8
  */
9
9
  export {};
10
- //# sourceMappingURL=cli.d.ts.map
10
+ //# sourceMappingURL=index.d.ts.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"index.d.ts","sourceRoot":"","sources":["../../src/cli/index.ts"],"names":[],"mappings":";AACA;;;;;;GAMG"}
@@ -1,9 +1,9 @@
1
1
  #!/usr/bin/env node
2
2
  import {
3
3
  createAgentLlM
4
- } from "./chunk-HKET766S.js";
4
+ } from "../chunk-MYGWM3YH.js";
5
5
 
6
- // src/cli.ts
6
+ // src/cli/index.ts
7
7
  import { tool } from "@langchain/core/tools";
8
8
  import { z } from "zod";
9
9
  import { HumanMessage, ToolMessage } from "@langchain/core/messages";
@@ -111,4 +111,4 @@ main().catch((err) => {
111
111
  console.error(err);
112
112
  process.exit(1);
113
113
  });
114
- //# sourceMappingURL=cli.js.map
114
+ //# sourceMappingURL=index.js.map