scrapex 1.0.0-alpha.1 → 1.0.0-beta.2

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (64) hide show
  1. package/README.md +164 -5
  2. package/dist/embeddings/index.cjs +52 -0
  3. package/dist/embeddings/index.d.cts +3 -0
  4. package/dist/embeddings/index.d.mts +3 -0
  5. package/dist/embeddings/index.mjs +4 -0
  6. package/dist/embeddings-BjNTQSG9.cjs +1455 -0
  7. package/dist/embeddings-BjNTQSG9.cjs.map +1 -0
  8. package/dist/embeddings-Bsymy_jA.mjs +1215 -0
  9. package/dist/embeddings-Bsymy_jA.mjs.map +1 -0
  10. package/dist/{enhancer-oM4BhYYS.cjs → enhancer-Cs_WyWtJ.cjs} +2 -51
  11. package/dist/enhancer-Cs_WyWtJ.cjs.map +1 -0
  12. package/dist/{enhancer-Q6CSc1gA.mjs → enhancer-INx5NlgO.mjs} +2 -45
  13. package/dist/enhancer-INx5NlgO.mjs.map +1 -0
  14. package/dist/http-base-CHLf-Tco.cjs +684 -0
  15. package/dist/http-base-CHLf-Tco.cjs.map +1 -0
  16. package/dist/http-base-DM7YNo6X.mjs +618 -0
  17. package/dist/http-base-DM7YNo6X.mjs.map +1 -0
  18. package/dist/index-Bvseqli-.d.cts +268 -0
  19. package/dist/index-Bvseqli-.d.cts.map +1 -0
  20. package/dist/index-CIFjNySr.d.mts +268 -0
  21. package/dist/index-CIFjNySr.d.mts.map +1 -0
  22. package/dist/index-D6qfjmZQ.d.mts +401 -0
  23. package/dist/index-D6qfjmZQ.d.mts.map +1 -0
  24. package/dist/index-RFSpP5g8.d.cts +401 -0
  25. package/dist/index-RFSpP5g8.d.cts.map +1 -0
  26. package/dist/index.cjs +171 -51
  27. package/dist/index.cjs.map +1 -1
  28. package/dist/index.d.cts +61 -2
  29. package/dist/index.d.cts.map +1 -1
  30. package/dist/index.d.mts +61 -2
  31. package/dist/index.d.mts.map +1 -1
  32. package/dist/index.mjs +129 -6
  33. package/dist/index.mjs.map +1 -1
  34. package/dist/llm/index.cjs +252 -233
  35. package/dist/llm/index.cjs.map +1 -1
  36. package/dist/llm/index.d.cts +132 -85
  37. package/dist/llm/index.d.cts.map +1 -1
  38. package/dist/llm/index.d.mts +132 -85
  39. package/dist/llm/index.d.mts.map +1 -1
  40. package/dist/llm/index.mjs +244 -236
  41. package/dist/llm/index.mjs.map +1 -1
  42. package/dist/parsers/index.cjs +10 -199
  43. package/dist/parsers/index.d.cts +2 -133
  44. package/dist/parsers/index.d.mts +2 -133
  45. package/dist/parsers/index.mjs +2 -191
  46. package/dist/parsers-Bneuws8x.cjs +569 -0
  47. package/dist/parsers-Bneuws8x.cjs.map +1 -0
  48. package/dist/parsers-DsawHeo0.mjs +482 -0
  49. package/dist/parsers-DsawHeo0.mjs.map +1 -0
  50. package/dist/types-BOcHQU9s.d.mts +831 -0
  51. package/dist/types-BOcHQU9s.d.mts.map +1 -0
  52. package/dist/types-DutdBpqd.d.cts +831 -0
  53. package/dist/types-DutdBpqd.d.cts.map +1 -0
  54. package/package.json +15 -16
  55. package/dist/enhancer-Q6CSc1gA.mjs.map +0 -1
  56. package/dist/enhancer-oM4BhYYS.cjs.map +0 -1
  57. package/dist/parsers/index.cjs.map +0 -1
  58. package/dist/parsers/index.d.cts.map +0 -1
  59. package/dist/parsers/index.d.mts.map +0 -1
  60. package/dist/parsers/index.mjs.map +0 -1
  61. package/dist/types-CNQZVW36.d.mts +0 -150
  62. package/dist/types-CNQZVW36.d.mts.map +0 -1
  63. package/dist/types-D0HYR95H.d.cts +0 -150
  64. package/dist/types-D0HYR95H.d.cts.map +0 -1
@@ -1,75 +1,99 @@
1
- import { a as EntitiesSchema, c as ScrapeError, i as ClassifySchema, n as enhance, o as SummarySchema, r as extract, s as TagsSchema, t as ask } from "../enhancer-Q6CSc1gA.mjs";
2
- import { createRequire } from "node:module";
1
+ import { d as ScrapeError, t as BaseHttpProvider } from "../http-base-DM7YNo6X.mjs";
2
+ import { a as EntitiesSchema, i as ClassifySchema, n as enhance, o as SummarySchema, r as extract, s as TagsSchema, t as ask } from "../enhancer-INx5NlgO.mjs";
3
+ import { z } from "zod";
3
4
 
4
- //#region rolldown:runtime
5
- var __require = /* @__PURE__ */ createRequire(import.meta.url);
6
-
7
- //#endregion
8
- //#region src/llm/anthropic.ts
9
- const DEFAULT_MODEL$1 = "claude-3-5-haiku-20241022";
10
- const DEFAULT_MAX_TOKENS$1 = 1024;
5
+ //#region src/llm/http.ts
11
6
  /**
12
- * Anthropic Claude provider
13
- *
14
- * Requires @anthropic-ai/sdk as a peer dependency.
15
- *
16
- * @example
17
- * ```ts
18
- * const provider = new AnthropicProvider({ apiKey: 'sk-...' });
19
- * const result = await scrape(url, { llm: provider, enhance: ['summarize'] });
20
- * ```
7
+ * HTTP-based LLM Provider using native fetch.
8
+ * Provides a unified interface for any REST-based LLM API.
21
9
  */
22
- var AnthropicProvider = class {
23
- name = "anthropic";
24
- client;
25
- model;
26
- constructor(config = {}) {
27
- const apiKey = config.apiKey ?? process.env.ANTHROPIC_API_KEY;
28
- if (!apiKey) throw new ScrapeError("Anthropic API key required. Set ANTHROPIC_API_KEY env var or pass apiKey in config.", "LLM_ERROR");
29
- this.model = config.model ?? DEFAULT_MODEL$1;
30
- try {
31
- const { Anthropic } = __require("@anthropic-ai/sdk");
32
- this.client = new Anthropic({
33
- apiKey,
34
- baseURL: config.baseUrl
10
+ /**
11
+ * HTTP-based LLM provider.
12
+ * Works with any REST API using native fetch.
13
+ */
14
+ var HttpLLMProvider = class extends BaseHttpProvider {
15
+ name;
16
+ requestBuilder;
17
+ responseMapper;
18
+ jsonMode;
19
+ constructor(config) {
20
+ super(config);
21
+ this.name = "http-llm";
22
+ this.jsonMode = config.jsonMode ?? false;
23
+ this.requestBuilder = config.requestBuilder ?? ((prompt, opts) => {
24
+ const messages = [];
25
+ if (opts.systemPrompt) messages.push({
26
+ role: "system",
27
+ content: opts.systemPrompt
35
28
  });
36
- } catch {
37
- throw new ScrapeError("@anthropic-ai/sdk is required for Anthropic provider. Install with: npm install @anthropic-ai/sdk", "LLM_ERROR");
38
- }
29
+ messages.push({
30
+ role: "user",
31
+ content: prompt
32
+ });
33
+ const request = {
34
+ model: this.model,
35
+ messages
36
+ };
37
+ if (opts.temperature !== void 0) request.temperature = opts.temperature;
38
+ if (opts.maxTokens !== void 0) request.max_tokens = opts.maxTokens;
39
+ return request;
40
+ });
41
+ this.responseMapper = config.responseMapper ?? ((response) => {
42
+ const resp = response;
43
+ if (Array.isArray(resp.choices) && resp.choices.length > 0) {
44
+ const choice = resp.choices[0];
45
+ if (choice.message?.content) return choice.message.content;
46
+ }
47
+ if (Array.isArray(resp.content)) {
48
+ const textBlock = resp.content.find((c) => c.type === "text");
49
+ if (textBlock?.text) return textBlock.text;
50
+ }
51
+ throw new ScrapeError("Unable to parse LLM response. Provide a custom responseMapper.", "VALIDATION_ERROR");
52
+ });
39
53
  }
54
+ /**
55
+ * Generate a text completion.
56
+ */
40
57
  async complete(prompt, options = {}) {
41
- try {
42
- const content = (await this.client.messages.create({
43
- model: this.model,
44
- max_tokens: options.maxTokens ?? DEFAULT_MAX_TOKENS$1,
45
- messages: [{
46
- role: "user",
47
- content: prompt
48
- }],
49
- system: options.systemPrompt,
50
- temperature: options.temperature
51
- })).content[0];
52
- if (content?.type === "text" && content.text) return content.text;
53
- throw new ScrapeError("Unexpected or empty response from Anthropic", "LLM_ERROR");
54
- } catch (error) {
55
- if (error instanceof ScrapeError) throw error;
56
- throw new ScrapeError(`Anthropic API error: ${error instanceof Error ? error.message : String(error)}`, "LLM_ERROR", void 0, error instanceof Error ? error : void 0);
57
- }
58
+ let body = this.requestBuilder(prompt, options);
59
+ if (this.jsonMode && typeof body === "object" && body !== null) body = {
60
+ ...body,
61
+ response_format: { type: "json_object" }
62
+ };
63
+ const { data } = await this.fetch(this.baseUrl, { body });
64
+ const content = this.responseMapper(data);
65
+ if (!content) throw new ScrapeError("Empty response from LLM", "LLM_ERROR");
66
+ return content;
58
67
  }
68
+ /**
69
+ * Generate a structured JSON completion with Zod validation.
70
+ */
59
71
  async completeJSON(prompt, schema, options = {}) {
60
72
  const jsonPrompt = `${prompt}
61
73
 
62
74
  Respond ONLY with valid JSON matching this schema:
63
- ${JSON.stringify(zodToJsonSchema$1(schema), null, 2)}
75
+ ${JSON.stringify(zodToJsonSchema(schema), null, 2)}
64
76
 
65
77
  Do not include any explanation or markdown formatting. Just the JSON object.`;
66
- const response = await this.complete(jsonPrompt, {
78
+ const useJsonMode = this.jsonMode;
79
+ let body = this.requestBuilder(jsonPrompt, {
67
80
  ...options,
68
81
  systemPrompt: options.systemPrompt ?? "You are a helpful assistant that responds only with valid JSON."
69
82
  });
83
+ if (useJsonMode && typeof body === "object" && body !== null) body = {
84
+ ...body,
85
+ response_format: { type: "json_object" }
86
+ };
87
+ const { data } = await this.fetch(this.baseUrl, { body });
88
+ const content = this.responseMapper(data);
89
+ if (!content) throw new ScrapeError("Empty response from LLM", "LLM_ERROR");
70
90
  try {
71
- const jsonMatch = response.match(/\{[\s\S]*\}/);
72
- if (!jsonMatch) throw new Error("No JSON object found in response");
91
+ const trimmed = content.trim();
92
+ try {
93
+ return schema.parse(JSON.parse(trimmed));
94
+ } catch {}
95
+ const jsonMatch = content.match(/[[{][\s\S]*[\]}]/);
96
+ if (!jsonMatch) throw new Error("No JSON found in response");
73
97
  const parsed = JSON.parse(jsonMatch[0]);
74
98
  return schema.parse(parsed);
75
99
  } catch (error) {
@@ -78,233 +102,217 @@ Do not include any explanation or markdown formatting. Just the JSON object.`;
78
102
  }
79
103
  };
80
104
  /**
81
- * Convert a Zod schema to a simple JSON Schema representation
82
- * (simplified version for prompt engineering)
105
+ * Convert a Zod schema to a JSON Schema representation.
106
+ * Uses Zod's built-in toJSONSchema method (Zod 4+).
107
+ * Used for prompting LLMs to return structured data.
83
108
  */
84
- function zodToJsonSchema$1(schema) {
109
+ function zodToJsonSchema(schema) {
110
+ if (typeof z.toJSONSchema === "function") {
111
+ const { $schema, ...rest } = z.toJSONSchema(schema);
112
+ return rest;
113
+ }
85
114
  const def = schema._def;
86
- switch (def.typeName) {
87
- case "ZodObject": {
115
+ switch (def.type) {
116
+ case "object": {
88
117
  const shape = schema.shape;
89
118
  const properties = {};
90
- for (const [key, value] of Object.entries(shape)) properties[key] = zodToJsonSchema$1(value);
119
+ const required = [];
120
+ for (const [key, value] of Object.entries(shape)) {
121
+ properties[key] = zodToJsonSchema(value);
122
+ if (value._def.type !== "optional") required.push(key);
123
+ }
91
124
  return {
92
125
  type: "object",
93
- properties
126
+ properties,
127
+ required
94
128
  };
95
129
  }
96
- case "ZodArray": return {
130
+ case "array": return {
97
131
  type: "array",
98
- items: zodToJsonSchema$1(def.type)
99
- };
100
- case "ZodString": return { type: "string" };
101
- case "ZodNumber": return { type: "number" };
102
- case "ZodBoolean": return { type: "boolean" };
103
- case "ZodEnum": return {
104
- type: "string",
105
- enum: def.values
132
+ items: zodToJsonSchema(def.element)
106
133
  };
134
+ case "string": return { type: "string" };
135
+ case "number": return { type: "number" };
136
+ case "boolean": return { type: "boolean" };
137
+ case "enum": {
138
+ const enumDef = def;
139
+ return {
140
+ type: "string",
141
+ enum: Object.values(enumDef.entries)
142
+ };
143
+ }
144
+ case "optional": return zodToJsonSchema(def.innerType);
107
145
  default: return { type: "string" };
108
146
  }
109
147
  }
110
148
 
111
149
  //#endregion
112
- //#region src/llm/openai.ts
113
- const DEFAULT_MODEL = "gpt-4o-mini";
114
- const DEFAULT_MAX_TOKENS = 1024;
115
- const DEFAULT_BASE_URL = "https://api.openai.com/v1";
150
+ //#region src/llm/presets.ts
116
151
  /**
117
- * OpenAI-compatible provider
118
- *
119
- * Works with:
120
- * - OpenAI API
121
- * - Ollama (http://localhost:11434/v1)
122
- * - LM Studio (http://localhost:1234/v1)
123
- * - LocalAI
124
- * - vLLM
125
- * - Any OpenAI-compatible API
126
- *
127
- * Requires `openai` as a peer dependency.
152
+ * Preset factory functions for common LLM providers.
153
+ * All presets use the HttpLLMProvider with appropriate configuration.
154
+ */
155
+ /**
156
+ * Create an OpenAI LLM provider.
128
157
  *
129
158
  * @example
130
159
  * ```ts
131
- * // OpenAI
132
- * const provider = new OpenAIProvider({ apiKey: 'sk-...' });
133
- *
134
- * // Ollama
135
- * const provider = new OpenAIProvider({
136
- * baseUrl: 'http://localhost:11434/v1',
137
- * model: 'llama3.2',
138
- * apiKey: 'ollama' // Ollama doesn't require a real key
139
- * });
160
+ * const provider = createOpenAI({ apiKey: 'sk-...' });
161
+ * const result = await scrape(url, { llm: provider, enhance: ['summarize'] });
162
+ * ```
163
+ */
164
+ function createOpenAI(options) {
165
+ const apiKey = options?.apiKey ?? process.env.OPENAI_API_KEY;
166
+ if (!apiKey) throw new Error("OpenAI API key required. Set OPENAI_API_KEY env var or pass apiKey option.");
167
+ return new HttpLLMProvider({
168
+ baseUrl: options?.baseUrl ?? "https://api.openai.com/v1/chat/completions",
169
+ model: options?.model ?? "gpt-4o-mini",
170
+ headers: { Authorization: `Bearer ${apiKey}` },
171
+ jsonMode: true
172
+ });
173
+ }
174
+ /**
175
+ * Create an Anthropic Claude LLM provider.
140
176
  *
141
- * // LM Studio
142
- * const provider = new OpenAIProvider({
143
- * baseUrl: 'http://localhost:1234/v1',
144
- * model: 'local-model',
145
- * apiKey: 'lm-studio'
146
- * });
177
+ * @example
178
+ * ```ts
179
+ * const provider = createAnthropic({ apiKey: 'sk-...' });
180
+ * const result = await scrape(url, { llm: provider, enhance: ['summarize'] });
147
181
  * ```
148
182
  */
149
- var OpenAIProvider = class {
150
- name = "openai";
151
- client;
152
- model;
153
- constructor(config = {}) {
154
- const apiKey = config.apiKey ?? process.env.OPENAI_API_KEY;
155
- const baseUrl = config.baseUrl ?? DEFAULT_BASE_URL;
156
- if (!apiKey && baseUrl === DEFAULT_BASE_URL) throw new ScrapeError("OpenAI API key required. Set OPENAI_API_KEY env var or pass apiKey in config.", "LLM_ERROR");
157
- this.model = config.model ?? DEFAULT_MODEL;
158
- try {
159
- const { OpenAI } = __require("openai");
160
- this.client = new OpenAI({
161
- apiKey: apiKey ?? "local",
162
- baseURL: baseUrl
163
- });
164
- } catch {
165
- throw new ScrapeError("openai package is required for OpenAI provider. Install with: npm install openai", "LLM_ERROR");
166
- }
167
- }
168
- async complete(prompt, options = {}) {
169
- try {
170
- const client = this.client;
171
- const messages = [];
172
- if (options.systemPrompt) messages.push({
173
- role: "system",
174
- content: options.systemPrompt
175
- });
176
- messages.push({
177
- role: "user",
178
- content: prompt
179
- });
180
- const content = (await client.chat.completions.create({
181
- model: this.model,
182
- max_tokens: options.maxTokens ?? DEFAULT_MAX_TOKENS,
183
- messages,
184
- temperature: options.temperature
185
- })).choices[0]?.message?.content;
186
- if (!content) throw new ScrapeError("Empty response from OpenAI", "LLM_ERROR");
187
- return content;
188
- } catch (error) {
189
- if (error instanceof ScrapeError) throw error;
190
- throw new ScrapeError(`OpenAI API error: ${error instanceof Error ? error.message : String(error)}`, "LLM_ERROR", void 0, error instanceof Error ? error : void 0);
191
- }
192
- }
193
- async completeJSON(prompt, schema, options = {}) {
194
- const client = this.client;
195
- try {
196
- const messages = [{
197
- role: "system",
198
- content: options.systemPrompt ?? "You are a helpful assistant that extracts information from content."
199
- }, {
183
+ function createAnthropic(options) {
184
+ const apiKey = options?.apiKey ?? process.env.ANTHROPIC_API_KEY;
185
+ if (!apiKey) throw new Error("Anthropic API key required. Set ANTHROPIC_API_KEY env var or pass apiKey option.");
186
+ const model = options?.model ?? "claude-3-5-haiku-20241022";
187
+ return new HttpLLMProvider({
188
+ baseUrl: "https://api.anthropic.com/v1/messages",
189
+ model,
190
+ headers: {
191
+ "x-api-key": apiKey,
192
+ "anthropic-version": "2023-06-01"
193
+ },
194
+ requestBuilder: (prompt, opts) => ({
195
+ model,
196
+ max_tokens: opts.maxTokens ?? 1024,
197
+ messages: [{
200
198
  role: "user",
201
199
  content: prompt
202
- }];
203
- const content = (await client.chat.completions.create({
204
- model: this.model,
205
- max_tokens: options.maxTokens ?? DEFAULT_MAX_TOKENS,
206
- messages,
207
- temperature: options.temperature,
208
- response_format: { type: "json_object" }
209
- })).choices[0]?.message?.content;
210
- if (!content) throw new ScrapeError("Empty response from OpenAI", "LLM_ERROR");
211
- const parsed = JSON.parse(content);
212
- return schema.parse(parsed);
213
- } catch (error) {
214
- if (error instanceof ScrapeError) throw error;
215
- const jsonPrompt = `${prompt}
216
-
217
- Respond ONLY with valid JSON matching this schema:
218
- ${JSON.stringify(zodToJsonSchema(schema), null, 2)}
219
-
220
- Do not include any explanation or markdown formatting. Just the JSON object.`;
221
- const response = await this.complete(jsonPrompt, {
222
- ...options,
223
- systemPrompt: "You respond only with valid JSON."
224
- });
225
- try {
226
- const jsonMatch = response.match(/\{[\s\S]*\}/);
227
- if (!jsonMatch) throw new Error("No JSON object found in response");
228
- const parsed = JSON.parse(jsonMatch[0]);
229
- return schema.parse(parsed);
230
- } catch (parseError) {
231
- throw new ScrapeError(`Failed to parse LLM response as JSON: ${parseError instanceof Error ? parseError.message : String(parseError)}`, "VALIDATION_ERROR", void 0, parseError instanceof Error ? parseError : void 0);
232
- }
233
- }
234
- }
235
- };
236
- /**
237
- * Convert a Zod schema to JSON Schema for structured outputs
238
- */
239
- function zodToJsonSchema(schema) {
240
- const def = schema._def;
241
- switch (def.typeName) {
242
- case "ZodObject": {
243
- const shape = schema.shape;
244
- const properties = {};
245
- const required = [];
246
- for (const [key, value] of Object.entries(shape)) {
247
- properties[key] = zodToJsonSchema(value);
248
- if (value._def.typeName !== "ZodOptional") required.push(key);
249
- }
250
- return {
251
- type: "object",
252
- properties,
253
- required
254
- };
255
- }
256
- case "ZodArray": return {
257
- type: "array",
258
- items: zodToJsonSchema(def.type)
259
- };
260
- case "ZodString": return { type: "string" };
261
- case "ZodNumber": return { type: "number" };
262
- case "ZodBoolean": return { type: "boolean" };
263
- case "ZodEnum": return {
264
- type: "string",
265
- enum: def.values
266
- };
267
- case "ZodOptional": return zodToJsonSchema(def.innerType);
268
- default: return { type: "string" };
269
- }
200
+ }],
201
+ ...opts.systemPrompt && { system: opts.systemPrompt },
202
+ ...opts.temperature !== void 0 && { temperature: opts.temperature }
203
+ }),
204
+ responseMapper: (res) => res.content.find((item) => item.type === "text")?.text ?? ""
205
+ });
270
206
  }
271
207
  /**
272
- * Create an OpenAI provider with default settings
208
+ * Create a Groq LLM provider.
209
+ * Groq provides fast inference for open-source models.
210
+ *
211
+ * @example
212
+ * ```ts
213
+ * const provider = createGroq({ model: 'llama-3.1-70b-versatile' });
214
+ * ```
273
215
  */
274
- function createOpenAI(config) {
275
- return new OpenAIProvider(config);
216
+ function createGroq(options) {
217
+ const apiKey = options?.apiKey ?? process.env.GROQ_API_KEY;
218
+ if (!apiKey) throw new Error("Groq API key required. Set GROQ_API_KEY env var or pass apiKey option.");
219
+ return new HttpLLMProvider({
220
+ baseUrl: "https://api.groq.com/openai/v1/chat/completions",
221
+ model: options?.model ?? "llama-3.1-70b-versatile",
222
+ headers: { Authorization: `Bearer ${apiKey}` },
223
+ jsonMode: true
224
+ });
276
225
  }
277
226
  /**
278
- * Create an Ollama provider
227
+ * Create an Ollama LLM provider for local models.
279
228
  *
280
229
  * @example
281
230
  * ```ts
282
231
  * const provider = createOllama({ model: 'llama3.2' });
283
232
  * ```
284
233
  */
285
- function createOllama(config = { model: "llama3.2" }) {
286
- return new OpenAIProvider({
287
- baseUrl: `http://localhost:${config.port ?? 11434}/v1`,
288
- model: config.model,
289
- apiKey: "ollama"
234
+ function createOllama(options) {
235
+ return new HttpLLMProvider({
236
+ baseUrl: options.baseUrl ?? "http://localhost:11434/v1/chat/completions",
237
+ model: options.model,
238
+ requireHttps: false,
239
+ allowPrivate: true
290
240
  });
291
241
  }
292
242
  /**
293
- * Create an LM Studio provider
243
+ * Create an LM Studio LLM provider for local models.
294
244
  *
295
245
  * @example
296
246
  * ```ts
297
247
  * const provider = createLMStudio({ model: 'local-model' });
298
248
  * ```
299
249
  */
300
- function createLMStudio(config = { model: "local-model" }) {
301
- return new OpenAIProvider({
302
- baseUrl: `http://localhost:${config.port ?? 1234}/v1`,
303
- model: config.model,
304
- apiKey: "lm-studio"
250
+ function createLMStudio(options) {
251
+ return new HttpLLMProvider({
252
+ baseUrl: options.baseUrl ?? "http://localhost:1234/v1/chat/completions",
253
+ model: options.model,
254
+ requireHttps: false,
255
+ allowPrivate: true
256
+ });
257
+ }
258
+ /**
259
+ * Create a Together AI LLM provider.
260
+ *
261
+ * @example
262
+ * ```ts
263
+ * const provider = createTogether({ model: 'meta-llama/Llama-3.2-3B-Instruct-Turbo' });
264
+ * ```
265
+ */
266
+ function createTogether(options) {
267
+ const apiKey = options?.apiKey ?? process.env.TOGETHER_API_KEY;
268
+ if (!apiKey) throw new Error("Together API key required. Set TOGETHER_API_KEY env var or pass apiKey option.");
269
+ return new HttpLLMProvider({
270
+ baseUrl: "https://api.together.xyz/v1/chat/completions",
271
+ model: options?.model ?? "meta-llama/Llama-3.2-3B-Instruct-Turbo",
272
+ headers: { Authorization: `Bearer ${apiKey}` },
273
+ jsonMode: true
274
+ });
275
+ }
276
+ /**
277
+ * Create an OpenRouter LLM provider.
278
+ * OpenRouter provides access to many models through a unified API.
279
+ *
280
+ * @example
281
+ * ```ts
282
+ * const provider = createOpenRouter({
283
+ * model: 'anthropic/claude-3.5-sonnet',
284
+ * });
285
+ * ```
286
+ */
287
+ function createOpenRouter(options) {
288
+ const apiKey = options.apiKey ?? process.env.OPENROUTER_API_KEY;
289
+ if (!apiKey) throw new Error("OpenRouter API key required. Set OPENROUTER_API_KEY env var or pass apiKey option.");
290
+ const headers = { Authorization: `Bearer ${apiKey}` };
291
+ if (options.siteUrl) headers["HTTP-Referer"] = options.siteUrl;
292
+ if (options.siteName) headers["X-Title"] = options.siteName;
293
+ return new HttpLLMProvider({
294
+ baseUrl: "https://openrouter.ai/api/v1/chat/completions",
295
+ model: options.model,
296
+ headers
305
297
  });
306
298
  }
299
+ /**
300
+ * Create a generic HTTP LLM provider.
301
+ * Use this for any OpenAI-compatible API.
302
+ *
303
+ * @example
304
+ * ```ts
305
+ * const provider = createHttpLLM({
306
+ * baseUrl: 'https://my-api.com/v1/chat/completions',
307
+ * model: 'my-model',
308
+ * headers: { Authorization: 'Bearer ...' },
309
+ * });
310
+ * ```
311
+ */
312
+ function createHttpLLM(config) {
313
+ return new HttpLLMProvider(config);
314
+ }
307
315
 
308
316
  //#endregion
309
- export { AnthropicProvider, ClassifySchema, EntitiesSchema, OpenAIProvider, SummarySchema, TagsSchema, ask, createLMStudio, createOllama, createOpenAI, enhance, extract };
317
+ export { ClassifySchema, EntitiesSchema, HttpLLMProvider, SummarySchema, TagsSchema, ask, createAnthropic, createGroq, createHttpLLM, createLMStudio, createOllama, createOpenAI, createOpenRouter, createTogether, enhance, extract, zodToJsonSchema };
310
318
  //# sourceMappingURL=index.mjs.map
@@ -1 +1 @@
1
- {"version":3,"file":"index.mjs","names":["DEFAULT_MODEL","DEFAULT_MAX_TOKENS","zodToJsonSchema","properties: Record<string, object>","messages: Array<{ role: 'system' | 'user'; content: string }>","properties: Record<string, object>","required: string[]"],"sources":["../../src/llm/anthropic.ts","../../src/llm/openai.ts"],"sourcesContent":["import type { z } from 'zod';\nimport { ScrapeError } from '@/core/errors.js';\nimport type { AnthropicConfig, CompletionOptions, LLMProvider } from './types.js';\n\nconst DEFAULT_MODEL = 'claude-3-5-haiku-20241022';\nconst DEFAULT_MAX_TOKENS = 1024;\n\n/**\n * Anthropic Claude provider\n *\n * Requires @anthropic-ai/sdk as a peer dependency.\n *\n * @example\n * ```ts\n * const provider = new AnthropicProvider({ apiKey: 'sk-...' });\n * const result = await scrape(url, { llm: provider, enhance: ['summarize'] });\n * ```\n */\nexport class AnthropicProvider implements LLMProvider {\n readonly name = 'anthropic';\n private client: unknown;\n private model: string;\n\n constructor(config: AnthropicConfig = {}) {\n const apiKey = config.apiKey ?? process.env.ANTHROPIC_API_KEY;\n if (!apiKey) {\n throw new ScrapeError(\n 'Anthropic API key required. Set ANTHROPIC_API_KEY env var or pass apiKey in config.',\n 'LLM_ERROR'\n );\n }\n\n this.model = config.model ?? DEFAULT_MODEL;\n\n // Dynamic import to avoid requiring the SDK if not used\n try {\n // eslint-disable-next-line @typescript-eslint/no-require-imports\n const { Anthropic } = require('@anthropic-ai/sdk') as typeof import('@anthropic-ai/sdk');\n this.client = new Anthropic({\n apiKey,\n baseURL: config.baseUrl,\n });\n } catch {\n throw new ScrapeError(\n '@anthropic-ai/sdk is required for Anthropic provider. Install with: npm install @anthropic-ai/sdk',\n 'LLM_ERROR'\n );\n }\n }\n\n async complete(prompt: string, options: CompletionOptions = {}): Promise<string> {\n try {\n const client = this.client as import('@anthropic-ai/sdk').Anthropic;\n const response = await client.messages.create({\n model: this.model,\n max_tokens: options.maxTokens ?? DEFAULT_MAX_TOKENS,\n messages: [{ role: 'user', content: prompt }],\n system: options.systemPrompt,\n temperature: options.temperature,\n });\n\n const content = response.content[0];\n if (content?.type === 'text' && content.text) {\n return content.text;\n }\n\n throw new ScrapeError('Unexpected or empty response from Anthropic', 'LLM_ERROR');\n } catch (error) {\n if (error instanceof ScrapeError) throw error;\n throw new ScrapeError(\n `Anthropic API error: ${error instanceof Error ? error.message : String(error)}`,\n 'LLM_ERROR',\n undefined,\n error instanceof Error ? error : undefined\n );\n }\n }\n\n async completeJSON<T>(\n prompt: string,\n schema: z.ZodType<T>,\n options: CompletionOptions = {}\n ): Promise<T> {\n const jsonPrompt = `${prompt}\n\nRespond ONLY with valid JSON matching this schema:\n${JSON.stringify(zodToJsonSchema(schema), null, 2)}\n\nDo not include any explanation or markdown formatting. Just the JSON object.`;\n\n const response = await this.complete(jsonPrompt, {\n ...options,\n systemPrompt:\n options.systemPrompt ?? 'You are a helpful assistant that responds only with valid JSON.',\n });\n\n try {\n // Try to extract JSON from the response\n const jsonMatch = response.match(/\\{[\\s\\S]*\\}/);\n if (!jsonMatch) {\n throw new Error('No JSON object found in response');\n }\n\n const parsed = JSON.parse(jsonMatch[0]);\n return schema.parse(parsed);\n } catch (error) {\n throw new ScrapeError(\n `Failed to parse LLM response as JSON: ${error instanceof Error ? error.message : String(error)}`,\n 'VALIDATION_ERROR',\n undefined,\n error instanceof Error ? error : undefined\n );\n }\n }\n}\n\n/**\n * Convert a Zod schema to a simple JSON Schema representation\n * (simplified version for prompt engineering)\n */\nfunction zodToJsonSchema(schema: z.ZodType<unknown>): object {\n const def = (schema as z.ZodType<unknown> & { _def: { typeName: string } })._def;\n\n switch (def.typeName) {\n case 'ZodObject': {\n const shape = (schema as z.ZodObject<z.ZodRawShape>).shape;\n const properties: Record<string, object> = {};\n for (const [key, value] of Object.entries(shape)) {\n properties[key] = zodToJsonSchema(value as z.ZodType<unknown>);\n }\n return { type: 'object', properties };\n }\n case 'ZodArray': {\n const arrayDef = def as unknown as { type: z.ZodType<unknown> };\n return { type: 'array', items: zodToJsonSchema(arrayDef.type) };\n }\n case 'ZodString':\n return { type: 'string' };\n case 'ZodNumber':\n return { type: 'number' };\n case 'ZodBoolean':\n return { type: 'boolean' };\n case 'ZodEnum': {\n const enumDef = def as unknown as { values: string[] };\n return { type: 'string', enum: enumDef.values };\n }\n default:\n return { type: 'string' };\n }\n}\n","import type { z } from 'zod';\nimport { ScrapeError } from '@/core/errors.js';\nimport type { CompletionOptions, LLMProvider, OpenAICompatibleConfig } from './types.js';\n\nconst DEFAULT_MODEL = 'gpt-4o-mini';\nconst DEFAULT_MAX_TOKENS = 1024;\nconst DEFAULT_BASE_URL = 'https://api.openai.com/v1';\n\n/**\n * OpenAI-compatible provider\n *\n * Works with:\n * - OpenAI API\n * - Ollama (http://localhost:11434/v1)\n * - LM Studio (http://localhost:1234/v1)\n * - LocalAI\n * - vLLM\n * - Any OpenAI-compatible API\n *\n * Requires `openai` as a peer dependency.\n *\n * @example\n * ```ts\n * // OpenAI\n * const provider = new OpenAIProvider({ apiKey: 'sk-...' });\n *\n * // Ollama\n * const provider = new OpenAIProvider({\n * baseUrl: 'http://localhost:11434/v1',\n * model: 'llama3.2',\n * apiKey: 'ollama' // Ollama doesn't require a real key\n * });\n *\n * // LM Studio\n * const provider = new OpenAIProvider({\n * baseUrl: 'http://localhost:1234/v1',\n * model: 'local-model',\n * apiKey: 'lm-studio'\n * });\n * ```\n */\nexport class OpenAIProvider implements LLMProvider {\n readonly name = 'openai';\n private client: unknown;\n private model: string;\n\n constructor(config: OpenAICompatibleConfig = {}) {\n const apiKey = config.apiKey ?? process.env.OPENAI_API_KEY;\n const baseUrl = config.baseUrl ?? DEFAULT_BASE_URL;\n\n // Only require API key for OpenAI (not for local providers)\n if (!apiKey && baseUrl === DEFAULT_BASE_URL) {\n throw new ScrapeError(\n 'OpenAI API key required. Set OPENAI_API_KEY env var or pass apiKey in config.',\n 'LLM_ERROR'\n );\n }\n\n this.model = config.model ?? DEFAULT_MODEL;\n\n // Dynamic import to avoid requiring the SDK if not used\n try {\n // eslint-disable-next-line @typescript-eslint/no-require-imports\n const { OpenAI } = require('openai') as typeof import('openai');\n this.client = new OpenAI({\n apiKey: apiKey ?? 'local', // Use 'local' as placeholder for local providers\n baseURL: baseUrl,\n });\n } catch {\n throw new ScrapeError(\n 'openai package is required for OpenAI provider. Install with: npm install openai',\n 'LLM_ERROR'\n );\n }\n }\n\n async complete(prompt: string, options: CompletionOptions = {}): Promise<string> {\n try {\n const client = this.client as import('openai').OpenAI;\n const messages: Array<{ role: 'system' | 'user'; content: string }> = [];\n\n if (options.systemPrompt) {\n messages.push({ role: 'system', content: options.systemPrompt });\n }\n messages.push({ role: 'user', content: prompt });\n\n const response = await client.chat.completions.create({\n model: this.model,\n max_tokens: options.maxTokens ?? DEFAULT_MAX_TOKENS,\n messages,\n temperature: options.temperature,\n });\n\n const content = response.choices[0]?.message?.content;\n if (!content) {\n throw new ScrapeError('Empty response from OpenAI', 'LLM_ERROR');\n }\n\n return content;\n } catch (error) {\n if (error instanceof ScrapeError) throw error;\n throw new ScrapeError(\n `OpenAI API error: ${error instanceof Error ? error.message : String(error)}`,\n 'LLM_ERROR',\n undefined,\n error instanceof Error ? error : undefined\n );\n }\n }\n\n async completeJSON<T>(\n prompt: string,\n schema: z.ZodType<T>,\n options: CompletionOptions = {}\n ): Promise<T> {\n const client = this.client as import('openai').OpenAI;\n\n try {\n // Use JSON mode for structured outputs\n const messages: Array<{ role: 'system' | 'user'; content: string }> = [\n {\n role: 'system',\n content:\n options.systemPrompt ??\n 'You are a helpful assistant that extracts information from content.',\n },\n { role: 'user', content: prompt },\n ];\n\n const response = await client.chat.completions.create({\n model: this.model,\n max_tokens: options.maxTokens ?? DEFAULT_MAX_TOKENS,\n messages,\n temperature: options.temperature,\n response_format: { type: 'json_object' },\n });\n\n const content = response.choices[0]?.message?.content;\n if (!content) {\n throw new ScrapeError('Empty response from OpenAI', 'LLM_ERROR');\n }\n\n const parsed = JSON.parse(content);\n return schema.parse(parsed);\n } catch (error) {\n // Fallback to regular completion with JSON instruction\n if (error instanceof ScrapeError) throw error;\n\n // If structured output failed, try regular completion\n const jsonPrompt = `${prompt}\n\nRespond ONLY with valid JSON matching this schema:\n${JSON.stringify(zodToJsonSchema(schema), null, 2)}\n\nDo not include any explanation or markdown formatting. Just the JSON object.`;\n\n const response = await this.complete(jsonPrompt, {\n ...options,\n systemPrompt: 'You respond only with valid JSON.',\n });\n\n try {\n const jsonMatch = response.match(/\\{[\\s\\S]*\\}/);\n if (!jsonMatch) {\n throw new Error('No JSON object found in response');\n }\n\n const parsed = JSON.parse(jsonMatch[0]);\n return schema.parse(parsed);\n } catch (parseError) {\n throw new ScrapeError(\n `Failed to parse LLM response as JSON: ${parseError instanceof Error ? parseError.message : String(parseError)}`,\n 'VALIDATION_ERROR',\n undefined,\n parseError instanceof Error ? parseError : undefined\n );\n }\n }\n }\n}\n\n/**\n * Convert a Zod schema to JSON Schema for structured outputs\n */\nfunction zodToJsonSchema(schema: z.ZodType<unknown>): object {\n const def = (schema as z.ZodType<unknown> & { _def: { typeName: string } })._def;\n\n switch (def.typeName) {\n case 'ZodObject': {\n const shape = (schema as z.ZodObject<z.ZodRawShape>).shape;\n const properties: Record<string, object> = {};\n const required: string[] = [];\n\n for (const [key, value] of Object.entries(shape)) {\n properties[key] = zodToJsonSchema(value as z.ZodType<unknown>);\n // Assume all fields are required unless wrapped in ZodOptional\n const valueDef = (value as z.ZodType<unknown> & { _def: { typeName: string } })._def;\n if (valueDef.typeName !== 'ZodOptional') {\n required.push(key);\n }\n }\n return { type: 'object', properties, required };\n }\n case 'ZodArray': {\n const arrayDef = def as unknown as { type: z.ZodType<unknown> };\n return { type: 'array', items: zodToJsonSchema(arrayDef.type) };\n }\n case 'ZodString':\n return { type: 'string' };\n case 'ZodNumber':\n return { type: 'number' };\n case 'ZodBoolean':\n return { type: 'boolean' };\n case 'ZodEnum': {\n const enumDef = def as unknown as { values: string[] };\n return { type: 'string', enum: enumDef.values };\n }\n case 'ZodOptional': {\n const optionalDef = def as unknown as { innerType: z.ZodType<unknown> };\n return zodToJsonSchema(optionalDef.innerType);\n }\n default:\n return { type: 'string' };\n }\n}\n\n// Convenience factory functions\n\n/**\n * Create an OpenAI provider with default settings\n */\nexport function createOpenAI(config?: OpenAICompatibleConfig): OpenAIProvider {\n return new OpenAIProvider(config);\n}\n\n/**\n * Create an Ollama provider\n *\n * @example\n * ```ts\n * const provider = createOllama({ model: 'llama3.2' });\n * ```\n */\nexport function createOllama(\n config: { model: string; port?: number } = { model: 'llama3.2' }\n): OpenAIProvider {\n return new OpenAIProvider({\n baseUrl: `http://localhost:${config.port ?? 11434}/v1`,\n model: config.model,\n apiKey: 'ollama',\n });\n}\n\n/**\n * Create an LM Studio provider\n *\n * @example\n * ```ts\n * const provider = createLMStudio({ model: 'local-model' });\n * ```\n */\nexport function createLMStudio(\n config: { model: string; port?: number } = { model: 'local-model' }\n): OpenAIProvider {\n return new OpenAIProvider({\n baseUrl: `http://localhost:${config.port ?? 1234}/v1`,\n model: config.model,\n apiKey: 'lm-studio',\n });\n}\n"],"mappings":";;;;;;;;AAIA,MAAMA,kBAAgB;AACtB,MAAMC,uBAAqB;;;;;;;;;;;;AAa3B,IAAa,oBAAb,MAAsD;CACpD,AAAS,OAAO;CAChB,AAAQ;CACR,AAAQ;CAER,YAAY,SAA0B,EAAE,EAAE;EACxC,MAAM,SAAS,OAAO,UAAU,QAAQ,IAAI;AAC5C,MAAI,CAAC,OACH,OAAM,IAAI,YACR,uFACA,YACD;AAGH,OAAK,QAAQ,OAAO,SAASD;AAG7B,MAAI;GAEF,MAAM,EAAE,wBAAsB,oBAAoB;AAClD,QAAK,SAAS,IAAI,UAAU;IAC1B;IACA,SAAS,OAAO;IACjB,CAAC;UACI;AACN,SAAM,IAAI,YACR,qGACA,YACD;;;CAIL,MAAM,SAAS,QAAgB,UAA6B,EAAE,EAAmB;AAC/E,MAAI;GAUF,MAAM,WARW,MADF,KAAK,OACU,SAAS,OAAO;IAC5C,OAAO,KAAK;IACZ,YAAY,QAAQ,aAAaC;IACjC,UAAU,CAAC;KAAE,MAAM;KAAQ,SAAS;KAAQ,CAAC;IAC7C,QAAQ,QAAQ;IAChB,aAAa,QAAQ;IACtB,CAAC,EAEuB,QAAQ;AACjC,OAAI,SAAS,SAAS,UAAU,QAAQ,KACtC,QAAO,QAAQ;AAGjB,SAAM,IAAI,YAAY,+CAA+C,YAAY;WAC1E,OAAO;AACd,OAAI,iBAAiB,YAAa,OAAM;AACxC,SAAM,IAAI,YACR,wBAAwB,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM,IAC9E,aACA,QACA,iBAAiB,QAAQ,QAAQ,OAClC;;;CAIL,MAAM,aACJ,QACA,QACA,UAA6B,EAAE,EACnB;EACZ,MAAM,aAAa,GAAG,OAAO;;;EAG/B,KAAK,UAAUC,kBAAgB,OAAO,EAAE,MAAM,EAAE,CAAC;;;EAI/C,MAAM,WAAW,MAAM,KAAK,SAAS,YAAY;GAC/C,GAAG;GACH,cACE,QAAQ,gBAAgB;GAC3B,CAAC;AAEF,MAAI;GAEF,MAAM,YAAY,SAAS,MAAM,cAAc;AAC/C,OAAI,CAAC,UACH,OAAM,IAAI,MAAM,mCAAmC;GAGrD,MAAM,SAAS,KAAK,MAAM,UAAU,GAAG;AACvC,UAAO,OAAO,MAAM,OAAO;WACpB,OAAO;AACd,SAAM,IAAI,YACR,yCAAyC,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM,IAC/F,oBACA,QACA,iBAAiB,QAAQ,QAAQ,OAClC;;;;;;;;AASP,SAASA,kBAAgB,QAAoC;CAC3D,MAAM,MAAO,OAA+D;AAE5E,SAAQ,IAAI,UAAZ;EACE,KAAK,aAAa;GAChB,MAAM,QAAS,OAAsC;GACrD,MAAMC,aAAqC,EAAE;AAC7C,QAAK,MAAM,CAAC,KAAK,UAAU,OAAO,QAAQ,MAAM,CAC9C,YAAW,OAAOD,kBAAgB,MAA4B;AAEhE,UAAO;IAAE,MAAM;IAAU;IAAY;;EAEvC,KAAK,WAEH,QAAO;GAAE,MAAM;GAAS,OAAOA,kBADd,IACuC,KAAK;GAAE;EAEjE,KAAK,YACH,QAAO,EAAE,MAAM,UAAU;EAC3B,KAAK,YACH,QAAO,EAAE,MAAM,UAAU;EAC3B,KAAK,aACH,QAAO,EAAE,MAAM,WAAW;EAC5B,KAAK,UAEH,QAAO;GAAE,MAAM;GAAU,MADT,IACuB;GAAQ;EAEjD,QACE,QAAO,EAAE,MAAM,UAAU;;;;;;AC/I/B,MAAM,gBAAgB;AACtB,MAAM,qBAAqB;AAC3B,MAAM,mBAAmB;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAmCzB,IAAa,iBAAb,MAAmD;CACjD,AAAS,OAAO;CAChB,AAAQ;CACR,AAAQ;CAER,YAAY,SAAiC,EAAE,EAAE;EAC/C,MAAM,SAAS,OAAO,UAAU,QAAQ,IAAI;EAC5C,MAAM,UAAU,OAAO,WAAW;AAGlC,MAAI,CAAC,UAAU,YAAY,iBACzB,OAAM,IAAI,YACR,iFACA,YACD;AAGH,OAAK,QAAQ,OAAO,SAAS;AAG7B,MAAI;GAEF,MAAM,EAAE,qBAAmB,SAAS;AACpC,QAAK,SAAS,IAAI,OAAO;IACvB,QAAQ,UAAU;IAClB,SAAS;IACV,CAAC;UACI;AACN,SAAM,IAAI,YACR,oFACA,YACD;;;CAIL,MAAM,SAAS,QAAgB,UAA6B,EAAE,EAAmB;AAC/E,MAAI;GACF,MAAM,SAAS,KAAK;GACpB,MAAME,WAAgE,EAAE;AAExE,OAAI,QAAQ,aACV,UAAS,KAAK;IAAE,MAAM;IAAU,SAAS,QAAQ;IAAc,CAAC;AAElE,YAAS,KAAK;IAAE,MAAM;IAAQ,SAAS;IAAQ,CAAC;GAShD,MAAM,WAPW,MAAM,OAAO,KAAK,YAAY,OAAO;IACpD,OAAO,KAAK;IACZ,YAAY,QAAQ,aAAa;IACjC;IACA,aAAa,QAAQ;IACtB,CAAC,EAEuB,QAAQ,IAAI,SAAS;AAC9C,OAAI,CAAC,QACH,OAAM,IAAI,YAAY,8BAA8B,YAAY;AAGlE,UAAO;WACA,OAAO;AACd,OAAI,iBAAiB,YAAa,OAAM;AACxC,SAAM,IAAI,YACR,qBAAqB,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM,IAC3E,aACA,QACA,iBAAiB,QAAQ,QAAQ,OAClC;;;CAIL,MAAM,aACJ,QACA,QACA,UAA6B,EAAE,EACnB;EACZ,MAAM,SAAS,KAAK;AAEpB,MAAI;GAEF,MAAMA,WAAgE,CACpE;IACE,MAAM;IACN,SACE,QAAQ,gBACR;IACH,EACD;IAAE,MAAM;IAAQ,SAAS;IAAQ,CAClC;GAUD,MAAM,WARW,MAAM,OAAO,KAAK,YAAY,OAAO;IACpD,OAAO,KAAK;IACZ,YAAY,QAAQ,aAAa;IACjC;IACA,aAAa,QAAQ;IACrB,iBAAiB,EAAE,MAAM,eAAe;IACzC,CAAC,EAEuB,QAAQ,IAAI,SAAS;AAC9C,OAAI,CAAC,QACH,OAAM,IAAI,YAAY,8BAA8B,YAAY;GAGlE,MAAM,SAAS,KAAK,MAAM,QAAQ;AAClC,UAAO,OAAO,MAAM,OAAO;WACpB,OAAO;AAEd,OAAI,iBAAiB,YAAa,OAAM;GAGxC,MAAM,aAAa,GAAG,OAAO;;;EAGjC,KAAK,UAAU,gBAAgB,OAAO,EAAE,MAAM,EAAE,CAAC;;;GAI7C,MAAM,WAAW,MAAM,KAAK,SAAS,YAAY;IAC/C,GAAG;IACH,cAAc;IACf,CAAC;AAEF,OAAI;IACF,MAAM,YAAY,SAAS,MAAM,cAAc;AAC/C,QAAI,CAAC,UACH,OAAM,IAAI,MAAM,mCAAmC;IAGrD,MAAM,SAAS,KAAK,MAAM,UAAU,GAAG;AACvC,WAAO,OAAO,MAAM,OAAO;YACpB,YAAY;AACnB,UAAM,IAAI,YACR,yCAAyC,sBAAsB,QAAQ,WAAW,UAAU,OAAO,WAAW,IAC9G,oBACA,QACA,sBAAsB,QAAQ,aAAa,OAC5C;;;;;;;;AAST,SAAS,gBAAgB,QAAoC;CAC3D,MAAM,MAAO,OAA+D;AAE5E,SAAQ,IAAI,UAAZ;EACE,KAAK,aAAa;GAChB,MAAM,QAAS,OAAsC;GACrD,MAAMC,aAAqC,EAAE;GAC7C,MAAMC,WAAqB,EAAE;AAE7B,QAAK,MAAM,CAAC,KAAK,UAAU,OAAO,QAAQ,MAAM,EAAE;AAChD,eAAW,OAAO,gBAAgB,MAA4B;AAG9D,QADkB,MAA8D,KACnE,aAAa,cACxB,UAAS,KAAK,IAAI;;AAGtB,UAAO;IAAE,MAAM;IAAU;IAAY;IAAU;;EAEjD,KAAK,WAEH,QAAO;GAAE,MAAM;GAAS,OAAO,gBADd,IACuC,KAAK;GAAE;EAEjE,KAAK,YACH,QAAO,EAAE,MAAM,UAAU;EAC3B,KAAK,YACH,QAAO,EAAE,MAAM,UAAU;EAC3B,KAAK,aACH,QAAO,EAAE,MAAM,WAAW;EAC5B,KAAK,UAEH,QAAO;GAAE,MAAM;GAAU,MADT,IACuB;GAAQ;EAEjD,KAAK,cAEH,QAAO,gBADa,IACe,UAAU;EAE/C,QACE,QAAO,EAAE,MAAM,UAAU;;;;;;AAS/B,SAAgB,aAAa,QAAiD;AAC5E,QAAO,IAAI,eAAe,OAAO;;;;;;;;;;AAWnC,SAAgB,aACd,SAA2C,EAAE,OAAO,YAAY,EAChD;AAChB,QAAO,IAAI,eAAe;EACxB,SAAS,oBAAoB,OAAO,QAAQ,MAAM;EAClD,OAAO,OAAO;EACd,QAAQ;EACT,CAAC;;;;;;;;;;AAWJ,SAAgB,eACd,SAA2C,EAAE,OAAO,eAAe,EACnD;AAChB,QAAO,IAAI,eAAe;EACxB,SAAS,oBAAoB,OAAO,QAAQ,KAAK;EACjD,OAAO,OAAO;EACd,QAAQ;EACT,CAAC"}
1
+ {"version":3,"file":"index.mjs","names":["messages: Array<{ role: string; content: string }>","request: Record<string, unknown>","properties: Record<string, object>","required: string[]","headers: Record<string, string>"],"sources":["../../src/llm/http.ts","../../src/llm/presets.ts"],"sourcesContent":["/**\n * HTTP-based LLM Provider using native fetch.\n * Provides a unified interface for any REST-based LLM API.\n */\n\nimport { z } from 'zod';\nimport { type BaseHttpConfig, BaseHttpProvider } from '../common/http-base.js';\nimport { ScrapeError } from '../core/errors.js';\nimport type { CompletionOptions, LLMProvider } from './types.js';\n\n// ─────────────────────────────────────────────────────────────\n// Types\n// ─────────────────────────────────────────────────────────────\n\n/**\n * HTTP LLM provider configuration.\n */\nexport interface HttpLLMConfig<TRequest = unknown, TResponse = unknown, TError = unknown>\n extends BaseHttpConfig<TError> {\n /**\n * Build request body from prompt and options.\n * @default OpenAI-compatible format with messages array\n */\n requestBuilder?: (prompt: string, options: CompletionOptions) => TRequest;\n /**\n * Extract completion text from response.\n * @default (res) => res.choices[0].message.content\n */\n responseMapper?: (response: TResponse) => string;\n /**\n * Enable JSON mode - adds response_format to request.\n * For OpenAI-compatible APIs, this adds { response_format: { type: \"json_object\" } }\n */\n jsonMode?: boolean;\n}\n\n// ─────────────────────────────────────────────────────────────\n// HTTP LLM Provider\n// ─────────────────────────────────────────────────────────────\n\n/**\n * HTTP-based LLM provider.\n * Works with any REST API using native fetch.\n */\nexport class HttpLLMProvider<TRequest = unknown, TResponse = unknown, TError = unknown>\n extends BaseHttpProvider<TError>\n implements LLMProvider\n{\n readonly name: string;\n\n private readonly requestBuilder: (prompt: string, options: CompletionOptions) => TRequest;\n private readonly responseMapper: (response: TResponse) => string;\n private readonly jsonMode: boolean;\n\n constructor(config: HttpLLMConfig<TRequest, TResponse, TError>) {\n super(config);\n this.name = 'http-llm';\n this.jsonMode = config.jsonMode ?? false;\n\n // Default request builder: OpenAI-compatible format\n this.requestBuilder =\n config.requestBuilder ??\n ((prompt: string, opts: CompletionOptions) => {\n const messages: Array<{ role: string; content: string }> = [];\n\n if (opts.systemPrompt) {\n messages.push({ role: 'system', content: opts.systemPrompt });\n }\n messages.push({ role: 'user', content: prompt });\n\n const request: Record<string, unknown> = {\n model: this.model,\n messages,\n };\n\n if (opts.temperature !== undefined) {\n request.temperature = opts.temperature;\n }\n if (opts.maxTokens !== undefined) {\n request.max_tokens = opts.maxTokens;\n }\n\n return request as TRequest;\n });\n\n // Default response mapper: OpenAI-compatible format\n this.responseMapper =\n config.responseMapper ??\n ((response: TResponse) => {\n const resp = response as Record<string, unknown>;\n\n // OpenAI format: { choices: [{ message: { content: \"...\" } }] }\n if (Array.isArray(resp.choices) && resp.choices.length > 0) {\n const choice = resp.choices[0] as { message?: { content?: string } };\n if (choice.message?.content) {\n return choice.message.content;\n }\n }\n\n // Anthropic format: { content: [{ type: \"text\", text: \"...\" }] }\n if (Array.isArray(resp.content)) {\n const textBlock = resp.content.find((c: { type?: string }) => c.type === 'text') as\n | { text?: string }\n | undefined;\n if (textBlock?.text) {\n return textBlock.text;\n }\n }\n\n throw new ScrapeError(\n 'Unable to parse LLM response. Provide a custom responseMapper.',\n 'VALIDATION_ERROR'\n );\n });\n }\n\n /**\n * Generate a text completion.\n */\n async complete(prompt: string, options: CompletionOptions = {}): Promise<string> {\n let body = this.requestBuilder(prompt, options);\n\n // Add JSON mode if enabled\n if (this.jsonMode && typeof body === 'object' && body !== null) {\n body = {\n ...body,\n response_format: { type: 'json_object' },\n } as TRequest;\n }\n\n const { data } = await this.fetch<TResponse>(this.baseUrl, { body });\n\n const content = this.responseMapper(data);\n if (!content) {\n throw new ScrapeError('Empty response from LLM', 'LLM_ERROR');\n }\n\n return content;\n }\n\n /**\n * Generate a structured JSON completion with Zod validation.\n */\n async completeJSON<T>(\n prompt: string,\n schema: z.ZodType<T>,\n options: CompletionOptions = {}\n ): Promise<T> {\n // Build a prompt that requests JSON output\n const jsonPrompt = `${prompt}\n\nRespond ONLY with valid JSON matching this schema:\n${JSON.stringify(zodToJsonSchema(schema), null, 2)}\n\nDo not include any explanation or markdown formatting. Just the JSON object.`;\n\n // Use JSON mode if available\n const useJsonMode = this.jsonMode;\n let body = this.requestBuilder(jsonPrompt, {\n ...options,\n systemPrompt:\n options.systemPrompt ?? 'You are a helpful assistant that responds only with valid JSON.',\n });\n\n if (useJsonMode && typeof body === 'object' && body !== null) {\n body = {\n ...body,\n response_format: { type: 'json_object' },\n } as TRequest;\n }\n\n const { data } = await this.fetch<TResponse>(this.baseUrl, { body });\n const content = this.responseMapper(data);\n\n if (!content) {\n throw new ScrapeError('Empty response from LLM', 'LLM_ERROR');\n }\n\n try {\n const trimmed = content.trim();\n try {\n return schema.parse(JSON.parse(trimmed));\n } catch {\n // Fall back to extracting JSON from markdown or surrounding text\n }\n\n const jsonMatch = content.match(/[[{][\\s\\S]*[\\]}]/);\n if (!jsonMatch) {\n throw new Error('No JSON found in response');\n }\n\n const parsed = JSON.parse(jsonMatch[0]);\n return schema.parse(parsed);\n } catch (error) {\n throw new ScrapeError(\n `Failed to parse LLM response as JSON: ${error instanceof Error ? error.message : String(error)}`,\n 'VALIDATION_ERROR',\n undefined,\n error instanceof Error ? error : undefined\n );\n }\n }\n}\n\n// ─────────────────────────────────────────────────────────────\n// Utilities\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Convert a Zod schema to a JSON Schema representation.\n * Uses Zod's built-in toJSONSchema method (Zod 4+).\n * Used for prompting LLMs to return structured data.\n */\nexport function zodToJsonSchema(schema: z.ZodType<unknown>): object {\n // Zod 4+ has built-in static toJSONSchema method\n if (typeof z.toJSONSchema === 'function') {\n const jsonSchema = z.toJSONSchema(schema);\n // Remove $schema key as it's not needed for LLM prompting\n const { $schema, ...rest } = jsonSchema as { $schema?: string; [key: string]: unknown };\n return rest;\n }\n\n // Fallback for older Zod versions using _def.type\n const def = (schema as z.ZodType<unknown> & { _def: { type: string } })._def;\n const type = def.type;\n\n switch (type) {\n case 'object': {\n const shape = (schema as z.ZodObject<z.ZodRawShape>).shape;\n const properties: Record<string, object> = {};\n const required: string[] = [];\n\n for (const [key, value] of Object.entries(shape)) {\n properties[key] = zodToJsonSchema(value as z.ZodType<unknown>);\n const valueDef = (value as z.ZodType<unknown> & { _def: { type: string } })._def;\n if (valueDef.type !== 'optional') {\n required.push(key);\n }\n }\n return { type: 'object', properties, required };\n }\n case 'array': {\n const arrayDef = def as unknown as { element: z.ZodType<unknown> };\n return { type: 'array', items: zodToJsonSchema(arrayDef.element) };\n }\n case 'string':\n return { type: 'string' };\n case 'number':\n return { type: 'number' };\n case 'boolean':\n return { type: 'boolean' };\n case 'enum': {\n const enumDef = def as unknown as { entries: Record<string, string> };\n return { type: 'string', enum: Object.values(enumDef.entries) };\n }\n case 'optional': {\n const optionalDef = def as unknown as { innerType: z.ZodType<unknown> };\n return zodToJsonSchema(optionalDef.innerType);\n }\n default:\n return { type: 'string' };\n }\n}\n\n// Re-export types for convenience\nexport type { ZodType } from 'zod';\n","/**\n * Preset factory functions for common LLM providers.\n * All presets use the HttpLLMProvider with appropriate configuration.\n */\n\nimport { type HttpLLMConfig, HttpLLMProvider } from './http.js';\nimport type { LLMProvider } from './types.js';\n\n// ─────────────────────────────────────────────────────────────\n// OpenAI\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create an OpenAI LLM provider.\n *\n * @example\n * ```ts\n * const provider = createOpenAI({ apiKey: 'sk-...' });\n * const result = await scrape(url, { llm: provider, enhance: ['summarize'] });\n * ```\n */\nexport function createOpenAI(options?: {\n apiKey?: string;\n model?: string;\n baseUrl?: string;\n}): LLMProvider {\n const apiKey = options?.apiKey ?? process.env.OPENAI_API_KEY;\n if (!apiKey) {\n throw new Error('OpenAI API key required. Set OPENAI_API_KEY env var or pass apiKey option.');\n }\n\n return new HttpLLMProvider({\n baseUrl: options?.baseUrl ?? 'https://api.openai.com/v1/chat/completions',\n model: options?.model ?? 'gpt-4o-mini',\n headers: { Authorization: `Bearer ${apiKey}` },\n jsonMode: true,\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// Anthropic\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Anthropic API response shape.\n */\ninterface AnthropicResponse {\n content: Array<{ type: string; text?: string }>;\n}\n\n/**\n * Create an Anthropic Claude LLM provider.\n *\n * @example\n * ```ts\n * const provider = createAnthropic({ apiKey: 'sk-...' });\n * const result = await scrape(url, { llm: provider, enhance: ['summarize'] });\n * ```\n */\nexport function createAnthropic(options?: { apiKey?: string; model?: string }): LLMProvider {\n const apiKey = options?.apiKey ?? process.env.ANTHROPIC_API_KEY;\n if (!apiKey) {\n throw new Error(\n 'Anthropic API key required. Set ANTHROPIC_API_KEY env var or pass apiKey option.'\n );\n }\n\n const model = options?.model ?? 'claude-3-5-haiku-20241022';\n\n return new HttpLLMProvider<unknown, AnthropicResponse>({\n baseUrl: 'https://api.anthropic.com/v1/messages',\n model,\n headers: {\n 'x-api-key': apiKey,\n 'anthropic-version': '2023-06-01',\n },\n requestBuilder: (prompt, opts) => ({\n model,\n max_tokens: opts.maxTokens ?? 1024,\n messages: [{ role: 'user', content: prompt }],\n ...(opts.systemPrompt && { system: opts.systemPrompt }),\n ...(opts.temperature !== undefined && { temperature: opts.temperature }),\n }),\n responseMapper: (res) => res.content.find((item) => item.type === 'text')?.text ?? '',\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// Groq\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create a Groq LLM provider.\n * Groq provides fast inference for open-source models.\n *\n * @example\n * ```ts\n * const provider = createGroq({ model: 'llama-3.1-70b-versatile' });\n * ```\n */\nexport function createGroq(options?: { apiKey?: string; model?: string }): LLMProvider {\n const apiKey = options?.apiKey ?? process.env.GROQ_API_KEY;\n if (!apiKey) {\n throw new Error('Groq API key required. Set GROQ_API_KEY env var or pass apiKey option.');\n }\n\n return new HttpLLMProvider({\n baseUrl: 'https://api.groq.com/openai/v1/chat/completions',\n model: options?.model ?? 'llama-3.1-70b-versatile',\n headers: { Authorization: `Bearer ${apiKey}` },\n jsonMode: true,\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// Ollama (Local)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create an Ollama LLM provider for local models.\n *\n * @example\n * ```ts\n * const provider = createOllama({ model: 'llama3.2' });\n * ```\n */\nexport function createOllama(options: { model: string; baseUrl?: string }): LLMProvider {\n return new HttpLLMProvider({\n baseUrl: options.baseUrl ?? 'http://localhost:11434/v1/chat/completions',\n model: options.model,\n requireHttps: false,\n allowPrivate: true,\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// LM Studio (Local)\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create an LM Studio LLM provider for local models.\n *\n * @example\n * ```ts\n * const provider = createLMStudio({ model: 'local-model' });\n * ```\n */\nexport function createLMStudio(options: { model: string; baseUrl?: string }): LLMProvider {\n return new HttpLLMProvider({\n baseUrl: options.baseUrl ?? 'http://localhost:1234/v1/chat/completions',\n model: options.model,\n requireHttps: false,\n allowPrivate: true,\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// Together AI\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create a Together AI LLM provider.\n *\n * @example\n * ```ts\n * const provider = createTogether({ model: 'meta-llama/Llama-3.2-3B-Instruct-Turbo' });\n * ```\n */\nexport function createTogether(options?: { apiKey?: string; model?: string }): LLMProvider {\n const apiKey = options?.apiKey ?? process.env.TOGETHER_API_KEY;\n if (!apiKey) {\n throw new Error(\n 'Together API key required. Set TOGETHER_API_KEY env var or pass apiKey option.'\n );\n }\n\n return new HttpLLMProvider({\n baseUrl: 'https://api.together.xyz/v1/chat/completions',\n model: options?.model ?? 'meta-llama/Llama-3.2-3B-Instruct-Turbo',\n headers: { Authorization: `Bearer ${apiKey}` },\n jsonMode: true,\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// OpenRouter\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create an OpenRouter LLM provider.\n * OpenRouter provides access to many models through a unified API.\n *\n * @example\n * ```ts\n * const provider = createOpenRouter({\n * model: 'anthropic/claude-3.5-sonnet',\n * });\n * ```\n */\nexport function createOpenRouter(options: {\n apiKey?: string;\n model: string;\n siteUrl?: string;\n siteName?: string;\n}): LLMProvider {\n const apiKey = options.apiKey ?? process.env.OPENROUTER_API_KEY;\n if (!apiKey) {\n throw new Error(\n 'OpenRouter API key required. Set OPENROUTER_API_KEY env var or pass apiKey option.'\n );\n }\n\n const headers: Record<string, string> = {\n Authorization: `Bearer ${apiKey}`,\n };\n\n if (options.siteUrl) {\n headers['HTTP-Referer'] = options.siteUrl;\n }\n if (options.siteName) {\n headers['X-Title'] = options.siteName;\n }\n\n return new HttpLLMProvider({\n baseUrl: 'https://openrouter.ai/api/v1/chat/completions',\n model: options.model,\n headers,\n });\n}\n\n// ─────────────────────────────────────────────────────────────\n// Generic HTTP Provider\n// ─────────────────────────────────────────────────────────────\n\n/**\n * Create a generic HTTP LLM provider.\n * Use this for any OpenAI-compatible API.\n *\n * @example\n * ```ts\n * const provider = createHttpLLM({\n * baseUrl: 'https://my-api.com/v1/chat/completions',\n * model: 'my-model',\n * headers: { Authorization: 'Bearer ...' },\n * });\n * ```\n */\nexport function createHttpLLM<TRequest = unknown, TResponse = unknown, TError = unknown>(\n config: HttpLLMConfig<TRequest, TResponse, TError>\n): LLMProvider {\n return new HttpLLMProvider(config);\n}\n"],"mappings":";;;;;;;;;;;;;AA4CA,IAAa,kBAAb,cACU,iBAEV;CACE,AAAS;CAET,AAAiB;CACjB,AAAiB;CACjB,AAAiB;CAEjB,YAAY,QAAoD;AAC9D,QAAM,OAAO;AACb,OAAK,OAAO;AACZ,OAAK,WAAW,OAAO,YAAY;AAGnC,OAAK,iBACH,OAAO,oBACL,QAAgB,SAA4B;GAC5C,MAAMA,WAAqD,EAAE;AAE7D,OAAI,KAAK,aACP,UAAS,KAAK;IAAE,MAAM;IAAU,SAAS,KAAK;IAAc,CAAC;AAE/D,YAAS,KAAK;IAAE,MAAM;IAAQ,SAAS;IAAQ,CAAC;GAEhD,MAAMC,UAAmC;IACvC,OAAO,KAAK;IACZ;IACD;AAED,OAAI,KAAK,gBAAgB,OACvB,SAAQ,cAAc,KAAK;AAE7B,OAAI,KAAK,cAAc,OACrB,SAAQ,aAAa,KAAK;AAG5B,UAAO;;AAIX,OAAK,iBACH,OAAO,oBACL,aAAwB;GACxB,MAAM,OAAO;AAGb,OAAI,MAAM,QAAQ,KAAK,QAAQ,IAAI,KAAK,QAAQ,SAAS,GAAG;IAC1D,MAAM,SAAS,KAAK,QAAQ;AAC5B,QAAI,OAAO,SAAS,QAClB,QAAO,OAAO,QAAQ;;AAK1B,OAAI,MAAM,QAAQ,KAAK,QAAQ,EAAE;IAC/B,MAAM,YAAY,KAAK,QAAQ,MAAM,MAAyB,EAAE,SAAS,OAAO;AAGhF,QAAI,WAAW,KACb,QAAO,UAAU;;AAIrB,SAAM,IAAI,YACR,kEACA,mBACD;;;;;;CAOP,MAAM,SAAS,QAAgB,UAA6B,EAAE,EAAmB;EAC/E,IAAI,OAAO,KAAK,eAAe,QAAQ,QAAQ;AAG/C,MAAI,KAAK,YAAY,OAAO,SAAS,YAAY,SAAS,KACxD,QAAO;GACL,GAAG;GACH,iBAAiB,EAAE,MAAM,eAAe;GACzC;EAGH,MAAM,EAAE,SAAS,MAAM,KAAK,MAAiB,KAAK,SAAS,EAAE,MAAM,CAAC;EAEpE,MAAM,UAAU,KAAK,eAAe,KAAK;AACzC,MAAI,CAAC,QACH,OAAM,IAAI,YAAY,2BAA2B,YAAY;AAG/D,SAAO;;;;;CAMT,MAAM,aACJ,QACA,QACA,UAA6B,EAAE,EACnB;EAEZ,MAAM,aAAa,GAAG,OAAO;;;EAG/B,KAAK,UAAU,gBAAgB,OAAO,EAAE,MAAM,EAAE,CAAC;;;EAK/C,MAAM,cAAc,KAAK;EACzB,IAAI,OAAO,KAAK,eAAe,YAAY;GACzC,GAAG;GACH,cACE,QAAQ,gBAAgB;GAC3B,CAAC;AAEF,MAAI,eAAe,OAAO,SAAS,YAAY,SAAS,KACtD,QAAO;GACL,GAAG;GACH,iBAAiB,EAAE,MAAM,eAAe;GACzC;EAGH,MAAM,EAAE,SAAS,MAAM,KAAK,MAAiB,KAAK,SAAS,EAAE,MAAM,CAAC;EACpE,MAAM,UAAU,KAAK,eAAe,KAAK;AAEzC,MAAI,CAAC,QACH,OAAM,IAAI,YAAY,2BAA2B,YAAY;AAG/D,MAAI;GACF,MAAM,UAAU,QAAQ,MAAM;AAC9B,OAAI;AACF,WAAO,OAAO,MAAM,KAAK,MAAM,QAAQ,CAAC;WAClC;GAIR,MAAM,YAAY,QAAQ,MAAM,mBAAmB;AACnD,OAAI,CAAC,UACH,OAAM,IAAI,MAAM,4BAA4B;GAG9C,MAAM,SAAS,KAAK,MAAM,UAAU,GAAG;AACvC,UAAO,OAAO,MAAM,OAAO;WACpB,OAAO;AACd,SAAM,IAAI,YACR,yCAAyC,iBAAiB,QAAQ,MAAM,UAAU,OAAO,MAAM,IAC/F,oBACA,QACA,iBAAiB,QAAQ,QAAQ,OAClC;;;;;;;;;AAcP,SAAgB,gBAAgB,QAAoC;AAElE,KAAI,OAAO,EAAE,iBAAiB,YAAY;EAGxC,MAAM,EAAE,SAAS,GAAG,SAFD,EAAE,aAAa,OAAO;AAGzC,SAAO;;CAIT,MAAM,MAAO,OAA2D;AAGxE,SAFa,IAAI,MAEjB;EACE,KAAK,UAAU;GACb,MAAM,QAAS,OAAsC;GACrD,MAAMC,aAAqC,EAAE;GAC7C,MAAMC,WAAqB,EAAE;AAE7B,QAAK,MAAM,CAAC,KAAK,UAAU,OAAO,QAAQ,MAAM,EAAE;AAChD,eAAW,OAAO,gBAAgB,MAA4B;AAE9D,QADkB,MAA0D,KAC/D,SAAS,WACpB,UAAS,KAAK,IAAI;;AAGtB,UAAO;IAAE,MAAM;IAAU;IAAY;IAAU;;EAEjD,KAAK,QAEH,QAAO;GAAE,MAAM;GAAS,OAAO,gBADd,IACuC,QAAQ;GAAE;EAEpE,KAAK,SACH,QAAO,EAAE,MAAM,UAAU;EAC3B,KAAK,SACH,QAAO,EAAE,MAAM,UAAU;EAC3B,KAAK,UACH,QAAO,EAAE,MAAM,WAAW;EAC5B,KAAK,QAAQ;GACX,MAAM,UAAU;AAChB,UAAO;IAAE,MAAM;IAAU,MAAM,OAAO,OAAO,QAAQ,QAAQ;IAAE;;EAEjE,KAAK,WAEH,QAAO,gBADa,IACe,UAAU;EAE/C,QACE,QAAO,EAAE,MAAM,UAAU;;;;;;;;;;;;;;;;;;;AC/O/B,SAAgB,aAAa,SAIb;CACd,MAAM,SAAS,SAAS,UAAU,QAAQ,IAAI;AAC9C,KAAI,CAAC,OACH,OAAM,IAAI,MAAM,6EAA6E;AAG/F,QAAO,IAAI,gBAAgB;EACzB,SAAS,SAAS,WAAW;EAC7B,OAAO,SAAS,SAAS;EACzB,SAAS,EAAE,eAAe,UAAU,UAAU;EAC9C,UAAU;EACX,CAAC;;;;;;;;;;;AAuBJ,SAAgB,gBAAgB,SAA4D;CAC1F,MAAM,SAAS,SAAS,UAAU,QAAQ,IAAI;AAC9C,KAAI,CAAC,OACH,OAAM,IAAI,MACR,mFACD;CAGH,MAAM,QAAQ,SAAS,SAAS;AAEhC,QAAO,IAAI,gBAA4C;EACrD,SAAS;EACT;EACA,SAAS;GACP,aAAa;GACb,qBAAqB;GACtB;EACD,iBAAiB,QAAQ,UAAU;GACjC;GACA,YAAY,KAAK,aAAa;GAC9B,UAAU,CAAC;IAAE,MAAM;IAAQ,SAAS;IAAQ,CAAC;GAC7C,GAAI,KAAK,gBAAgB,EAAE,QAAQ,KAAK,cAAc;GACtD,GAAI,KAAK,gBAAgB,UAAa,EAAE,aAAa,KAAK,aAAa;GACxE;EACD,iBAAiB,QAAQ,IAAI,QAAQ,MAAM,SAAS,KAAK,SAAS,OAAO,EAAE,QAAQ;EACpF,CAAC;;;;;;;;;;;AAgBJ,SAAgB,WAAW,SAA4D;CACrF,MAAM,SAAS,SAAS,UAAU,QAAQ,IAAI;AAC9C,KAAI,CAAC,OACH,OAAM,IAAI,MAAM,yEAAyE;AAG3F,QAAO,IAAI,gBAAgB;EACzB,SAAS;EACT,OAAO,SAAS,SAAS;EACzB,SAAS,EAAE,eAAe,UAAU,UAAU;EAC9C,UAAU;EACX,CAAC;;;;;;;;;;AAeJ,SAAgB,aAAa,SAA2D;AACtF,QAAO,IAAI,gBAAgB;EACzB,SAAS,QAAQ,WAAW;EAC5B,OAAO,QAAQ;EACf,cAAc;EACd,cAAc;EACf,CAAC;;;;;;;;;;AAeJ,SAAgB,eAAe,SAA2D;AACxF,QAAO,IAAI,gBAAgB;EACzB,SAAS,QAAQ,WAAW;EAC5B,OAAO,QAAQ;EACf,cAAc;EACd,cAAc;EACf,CAAC;;;;;;;;;;AAeJ,SAAgB,eAAe,SAA4D;CACzF,MAAM,SAAS,SAAS,UAAU,QAAQ,IAAI;AAC9C,KAAI,CAAC,OACH,OAAM,IAAI,MACR,iFACD;AAGH,QAAO,IAAI,gBAAgB;EACzB,SAAS;EACT,OAAO,SAAS,SAAS;EACzB,SAAS,EAAE,eAAe,UAAU,UAAU;EAC9C,UAAU;EACX,CAAC;;;;;;;;;;;;;AAkBJ,SAAgB,iBAAiB,SAKjB;CACd,MAAM,SAAS,QAAQ,UAAU,QAAQ,IAAI;AAC7C,KAAI,CAAC,OACH,OAAM,IAAI,MACR,qFACD;CAGH,MAAMC,UAAkC,EACtC,eAAe,UAAU,UAC1B;AAED,KAAI,QAAQ,QACV,SAAQ,kBAAkB,QAAQ;AAEpC,KAAI,QAAQ,SACV,SAAQ,aAAa,QAAQ;AAG/B,QAAO,IAAI,gBAAgB;EACzB,SAAS;EACT,OAAO,QAAQ;EACf;EACD,CAAC;;;;;;;;;;;;;;;AAoBJ,SAAgB,cACd,QACa;AACb,QAAO,IAAI,gBAAgB,OAAO"}