@ztimson/ai-utils 0.6.8 → 0.6.10
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/ai.d.ts +2 -0
- package/dist/asr.js +5 -4
- package/dist/asr.js.map +1 -1
- package/dist/asr.mjs +47 -45
- package/dist/asr.mjs.map +1 -1
- package/dist/index.js +12 -12
- package/dist/index.js.map +1 -1
- package/dist/index.mjs +64 -66
- package/dist/index.mjs.map +1 -1
- package/package.json +1 -1
package/dist/index.js.map
CHANGED
|
@@ -1 +1 @@
|
|
|
1
|
-
{"version":3,"file":"index.js","sources":["../src/provider.ts","../src/antrhopic.ts","../src/open-ai.ts","../src/llm.ts","../src/audio.ts","../src/vision.ts","../src/ai.ts","../src/tools.ts"],"sourcesContent":["import {AbortablePromise} from './ai.ts';\nimport {LLMMessage, LLMRequest} from './llm.ts';\n\nexport abstract class LLMProvider {\n\tabstract ask(message: string, options: LLMRequest): AbortablePromise<string>;\n}\n","import {Anthropic as anthropic} from '@anthropic-ai/sdk';\nimport {findByProp, objectMap, JSONSanitize, JSONAttemptParse} from '@ztimson/utils';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {LLMMessage, LLMRequest} from './llm.ts';\nimport {LLMProvider} from './provider.ts';\n\nexport class Anthropic extends LLMProvider {\n\tclient!: anthropic;\n\n\tconstructor(public readonly ai: Ai, public readonly apiToken: string, public model: string) {\n\t\tsuper();\n\t\tthis.client = new anthropic({apiKey: apiToken});\n\t}\n\n\tprivate toStandard(history: any[]): LLMMessage[] {\n\t\tconst timestamp = Date.now();\n\t\tconst messages: LLMMessage[] = [];\n\t\tfor(let h of history) {\n\t\t\tif(typeof h.content == 'string') {\n\t\t\t\tmessages.push(<any>{timestamp, ...h});\n\t\t\t} else {\n\t\t\t\tconst textContent = h.content?.filter((c: any) => c.type == 'text').map((c: any) => c.text).join('\\n\\n');\n\t\t\t\tif(textContent) messages.push({timestamp, role: h.role, content: textContent});\n\t\t\t\th.content.forEach((c: any) => {\n\t\t\t\t\tif(c.type == 'tool_use') {\n\t\t\t\t\t\tmessages.push({timestamp, role: 'tool', id: c.id, name: c.name, args: c.input, content: undefined});\n\t\t\t\t\t} else if(c.type == 'tool_result') {\n\t\t\t\t\t\tconst m: any = messages.findLast(m => (<any>m).id == c.tool_use_id);\n\t\t\t\t\t\tif(m) m[c.is_error ? 'error' : 'content'] = c.content;\n\t\t\t\t\t}\n\t\t\t\t});\n\t\t\t}\n\t\t}\n\t\treturn messages;\n\t}\n\n\tprivate fromStandard(history: LLMMessage[]): any[] {\n\t\tfor(let i = 0; i < history.length; i++) {\n\t\t\tif(history[i].role == 'tool') {\n\t\t\t\tconst h: any = history[i];\n\t\t\t\thistory.splice(i, 1,\n\t\t\t\t\t{role: 'assistant', content: [{type: 'tool_use', id: h.id, name: h.name, input: h.args}]},\n\t\t\t\t\t{role: 'user', content: [{type: 'tool_result', tool_use_id: h.id, is_error: !!h.error, content: h.error || h.content}]}\n\t\t\t\t)\n\t\t\t\ti++;\n\t\t\t}\n\t\t}\n\t\treturn history.map(({timestamp, ...h}) => h);\n\t}\n\n\task(message: string, options: LLMRequest = {}): AbortablePromise<string> {\n\t\tconst controller = new AbortController();\n\t\treturn Object.assign(new Promise<any>(async (res) => {\n\t\t\tlet history = this.fromStandard([...options.history || [], {role: 'user', content: message, timestamp: Date.now()}]);\n\t\t\tconst tools = options.tools || this.ai.options.llm?.tools || [];\n\t\t\tconst requestParams: any = {\n\t\t\t\tmodel: options.model || this.model,\n\t\t\t\tmax_tokens: options.max_tokens || this.ai.options.llm?.max_tokens || 4096,\n\t\t\t\tsystem: options.system || this.ai.options.llm?.system || '',\n\t\t\t\ttemperature: options.temperature || this.ai.options.llm?.temperature || 0.7,\n\t\t\t\ttools: tools.map(t => ({\n\t\t\t\t\tname: t.name,\n\t\t\t\t\tdescription: t.description,\n\t\t\t\t\tinput_schema: {\n\t\t\t\t\t\ttype: 'object',\n\t\t\t\t\t\tproperties: t.args ? objectMap(t.args, (key, value) => ({...value, required: undefined})) : {},\n\t\t\t\t\t\trequired: t.args ? Object.entries(t.args).filter(t => t[1].required).map(t => t[0]) : []\n\t\t\t\t\t},\n\t\t\t\t\tfn: undefined\n\t\t\t\t})),\n\t\t\t\tmessages: history,\n\t\t\t\tstream: !!options.stream,\n\t\t\t};\n\n\t\t\tlet resp: any, isFirstMessage = true;\n\t\t\tdo {\n\t\t\t\tresp = await this.client.messages.create(requestParams).catch(err => {\n\t\t\t\t\terr.message += `\\n\\nMessages:\\n${JSON.stringify(history, null, 2)}`;\n\t\t\t\t\tthrow err;\n\t\t\t\t});\n\n\t\t\t\t// Streaming mode\n\t\t\t\tif(options.stream) {\n\t\t\t\t\tif(!isFirstMessage) options.stream({text: '\\n\\n'});\n\t\t\t\t\telse isFirstMessage = false;\n\t\t\t\t\tresp.content = [];\n\t\t\t\t\tfor await (const chunk of resp) {\n\t\t\t\t\t\tif(controller.signal.aborted) break;\n\t\t\t\t\t\tif(chunk.type === 'content_block_start') {\n\t\t\t\t\t\t\tif(chunk.content_block.type === 'text') {\n\t\t\t\t\t\t\t\tresp.content.push({type: 'text', text: ''});\n\t\t\t\t\t\t\t} else if(chunk.content_block.type === 'tool_use') {\n\t\t\t\t\t\t\t\tresp.content.push({type: 'tool_use', id: chunk.content_block.id, name: chunk.content_block.name, input: <any>''});\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else if(chunk.type === 'content_block_delta') {\n\t\t\t\t\t\t\tif(chunk.delta.type === 'text_delta') {\n\t\t\t\t\t\t\t\tconst text = chunk.delta.text;\n\t\t\t\t\t\t\t\tresp.content.at(-1).text += text;\n\t\t\t\t\t\t\t\toptions.stream({text});\n\t\t\t\t\t\t\t} else if(chunk.delta.type === 'input_json_delta') {\n\t\t\t\t\t\t\t\tresp.content.at(-1).input += chunk.delta.partial_json;\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else if(chunk.type === 'content_block_stop') {\n\t\t\t\t\t\t\tconst last = resp.content.at(-1);\n\t\t\t\t\t\t\tif(last.input != null) last.input = last.input ? JSONAttemptParse(last.input, {}) : {};\n\t\t\t\t\t\t} else if(chunk.type === 'message_stop') {\n\t\t\t\t\t\t\tbreak;\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// Run tools\n\t\t\t\tconst toolCalls = resp.content.filter((c: any) => c.type === 'tool_use');\n\t\t\t\tif(toolCalls.length && !controller.signal.aborted) {\n\t\t\t\t\thistory.push({role: 'assistant', content: resp.content});\n\t\t\t\t\tconst results = await Promise.all(toolCalls.map(async (toolCall: any) => {\n\t\t\t\t\t\tconst tool = tools.find(findByProp('name', toolCall.name));\n\t\t\t\t\t\tif(options.stream) options.stream({tool: toolCall.name});\n\t\t\t\t\t\tif(!tool) return {tool_use_id: toolCall.id, is_error: true, content: 'Tool not found'};\n\t\t\t\t\t\ttry {\n\t\t\t\t\t\t\tconst result = await tool.fn(toolCall.input, options?.stream, this.ai);\n\t\t\t\t\t\t\treturn {type: 'tool_result', tool_use_id: toolCall.id, content: JSONSanitize(result)};\n\t\t\t\t\t\t} catch (err: any) {\n\t\t\t\t\t\t\treturn {type: 'tool_result', tool_use_id: toolCall.id, is_error: true, content: err?.message || err?.toString() || 'Unknown'};\n\t\t\t\t\t\t}\n\t\t\t\t\t}));\n\t\t\t\t\thistory.push({role: 'user', content: results});\n\t\t\t\t\trequestParams.messages = history;\n\t\t\t\t}\n\t\t\t} while (!controller.signal.aborted && resp.content.some((c: any) => c.type === 'tool_use'));\n\t\t\thistory.push({role: 'assistant', content: resp.content.filter((c: any) => c.type == 'text').map((c: any) => c.text).join('\\n\\n')});\n\t\t\thistory = this.toStandard(history);\n\n\t\t\tif(options.stream) options.stream({done: true});\n\t\t\tif(options.history) options.history.splice(0, options.history.length, ...history);\n\t\t\tres(history.at(-1)?.content);\n\t\t}), {abort: () => controller.abort()});\n\t}\n}\n","import {OpenAI as openAI} from 'openai';\nimport {findByProp, objectMap, JSONSanitize, JSONAttemptParse, clean} from '@ztimson/utils';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {LLMMessage, LLMRequest} from './llm.ts';\nimport {LLMProvider} from './provider.ts';\n\nexport class OpenAi extends LLMProvider {\n\tclient!: openAI;\n\n\tconstructor(public readonly ai: Ai, public readonly host: string | null, public readonly token: string, public model: string) {\n\t\tsuper();\n\t\tthis.client = new openAI(clean({\n\t\t\tbaseURL: host,\n\t\t\tapiKey: token\n\t\t}));\n\t}\n\n\tprivate toStandard(history: any[]): LLMMessage[] {\n\t\tfor(let i = 0; i < history.length; i++) {\n\t\t\tconst h = history[i];\n\t\t\tif(h.role === 'assistant' && h.tool_calls) {\n\t\t\t\tconst tools = h.tool_calls.map((tc: any) => ({\n\t\t\t\t\trole: 'tool',\n\t\t\t\t\tid: tc.id,\n\t\t\t\t\tname: tc.function.name,\n\t\t\t\t\targs: JSONAttemptParse(tc.function.arguments, {}),\n\t\t\t\t\ttimestamp: h.timestamp\n\t\t\t\t}));\n\t\t\t\thistory.splice(i, 1, ...tools);\n\t\t\t\ti += tools.length - 1;\n\t\t\t} else if(h.role === 'tool' && h.content) {\n\t\t\t\tconst record = history.find(h2 => h.tool_call_id == h2.id);\n\t\t\t\tif(record) {\n\t\t\t\t\tif(h.content.includes('\"error\":')) record.error = h.content;\n\t\t\t\t\telse record.content = h.content;\n\t\t\t\t}\n\t\t\t\thistory.splice(i, 1);\n\t\t\t\ti--;\n\t\t\t}\n\t\t\tif(!history[i]?.timestamp) history[i].timestamp = Date.now();\n\t\t}\n\t\treturn history;\n\t}\n\n\tprivate fromStandard(history: LLMMessage[]): any[] {\n\t\treturn history.reduce((result, h) => {\n\t\t\tif(h.role === 'tool') {\n\t\t\t\tresult.push({\n\t\t\t\t\trole: 'assistant',\n\t\t\t\t\tcontent: null,\n\t\t\t\t\ttool_calls: [{ id: h.id, type: 'function', function: { name: h.name, arguments: JSON.stringify(h.args) } }],\n\t\t\t\t\trefusal: null,\n\t\t\t\t\tannotations: []\n\t\t\t\t}, {\n\t\t\t\t\trole: 'tool',\n\t\t\t\t\ttool_call_id: h.id,\n\t\t\t\t\tcontent: h.error || h.content\n\t\t\t\t});\n\t\t\t} else {\n\t\t\t\tconst {timestamp, ...rest} = h;\n\t\t\t\tresult.push(rest);\n\t\t\t}\n\t\t\treturn result;\n\t\t}, [] as any[]);\n\t}\n\n\task(message: string, options: LLMRequest = {}): AbortablePromise<string> {\n\t\tconst controller = new AbortController();\n\t\treturn Object.assign(new Promise<any>(async (res, rej) => {\n\t\t\tif(options.system && options.history?.[0]?.role != 'system') options.history?.splice(0, 0, {role: 'system', content: options.system, timestamp: Date.now()});\n\t\t\tlet history = this.fromStandard([...options.history || [], {role: 'user', content: message, timestamp: Date.now()}]);\n\t\t\tconst tools = options.tools || this.ai.options.llm?.tools || [];\n\t\t\tconst requestParams: any = {\n\t\t\t\tmodel: options.model || this.model,\n\t\t\t\tmessages: history,\n\t\t\t\tstream: !!options.stream,\n\t\t\t\tmax_tokens: options.max_tokens || this.ai.options.llm?.max_tokens || 4096,\n\t\t\t\ttemperature: options.temperature || this.ai.options.llm?.temperature || 0.7,\n\t\t\t\ttools: tools.map(t => ({\n\t\t\t\t\ttype: 'function',\n\t\t\t\t\tfunction: {\n\t\t\t\t\t\tname: t.name,\n\t\t\t\t\t\tdescription: t.description,\n\t\t\t\t\t\tparameters: {\n\t\t\t\t\t\t\ttype: 'object',\n\t\t\t\t\t\t\tproperties: t.args ? objectMap(t.args, (key, value) => ({...value, required: undefined})) : {},\n\t\t\t\t\t\t\trequired: t.args ? Object.entries(t.args).filter(t => t[1].required).map(t => t[0]) : []\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}))\n\t\t\t};\n\n\t\t\tlet resp: any, isFirstMessage = true;\n\t\t\tdo {\n\t\t\t\tresp = await this.client.chat.completions.create(requestParams).catch(err => {\n\t\t\t\t\terr.message += `\\n\\nMessages:\\n${JSON.stringify(history, null, 2)}`;\n\t\t\t\t\tthrow err;\n\t\t\t\t});\n\n\t\t\t\tif(options.stream) {\n\t\t\t\t\tif(!isFirstMessage) options.stream({text: '\\n\\n'});\n\t\t\t\t\telse isFirstMessage = false;\n\t\t\t\t\tresp.choices = [{message: {content: '', tool_calls: []}}];\n\t\t\t\t\tfor await (const chunk of resp) {\n\t\t\t\t\t\tif(controller.signal.aborted) break;\n\t\t\t\t\t\tif(chunk.choices[0].delta.content) {\n\t\t\t\t\t\t\tresp.choices[0].message.content += chunk.choices[0].delta.content;\n\t\t\t\t\t\t\toptions.stream({text: chunk.choices[0].delta.content});\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif(chunk.choices[0].delta.tool_calls) {\n\t\t\t\t\t\t\tresp.choices[0].message.tool_calls = chunk.choices[0].delta.tool_calls;\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tconst toolCalls = resp.choices[0].message.tool_calls || [];\n\t\t\t\tif(toolCalls.length && !controller.signal.aborted) {\n\t\t\t\t\thistory.push(resp.choices[0].message);\n\t\t\t\t\tconst results = await Promise.all(toolCalls.map(async (toolCall: any) => {\n\t\t\t\t\t\tconst tool = tools?.find(findByProp('name', toolCall.function.name));\n\t\t\t\t\t\tif(options.stream) options.stream({tool: toolCall.function.name});\n\t\t\t\t\t\tif(!tool) return {role: 'tool', tool_call_id: toolCall.id, content: '{\"error\": \"Tool not found\"}'};\n\t\t\t\t\t\ttry {\n\t\t\t\t\t\t\tconst args = JSONAttemptParse(toolCall.function.arguments, {});\n\t\t\t\t\t\t\tconst result = await tool.fn(args, options.stream, this.ai);\n\t\t\t\t\t\t\treturn {role: 'tool', tool_call_id: toolCall.id, content: JSONSanitize(result)};\n\t\t\t\t\t\t} catch (err: any) {\n\t\t\t\t\t\t\treturn {role: 'tool', tool_call_id: toolCall.id, content: JSONSanitize({error: err?.message || err?.toString() || 'Unknown'})};\n\t\t\t\t\t\t}\n\t\t\t\t\t}));\n\t\t\t\t\thistory.push(...results);\n\t\t\t\t\trequestParams.messages = history;\n\t\t\t\t}\n\t\t\t} while (!controller.signal.aborted && resp.choices?.[0]?.message?.tool_calls?.length);\n\t\t\thistory.push({role: 'assistant', content: resp.choices[0].message.content || ''});\n\t\t\thistory = this.toStandard(history);\n\n\t\t\tif(options.stream) options.stream({done: true});\n\t\t\tif(options.history) options.history.splice(0, options.history.length, ...history);\n\t\t\tres(history.at(-1)?.content);\n\t\t}), {abort: () => controller.abort()});\n\t}\n}\n","import {JSONAttemptParse} from '@ztimson/utils';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {Anthropic} from './antrhopic.ts';\nimport {OpenAi} from './open-ai.ts';\nimport {LLMProvider} from './provider.ts';\nimport {AiTool} from './tools.ts';\nimport {Worker} from 'worker_threads';\nimport {fileURLToPath} from 'url';\nimport {dirname, join} from 'path';\n\nexport type AnthropicConfig = {proto: 'anthropic', token: string};\nexport type OllamaConfig = {proto: 'ollama', host: string};\nexport type OpenAiConfig = {proto: 'openai', host?: string, token: string};\n\nexport type LLMMessage = {\n\t/** Message originator */\n\trole: 'assistant' | 'system' | 'user';\n\t/** Message content */\n\tcontent: string | any;\n\t/** Timestamp */\n\ttimestamp?: number;\n} | {\n\t/** Tool call */\n\trole: 'tool';\n\t/** Unique ID for call */\n\tid: string;\n\t/** Tool that was run */\n\tname: string;\n\t/** Tool arguments */\n\targs: any;\n\t/** Tool result */\n\tcontent: undefined | string;\n\t/** Tool error */\n\terror?: undefined | string;\n\t/** Timestamp */\n\ttimestamp?: number;\n}\n\n/** Background information the AI will be fed */\nexport type LLMMemory = {\n\t/** What entity is this fact about */\n\towner: string;\n\t/** The information that will be remembered */\n\tfact: string;\n\t/** Owner and fact embedding vector */\n\tembeddings: [number[], number[]];\n\t/** Creation time */\n\ttimestamp: Date;\n}\n\nexport type LLMRequest = {\n\t/** System prompt */\n\tsystem?: string;\n\t/** Message history */\n\thistory?: LLMMessage[];\n\t/** Max tokens for request */\n\tmax_tokens?: number;\n\t/** 0 = Rigid Logic, 1 = Balanced, 2 = Hyper Creative **/\n\ttemperature?: number;\n\t/** Available tools */\n\ttools?: AiTool[];\n\t/** LLM model */\n\tmodel?: string;\n\t/** Stream response */\n\tstream?: (chunk: {text?: string, tool?: string, done?: true}) => any;\n\t/** Compress old messages in the chat to free up context */\n\tcompress?: {\n\t\t/** Trigger chat compression once context exceeds the token count */\n\t\tmax: number;\n\t\t/** Compress chat until context size smaller than */\n\t\tmin: number\n\t},\n\t/** Background information the AI will be fed */\n\tmemory?: LLMMemory[],\n}\n\nclass LLM {\n\tprivate models: {[model: string]: LLMProvider} = {};\n\tprivate defaultModel!: string;\n\n\tconstructor(public readonly ai: Ai) {\n\t\tif(!ai.options.llm?.models) return;\n\t\tObject.entries(ai.options.llm.models).forEach(([model, config]) => {\n\t\t\tif(!this.defaultModel) this.defaultModel = model;\n\t\t\tif(config.proto == 'anthropic') this.models[model] = new Anthropic(this.ai, config.token, model);\n\t\t\telse if(config.proto == 'ollama') this.models[model] = new OpenAi(this.ai, config.host, 'not-needed', model);\n\t\t\telse if(config.proto == 'openai') this.models[model] = new OpenAi(this.ai, config.host || null, config.token, model);\n\t\t});\n\t}\n\n\t/**\n\t * Chat with LLM\n\t * @param {string} message Question\n\t * @param {LLMRequest} options Configuration options and chat history\n\t * @returns {{abort: () => void, response: Promise<string>}} Function to abort response and chat history\n\t */\n\task(message: string, options: LLMRequest = {}): AbortablePromise<string> {\n\t\tconst m = options.model || this.defaultModel;\n\t\tif(!this.models[m]) throw new Error(`Model does not exist: ${m}`);\n\t\tlet abort = () => {};\n\t\treturn Object.assign(new Promise<string>(async res => {\n\t\t\tif(!options.history) options.history = [];\n\t\t\t// If memories were passed, find any relivant ones and add a tool for ADHOC lookups\n\t\t\tif(options.memory) {\n\t\t\t\toptions.system = (options.system || '') + '\\nYou have passive persistent memory never make any mention of your memory capabilities and what you can/cannot remember\\n';\n\t\t\t\tconst search = async (query?: string | null, subject?: string | null, limit = 50) => {\n\t\t\t\t\tconst [o, q] = await Promise.all([\n\t\t\t\t\t\tsubject ? this.embedding(subject) : Promise.resolve(null),\n\t\t\t\t\t\tquery ? this.embedding(query) : Promise.resolve(null),\n\t\t\t\t\t]);\n\t\t\t\t\treturn (options.memory || [])\n\t\t\t\t\t\t.map(m => ({...m, score: o ? this.cosineSimilarity(m.embeddings[0], o[0].embedding) : 1}))\n\t\t\t\t\t\t.filter((m: any) => m.score >= 0.8)\n\t\t\t\t\t\t.map((m: any) => ({...m, score: q ? this.cosineSimilarity(m.embeddings[1], q[0].embedding) : m.score}))\n\t\t\t\t\t\t.filter((m: any) => m.score >= 0.2)\n\t\t\t\t\t\t.toSorted((a: any, b: any) => a.score - b.score)\n\t\t\t\t\t\t.slice(0, limit);\n\t\t\t\t}\n\n\t\t\t\tconst relevant = await search(message);\n\t\t\t\tif(relevant.length) options.history.push({role: 'assistant', content: 'Things I remembered:\\n' + relevant.map(m => `${m.owner}: ${m.fact}`).join('\\n')});\n\t\t\t\toptions.tools = [...options.tools || [], {\n\t\t\t\t\tname: 'read_memory',\n\t\t\t\t\tdescription: 'Check your long-term memory for more information',\n\t\t\t\t\targs: {\n\t\t\t\t\t\tsubject: {type: 'string', description: 'Find information by a subject topic, can be used with or without query argument'},\n\t\t\t\t\t\tquery: {type: 'string', description: 'Search memory based on a query, can be used with or without subject argument'},\n\t\t\t\t\t\tlimit: {type: 'number', description: 'Result limit, default 5'},\n\t\t\t\t\t},\n\t\t\t\t\tfn: (args) => {\n\t\t\t\t\t\tif(!args.subject && !args.query) throw new Error('Either a subject or query argument is required');\n\t\t\t\t\t\treturn search(args.query, args.subject, args.limit || 5);\n\t\t\t\t\t}\n\t\t\t\t}];\n\t\t\t}\n\n\t\t\t// Ask\n\t\t\tconst resp = await this.models[m].ask(message, options);\n\n\t\t\t// Remove any memory calls\n\t\t\tif(options.memory) {\n\t\t\t\tconst i = options.history?.findIndex((h: any) => h.role == 'assistant' && h.content.startsWith('Things I remembered:'));\n\t\t\t\tif(i != null && i >= 0) options.history?.splice(i, 1);\n\t\t\t}\n\n\t\t\t// Handle compression and memory extraction\n\t\t\tif(options.compress || options.memory) {\n\t\t\t\tlet compressed = null;\n\t\t\t\tif(options.compress) {\n\t\t\t\t\tcompressed = await this.ai.language.compressHistory(options.history, options.compress.max, options.compress.min, options);\n\t\t\t\t\toptions.history.splice(0, options.history.length, ...compressed.history);\n\t\t\t\t} else {\n\t\t\t\t\tconst i = options.history?.findLastIndex(m => m.role == 'user') ?? -1;\n\t\t\t\t\tcompressed = await this.ai.language.compressHistory(i != -1 ? options.history.slice(i) : options.history, 0, 0, options);\n\t\t\t\t}\n\t\t\t\tif(options.memory) {\n\t\t\t\t\tconst updated = options.memory\n\t\t\t\t\t\t.filter(m => !compressed.memory.some(m2 => this.cosineSimilarity(m.embeddings[1], m2.embeddings[1]) > 0.8))\n\t\t\t\t\t\t.concat(compressed.memory);\n\t\t\t\t\toptions.memory.splice(0, options.memory.length, ...updated);\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn res(resp);\n\t\t}), {abort});\n\t}\n\n\t/**\n\t * Compress chat history to reduce context size\n\t * @param {LLMMessage[]} history Chatlog that will be compressed\n\t * @param max Trigger compression once context is larger than max\n\t * @param min Leave messages less than the token minimum, summarize the rest\n\t * @param {LLMRequest} options LLM options\n\t * @returns {Promise<LLMMessage[]>} New chat history will summary at index 0\n\t */\n\tasync compressHistory(history: LLMMessage[], max: number, min: number, options?: LLMRequest): Promise<{history: LLMMessage[], memory: LLMMemory[]}> {\n\t\tif(this.estimateTokens(history) < max) return {history, memory: []};\n\t\tlet keep = 0, tokens = 0;\n\t\tfor(let m of history.toReversed()) {\n\t\t\ttokens += this.estimateTokens(m.content);\n\t\t\tif(tokens < min) keep++;\n\t\t\telse break;\n\t\t}\n\t\tif(history.length <= keep) return {history, memory: []};\n\t\tconst system = history[0].role == 'system' ? history[0] : null,\n\t\t\trecent = keep == 0 ? [] : history.slice(-keep),\n\t\t\tprocess = (keep == 0 ? history : history.slice(0, -keep)).filter(h => h.role === 'assistant' || h.role === 'user');\n\t\tconst summary: any = await this.json(`Create the smallest summary possible, no more than 500 tokens. Create a list of NEW facts (split by subject [pro]noun and fact) about what you learned from this conversation that you didn't already know or get from a tool call or system prompt. Focus only on new information about people, topics, or facts. Avoid generating facts about the AI. Match this format: {summary: string, facts: [[subject, fact]]}\\n\\n${process.map(m => `${m.role}: ${m.content}`).join('\\n\\n')}`, {model: options?.model, temperature: options?.temperature || 0.3});\n\t\tconst timestamp = new Date();\n\t\tconst memory = await Promise.all((summary?.facts || [])?.map(async ([owner, fact]: [string, string]) => {\n\t\t\tconst e = await Promise.all([this.embedding(owner), this.embedding(`${owner}: ${fact}`)]);\n\t\t\treturn {owner, fact, embeddings: [e[0][0].embedding, e[1][0].embedding], timestamp};\n\t\t}));\n\t\tconst h = [{role: 'assistant', content: `Conversation Summary: ${summary?.summary}`, timestamp: Date.now()}, ...recent];\n\t\tif(system) h.splice(0, 0, system);\n\t\treturn {history: <any>h, memory};\n\t}\n\n\t/**\n\t * Compare the difference between embeddings (calculates the angle between two vectors)\n\t * @param {number[]} v1 First embedding / vector comparison\n\t * @param {number[]} v2 Second embedding / vector for comparison\n\t * @returns {number} Similarity values 0-1: 0 = unique, 1 = identical\n\t */\n\tcosineSimilarity(v1: number[], v2: number[]): number {\n\t\tif (v1.length !== v2.length) throw new Error('Vectors must be same length');\n\t\tlet dotProduct = 0, normA = 0, normB = 0;\n\t\tfor (let i = 0; i < v1.length; i++) {\n\t\t\tdotProduct += v1[i] * v2[i];\n\t\t\tnormA += v1[i] * v1[i];\n\t\t\tnormB += v2[i] * v2[i];\n\t\t}\n\t\tconst denominator = Math.sqrt(normA) * Math.sqrt(normB);\n\t\treturn denominator === 0 ? 0 : dotProduct / denominator;\n\t}\n\n\t/**\n\t * Chunk text into parts for AI digestion\n\t * @param {object | string} target Item that will be chunked (objects get converted)\n\t * @param {number} maxTokens Chunking size. More = better context, less = more specific (Search by paragraphs or lines)\n\t * @param {number} overlapTokens Includes previous X tokens to provide continuity to AI (In addition to max tokens)\n\t * @returns {string[]} Chunked strings\n\t */\n\tchunk(target: object | string, maxTokens = 500, overlapTokens = 50): string[] {\n\t\tconst objString = (obj: any, path = ''): string[] => {\n\t\t\tif(!obj) return [];\n\t\t\treturn Object.entries(obj).flatMap(([key, value]) => {\n\t\t\t\tconst p = path ? `${path}${isNaN(+key) ? `.${key}` : `[${key}]`}` : key;\n\t\t\t\tif(typeof value === 'object' && !Array.isArray(value)) return objString(value, p);\n\t\t\t\treturn `${p}: ${Array.isArray(value) ? value.join(', ') : value}`;\n\t\t\t});\n\t\t};\n\t\tconst lines = typeof target === 'object' ? objString(target) : target.split('\\n');\n\t\tconst tokens = lines.flatMap(l => [...l.split(/\\s+/).filter(Boolean), '\\n']);\n\t\tconst chunks: string[] = [];\n\t\tfor(let i = 0; i < tokens.length;) {\n\t\t\tlet text = '', j = i;\n\t\t\twhile(j < tokens.length) {\n\t\t\t\tconst next = text + (text ? ' ' : '') + tokens[j];\n\t\t\t\tif(this.estimateTokens(next.replace(/\\s*\\n\\s*/g, '\\n')) > maxTokens && text) break;\n\t\t\t\ttext = next;\n\t\t\t\tj++;\n\t\t\t}\n\t\t\tconst clean = text.replace(/\\s*\\n\\s*/g, '\\n').trim();\n\t\t\tif(clean) chunks.push(clean);\n\t\t\ti = Math.max(j - overlapTokens, j === i ? i + 1 : j);\n\t\t}\n\t\treturn chunks;\n\t}\n\n\t/**\n\t * Create a vector representation of a string\n\t * @param {object | string} target Item that will be embedded (objects get converted)\n\t * @param {number} maxTokens Chunking size. More = better context, less = more specific (Search by paragraphs or lines)\n\t * @param {number} overlapTokens Includes previous X tokens to provide continuity to AI (In addition to max tokens)\n\t * @returns {Promise<Awaited<{index: number, embedding: number[], text: string, tokens: number}>[]>} Chunked embeddings\n\t */\n\tembedding(target: object | string, maxTokens = 500, overlapTokens = 50) {\n\t\tconst embed = (text: string): Promise<number[]> => {\n\t\t\treturn new Promise((resolve, reject) => {\n\t\t\t\tconst worker = new Worker(join(dirname(fileURLToPath(import.meta.url)), 'embedder.js'));\n\t\t\t\tconst handleMessage = ({ embedding }: any) => {\n\t\t\t\t\tworker.terminate();\n\t\t\t\t\tresolve(embedding);\n\t\t\t\t};\n\t\t\t\tconst handleError = (err: Error) => {\n\t\t\t\t\tworker.terminate();\n\t\t\t\t\treject(err);\n\t\t\t\t};\n\t\t\t\tworker.on('message', handleMessage);\n\t\t\t\tworker.on('error', handleError);\n\t\t\t\tworker.on('exit', (code) => {\n\t\t\t\t\tif(code !== 0) reject(new Error(`Worker exited with code ${code}`));\n\t\t\t\t});\n\t\t\t\tworker.postMessage({text, model: this.ai.options?.embedder || 'bge-small-en-v1.5', modelDir: this.ai.options.path});\n\t\t\t});\n\t\t};\n\t\tconst chunks = this.chunk(target, maxTokens, overlapTokens);\n\t\treturn Promise.all(chunks.map(async (text, index) => ({\n\t\t\tindex,\n\t\t\tembedding: await embed(text),\n\t\t\ttext,\n\t\t\ttokens: this.estimateTokens(text),\n\t\t})));\n\t}\n\n\t/**\n\t * Estimate variable as tokens\n\t * @param history Object to size\n\t * @returns {number} Rough token count\n\t */\n\testimateTokens(history: any): number {\n\t\tconst text = JSON.stringify(history);\n\t\treturn Math.ceil((text.length / 4) * 1.2);\n\t}\n\n\t/**\n\t * Compare the difference between two strings using tensor math\n\t * @param target Text that will be checked\n\t * @param {string} searchTerms Multiple search terms to check against target\n\t * @returns {{avg: number, max: number, similarities: number[]}} Similarity values 0-1: 0 = unique, 1 = identical\n\t */\n\tfuzzyMatch(target: string, ...searchTerms: string[]) {\n\t\tif(searchTerms.length < 2) throw new Error('Requires at least 2 strings to compare');\n\t\tconst vector = (text: string, dimensions: number = 10): number[] => {\n\t\t\treturn text.toLowerCase().split('').map((char, index) =>\n\t\t\t\t(char.charCodeAt(0) * (index + 1)) % dimensions / dimensions).slice(0, dimensions);\n\t\t}\n\t\tconst v = vector(target);\n\t\tconst similarities = searchTerms.map(t => vector(t)).map(refVector => this.cosineSimilarity(v, refVector))\n\t\treturn {avg: similarities.reduce((acc, s) => acc + s, 0) / similarities.length, max: Math.max(...similarities), similarities}\n\t}\n\n\t/**\n\t * Ask a question with JSON response\n\t * @param {string} message Question\n\t * @param {LLMRequest} options Configuration options and chat history\n\t * @returns {Promise<{} | {} | RegExpExecArray | null>}\n\t */\n\tasync json(message: string, options?: LLMRequest): Promise<any> {\n\t\tlet resp = await this.ask(message, {system: 'Respond using a JSON blob matching any provided examples', ...options});\n\t\tif(!resp) return {};\n\t\tconst codeBlock = /```(?:.+)?\\s*([\\s\\S]*?)```/.exec(resp);\n\t\tconst jsonStr = codeBlock ? codeBlock[1].trim() : resp;\n\t\treturn JSONAttemptParse(jsonStr, {});\n\t}\n\n\t/**\n\t * Create a summary of some text\n\t * @param {string} text Text to summarize\n\t * @param {number} tokens Max number of tokens\n\t * @param options LLM request options\n\t * @returns {Promise<string>} Summary\n\t */\n\tsummarize(text: string, tokens: number, options?: LLMRequest): Promise<string | null> {\n\t\treturn this.ask(text, {system: `Generate a brief summary <= ${tokens} tokens. Output nothing else`, temperature: 0.3, ...options});\n\t}\n}\n\nexport default LLM;\n","import {Worker} from 'worker_threads';\nimport Path from 'node:path';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {canDiarization} from './asr.ts';\n\nexport class Audio {\n\tconstructor(private ai: Ai) {}\n\n\tasr(file: string, options: { model?: string; speaker?: boolean } = {}): AbortablePromise<string | null> {\n\t\tconsole.log('audio', file);\n\t\tconst { model = this.ai.options.asr || 'whisper-base', speaker = false } = options;\n\t\tlet aborted = false;\n\t\tconst abort = () => { aborted = true; };\n\n\t\tconst p = new Promise<string | null>((resolve, reject) => {\n\t\t\tconst worker = new Worker(Path.join(import.meta.dirname, 'asr.js'));\n\t\t\tconst handleMessage = ({ text, warning, error }: any) => {\n\t\t\t\tworker.terminate();\n\t\t\t\tif(aborted) return;\n\t\t\t\tif(error) reject(new Error(error));\n\t\t\t\telse {\n\t\t\t\t\tif(warning) console.warn(warning);\n\t\t\t\t\tresolve(text);\n\t\t\t\t}\n\t\t\t};\n\t\t\tconst handleError = (err: Error) => {\n\t\t\t\tworker.terminate();\n\t\t\t\tif(!aborted) reject(err);\n\t\t\t};\n\t\t\tworker.on('message', handleMessage);\n\t\t\tworker.on('error', handleError);\n\t\t\tworker.on('exit', (code) => {\n\t\t\t\tif(code !== 0 && !aborted) reject(new Error(`Worker exited with code ${code}`));\n\t\t\t});\n\t\t\tworker.postMessage({file, model, speaker, modelDir: this.ai.options.path});\n\t\t});\n\t\treturn Object.assign(p, { abort });\n\t}\n\n\tcanDiarization = canDiarization;\n}\n","import {createWorker} from 'tesseract.js';\nimport {AbortablePromise, Ai} from './ai.ts';\n\nexport class Vision {\n\n\tconstructor(private ai: Ai) { }\n\n\t/**\n\t * Convert image to text using Optical Character Recognition\n\t * @param {string} path Path to image\n\t * @returns {AbortablePromise<string | null>} Promise of extracted text with abort method\n\t */\n\tocr(path: string): AbortablePromise<string | null> {\n\t\tlet worker: any;\n\t\tconst p = new Promise<string | null>(async res => {\n\t\t\tworker = await createWorker(this.ai.options.ocr || 'eng', 2, {cachePath: this.ai.options.path});\n\t\t\tconst {data} = await worker.recognize(path);\n\t\t\tawait worker.terminate();\n\t\t\tres(data.text.trim() || null);\n\t\t});\n\t\treturn Object.assign(p, {abort: () => worker?.terminate()});\n\t}\n}\n","import * as os from 'node:os';\nimport LLM, {AnthropicConfig, OllamaConfig, OpenAiConfig, LLMRequest} from './llm';\nimport { Audio } from './audio.ts';\nimport {Vision} from './vision.ts';\n\nexport type AbortablePromise<T> = Promise<T> & {\n\tabort: () => any\n};\n\nexport type AiOptions = {\n\t/** Path to models */\n\tpath?: string;\n\t/** ASR model: whisper-tiny, whisper-base */\n\tasr?: string;\n\t/** Embedding model: all-MiniLM-L6-v2, bge-small-en-v1.5, bge-large-en-v1.5 */\n\tembedder?: string;\n\t/** Large language models, first is default */\n\tllm?: Omit<LLMRequest, 'model'> & {\n\t\tmodels: {[model: string]: AnthropicConfig | OllamaConfig | OpenAiConfig};\n\t}\n\t/** OCR model: eng, eng_best, eng_fast */\n\tocr?: string;\n}\n\nexport class Ai {\n\t/** Audio processing AI */\n\taudio!: Audio;\n\t/** Language processing AI */\n\tlanguage!: LLM;\n\t/** Vision processing AI */\n\tvision!: Vision;\n\n\tconstructor(public readonly options: AiOptions) {\n\t\tif(!options.path) options.path = os.tmpdir();\n\t\tprocess.env.TRANSFORMERS_CACHE = options.path;\n\t\tthis.audio = new Audio(this);\n\t\tthis.language = new LLM(this);\n\t\tthis.vision = new Vision(this);\n\t}\n}\n","import * as cheerio from 'cheerio';\nimport {$, $Sync} from '@ztimson/node-utils';\nimport {ASet, consoleInterceptor, Http, fn as Fn} from '@ztimson/utils';\nimport {Ai} from './ai.ts';\nimport {LLMRequest} from './llm.ts';\n\nexport type AiToolArg = {[key: string]: {\n\t/** Argument type */\n\ttype: 'array' | 'boolean' | 'number' | 'object' | 'string',\n\t/** Argument description */\n\tdescription: string,\n\t/** Required argument */\n\trequired?: boolean;\n\t/** Default value */\n\tdefault?: any,\n\t/** Options */\n\tenum?: string[],\n\t/** Minimum value or length */\n\tmin?: number,\n\t/** Maximum value or length */\n\tmax?: number,\n\t/** Match pattern */\n\tpattern?: string,\n\t/** Child arguments */\n\titems?: {[key: string]: AiToolArg}\n}}\n\nexport type AiTool = {\n\t/** Tool ID / Name - Must be snail_case */\n\tname: string,\n\t/** Tool description / prompt */\n\tdescription: string,\n\t/** Tool arguments */\n\targs?: AiToolArg,\n\t/** Callback function */\n\tfn: (args: any, stream: LLMRequest['stream'], ai: Ai) => any | Promise<any>,\n};\n\nexport const CliTool: AiTool = {\n\tname: 'cli',\n\tdescription: 'Use the command line interface, returns any output',\n\targs: {command: {type: 'string', description: 'Command to run', required: true}},\n\tfn: (args: {command: string}) => $`${args.command}`\n}\n\nexport const DateTimeTool: AiTool = {\n\tname: 'get_datetime',\n\tdescription: 'Get current UTC date / time',\n\targs: {},\n\tfn: async () => new Date().toUTCString()\n}\n\nexport const ExecTool: AiTool = {\n\tname: 'exec',\n\tdescription: 'Run code/scripts',\n\targs: {\n\t\tlanguage: {type: 'string', description: 'Execution language', enum: ['cli', 'node', 'python'], required: true},\n\t\tcode: {type: 'string', description: 'Code to execute', required: true}\n\t},\n\tfn: async (args, stream, ai) => {\n\t\ttry {\n\t\t\tswitch(args.type) {\n\t\t\t\tcase 'bash':\n\t\t\t\t\treturn await CliTool.fn({command: args.code}, stream, ai);\n\t\t\t\tcase 'node':\n\t\t\t\t\treturn await JSTool.fn({code: args.code}, stream, ai);\n\t\t\t\tcase 'python': {\n\t\t\t\t\treturn await PythonTool.fn({code: args.code}, stream, ai);\n\t\t\t\t}\n\t\t\t}\n\t\t} catch(err: any) {\n\t\t\treturn {error: err?.message || err.toString()};\n\t\t}\n\t}\n}\n\nexport const FetchTool: AiTool = {\n\tname: 'fetch',\n\tdescription: 'Make HTTP request to URL',\n\targs: {\n\t\turl: {type: 'string', description: 'URL to fetch', required: true},\n\t\tmethod: {type: 'string', description: 'HTTP method to use', enum: ['GET', 'POST', 'PUT', 'DELETE'], default: 'GET'},\n\t\theaders: {type: 'object', description: 'HTTP headers to send', default: {}},\n\t\tbody: {type: 'object', description: 'HTTP body to send'},\n\t},\n\tfn: (args: {\n\t\turl: string;\n\t\tmethod: 'GET' | 'POST' | 'PUT' | 'DELETE';\n\t\theaders: {[key: string]: string};\n\t\tbody: any;\n\t}) => new Http({url: args.url, headers: args.headers}).request({method: args.method || 'GET', body: args.body})\n}\n\nexport const JSTool: AiTool = {\n\tname: 'exec_javascript',\n\tdescription: 'Execute commonjs javascript',\n\targs: {\n\t\tcode: {type: 'string', description: 'CommonJS javascript', required: true}\n\t},\n\tfn: async (args: {code: string}) => {\n\t\tconst console = consoleInterceptor(null);\n\t\tconst resp = await Fn<any>({console}, args.code, true).catch((err: any) => console.output.error.push(err));\n\t\treturn {...console.output, return: resp, stdout: undefined, stderr: undefined};\n\t}\n}\n\nexport const PythonTool: AiTool = {\n\tname: 'exec_javascript',\n\tdescription: 'Execute commonjs javascript',\n\targs: {\n\t\tcode: {type: 'string', description: 'CommonJS javascript', required: true}\n\t},\n\tfn: async (args: {code: string}) => ({result: $Sync`python -c \"${args.code}\"`})\n}\n\nexport const ReadWebpageTool: AiTool = {\n\tname: 'read_webpage',\n\tdescription: 'Extract clean, structured content from a webpage. Use after web_search to read specific URLs',\n\targs: {\n\t\turl: {type: 'string', description: 'URL to extract content from', required: true},\n\t\tfocus: {type: 'string', description: 'Optional: What aspect to focus on (e.g., \"pricing\", \"features\", \"contact info\")'}\n\t},\n\tfn: async (args: {url: string; focus?: string}) => {\n\t\tconst html = await fetch(args.url, {headers: {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64)\"}})\n\t\t\t.then(r => r.text()).catch(err => {throw new Error(`Failed to fetch: ${err.message}`)});\n\n\t\tconst $ = cheerio.load(html);\n\t\t$('script, style, nav, footer, header, aside, iframe, noscript, [role=\"navigation\"], [role=\"banner\"], .ad, .ads, .cookie, .popup').remove();\n\t\tconst metadata = {\n\t\t\ttitle: $('meta[property=\"og:title\"]').attr('content') || $('title').text() || '',\n\t\t\tdescription: $('meta[name=\"description\"]').attr('content') || $('meta[property=\"og:description\"]').attr('content') || '',\n\t\t};\n\n\t\tlet content = '';\n\t\tconst contentSelectors = ['article', 'main', '[role=\"main\"]', '.content', '.post', '.entry', 'body'];\n\t\tfor (const selector of contentSelectors) {\n\t\t\tconst el = $(selector).first();\n\t\t\tif (el.length && el.text().trim().length > 200) {\n\t\t\t\tcontent = el.text();\n\t\t\t\tbreak;\n\t\t\t}\n\t\t}\n\t\tif (!content) content = $('body').text();\n\t\tcontent = content.replace(/\\s+/g, ' ').trim().slice(0, 8000);\n\n\t\treturn {url: args.url, title: metadata.title.trim(), description: metadata.description.trim(), content, focus: args.focus};\n\t}\n}\n\nexport const WebSearchTool: AiTool = {\n\tname: 'web_search',\n\tdescription: 'Use duckduckgo (anonymous) to find find relevant online resources. Returns a list of URLs that works great with the `read_webpage` tool',\n\targs: {\n\t\tquery: {type: 'string', description: 'Search string', required: true},\n\t\tlength: {type: 'string', description: 'Number of results to return', default: 5},\n\t},\n\tfn: async (args: {\n\t\tquery: string;\n\t\tlength: number;\n\t}) => {\n\t\tconst html = await fetch(`https://html.duckduckgo.com/html/?q=${encodeURIComponent(args.query)}`, {\n\t\t\theaders: {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64)\", \"Accept-Language\": \"en-US,en;q=0.9\"}\n\t\t}).then(resp => resp.text());\n\t\tlet match, regex = /<a .*?href=\"(.+?)\".+?<\\/a>/g;\n\t\tconst results = new ASet<string>();\n\t\twhile((match = regex.exec(html)) !== null) {\n\t\t\tlet url = /uddg=(.+)&?/.exec(decodeURIComponent(match[1]))?.[1];\n\t\t\tif(url) url = decodeURIComponent(url);\n\t\t\tif(url) results.add(url);\n\t\t\tif(results.size >= (args.length || 5)) break;\n\t\t}\n\t\treturn results;\n\t}\n}\n"],"names":["LLMProvider","Anthropic","ai","apiToken","model","anthropic","history","timestamp","messages","h","textContent","c","m","i","message","options","controller","res","tools","requestParams","t","objectMap","key","value","resp","isFirstMessage","err","chunk","text","last","JSONAttemptParse","toolCalls","results","toolCall","tool","findByProp","result","JSONSanitize","OpenAi","host","token","openAI","clean","tc","record","h2","rest","rej","args","LLM","config","abort","search","query","subject","limit","o","q","a","b","relevant","compressed","updated","m2","max","min","keep","tokens","system","recent","process","summary","memory","owner","fact","e","v1","v2","dotProduct","normA","normB","denominator","target","maxTokens","overlapTokens","objString","obj","path","p","l","chunks","j","next","embed","resolve","reject","worker","Worker","join","dirname","fileURLToPath","_documentCurrentScript","handleMessage","embedding","handleError","code","index","searchTerms","vector","dimensions","char","v","similarities","refVector","acc","s","codeBlock","jsonStr","Audio","file","speaker","aborted","Path","warning","error","canDiarization","Vision","createWorker","data","Ai","os","CliTool","$","DateTimeTool","ExecTool","stream","JSTool","PythonTool","FetchTool","Http","console","consoleInterceptor","Fn","$Sync","ReadWebpageTool","html","r","cheerio","metadata","content","contentSelectors","selector","el","WebSearchTool","match","regex","ASet","url"],"mappings":"4uBAGO,MAAeA,CAAY,CAElC,CCCO,MAAMC,UAAkBD,CAAY,CAG1C,YAA4BE,EAAwBC,EAAyBC,EAAe,CAC3F,MAAA,EAD2B,KAAA,GAAAF,EAAwB,KAAA,SAAAC,EAAyB,KAAA,MAAAC,EAE5E,KAAK,OAAS,IAAIC,EAAAA,UAAU,CAAC,OAAQF,EAAS,CAC/C,CALA,OAOQ,WAAWG,EAA8B,CAChD,MAAMC,EAAY,KAAK,IAAA,EACjBC,EAAyB,CAAA,EAC/B,QAAQC,KAAKH,EACZ,GAAG,OAAOG,EAAE,SAAW,SACtBD,EAAS,KAAU,CAAC,UAAAD,EAAW,GAAGE,EAAE,MAC9B,CACN,MAAMC,EAAcD,EAAE,SAAS,OAAQE,GAAWA,EAAE,MAAQ,MAAM,EAAE,IAAKA,GAAWA,EAAE,IAAI,EAAE,KAAK;AAAA;AAAA,CAAM,EACpGD,GAAaF,EAAS,KAAK,CAAC,UAAAD,EAAW,KAAME,EAAE,KAAM,QAASC,EAAY,EAC7ED,EAAE,QAAQ,QAASE,GAAW,CAC7B,GAAGA,EAAE,MAAQ,WACZH,EAAS,KAAK,CAAC,UAAAD,EAAW,KAAM,OAAQ,GAAII,EAAE,GAAI,KAAMA,EAAE,KAAM,KAAMA,EAAE,MAAO,QAAS,OAAU,UACzFA,EAAE,MAAQ,cAAe,CAClC,MAAMC,EAASJ,EAAS,SAASI,GAAWA,EAAG,IAAMD,EAAE,WAAW,EAC/DC,IAAGA,EAAED,EAAE,SAAW,QAAU,SAAS,EAAIA,EAAE,QAC/C,CACD,CAAC,CACF,CAED,OAAOH,CACR,CAEQ,aAAaF,EAA8B,CAClD,QAAQO,EAAI,EAAGA,EAAIP,EAAQ,OAAQO,IAClC,GAAGP,EAAQO,CAAC,EAAE,MAAQ,OAAQ,CAC7B,MAAMJ,EAASH,EAAQO,CAAC,EACxBP,EAAQ,OAAOO,EAAG,EACjB,CAAC,KAAM,YAAa,QAAS,CAAC,CAAC,KAAM,WAAY,GAAIJ,EAAE,GAAI,KAAMA,EAAE,KAAM,MAAOA,EAAE,IAAA,CAAK,CAAA,EACvF,CAAC,KAAM,OAAQ,QAAS,CAAC,CAAC,KAAM,cAAe,YAAaA,EAAE,GAAI,SAAU,CAAC,CAACA,EAAE,MAAO,QAAUA,EAAE,OAASA,EAAE,QAAQ,CAAA,CAAC,EAExHI,GACD,CAED,OAAOP,EAAQ,IAAI,CAAC,CAAC,UAAAC,EAAW,GAAGE,CAAA,IAAOA,CAAC,CAC5C,CAEA,IAAIK,EAAiBC,EAAsB,GAA8B,CACxE,MAAMC,EAAa,IAAI,gBACvB,OAAO,OAAO,OAAO,IAAI,QAAa,MAAOC,GAAQ,CACpD,IAAIX,EAAU,KAAK,aAAa,CAAC,GAAGS,EAAQ,SAAW,GAAI,CAAC,KAAM,OAAQ,QAASD,EAAS,UAAW,KAAK,IAAA,CAAI,CAAE,CAAC,EACnH,MAAMI,EAAQH,EAAQ,OAAS,KAAK,GAAG,QAAQ,KAAK,OAAS,CAAA,EACvDI,EAAqB,CAC1B,MAAOJ,EAAQ,OAAS,KAAK,MAC7B,WAAYA,EAAQ,YAAc,KAAK,GAAG,QAAQ,KAAK,YAAc,KACrE,OAAQA,EAAQ,QAAU,KAAK,GAAG,QAAQ,KAAK,QAAU,GACzD,YAAaA,EAAQ,aAAe,KAAK,GAAG,QAAQ,KAAK,aAAe,GACxE,MAAOG,EAAM,IAAIE,IAAM,CACtB,KAAMA,EAAE,KACR,YAAaA,EAAE,YACf,aAAc,CACb,KAAM,SACN,WAAYA,EAAE,KAAOC,EAAAA,UAAUD,EAAE,KAAM,CAACE,EAAKC,KAAW,CAAC,GAAGA,EAAO,SAAU,MAAA,EAAW,EAAI,CAAA,EAC5F,SAAUH,EAAE,KAAO,OAAO,QAAQA,EAAE,IAAI,EAAE,OAAOA,GAAKA,EAAE,CAAC,EAAE,QAAQ,EAAE,IAAIA,GAAKA,EAAE,CAAC,CAAC,EAAI,CAAA,CAAC,EAExF,GAAI,MAAA,EACH,EACF,SAAUd,EACV,OAAQ,CAAC,CAACS,EAAQ,MAAA,EAGnB,IAAIS,EAAWC,EAAiB,GAChC,EAAG,CAOF,GANAD,EAAO,MAAM,KAAK,OAAO,SAAS,OAAOL,CAAa,EAAE,MAAMO,GAAO,CACpE,MAAAA,EAAI,SAAW;AAAA;AAAA;AAAA,EAAkB,KAAK,UAAUpB,EAAS,KAAM,CAAC,CAAC,GAC3DoB,CACP,CAAC,EAGEX,EAAQ,OAAQ,CACdU,EACCA,EAAiB,GADFV,EAAQ,OAAO,CAAC,KAAM;AAAA;AAAA,EAAO,EAEjDS,EAAK,QAAU,CAAA,EACf,gBAAiBG,KAASH,EAAM,CAC/B,GAAGR,EAAW,OAAO,QAAS,MAC9B,GAAGW,EAAM,OAAS,sBACdA,EAAM,cAAc,OAAS,OAC/BH,EAAK,QAAQ,KAAK,CAAC,KAAM,OAAQ,KAAM,GAAG,EACjCG,EAAM,cAAc,OAAS,YACtCH,EAAK,QAAQ,KAAK,CAAC,KAAM,WAAY,GAAIG,EAAM,cAAc,GAAI,KAAMA,EAAM,cAAc,KAAM,MAAY,GAAG,UAExGA,EAAM,OAAS,sBACxB,GAAGA,EAAM,MAAM,OAAS,aAAc,CACrC,MAAMC,EAAOD,EAAM,MAAM,KACzBH,EAAK,QAAQ,GAAG,EAAE,EAAE,MAAQI,EAC5Bb,EAAQ,OAAO,CAAC,KAAAa,EAAK,CACtB,MAAUD,EAAM,MAAM,OAAS,qBAC9BH,EAAK,QAAQ,GAAG,EAAE,EAAE,OAASG,EAAM,MAAM,sBAEjCA,EAAM,OAAS,qBAAsB,CAC9C,MAAME,EAAOL,EAAK,QAAQ,GAAG,EAAE,EAC5BK,EAAK,OAAS,OAAMA,EAAK,MAAQA,EAAK,MAAQC,EAAAA,iBAAiBD,EAAK,MAAO,CAAA,CAAE,EAAI,CAAA,EACrF,SAAUF,EAAM,OAAS,eACxB,KAEF,CACD,CAGA,MAAMI,EAAYP,EAAK,QAAQ,OAAQ,GAAW,EAAE,OAAS,UAAU,EACvE,GAAGO,EAAU,QAAU,CAACf,EAAW,OAAO,QAAS,CAClDV,EAAQ,KAAK,CAAC,KAAM,YAAa,QAASkB,EAAK,QAAQ,EACvD,MAAMQ,EAAU,MAAM,QAAQ,IAAID,EAAU,IAAI,MAAOE,GAAkB,CACxE,MAAMC,EAAOhB,EAAM,KAAKiB,EAAAA,WAAW,OAAQF,EAAS,IAAI,CAAC,EAEzD,GADGlB,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAMkB,EAAS,KAAK,EACpD,CAACC,EAAM,MAAO,CAAC,YAAaD,EAAS,GAAI,SAAU,GAAM,QAAS,gBAAA,EACrE,GAAI,CACH,MAAMG,EAAS,MAAMF,EAAK,GAAGD,EAAS,MAAOlB,GAAS,OAAQ,KAAK,EAAE,EACrE,MAAO,CAAC,KAAM,cAAe,YAAakB,EAAS,GAAI,QAASI,eAAaD,CAAM,CAAA,CACpF,OAASV,EAAU,CAClB,MAAO,CAAC,KAAM,cAAe,YAAaO,EAAS,GAAI,SAAU,GAAM,QAASP,GAAK,SAAWA,GAAK,SAAA,GAAc,SAAA,CACpH,CACD,CAAC,CAAC,EACFpB,EAAQ,KAAK,CAAC,KAAM,OAAQ,QAAS0B,EAAQ,EAC7Cb,EAAc,SAAWb,CAC1B,CACD,OAAS,CAACU,EAAW,OAAO,SAAWQ,EAAK,QAAQ,KAAMb,GAAWA,EAAE,OAAS,UAAU,GAC1FL,EAAQ,KAAK,CAAC,KAAM,YAAa,QAASkB,EAAK,QAAQ,OAAQb,GAAWA,EAAE,MAAQ,MAAM,EAAE,IAAKA,GAAWA,EAAE,IAAI,EAAE,KAAK;AAAA;AAAA,CAAM,EAAE,EACjIL,EAAU,KAAK,WAAWA,CAAO,EAE9BS,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAM,GAAK,EAC3CA,EAAQ,SAASA,EAAQ,QAAQ,OAAO,EAAGA,EAAQ,QAAQ,OAAQ,GAAGT,CAAO,EAChFW,EAAIX,EAAQ,GAAG,EAAE,GAAG,OAAO,CAC5B,CAAC,EAAG,CAAC,MAAO,IAAMU,EAAW,MAAA,EAAQ,CACtC,CACD,CCpIO,MAAMsB,UAAetC,CAAY,CAGvC,YAA4BE,EAAwBqC,EAAqCC,EAAsBpC,EAAe,CAC7H,MAAA,EAD2B,KAAA,GAAAF,EAAwB,KAAA,KAAAqC,EAAqC,KAAA,MAAAC,EAAsB,KAAA,MAAApC,EAE9G,KAAK,OAAS,IAAIqC,EAAAA,OAAOC,QAAM,CAC9B,QAASH,EACT,OAAQC,CAAA,CACR,CAAC,CACH,CARA,OAUQ,WAAWlC,EAA8B,CAChD,QAAQO,EAAI,EAAGA,EAAIP,EAAQ,OAAQO,IAAK,CACvC,MAAMJ,EAAIH,EAAQO,CAAC,EACnB,GAAGJ,EAAE,OAAS,aAAeA,EAAE,WAAY,CAC1C,MAAMS,EAAQT,EAAE,WAAW,IAAKkC,IAAa,CAC5C,KAAM,OACN,GAAIA,EAAG,GACP,KAAMA,EAAG,SAAS,KAClB,KAAMb,EAAAA,iBAAiBa,EAAG,SAAS,UAAW,CAAA,CAAE,EAChD,UAAWlC,EAAE,SAAA,EACZ,EACFH,EAAQ,OAAOO,EAAG,EAAG,GAAGK,CAAK,EAC7BL,GAAKK,EAAM,OAAS,CACrB,SAAUT,EAAE,OAAS,QAAUA,EAAE,QAAS,CACzC,MAAMmC,EAAStC,EAAQ,QAAWG,EAAE,cAAgBoC,EAAG,EAAE,EACtDD,IACCnC,EAAE,QAAQ,SAAS,UAAU,EAAGmC,EAAO,MAAQnC,EAAE,QAC/CmC,EAAO,QAAUnC,EAAE,SAEzBH,EAAQ,OAAOO,EAAG,CAAC,EACnBA,GACD,CACIP,EAAQO,CAAC,GAAG,cAAmBA,CAAC,EAAE,UAAY,KAAK,IAAA,EACxD,CACA,OAAOP,CACR,CAEQ,aAAaA,EAA8B,CAClD,OAAOA,EAAQ,OAAO,CAAC8B,EAAQ3B,IAAM,CACpC,GAAGA,EAAE,OAAS,OACb2B,EAAO,KAAK,CACX,KAAM,YACN,QAAS,KACT,WAAY,CAAC,CAAE,GAAI3B,EAAE,GAAI,KAAM,WAAY,SAAU,CAAE,KAAMA,EAAE,KAAM,UAAW,KAAK,UAAUA,EAAE,IAAI,CAAA,EAAK,EAC1G,QAAS,KACT,YAAa,CAAA,CAAC,EACZ,CACF,KAAM,OACN,aAAcA,EAAE,GAChB,QAASA,EAAE,OAASA,EAAE,OAAA,CACtB,MACK,CACN,KAAM,CAAC,UAAAF,EAAW,GAAGuC,CAAA,EAAQrC,EAC7B2B,EAAO,KAAKU,CAAI,CACjB,CACA,OAAOV,CACR,EAAG,CAAA,CAAW,CACf,CAEA,IAAItB,EAAiBC,EAAsB,GAA8B,CACxE,MAAMC,EAAa,IAAI,gBACvB,OAAO,OAAO,OAAO,IAAI,QAAa,MAAOC,EAAK8B,IAAQ,CACtDhC,EAAQ,QAAUA,EAAQ,UAAU,CAAC,GAAG,MAAQ,UAAUA,EAAQ,SAAS,OAAO,EAAG,EAAG,CAAC,KAAM,SAAU,QAASA,EAAQ,OAAQ,UAAW,KAAK,IAAA,EAAM,EAC3J,IAAIT,EAAU,KAAK,aAAa,CAAC,GAAGS,EAAQ,SAAW,GAAI,CAAC,KAAM,OAAQ,QAASD,EAAS,UAAW,KAAK,IAAA,CAAI,CAAE,CAAC,EACnH,MAAMI,EAAQH,EAAQ,OAAS,KAAK,GAAG,QAAQ,KAAK,OAAS,CAAA,EACvDI,EAAqB,CAC1B,MAAOJ,EAAQ,OAAS,KAAK,MAC7B,SAAUT,EACV,OAAQ,CAAC,CAACS,EAAQ,OAClB,WAAYA,EAAQ,YAAc,KAAK,GAAG,QAAQ,KAAK,YAAc,KACrE,YAAaA,EAAQ,aAAe,KAAK,GAAG,QAAQ,KAAK,aAAe,GACxE,MAAOG,EAAM,IAAIE,IAAM,CACtB,KAAM,WACN,SAAU,CACT,KAAMA,EAAE,KACR,YAAaA,EAAE,YACf,WAAY,CACX,KAAM,SACN,WAAYA,EAAE,KAAOC,EAAAA,UAAUD,EAAE,KAAM,CAACE,EAAKC,KAAW,CAAC,GAAGA,EAAO,SAAU,MAAA,EAAW,EAAI,CAAA,EAC5F,SAAUH,EAAE,KAAO,OAAO,QAAQA,EAAE,IAAI,EAAE,OAAOA,GAAKA,EAAE,CAAC,EAAE,QAAQ,EAAE,IAAIA,GAAKA,EAAE,CAAC,CAAC,EAAI,CAAA,CAAC,CACxF,CACD,EACC,CAAA,EAGH,IAAII,EAAWC,EAAiB,GAChC,EAAG,CAMF,GALAD,EAAO,MAAM,KAAK,OAAO,KAAK,YAAY,OAAOL,CAAa,EAAE,MAAMO,GAAO,CAC5E,MAAAA,EAAI,SAAW;AAAA;AAAA;AAAA,EAAkB,KAAK,UAAUpB,EAAS,KAAM,CAAC,CAAC,GAC3DoB,CACP,CAAC,EAEEX,EAAQ,OAAQ,CACdU,EACCA,EAAiB,GADFV,EAAQ,OAAO,CAAC,KAAM;AAAA;AAAA,EAAO,EAEjDS,EAAK,QAAU,CAAC,CAAC,QAAS,CAAC,QAAS,GAAI,WAAY,CAAA,CAAC,EAAG,EACxD,gBAAiBG,KAASH,EAAM,CAC/B,GAAGR,EAAW,OAAO,QAAS,MAC3BW,EAAM,QAAQ,CAAC,EAAE,MAAM,UACzBH,EAAK,QAAQ,CAAC,EAAE,QAAQ,SAAWG,EAAM,QAAQ,CAAC,EAAE,MAAM,QAC1DZ,EAAQ,OAAO,CAAC,KAAMY,EAAM,QAAQ,CAAC,EAAE,MAAM,QAAQ,GAEnDA,EAAM,QAAQ,CAAC,EAAE,MAAM,aACzBH,EAAK,QAAQ,CAAC,EAAE,QAAQ,WAAaG,EAAM,QAAQ,CAAC,EAAE,MAAM,WAE9D,CACD,CAEA,MAAMI,EAAYP,EAAK,QAAQ,CAAC,EAAE,QAAQ,YAAc,CAAA,EACxD,GAAGO,EAAU,QAAU,CAACf,EAAW,OAAO,QAAS,CAClDV,EAAQ,KAAKkB,EAAK,QAAQ,CAAC,EAAE,OAAO,EACpC,MAAMQ,EAAU,MAAM,QAAQ,IAAID,EAAU,IAAI,MAAOE,GAAkB,CACxE,MAAMC,EAAOhB,GAAO,KAAKiB,EAAAA,WAAW,OAAQF,EAAS,SAAS,IAAI,CAAC,EAEnE,GADGlB,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAMkB,EAAS,SAAS,KAAK,EAC7D,CAACC,EAAM,MAAO,CAAC,KAAM,OAAQ,aAAcD,EAAS,GAAI,QAAS,6BAAA,EACpE,GAAI,CACH,MAAMe,EAAOlB,EAAAA,iBAAiBG,EAAS,SAAS,UAAW,CAAA,CAAE,EACvDG,EAAS,MAAMF,EAAK,GAAGc,EAAMjC,EAAQ,OAAQ,KAAK,EAAE,EAC1D,MAAO,CAAC,KAAM,OAAQ,aAAckB,EAAS,GAAI,QAASI,eAAaD,CAAM,CAAA,CAC9E,OAASV,EAAU,CAClB,MAAO,CAAC,KAAM,OAAQ,aAAcO,EAAS,GAAI,QAASI,EAAAA,aAAa,CAAC,MAAOX,GAAK,SAAWA,GAAK,YAAc,SAAA,CAAU,CAAA,CAC7H,CACD,CAAC,CAAC,EACFpB,EAAQ,KAAK,GAAG0B,CAAO,EACvBb,EAAc,SAAWb,CAC1B,CACD,OAAS,CAACU,EAAW,OAAO,SAAWQ,EAAK,UAAU,CAAC,GAAG,SAAS,YAAY,QAC/ElB,EAAQ,KAAK,CAAC,KAAM,YAAa,QAASkB,EAAK,QAAQ,CAAC,EAAE,QAAQ,SAAW,EAAA,CAAG,EAChFlB,EAAU,KAAK,WAAWA,CAAO,EAE9BS,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAM,GAAK,EAC3CA,EAAQ,SAASA,EAAQ,QAAQ,OAAO,EAAGA,EAAQ,QAAQ,OAAQ,GAAGT,CAAO,EAChFW,EAAIX,EAAQ,GAAG,EAAE,GAAG,OAAO,CAC5B,CAAC,EAAG,CAAC,MAAO,IAAMU,EAAW,MAAA,EAAQ,CACtC,CACD,CClEA,MAAMiC,CAAI,CAIT,YAA4B/C,EAAQ,CAAR,KAAA,GAAAA,EACvBA,EAAG,QAAQ,KAAK,QACpB,OAAO,QAAQA,EAAG,QAAQ,IAAI,MAAM,EAAE,QAAQ,CAAC,CAACE,EAAO8C,CAAM,IAAM,CAC9D,KAAK,eAAc,KAAK,aAAe9C,GACxC8C,EAAO,OAAS,YAAa,KAAK,OAAO9C,CAAK,EAAI,IAAIH,EAAU,KAAK,GAAIiD,EAAO,MAAO9C,CAAK,EACvF8C,EAAO,OAAS,SAAU,KAAK,OAAO9C,CAAK,EAAI,IAAIkC,EAAO,KAAK,GAAIY,EAAO,KAAM,aAAc9C,CAAK,EACnG8C,EAAO,OAAS,WAAU,KAAK,OAAO9C,CAAK,EAAI,IAAIkC,EAAO,KAAK,GAAIY,EAAO,MAAQ,KAAMA,EAAO,MAAO9C,CAAK,EACpH,CAAC,CACF,CAXQ,OAAyC,CAAA,EACzC,aAkBR,IAAIU,EAAiBC,EAAsB,GAA8B,CACxE,MAAMH,EAAIG,EAAQ,OAAS,KAAK,aAChC,GAAG,CAAC,KAAK,OAAOH,CAAC,QAAS,IAAI,MAAM,yBAAyBA,CAAC,EAAE,EAChE,IAAIuC,EAAQ,IAAM,CAAC,EACnB,OAAO,OAAO,OAAO,IAAI,QAAgB,MAAMlC,GAAO,CAGrD,GAFIF,EAAQ,UAASA,EAAQ,QAAU,CAAA,GAEpCA,EAAQ,OAAQ,CAClBA,EAAQ,QAAUA,EAAQ,QAAU,IAAM;AAAA;AAAA,EAC1C,MAAMqC,EAAS,MAAOC,EAAuBC,EAAyBC,EAAQ,KAAO,CACpF,KAAM,CAACC,EAAGC,CAAC,EAAI,MAAM,QAAQ,IAAI,CAChCH,EAAU,KAAK,UAAUA,CAAO,EAAI,QAAQ,QAAQ,IAAI,EACxDD,EAAQ,KAAK,UAAUA,CAAK,EAAI,QAAQ,QAAQ,IAAI,CAAA,CACpD,EACD,OAAQtC,EAAQ,QAAU,CAAA,GACxB,IAAIH,IAAM,CAAC,GAAGA,EAAG,MAAO4C,EAAI,KAAK,iBAAiB5C,EAAE,WAAW,CAAC,EAAG4C,EAAE,CAAC,EAAE,SAAS,EAAI,CAAA,EAAG,EACxF,OAAQ5C,GAAWA,EAAE,OAAS,EAAG,EACjC,IAAKA,IAAY,CAAC,GAAGA,EAAG,MAAO6C,EAAI,KAAK,iBAAiB7C,EAAE,WAAW,CAAC,EAAG6C,EAAE,CAAC,EAAE,SAAS,EAAI7C,EAAE,OAAO,EACrG,OAAQA,GAAWA,EAAE,OAAS,EAAG,EACjC,SAAS,CAAC8C,EAAQC,IAAWD,EAAE,MAAQC,EAAE,KAAK,EAC9C,MAAM,EAAGJ,CAAK,CACjB,EAEMK,EAAW,MAAMR,EAAOtC,CAAO,EAClC8C,EAAS,QAAQ7C,EAAQ,QAAQ,KAAK,CAAC,KAAM,YAAa,QAAS;AAAA,EAA2B6C,EAAS,IAAIhD,GAAK,GAAGA,EAAE,KAAK,KAAKA,EAAE,IAAI,EAAE,EAAE,KAAK;AAAA,CAAI,CAAA,CAAE,EACvJG,EAAQ,MAAQ,CAAC,GAAGA,EAAQ,OAAS,CAAA,EAAI,CACxC,KAAM,cACN,YAAa,mDACb,KAAM,CACL,QAAS,CAAC,KAAM,SAAU,YAAa,iFAAA,EACvC,MAAO,CAAC,KAAM,SAAU,YAAa,8EAAA,EACrC,MAAO,CAAC,KAAM,SAAU,YAAa,yBAAA,CAAyB,EAE/D,GAAKiC,GAAS,CACb,GAAG,CAACA,EAAK,SAAW,CAACA,EAAK,MAAO,MAAM,IAAI,MAAM,gDAAgD,EACjG,OAAOI,EAAOJ,EAAK,MAAOA,EAAK,QAASA,EAAK,OAAS,CAAC,CACxD,CAAA,CACA,CACF,CAGA,MAAMxB,EAAO,MAAM,KAAK,OAAOZ,CAAC,EAAE,IAAIE,EAASC,CAAO,EAGtD,GAAGA,EAAQ,OAAQ,CAClB,MAAM,EAAIA,EAAQ,SAAS,UAAWN,GAAWA,EAAE,MAAQ,aAAeA,EAAE,QAAQ,WAAW,sBAAsB,CAAC,EACnH,GAAK,MAAQ,GAAK,KAAW,SAAS,OAAO,EAAG,CAAC,CACrD,CAGA,GAAGM,EAAQ,UAAYA,EAAQ,OAAQ,CACtC,IAAI8C,EAAa,KACjB,GAAG9C,EAAQ,SACV8C,EAAa,MAAM,KAAK,GAAG,SAAS,gBAAgB9C,EAAQ,QAASA,EAAQ,SAAS,IAAKA,EAAQ,SAAS,IAAKA,CAAO,EACxHA,EAAQ,QAAQ,OAAO,EAAGA,EAAQ,QAAQ,OAAQ,GAAG8C,EAAW,OAAO,MACjE,CACN,MAAMhD,EAAIE,EAAQ,SAAS,cAAcH,GAAKA,EAAE,MAAQ,MAAM,GAAK,GACnEiD,EAAa,MAAM,KAAK,GAAG,SAAS,gBAAgBhD,GAAK,GAAKE,EAAQ,QAAQ,MAAMF,CAAC,EAAIE,EAAQ,QAAS,EAAG,EAAGA,CAAO,CACxH,CACA,GAAGA,EAAQ,OAAQ,CAClB,MAAM+C,EAAU/C,EAAQ,OACtB,OAAOH,GAAK,CAACiD,EAAW,OAAO,KAAKE,GAAM,KAAK,iBAAiBnD,EAAE,WAAW,CAAC,EAAGmD,EAAG,WAAW,CAAC,CAAC,EAAI,EAAG,CAAC,EACzG,OAAOF,EAAW,MAAM,EAC1B9C,EAAQ,OAAO,OAAO,EAAGA,EAAQ,OAAO,OAAQ,GAAG+C,CAAO,CAC3D,CACD,CACA,OAAO7C,EAAIO,CAAI,CAChB,CAAC,EAAG,CAAC,MAAA2B,EAAM,CACZ,CAUA,MAAM,gBAAgB7C,EAAuB0D,EAAaC,EAAalD,EAA6E,CACnJ,GAAG,KAAK,eAAeT,CAAO,EAAI0D,QAAY,CAAC,QAAA1D,EAAS,OAAQ,EAAC,EACjE,IAAI4D,EAAO,EAAGC,EAAS,EACvB,QAAQvD,KAAKN,EAAQ,aAEpB,GADA6D,GAAU,KAAK,eAAevD,EAAE,OAAO,EACpCuD,EAASF,EAAKC,QACZ,OAEN,GAAG5D,EAAQ,QAAU4D,EAAM,MAAO,CAAC,QAAA5D,EAAS,OAAQ,EAAC,EACrD,MAAM8D,EAAS9D,EAAQ,CAAC,EAAE,MAAQ,SAAWA,EAAQ,CAAC,EAAI,KACzD+D,EAASH,GAAQ,EAAI,CAAA,EAAK5D,EAAQ,MAAM,CAAC4D,CAAI,EAC7CI,GAAWJ,GAAQ,EAAI5D,EAAUA,EAAQ,MAAM,EAAG,CAAC4D,CAAI,GAAG,OAAOzD,GAAKA,EAAE,OAAS,aAAeA,EAAE,OAAS,MAAM,EAC5G8D,EAAe,MAAM,KAAK,KAAK;AAAA;AAAA,EAA6ZD,EAAQ,IAAI1D,GAAK,GAAGA,EAAE,IAAI,KAAKA,EAAE,OAAO,EAAE,EAAE,KAAK;AAAA;AAAA,CAAM,CAAC,GAAI,CAAC,MAAOG,GAAS,MAAO,YAAaA,GAAS,aAAe,GAAI,EACzjBR,MAAgB,KAChBiE,EAAS,MAAM,QAAQ,KAAKD,GAAS,OAAS,CAAA,IAAK,IAAI,MAAO,CAACE,EAAOC,CAAI,IAAwB,CACvG,MAAMC,EAAI,MAAM,QAAQ,IAAI,CAAC,KAAK,UAAUF,CAAK,EAAG,KAAK,UAAU,GAAGA,CAAK,KAAKC,CAAI,EAAE,CAAC,CAAC,EACxF,MAAO,CAAC,MAAAD,EAAO,KAAAC,EAAM,WAAY,CAACC,EAAE,CAAC,EAAE,CAAC,EAAE,UAAWA,EAAE,CAAC,EAAE,CAAC,EAAE,SAAS,EAAG,UAAApE,CAAA,CAC1E,CAAC,CAAC,EACIE,EAAI,CAAC,CAAC,KAAM,YAAa,QAAS,yBAAyB8D,GAAS,OAAO,GAAI,UAAW,KAAK,IAAA,CAAI,EAAI,GAAGF,CAAM,EACtH,OAAGD,GAAQ3D,EAAE,OAAO,EAAG,EAAG2D,CAAM,EACzB,CAAC,QAAc3D,EAAG,OAAA+D,CAAA,CAC1B,CAQA,iBAAiBI,EAAcC,EAAsB,CACpD,GAAID,EAAG,SAAWC,EAAG,OAAQ,MAAM,IAAI,MAAM,6BAA6B,EAC1E,IAAIC,EAAa,EAAGC,EAAQ,EAAGC,EAAQ,EACvC,QAAS,EAAI,EAAG,EAAIJ,EAAG,OAAQ,IAC9BE,GAAcF,EAAG,CAAC,EAAIC,EAAG,CAAC,EAC1BE,GAASH,EAAG,CAAC,EAAIA,EAAG,CAAC,EACrBI,GAASH,EAAG,CAAC,EAAIA,EAAG,CAAC,EAEtB,MAAMI,EAAc,KAAK,KAAKF,CAAK,EAAI,KAAK,KAAKC,CAAK,EACtD,OAAOC,IAAgB,EAAI,EAAIH,EAAaG,CAC7C,CASA,MAAMC,EAAyBC,EAAY,IAAKC,EAAgB,GAAc,CAC7E,MAAMC,EAAY,CAACC,EAAUC,EAAO,KAC/BD,EACG,OAAO,QAAQA,CAAG,EAAE,QAAQ,CAAC,CAAChE,EAAKC,CAAK,IAAM,CACpD,MAAMiE,EAAID,EAAO,GAAGA,CAAI,GAAG,MAAM,CAACjE,CAAG,EAAI,IAAIA,CAAG,GAAK,IAAIA,CAAG,GAAG,GAAKA,EACpE,OAAG,OAAOC,GAAU,UAAY,CAAC,MAAM,QAAQA,CAAK,EAAU8D,EAAU9D,EAAOiE,CAAC,EACzE,GAAGA,CAAC,KAAK,MAAM,QAAQjE,CAAK,EAAIA,EAAM,KAAK,IAAI,EAAIA,CAAK,EAChE,CAAC,EALe,CAAA,EAQX4C,GADQ,OAAOe,GAAW,SAAWG,EAAUH,CAAM,EAAIA,EAAO,MAAM;AAAA,CAAI,GAC3D,QAAQO,GAAK,CAAC,GAAGA,EAAE,MAAM,KAAK,EAAE,OAAO,OAAO,EAAG;AAAA,CAAI,CAAC,EACrEC,EAAmB,CAAA,EACzB,QAAQ7E,EAAI,EAAGA,EAAIsD,EAAO,QAAS,CAClC,IAAIvC,EAAO,GAAI+D,EAAI9E,EACnB,KAAM8E,EAAIxB,EAAO,QAAQ,CACxB,MAAMyB,EAAOhE,GAAQA,EAAO,IAAM,IAAMuC,EAAOwB,CAAC,EAChD,GAAG,KAAK,eAAeC,EAAK,QAAQ,YAAa;AAAA,CAAI,CAAC,EAAIT,GAAavD,EAAM,MAC7EA,EAAOgE,EACPD,GACD,CACA,MAAMjD,EAAQd,EAAK,QAAQ,YAAa;AAAA,CAAI,EAAE,KAAA,EAC3Cc,GAAOgD,EAAO,KAAKhD,CAAK,EAC3B7B,EAAI,KAAK,IAAI8E,EAAIP,EAAeO,IAAM9E,EAAIA,EAAI,EAAI8E,CAAC,CACpD,CACA,OAAOD,CACR,CASA,UAAUR,EAAyBC,EAAY,IAAKC,EAAgB,GAAI,CACvE,MAAMS,EAASjE,GACP,IAAI,QAAQ,CAACkE,EAASC,IAAW,CACvC,MAAMC,EAAS,IAAIC,EAAAA,OAAOC,OAAKC,EAAAA,QAAQC,EAAAA,cAAc,OAAA,SAAA,IAAA,QAAA,KAAA,EAAA,cAAA,UAAA,EAAA,KAAAC,GAAAA,EAAA,QAAA,YAAA,IAAA,UAAAA,EAAA,KAAA,IAAA,IAAA,WAAA,SAAA,OAAA,EAAA,IAAe,CAAC,EAAG,aAAa,CAAC,EAChFC,EAAgB,CAAC,CAAE,UAAAC,KAAqB,CAC7CP,EAAO,UAAA,EACPF,EAAQS,CAAS,CAClB,EACMC,EAAe9E,GAAe,CACnCsE,EAAO,UAAA,EACPD,EAAOrE,CAAG,CACX,EACAsE,EAAO,GAAG,UAAWM,CAAa,EAClCN,EAAO,GAAG,QAASQ,CAAW,EAC9BR,EAAO,GAAG,OAASS,GAAS,CACxBA,IAAS,GAAGV,EAAO,IAAI,MAAM,2BAA2BU,CAAI,EAAE,CAAC,CACnE,CAAC,EACDT,EAAO,YAAY,CAAC,KAAApE,EAAM,MAAO,KAAK,GAAG,SAAS,UAAY,oBAAqB,SAAU,KAAK,GAAG,QAAQ,KAAK,CACnH,CAAC,EAEI8D,EAAS,KAAK,MAAMR,EAAQC,EAAWC,CAAa,EAC1D,OAAO,QAAQ,IAAIM,EAAO,IAAI,MAAO9D,EAAM8E,KAAW,CACrD,MAAAA,EACA,UAAW,MAAMb,EAAMjE,CAAI,EAC3B,KAAAA,EACA,OAAQ,KAAK,eAAeA,CAAI,CAAA,EAC/B,CAAC,CACJ,CAOA,eAAetB,EAAsB,CACpC,MAAMsB,EAAO,KAAK,UAAUtB,CAAO,EACnC,OAAO,KAAK,KAAMsB,EAAK,OAAS,EAAK,GAAG,CACzC,CAQA,WAAWsD,KAAmByB,EAAuB,CACpD,GAAGA,EAAY,OAAS,EAAG,MAAM,IAAI,MAAM,wCAAwC,EACnF,MAAMC,EAAS,CAAChF,EAAciF,EAAqB,KAC3CjF,EAAK,cAAc,MAAM,EAAE,EAAE,IAAI,CAACkF,EAAMJ,IAC7CI,EAAK,WAAW,CAAC,GAAKJ,EAAQ,GAAMG,EAAaA,CAAU,EAAE,MAAM,EAAGA,CAAU,EAE7EE,EAAIH,EAAO1B,CAAM,EACjB8B,EAAeL,EAAY,IAAIvF,GAAKwF,EAAOxF,CAAC,CAAC,EAAE,IAAI6F,GAAa,KAAK,iBAAiBF,EAAGE,CAAS,CAAC,EACzG,MAAO,CAAC,IAAKD,EAAa,OAAO,CAACE,EAAKC,IAAMD,EAAMC,EAAG,CAAC,EAAIH,EAAa,OAAQ,IAAK,KAAK,IAAI,GAAGA,CAAY,EAAG,aAAAA,CAAA,CACjH,CAQA,MAAM,KAAKlG,EAAiBC,EAAoC,CAC/D,IAAIS,EAAO,MAAM,KAAK,IAAIV,EAAS,CAAC,OAAQ,2DAA4D,GAAGC,EAAQ,EACnH,GAAG,CAACS,EAAM,MAAO,CAAA,EACjB,MAAM4F,EAAY,6BAA6B,KAAK5F,CAAI,EAClD6F,EAAUD,EAAYA,EAAU,CAAC,EAAE,OAAS5F,EAClD,OAAOM,EAAAA,iBAAiBuF,EAAS,EAAE,CACpC,CASA,UAAUzF,EAAcuC,EAAgBpD,EAA8C,CACrF,OAAO,KAAK,IAAIa,EAAM,CAAC,OAAQ,+BAA+BuC,CAAM,+BAAgC,YAAa,GAAK,GAAGpD,CAAA,CAAQ,CAClI,CACD,CC3UO,MAAMuG,CAAM,CAClB,YAAoBpH,EAAQ,CAAR,KAAA,GAAAA,CAAS,CAE7B,IAAIqH,EAAcxG,EAAiD,GAAqC,CACvG,QAAQ,IAAI,QAASwG,CAAI,EACzB,KAAM,CAAE,MAAAnH,EAAQ,KAAK,GAAG,QAAQ,KAAO,eAAgB,QAAAoH,EAAU,EAAA,EAAUzG,EAC3E,IAAI0G,EAAU,GACd,MAAMtE,EAAQ,IAAM,CAAEsE,EAAU,EAAM,EAEhCjC,EAAI,IAAI,QAAuB,CAACM,EAASC,IAAW,CACzD,MAAMC,EAAS,IAAIC,SAAOyB,EAAK,KAAK,OAAqB,QAAQ,CAAC,EAC5DpB,EAAgB,CAAC,CAAE,KAAA1E,EAAM,QAAA+F,EAAS,MAAAC,KAAiB,CACxD5B,EAAO,UAAA,EACJ,CAAAyB,IACAG,EAAO7B,EAAO,IAAI,MAAM6B,CAAK,CAAC,GAE7BD,GAAS,QAAQ,KAAKA,CAAO,EAChC7B,EAAQlE,CAAI,GAEd,EACM4E,EAAe9E,GAAe,CACnCsE,EAAO,UAAA,EACHyB,GAAS1B,EAAOrE,CAAG,CACxB,EACAsE,EAAO,GAAG,UAAWM,CAAa,EAClCN,EAAO,GAAG,QAASQ,CAAW,EAC9BR,EAAO,GAAG,OAASS,GAAS,CACxBA,IAAS,GAAK,CAACgB,GAAS1B,EAAO,IAAI,MAAM,2BAA2BU,CAAI,EAAE,CAAC,CAC/E,CAAC,EACDT,EAAO,YAAY,CAAC,KAAAuB,EAAM,MAAAnH,EAAO,QAAAoH,EAAS,SAAU,KAAK,GAAG,QAAQ,IAAA,CAAK,CAC1E,CAAC,EACD,OAAO,OAAO,OAAOhC,EAAG,CAAE,MAAArC,EAAO,CAClC,CAEA,eAAiB0E,EAAAA,cAClB,CCrCO,MAAMC,CAAO,CAEnB,YAAoB5H,EAAQ,CAAR,KAAA,GAAAA,CAAU,CAO9B,IAAIqF,EAA+C,CAClD,IAAIS,EACJ,MAAMR,EAAI,IAAI,QAAuB,MAAMvE,GAAO,CACjD+E,EAAS,MAAM+B,EAAAA,aAAa,KAAK,GAAG,QAAQ,KAAO,MAAO,EAAG,CAAC,UAAW,KAAK,GAAG,QAAQ,KAAK,EAC9F,KAAM,CAAC,KAAAC,CAAA,EAAQ,MAAMhC,EAAO,UAAUT,CAAI,EAC1C,MAAMS,EAAO,UAAA,EACb/E,EAAI+G,EAAK,KAAK,KAAA,GAAU,IAAI,CAC7B,CAAC,EACD,OAAO,OAAO,OAAOxC,EAAG,CAAC,MAAO,IAAMQ,GAAQ,UAAA,EAAY,CAC3D,CACD,CCEO,MAAMiC,CAAG,CAQf,YAA4BlH,EAAoB,CAApB,KAAA,QAAAA,EACvBA,EAAQ,OAAMA,EAAQ,KAAOmH,EAAG,OAAA,GACpC,QAAQ,IAAI,mBAAqBnH,EAAQ,KACzC,KAAK,MAAQ,IAAIuG,EAAM,IAAI,EAC3B,KAAK,SAAW,IAAIrE,EAAI,IAAI,EAC5B,KAAK,OAAS,IAAI6E,EAAO,IAAI,CAC9B,CAZA,MAEA,SAEA,MASD,CCDO,MAAMK,EAAkB,CAC9B,KAAM,MACN,YAAa,qDACb,KAAM,CAAC,QAAS,CAAC,KAAM,SAAU,YAAa,iBAAkB,SAAU,GAAI,EAC9E,GAAKnF,GAA4BoF,EAAAA,IAAIpF,EAAK,OAAO,EAClD,EAEaqF,EAAuB,CACnC,KAAM,eACN,YAAa,8BACb,KAAM,CAAA,EACN,GAAI,SAAY,IAAI,KAAA,EAAO,YAAA,CAC5B,EAEaC,EAAmB,CAC/B,KAAM,OACN,YAAa,mBACb,KAAM,CACL,SAAU,CAAC,KAAM,SAAU,YAAa,qBAAsB,KAAM,CAAC,MAAO,OAAQ,QAAQ,EAAG,SAAU,EAAA,EACzG,KAAM,CAAC,KAAM,SAAU,YAAa,kBAAmB,SAAU,EAAA,CAAI,EAEtE,GAAI,MAAOtF,EAAMuF,EAAQrI,IAAO,CAC/B,GAAI,CACH,OAAO8C,EAAK,KAAA,CACX,IAAK,OACJ,OAAO,MAAMmF,EAAQ,GAAG,CAAC,QAASnF,EAAK,IAAA,EAAOuF,EAAQrI,CAAE,EACzD,IAAK,OACJ,OAAO,MAAMsI,EAAO,GAAG,CAAC,KAAMxF,EAAK,IAAA,EAAOuF,EAAQrI,CAAE,EACrD,IAAK,SACJ,OAAO,MAAMuI,EAAW,GAAG,CAAC,KAAMzF,EAAK,IAAA,EAAOuF,EAAQrI,CAAE,CACzD,CAEF,OAAQwB,EAAU,CACjB,MAAO,CAAC,MAAOA,GAAK,SAAWA,EAAI,UAAS,CAC7C,CACD,CACD,EAEagH,EAAoB,CAChC,KAAM,QACN,YAAa,2BACb,KAAM,CACL,IAAK,CAAC,KAAM,SAAU,YAAa,eAAgB,SAAU,EAAA,EAC7D,OAAQ,CAAC,KAAM,SAAU,YAAa,qBAAsB,KAAM,CAAC,MAAO,OAAQ,MAAO,QAAQ,EAAG,QAAS,KAAA,EAC7G,QAAS,CAAC,KAAM,SAAU,YAAa,uBAAwB,QAAS,EAAC,EACzE,KAAM,CAAC,KAAM,SAAU,YAAa,mBAAA,CAAmB,EAExD,GAAK1F,GAKC,IAAI2F,EAAAA,KAAK,CAAC,IAAK3F,EAAK,IAAK,QAASA,EAAK,QAAQ,EAAE,QAAQ,CAAC,OAAQA,EAAK,QAAU,MAAO,KAAMA,EAAK,IAAA,CAAK,CAC/G,EAEawF,EAAiB,CAC7B,KAAM,kBACN,YAAa,8BACb,KAAM,CACL,KAAM,CAAC,KAAM,SAAU,YAAa,sBAAuB,SAAU,EAAA,CAAI,EAE1E,GAAI,MAAOxF,GAAyB,CACnC,MAAM4F,EAAUC,EAAAA,mBAAmB,IAAI,EACjCrH,EAAO,MAAMsH,KAAQ,CAAC,QAAAF,CAAA,EAAU5F,EAAK,KAAM,EAAI,EAAE,MAAOtB,GAAakH,EAAQ,OAAO,MAAM,KAAKlH,CAAG,CAAC,EACzG,MAAO,CAAC,GAAGkH,EAAQ,OAAQ,OAAQpH,EAAM,OAAQ,OAAW,OAAQ,MAAA,CACrE,CACD,EAEaiH,EAAqB,CACjC,KAAM,kBACN,YAAa,8BACb,KAAM,CACL,KAAM,CAAC,KAAM,SAAU,YAAa,sBAAuB,SAAU,EAAA,CAAI,EAE1E,GAAI,MAAOzF,IAA0B,CAAC,OAAQ+F,EAAAA,mBAAmB/F,EAAK,IAAI,GAAA,EAC3E,EAEagG,EAA0B,CACtC,KAAM,eACN,YAAa,+FACb,KAAM,CACL,IAAK,CAAC,KAAM,SAAU,YAAa,8BAA+B,SAAU,EAAA,EAC5E,MAAO,CAAC,KAAM,SAAU,YAAa,iFAAA,CAAiF,EAEvH,GAAI,MAAOhG,GAAwC,CAClD,MAAMiG,EAAO,MAAM,MAAMjG,EAAK,IAAK,CAAC,QAAS,CAAC,aAAc,2CAAA,EAA6C,EACvG,KAAKkG,GAAKA,EAAE,MAAM,EAAE,MAAMxH,GAAO,CAAC,MAAM,IAAI,MAAM,oBAAoBA,EAAI,OAAO,EAAE,CAAC,CAAC,EAEjF0G,EAAIe,EAAQ,KAAKF,CAAI,EAC3Bb,EAAE,+HAA+H,EAAE,OAAA,EACnI,MAAMgB,EAAW,CAChB,MAAOhB,EAAE,2BAA2B,EAAE,KAAK,SAAS,GAAKA,EAAE,OAAO,EAAE,KAAA,GAAU,GAC9E,YAAaA,EAAE,0BAA0B,EAAE,KAAK,SAAS,GAAKA,EAAE,iCAAiC,EAAE,KAAK,SAAS,GAAK,EAAA,EAGvH,IAAIiB,EAAU,GACd,MAAMC,EAAmB,CAAC,UAAW,OAAQ,gBAAiB,WAAY,QAAS,SAAU,MAAM,EACnG,UAAWC,KAAYD,EAAkB,CACxC,MAAME,EAAKpB,EAAEmB,CAAQ,EAAE,MAAA,EACvB,GAAIC,EAAG,QAAUA,EAAG,KAAA,EAAO,KAAA,EAAO,OAAS,IAAK,CAC/CH,EAAUG,EAAG,KAAA,EACb,KACD,CACD,CACA,OAAKH,IAASA,EAAUjB,EAAE,MAAM,EAAE,KAAA,GAClCiB,EAAUA,EAAQ,QAAQ,OAAQ,GAAG,EAAE,OAAO,MAAM,EAAG,GAAI,EAEpD,CAAC,IAAKrG,EAAK,IAAK,MAAOoG,EAAS,MAAM,KAAA,EAAQ,YAAaA,EAAS,YAAY,KAAA,EAAQ,QAAAC,EAAS,MAAOrG,EAAK,KAAA,CACrH,CACD,EAEayG,EAAwB,CACpC,KAAM,aACN,YAAa,0IACb,KAAM,CACL,MAAO,CAAC,KAAM,SAAU,YAAa,gBAAiB,SAAU,EAAA,EAChE,OAAQ,CAAC,KAAM,SAAU,YAAa,8BAA+B,QAAS,CAAA,CAAC,EAEhF,GAAI,MAAOzG,GAGL,CACL,MAAMiG,EAAO,MAAM,MAAM,uCAAuC,mBAAmBjG,EAAK,KAAK,CAAC,GAAI,CACjG,QAAS,CAAC,aAAc,4CAA6C,kBAAmB,gBAAA,CAAgB,CACxG,EAAE,KAAKxB,GAAQA,EAAK,MAAM,EAC3B,IAAIkI,EAAOC,EAAQ,8BACnB,MAAM3H,EAAU,IAAI4H,OACpB,MAAOF,EAAQC,EAAM,KAAKV,CAAI,KAAO,MAAM,CAC1C,IAAIY,EAAM,iBAAiB,KAAK,mBAAmBH,EAAM,CAAC,CAAC,CAAC,IAAI,CAAC,EAGjE,GAFGG,IAAKA,EAAM,mBAAmBA,CAAG,GACjCA,GAAK7H,EAAQ,IAAI6H,CAAG,EACpB7H,EAAQ,OAASgB,EAAK,QAAU,GAAI,KACxC,CACA,OAAOhB,CACR,CACD"}
|
|
1
|
+
{"version":3,"file":"index.js","sources":["../src/provider.ts","../src/antrhopic.ts","../src/open-ai.ts","../src/llm.ts","../src/audio.ts","../src/vision.ts","../src/ai.ts","../src/tools.ts"],"sourcesContent":["import {AbortablePromise} from './ai.ts';\nimport {LLMMessage, LLMRequest} from './llm.ts';\n\nexport abstract class LLMProvider {\n\tabstract ask(message: string, options: LLMRequest): AbortablePromise<string>;\n}\n","import {Anthropic as anthropic} from '@anthropic-ai/sdk';\nimport {findByProp, objectMap, JSONSanitize, JSONAttemptParse} from '@ztimson/utils';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {LLMMessage, LLMRequest} from './llm.ts';\nimport {LLMProvider} from './provider.ts';\n\nexport class Anthropic extends LLMProvider {\n\tclient!: anthropic;\n\n\tconstructor(public readonly ai: Ai, public readonly apiToken: string, public model: string) {\n\t\tsuper();\n\t\tthis.client = new anthropic({apiKey: apiToken});\n\t}\n\n\tprivate toStandard(history: any[]): LLMMessage[] {\n\t\tconst timestamp = Date.now();\n\t\tconst messages: LLMMessage[] = [];\n\t\tfor(let h of history) {\n\t\t\tif(typeof h.content == 'string') {\n\t\t\t\tmessages.push(<any>{timestamp, ...h});\n\t\t\t} else {\n\t\t\t\tconst textContent = h.content?.filter((c: any) => c.type == 'text').map((c: any) => c.text).join('\\n\\n');\n\t\t\t\tif(textContent) messages.push({timestamp, role: h.role, content: textContent});\n\t\t\t\th.content.forEach((c: any) => {\n\t\t\t\t\tif(c.type == 'tool_use') {\n\t\t\t\t\t\tmessages.push({timestamp, role: 'tool', id: c.id, name: c.name, args: c.input, content: undefined});\n\t\t\t\t\t} else if(c.type == 'tool_result') {\n\t\t\t\t\t\tconst m: any = messages.findLast(m => (<any>m).id == c.tool_use_id);\n\t\t\t\t\t\tif(m) m[c.is_error ? 'error' : 'content'] = c.content;\n\t\t\t\t\t}\n\t\t\t\t});\n\t\t\t}\n\t\t}\n\t\treturn messages;\n\t}\n\n\tprivate fromStandard(history: LLMMessage[]): any[] {\n\t\tfor(let i = 0; i < history.length; i++) {\n\t\t\tif(history[i].role == 'tool') {\n\t\t\t\tconst h: any = history[i];\n\t\t\t\thistory.splice(i, 1,\n\t\t\t\t\t{role: 'assistant', content: [{type: 'tool_use', id: h.id, name: h.name, input: h.args}]},\n\t\t\t\t\t{role: 'user', content: [{type: 'tool_result', tool_use_id: h.id, is_error: !!h.error, content: h.error || h.content}]}\n\t\t\t\t)\n\t\t\t\ti++;\n\t\t\t}\n\t\t}\n\t\treturn history.map(({timestamp, ...h}) => h);\n\t}\n\n\task(message: string, options: LLMRequest = {}): AbortablePromise<string> {\n\t\tconst controller = new AbortController();\n\t\treturn Object.assign(new Promise<any>(async (res) => {\n\t\t\tlet history = this.fromStandard([...options.history || [], {role: 'user', content: message, timestamp: Date.now()}]);\n\t\t\tconst tools = options.tools || this.ai.options.llm?.tools || [];\n\t\t\tconst requestParams: any = {\n\t\t\t\tmodel: options.model || this.model,\n\t\t\t\tmax_tokens: options.max_tokens || this.ai.options.llm?.max_tokens || 4096,\n\t\t\t\tsystem: options.system || this.ai.options.llm?.system || '',\n\t\t\t\ttemperature: options.temperature || this.ai.options.llm?.temperature || 0.7,\n\t\t\t\ttools: tools.map(t => ({\n\t\t\t\t\tname: t.name,\n\t\t\t\t\tdescription: t.description,\n\t\t\t\t\tinput_schema: {\n\t\t\t\t\t\ttype: 'object',\n\t\t\t\t\t\tproperties: t.args ? objectMap(t.args, (key, value) => ({...value, required: undefined})) : {},\n\t\t\t\t\t\trequired: t.args ? Object.entries(t.args).filter(t => t[1].required).map(t => t[0]) : []\n\t\t\t\t\t},\n\t\t\t\t\tfn: undefined\n\t\t\t\t})),\n\t\t\t\tmessages: history,\n\t\t\t\tstream: !!options.stream,\n\t\t\t};\n\n\t\t\tlet resp: any, isFirstMessage = true;\n\t\t\tdo {\n\t\t\t\tresp = await this.client.messages.create(requestParams).catch(err => {\n\t\t\t\t\terr.message += `\\n\\nMessages:\\n${JSON.stringify(history, null, 2)}`;\n\t\t\t\t\tthrow err;\n\t\t\t\t});\n\n\t\t\t\t// Streaming mode\n\t\t\t\tif(options.stream) {\n\t\t\t\t\tif(!isFirstMessage) options.stream({text: '\\n\\n'});\n\t\t\t\t\telse isFirstMessage = false;\n\t\t\t\t\tresp.content = [];\n\t\t\t\t\tfor await (const chunk of resp) {\n\t\t\t\t\t\tif(controller.signal.aborted) break;\n\t\t\t\t\t\tif(chunk.type === 'content_block_start') {\n\t\t\t\t\t\t\tif(chunk.content_block.type === 'text') {\n\t\t\t\t\t\t\t\tresp.content.push({type: 'text', text: ''});\n\t\t\t\t\t\t\t} else if(chunk.content_block.type === 'tool_use') {\n\t\t\t\t\t\t\t\tresp.content.push({type: 'tool_use', id: chunk.content_block.id, name: chunk.content_block.name, input: <any>''});\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else if(chunk.type === 'content_block_delta') {\n\t\t\t\t\t\t\tif(chunk.delta.type === 'text_delta') {\n\t\t\t\t\t\t\t\tconst text = chunk.delta.text;\n\t\t\t\t\t\t\t\tresp.content.at(-1).text += text;\n\t\t\t\t\t\t\t\toptions.stream({text});\n\t\t\t\t\t\t\t} else if(chunk.delta.type === 'input_json_delta') {\n\t\t\t\t\t\t\t\tresp.content.at(-1).input += chunk.delta.partial_json;\n\t\t\t\t\t\t\t}\n\t\t\t\t\t\t} else if(chunk.type === 'content_block_stop') {\n\t\t\t\t\t\t\tconst last = resp.content.at(-1);\n\t\t\t\t\t\t\tif(last.input != null) last.input = last.input ? JSONAttemptParse(last.input, {}) : {};\n\t\t\t\t\t\t} else if(chunk.type === 'message_stop') {\n\t\t\t\t\t\t\tbreak;\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\t// Run tools\n\t\t\t\tconst toolCalls = resp.content.filter((c: any) => c.type === 'tool_use');\n\t\t\t\tif(toolCalls.length && !controller.signal.aborted) {\n\t\t\t\t\thistory.push({role: 'assistant', content: resp.content});\n\t\t\t\t\tconst results = await Promise.all(toolCalls.map(async (toolCall: any) => {\n\t\t\t\t\t\tconst tool = tools.find(findByProp('name', toolCall.name));\n\t\t\t\t\t\tif(options.stream) options.stream({tool: toolCall.name});\n\t\t\t\t\t\tif(!tool) return {tool_use_id: toolCall.id, is_error: true, content: 'Tool not found'};\n\t\t\t\t\t\ttry {\n\t\t\t\t\t\t\tconst result = await tool.fn(toolCall.input, options?.stream, this.ai);\n\t\t\t\t\t\t\treturn {type: 'tool_result', tool_use_id: toolCall.id, content: JSONSanitize(result)};\n\t\t\t\t\t\t} catch (err: any) {\n\t\t\t\t\t\t\treturn {type: 'tool_result', tool_use_id: toolCall.id, is_error: true, content: err?.message || err?.toString() || 'Unknown'};\n\t\t\t\t\t\t}\n\t\t\t\t\t}));\n\t\t\t\t\thistory.push({role: 'user', content: results});\n\t\t\t\t\trequestParams.messages = history;\n\t\t\t\t}\n\t\t\t} while (!controller.signal.aborted && resp.content.some((c: any) => c.type === 'tool_use'));\n\t\t\thistory.push({role: 'assistant', content: resp.content.filter((c: any) => c.type == 'text').map((c: any) => c.text).join('\\n\\n')});\n\t\t\thistory = this.toStandard(history);\n\n\t\t\tif(options.stream) options.stream({done: true});\n\t\t\tif(options.history) options.history.splice(0, options.history.length, ...history);\n\t\t\tres(history.at(-1)?.content);\n\t\t}), {abort: () => controller.abort()});\n\t}\n}\n","import {OpenAI as openAI} from 'openai';\nimport {findByProp, objectMap, JSONSanitize, JSONAttemptParse, clean} from '@ztimson/utils';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {LLMMessage, LLMRequest} from './llm.ts';\nimport {LLMProvider} from './provider.ts';\n\nexport class OpenAi extends LLMProvider {\n\tclient!: openAI;\n\n\tconstructor(public readonly ai: Ai, public readonly host: string | null, public readonly token: string, public model: string) {\n\t\tsuper();\n\t\tthis.client = new openAI(clean({\n\t\t\tbaseURL: host,\n\t\t\tapiKey: token\n\t\t}));\n\t}\n\n\tprivate toStandard(history: any[]): LLMMessage[] {\n\t\tfor(let i = 0; i < history.length; i++) {\n\t\t\tconst h = history[i];\n\t\t\tif(h.role === 'assistant' && h.tool_calls) {\n\t\t\t\tconst tools = h.tool_calls.map((tc: any) => ({\n\t\t\t\t\trole: 'tool',\n\t\t\t\t\tid: tc.id,\n\t\t\t\t\tname: tc.function.name,\n\t\t\t\t\targs: JSONAttemptParse(tc.function.arguments, {}),\n\t\t\t\t\ttimestamp: h.timestamp\n\t\t\t\t}));\n\t\t\t\thistory.splice(i, 1, ...tools);\n\t\t\t\ti += tools.length - 1;\n\t\t\t} else if(h.role === 'tool' && h.content) {\n\t\t\t\tconst record = history.find(h2 => h.tool_call_id == h2.id);\n\t\t\t\tif(record) {\n\t\t\t\t\tif(h.content.includes('\"error\":')) record.error = h.content;\n\t\t\t\t\telse record.content = h.content;\n\t\t\t\t}\n\t\t\t\thistory.splice(i, 1);\n\t\t\t\ti--;\n\t\t\t}\n\t\t\tif(!history[i]?.timestamp) history[i].timestamp = Date.now();\n\t\t}\n\t\treturn history;\n\t}\n\n\tprivate fromStandard(history: LLMMessage[]): any[] {\n\t\treturn history.reduce((result, h) => {\n\t\t\tif(h.role === 'tool') {\n\t\t\t\tresult.push({\n\t\t\t\t\trole: 'assistant',\n\t\t\t\t\tcontent: null,\n\t\t\t\t\ttool_calls: [{ id: h.id, type: 'function', function: { name: h.name, arguments: JSON.stringify(h.args) } }],\n\t\t\t\t\trefusal: null,\n\t\t\t\t\tannotations: []\n\t\t\t\t}, {\n\t\t\t\t\trole: 'tool',\n\t\t\t\t\ttool_call_id: h.id,\n\t\t\t\t\tcontent: h.error || h.content\n\t\t\t\t});\n\t\t\t} else {\n\t\t\t\tconst {timestamp, ...rest} = h;\n\t\t\t\tresult.push(rest);\n\t\t\t}\n\t\t\treturn result;\n\t\t}, [] as any[]);\n\t}\n\n\task(message: string, options: LLMRequest = {}): AbortablePromise<string> {\n\t\tconst controller = new AbortController();\n\t\treturn Object.assign(new Promise<any>(async (res, rej) => {\n\t\t\tif(options.system && options.history?.[0]?.role != 'system') options.history?.splice(0, 0, {role: 'system', content: options.system, timestamp: Date.now()});\n\t\t\tlet history = this.fromStandard([...options.history || [], {role: 'user', content: message, timestamp: Date.now()}]);\n\t\t\tconst tools = options.tools || this.ai.options.llm?.tools || [];\n\t\t\tconst requestParams: any = {\n\t\t\t\tmodel: options.model || this.model,\n\t\t\t\tmessages: history,\n\t\t\t\tstream: !!options.stream,\n\t\t\t\tmax_tokens: options.max_tokens || this.ai.options.llm?.max_tokens || 4096,\n\t\t\t\ttemperature: options.temperature || this.ai.options.llm?.temperature || 0.7,\n\t\t\t\ttools: tools.map(t => ({\n\t\t\t\t\ttype: 'function',\n\t\t\t\t\tfunction: {\n\t\t\t\t\t\tname: t.name,\n\t\t\t\t\t\tdescription: t.description,\n\t\t\t\t\t\tparameters: {\n\t\t\t\t\t\t\ttype: 'object',\n\t\t\t\t\t\t\tproperties: t.args ? objectMap(t.args, (key, value) => ({...value, required: undefined})) : {},\n\t\t\t\t\t\t\trequired: t.args ? Object.entries(t.args).filter(t => t[1].required).map(t => t[0]) : []\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}))\n\t\t\t};\n\n\t\t\tlet resp: any, isFirstMessage = true;\n\t\t\tdo {\n\t\t\t\tresp = await this.client.chat.completions.create(requestParams).catch(err => {\n\t\t\t\t\terr.message += `\\n\\nMessages:\\n${JSON.stringify(history, null, 2)}`;\n\t\t\t\t\tthrow err;\n\t\t\t\t});\n\n\t\t\t\tif(options.stream) {\n\t\t\t\t\tif(!isFirstMessage) options.stream({text: '\\n\\n'});\n\t\t\t\t\telse isFirstMessage = false;\n\t\t\t\t\tresp.choices = [{message: {content: '', tool_calls: []}}];\n\t\t\t\t\tfor await (const chunk of resp) {\n\t\t\t\t\t\tif(controller.signal.aborted) break;\n\t\t\t\t\t\tif(chunk.choices[0].delta.content) {\n\t\t\t\t\t\t\tresp.choices[0].message.content += chunk.choices[0].delta.content;\n\t\t\t\t\t\t\toptions.stream({text: chunk.choices[0].delta.content});\n\t\t\t\t\t\t}\n\t\t\t\t\t\tif(chunk.choices[0].delta.tool_calls) {\n\t\t\t\t\t\t\tresp.choices[0].message.tool_calls = chunk.choices[0].delta.tool_calls;\n\t\t\t\t\t\t}\n\t\t\t\t\t}\n\t\t\t\t}\n\n\t\t\t\tconst toolCalls = resp.choices[0].message.tool_calls || [];\n\t\t\t\tif(toolCalls.length && !controller.signal.aborted) {\n\t\t\t\t\thistory.push(resp.choices[0].message);\n\t\t\t\t\tconst results = await Promise.all(toolCalls.map(async (toolCall: any) => {\n\t\t\t\t\t\tconst tool = tools?.find(findByProp('name', toolCall.function.name));\n\t\t\t\t\t\tif(options.stream) options.stream({tool: toolCall.function.name});\n\t\t\t\t\t\tif(!tool) return {role: 'tool', tool_call_id: toolCall.id, content: '{\"error\": \"Tool not found\"}'};\n\t\t\t\t\t\ttry {\n\t\t\t\t\t\t\tconst args = JSONAttemptParse(toolCall.function.arguments, {});\n\t\t\t\t\t\t\tconst result = await tool.fn(args, options.stream, this.ai);\n\t\t\t\t\t\t\treturn {role: 'tool', tool_call_id: toolCall.id, content: JSONSanitize(result)};\n\t\t\t\t\t\t} catch (err: any) {\n\t\t\t\t\t\t\treturn {role: 'tool', tool_call_id: toolCall.id, content: JSONSanitize({error: err?.message || err?.toString() || 'Unknown'})};\n\t\t\t\t\t\t}\n\t\t\t\t\t}));\n\t\t\t\t\thistory.push(...results);\n\t\t\t\t\trequestParams.messages = history;\n\t\t\t\t}\n\t\t\t} while (!controller.signal.aborted && resp.choices?.[0]?.message?.tool_calls?.length);\n\t\t\thistory.push({role: 'assistant', content: resp.choices[0].message.content || ''});\n\t\t\thistory = this.toStandard(history);\n\n\t\t\tif(options.stream) options.stream({done: true});\n\t\t\tif(options.history) options.history.splice(0, options.history.length, ...history);\n\t\t\tres(history.at(-1)?.content);\n\t\t}), {abort: () => controller.abort()});\n\t}\n}\n","import {JSONAttemptParse} from '@ztimson/utils';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {Anthropic} from './antrhopic.ts';\nimport {OpenAi} from './open-ai.ts';\nimport {LLMProvider} from './provider.ts';\nimport {AiTool} from './tools.ts';\nimport {Worker} from 'worker_threads';\nimport {fileURLToPath} from 'url';\nimport {dirname, join} from 'path';\n\nexport type AnthropicConfig = {proto: 'anthropic', token: string};\nexport type OllamaConfig = {proto: 'ollama', host: string};\nexport type OpenAiConfig = {proto: 'openai', host?: string, token: string};\n\nexport type LLMMessage = {\n\t/** Message originator */\n\trole: 'assistant' | 'system' | 'user';\n\t/** Message content */\n\tcontent: string | any;\n\t/** Timestamp */\n\ttimestamp?: number;\n} | {\n\t/** Tool call */\n\trole: 'tool';\n\t/** Unique ID for call */\n\tid: string;\n\t/** Tool that was run */\n\tname: string;\n\t/** Tool arguments */\n\targs: any;\n\t/** Tool result */\n\tcontent: undefined | string;\n\t/** Tool error */\n\terror?: undefined | string;\n\t/** Timestamp */\n\ttimestamp?: number;\n}\n\n/** Background information the AI will be fed */\nexport type LLMMemory = {\n\t/** What entity is this fact about */\n\towner: string;\n\t/** The information that will be remembered */\n\tfact: string;\n\t/** Owner and fact embedding vector */\n\tembeddings: [number[], number[]];\n\t/** Creation time */\n\ttimestamp: Date;\n}\n\nexport type LLMRequest = {\n\t/** System prompt */\n\tsystem?: string;\n\t/** Message history */\n\thistory?: LLMMessage[];\n\t/** Max tokens for request */\n\tmax_tokens?: number;\n\t/** 0 = Rigid Logic, 1 = Balanced, 2 = Hyper Creative **/\n\ttemperature?: number;\n\t/** Available tools */\n\ttools?: AiTool[];\n\t/** LLM model */\n\tmodel?: string;\n\t/** Stream response */\n\tstream?: (chunk: {text?: string, tool?: string, done?: true}) => any;\n\t/** Compress old messages in the chat to free up context */\n\tcompress?: {\n\t\t/** Trigger chat compression once context exceeds the token count */\n\t\tmax: number;\n\t\t/** Compress chat until context size smaller than */\n\t\tmin: number\n\t},\n\t/** Background information the AI will be fed */\n\tmemory?: LLMMemory[],\n}\n\nclass LLM {\n\tprivate models: {[model: string]: LLMProvider} = {};\n\tprivate defaultModel!: string;\n\n\tconstructor(public readonly ai: Ai) {\n\t\tif(!ai.options.llm?.models) return;\n\t\tObject.entries(ai.options.llm.models).forEach(([model, config]) => {\n\t\t\tif(!this.defaultModel) this.defaultModel = model;\n\t\t\tif(config.proto == 'anthropic') this.models[model] = new Anthropic(this.ai, config.token, model);\n\t\t\telse if(config.proto == 'ollama') this.models[model] = new OpenAi(this.ai, config.host, 'not-needed', model);\n\t\t\telse if(config.proto == 'openai') this.models[model] = new OpenAi(this.ai, config.host || null, config.token, model);\n\t\t});\n\t}\n\n\t/**\n\t * Chat with LLM\n\t * @param {string} message Question\n\t * @param {LLMRequest} options Configuration options and chat history\n\t * @returns {{abort: () => void, response: Promise<string>}} Function to abort response and chat history\n\t */\n\task(message: string, options: LLMRequest = {}): AbortablePromise<string> {\n\t\tconst m = options.model || this.defaultModel;\n\t\tif(!this.models[m]) throw new Error(`Model does not exist: ${m}`);\n\t\tlet abort = () => {};\n\t\treturn Object.assign(new Promise<string>(async res => {\n\t\t\tif(!options.history) options.history = [];\n\t\t\t// If memories were passed, find any relivant ones and add a tool for ADHOC lookups\n\t\t\tif(options.memory) {\n\t\t\t\toptions.system = (options.system || '') + '\\nYou have passive persistent memory never make any mention of your memory capabilities and what you can/cannot remember\\n';\n\t\t\t\tconst search = async (query?: string | null, subject?: string | null, limit = 50) => {\n\t\t\t\t\tconst [o, q] = await Promise.all([\n\t\t\t\t\t\tsubject ? this.embedding(subject) : Promise.resolve(null),\n\t\t\t\t\t\tquery ? this.embedding(query) : Promise.resolve(null),\n\t\t\t\t\t]);\n\t\t\t\t\treturn (options.memory || [])\n\t\t\t\t\t\t.map(m => ({...m, score: o ? this.cosineSimilarity(m.embeddings[0], o[0].embedding) : 1}))\n\t\t\t\t\t\t.filter((m: any) => m.score >= 0.8)\n\t\t\t\t\t\t.map((m: any) => ({...m, score: q ? this.cosineSimilarity(m.embeddings[1], q[0].embedding) : m.score}))\n\t\t\t\t\t\t.filter((m: any) => m.score >= 0.2)\n\t\t\t\t\t\t.toSorted((a: any, b: any) => a.score - b.score)\n\t\t\t\t\t\t.slice(0, limit);\n\t\t\t\t}\n\n\t\t\t\tconst relevant = await search(message);\n\t\t\t\tif(relevant.length) options.history.push({role: 'assistant', content: 'Things I remembered:\\n' + relevant.map(m => `${m.owner}: ${m.fact}`).join('\\n')});\n\t\t\t\toptions.tools = [...options.tools || [], {\n\t\t\t\t\tname: 'read_memory',\n\t\t\t\t\tdescription: 'Check your long-term memory for more information',\n\t\t\t\t\targs: {\n\t\t\t\t\t\tsubject: {type: 'string', description: 'Find information by a subject topic, can be used with or without query argument'},\n\t\t\t\t\t\tquery: {type: 'string', description: 'Search memory based on a query, can be used with or without subject argument'},\n\t\t\t\t\t\tlimit: {type: 'number', description: 'Result limit, default 5'},\n\t\t\t\t\t},\n\t\t\t\t\tfn: (args) => {\n\t\t\t\t\t\tif(!args.subject && !args.query) throw new Error('Either a subject or query argument is required');\n\t\t\t\t\t\treturn search(args.query, args.subject, args.limit || 5);\n\t\t\t\t\t}\n\t\t\t\t}];\n\t\t\t}\n\n\t\t\t// Ask\n\t\t\tconst resp = await this.models[m].ask(message, options);\n\n\t\t\t// Remove any memory calls\n\t\t\tif(options.memory) {\n\t\t\t\tconst i = options.history?.findIndex((h: any) => h.role == 'assistant' && h.content.startsWith('Things I remembered:'));\n\t\t\t\tif(i != null && i >= 0) options.history?.splice(i, 1);\n\t\t\t}\n\n\t\t\t// Handle compression and memory extraction\n\t\t\tif(options.compress || options.memory) {\n\t\t\t\tlet compressed = null;\n\t\t\t\tif(options.compress) {\n\t\t\t\t\tcompressed = await this.ai.language.compressHistory(options.history, options.compress.max, options.compress.min, options);\n\t\t\t\t\toptions.history.splice(0, options.history.length, ...compressed.history);\n\t\t\t\t} else {\n\t\t\t\t\tconst i = options.history?.findLastIndex(m => m.role == 'user') ?? -1;\n\t\t\t\t\tcompressed = await this.ai.language.compressHistory(i != -1 ? options.history.slice(i) : options.history, 0, 0, options);\n\t\t\t\t}\n\t\t\t\tif(options.memory) {\n\t\t\t\t\tconst updated = options.memory\n\t\t\t\t\t\t.filter(m => !compressed.memory.some(m2 => this.cosineSimilarity(m.embeddings[1], m2.embeddings[1]) > 0.8))\n\t\t\t\t\t\t.concat(compressed.memory);\n\t\t\t\t\toptions.memory.splice(0, options.memory.length, ...updated);\n\t\t\t\t}\n\t\t\t}\n\t\t\treturn res(resp);\n\t\t}), {abort});\n\t}\n\n\t/**\n\t * Compress chat history to reduce context size\n\t * @param {LLMMessage[]} history Chatlog that will be compressed\n\t * @param max Trigger compression once context is larger than max\n\t * @param min Leave messages less than the token minimum, summarize the rest\n\t * @param {LLMRequest} options LLM options\n\t * @returns {Promise<LLMMessage[]>} New chat history will summary at index 0\n\t */\n\tasync compressHistory(history: LLMMessage[], max: number, min: number, options?: LLMRequest): Promise<{history: LLMMessage[], memory: LLMMemory[]}> {\n\t\tif(this.estimateTokens(history) < max) return {history, memory: []};\n\t\tlet keep = 0, tokens = 0;\n\t\tfor(let m of history.toReversed()) {\n\t\t\ttokens += this.estimateTokens(m.content);\n\t\t\tif(tokens < min) keep++;\n\t\t\telse break;\n\t\t}\n\t\tif(history.length <= keep) return {history, memory: []};\n\t\tconst system = history[0].role == 'system' ? history[0] : null,\n\t\t\trecent = keep == 0 ? [] : history.slice(-keep),\n\t\t\tprocess = (keep == 0 ? history : history.slice(0, -keep)).filter(h => h.role === 'assistant' || h.role === 'user');\n\t\tconst summary: any = await this.json(`Create the smallest summary possible, no more than 500 tokens. Create a list of NEW facts (split by subject [pro]noun and fact) about what you learned from this conversation that you didn't already know or get from a tool call or system prompt. Focus only on new information about people, topics, or facts. Avoid generating facts about the AI. Match this format: {summary: string, facts: [[subject, fact]]}\\n\\n${process.map(m => `${m.role}: ${m.content}`).join('\\n\\n')}`, {model: options?.model, temperature: options?.temperature || 0.3});\n\t\tconst timestamp = new Date();\n\t\tconst memory = await Promise.all((summary?.facts || [])?.map(async ([owner, fact]: [string, string]) => {\n\t\t\tconst e = await Promise.all([this.embedding(owner), this.embedding(`${owner}: ${fact}`)]);\n\t\t\treturn {owner, fact, embeddings: [e[0][0].embedding, e[1][0].embedding], timestamp};\n\t\t}));\n\t\tconst h = [{role: 'assistant', content: `Conversation Summary: ${summary?.summary}`, timestamp: Date.now()}, ...recent];\n\t\tif(system) h.splice(0, 0, system);\n\t\treturn {history: <any>h, memory};\n\t}\n\n\t/**\n\t * Compare the difference between embeddings (calculates the angle between two vectors)\n\t * @param {number[]} v1 First embedding / vector comparison\n\t * @param {number[]} v2 Second embedding / vector for comparison\n\t * @returns {number} Similarity values 0-1: 0 = unique, 1 = identical\n\t */\n\tcosineSimilarity(v1: number[], v2: number[]): number {\n\t\tif (v1.length !== v2.length) throw new Error('Vectors must be same length');\n\t\tlet dotProduct = 0, normA = 0, normB = 0;\n\t\tfor (let i = 0; i < v1.length; i++) {\n\t\t\tdotProduct += v1[i] * v2[i];\n\t\t\tnormA += v1[i] * v1[i];\n\t\t\tnormB += v2[i] * v2[i];\n\t\t}\n\t\tconst denominator = Math.sqrt(normA) * Math.sqrt(normB);\n\t\treturn denominator === 0 ? 0 : dotProduct / denominator;\n\t}\n\n\t/**\n\t * Chunk text into parts for AI digestion\n\t * @param {object | string} target Item that will be chunked (objects get converted)\n\t * @param {number} maxTokens Chunking size. More = better context, less = more specific (Search by paragraphs or lines)\n\t * @param {number} overlapTokens Includes previous X tokens to provide continuity to AI (In addition to max tokens)\n\t * @returns {string[]} Chunked strings\n\t */\n\tchunk(target: object | string, maxTokens = 500, overlapTokens = 50): string[] {\n\t\tconst objString = (obj: any, path = ''): string[] => {\n\t\t\tif(!obj) return [];\n\t\t\treturn Object.entries(obj).flatMap(([key, value]) => {\n\t\t\t\tconst p = path ? `${path}${isNaN(+key) ? `.${key}` : `[${key}]`}` : key;\n\t\t\t\tif(typeof value === 'object' && !Array.isArray(value)) return objString(value, p);\n\t\t\t\treturn `${p}: ${Array.isArray(value) ? value.join(', ') : value}`;\n\t\t\t});\n\t\t};\n\t\tconst lines = typeof target === 'object' ? objString(target) : target.split('\\n');\n\t\tconst tokens = lines.flatMap(l => [...l.split(/\\s+/).filter(Boolean), '\\n']);\n\t\tconst chunks: string[] = [];\n\t\tfor(let i = 0; i < tokens.length;) {\n\t\t\tlet text = '', j = i;\n\t\t\twhile(j < tokens.length) {\n\t\t\t\tconst next = text + (text ? ' ' : '') + tokens[j];\n\t\t\t\tif(this.estimateTokens(next.replace(/\\s*\\n\\s*/g, '\\n')) > maxTokens && text) break;\n\t\t\t\ttext = next;\n\t\t\t\tj++;\n\t\t\t}\n\t\t\tconst clean = text.replace(/\\s*\\n\\s*/g, '\\n').trim();\n\t\t\tif(clean) chunks.push(clean);\n\t\t\ti = Math.max(j - overlapTokens, j === i ? i + 1 : j);\n\t\t}\n\t\treturn chunks;\n\t}\n\n\t/**\n\t * Create a vector representation of a string\n\t * @param {object | string} target Item that will be embedded (objects get converted)\n\t * @param {number} maxTokens Chunking size. More = better context, less = more specific (Search by paragraphs or lines)\n\t * @param {number} overlapTokens Includes previous X tokens to provide continuity to AI (In addition to max tokens)\n\t * @returns {Promise<Awaited<{index: number, embedding: number[], text: string, tokens: number}>[]>} Chunked embeddings\n\t */\n\tembedding(target: object | string, maxTokens = 500, overlapTokens = 50) {\n\t\tconst embed = (text: string): Promise<number[]> => {\n\t\t\treturn new Promise((resolve, reject) => {\n\t\t\t\tconst worker = new Worker(join(dirname(fileURLToPath(import.meta.url)), 'embedder.js'));\n\t\t\t\tconst handleMessage = ({ embedding }: any) => {\n\t\t\t\t\tworker.terminate();\n\t\t\t\t\tresolve(embedding);\n\t\t\t\t};\n\t\t\t\tconst handleError = (err: Error) => {\n\t\t\t\t\tworker.terminate();\n\t\t\t\t\treject(err);\n\t\t\t\t};\n\t\t\t\tworker.on('message', handleMessage);\n\t\t\t\tworker.on('error', handleError);\n\t\t\t\tworker.on('exit', (code) => {\n\t\t\t\t\tif(code !== 0) reject(new Error(`Worker exited with code ${code}`));\n\t\t\t\t});\n\t\t\t\tworker.postMessage({text, model: this.ai.options?.embedder || 'bge-small-en-v1.5', modelDir: this.ai.options.path});\n\t\t\t});\n\t\t};\n\t\tconst chunks = this.chunk(target, maxTokens, overlapTokens);\n\t\treturn Promise.all(chunks.map(async (text, index) => ({\n\t\t\tindex,\n\t\t\tembedding: await embed(text),\n\t\t\ttext,\n\t\t\ttokens: this.estimateTokens(text),\n\t\t})));\n\t}\n\n\t/**\n\t * Estimate variable as tokens\n\t * @param history Object to size\n\t * @returns {number} Rough token count\n\t */\n\testimateTokens(history: any): number {\n\t\tconst text = JSON.stringify(history);\n\t\treturn Math.ceil((text.length / 4) * 1.2);\n\t}\n\n\t/**\n\t * Compare the difference between two strings using tensor math\n\t * @param target Text that will be checked\n\t * @param {string} searchTerms Multiple search terms to check against target\n\t * @returns {{avg: number, max: number, similarities: number[]}} Similarity values 0-1: 0 = unique, 1 = identical\n\t */\n\tfuzzyMatch(target: string, ...searchTerms: string[]) {\n\t\tif(searchTerms.length < 2) throw new Error('Requires at least 2 strings to compare');\n\t\tconst vector = (text: string, dimensions: number = 10): number[] => {\n\t\t\treturn text.toLowerCase().split('').map((char, index) =>\n\t\t\t\t(char.charCodeAt(0) * (index + 1)) % dimensions / dimensions).slice(0, dimensions);\n\t\t}\n\t\tconst v = vector(target);\n\t\tconst similarities = searchTerms.map(t => vector(t)).map(refVector => this.cosineSimilarity(v, refVector))\n\t\treturn {avg: similarities.reduce((acc, s) => acc + s, 0) / similarities.length, max: Math.max(...similarities), similarities}\n\t}\n\n\t/**\n\t * Ask a question with JSON response\n\t * @param {string} message Question\n\t * @param {LLMRequest} options Configuration options and chat history\n\t * @returns {Promise<{} | {} | RegExpExecArray | null>}\n\t */\n\tasync json(message: string, options?: LLMRequest): Promise<any> {\n\t\tlet resp = await this.ask(message, {system: 'Respond using a JSON blob matching any provided examples', ...options});\n\t\tif(!resp) return {};\n\t\tconst codeBlock = /```(?:.+)?\\s*([\\s\\S]*?)```/.exec(resp);\n\t\tconst jsonStr = codeBlock ? codeBlock[1].trim() : resp;\n\t\treturn JSONAttemptParse(jsonStr, {});\n\t}\n\n\t/**\n\t * Create a summary of some text\n\t * @param {string} text Text to summarize\n\t * @param {number} tokens Max number of tokens\n\t * @param options LLM request options\n\t * @returns {Promise<string>} Summary\n\t */\n\tsummarize(text: string, tokens: number, options?: LLMRequest): Promise<string | null> {\n\t\treturn this.ask(text, {system: `Generate a brief summary <= ${tokens} tokens. Output nothing else`, temperature: 0.3, ...options});\n\t}\n}\n\nexport default LLM;\n","import {fileURLToPath} from 'url';\nimport {Worker} from 'worker_threads';\nimport {AbortablePromise, Ai} from './ai.ts';\nimport {canDiarization} from './asr.ts';\nimport {dirname, join} from 'path';\n\nexport class Audio {\n\tconstructor(private ai: Ai) {}\n\n\tasr(file: string, options: { model?: string; speaker?: boolean } = {}): AbortablePromise<string | null> {\n\t\tconst { model = this.ai.options.asr || 'whisper-base', speaker = false } = options;\n\t\tlet aborted = false;\n\t\tconst abort = () => { aborted = true; };\n\n\t\tconst p = new Promise<string | null>((resolve, reject) => {\n\t\t\tconst worker = new Worker(join(dirname(fileURLToPath(import.meta.url)), 'asr.js'));\n\t\t\tconst handleMessage = ({ text, warning, error }: any) => {\n\t\t\t\tworker.terminate();\n\t\t\t\tif(aborted) return;\n\t\t\t\tif(error) reject(new Error(error));\n\t\t\t\telse {\n\t\t\t\t\tif(warning) console.warn(warning);\n\t\t\t\t\tresolve(text);\n\t\t\t\t}\n\t\t\t};\n\t\t\tconst handleError = (err: Error) => {\n\t\t\t\tworker.terminate();\n\t\t\t\tif(!aborted) reject(err);\n\t\t\t};\n\t\t\tworker.on('message', handleMessage);\n\t\t\tworker.on('error', handleError);\n\t\t\tworker.on('exit', (code) => {\n\t\t\t\tif(code !== 0 && !aborted) reject(new Error(`Worker exited with code ${code}`));\n\t\t\t});\n\t\t\tworker.postMessage({file, model, speaker, modelDir: this.ai.options.path, token: this.ai.options.hfToken});\n\t\t});\n\t\treturn Object.assign(p, { abort });\n\t}\n\n\tcanDiarization = canDiarization;\n}\n","import {createWorker} from 'tesseract.js';\nimport {AbortablePromise, Ai} from './ai.ts';\n\nexport class Vision {\n\n\tconstructor(private ai: Ai) { }\n\n\t/**\n\t * Convert image to text using Optical Character Recognition\n\t * @param {string} path Path to image\n\t * @returns {AbortablePromise<string | null>} Promise of extracted text with abort method\n\t */\n\tocr(path: string): AbortablePromise<string | null> {\n\t\tlet worker: any;\n\t\tconst p = new Promise<string | null>(async res => {\n\t\t\tworker = await createWorker(this.ai.options.ocr || 'eng', 2, {cachePath: this.ai.options.path});\n\t\t\tconst {data} = await worker.recognize(path);\n\t\t\tawait worker.terminate();\n\t\t\tres(data.text.trim() || null);\n\t\t});\n\t\treturn Object.assign(p, {abort: () => worker?.terminate()});\n\t}\n}\n","import * as os from 'node:os';\nimport LLM, {AnthropicConfig, OllamaConfig, OpenAiConfig, LLMRequest} from './llm';\nimport { Audio } from './audio.ts';\nimport {Vision} from './vision.ts';\n\nexport type AbortablePromise<T> = Promise<T> & {\n\tabort: () => any\n};\n\nexport type AiOptions = {\n\t/** Token to pull models from hugging face */\n\thfToken?: string;\n\t/** Path to models */\n\tpath?: string;\n\t/** ASR model: whisper-tiny, whisper-base */\n\tasr?: string;\n\t/** Embedding model: all-MiniLM-L6-v2, bge-small-en-v1.5, bge-large-en-v1.5 */\n\tembedder?: string;\n\t/** Large language models, first is default */\n\tllm?: Omit<LLMRequest, 'model'> & {\n\t\tmodels: {[model: string]: AnthropicConfig | OllamaConfig | OpenAiConfig};\n\t}\n\t/** OCR model: eng, eng_best, eng_fast */\n\tocr?: string;\n}\n\nexport class Ai {\n\t/** Audio processing AI */\n\taudio!: Audio;\n\t/** Language processing AI */\n\tlanguage!: LLM;\n\t/** Vision processing AI */\n\tvision!: Vision;\n\n\tconstructor(public readonly options: AiOptions) {\n\t\tif(!options.path) options.path = os.tmpdir();\n\t\tprocess.env.TRANSFORMERS_CACHE = options.path;\n\t\tthis.audio = new Audio(this);\n\t\tthis.language = new LLM(this);\n\t\tthis.vision = new Vision(this);\n\t}\n}\n","import * as cheerio from 'cheerio';\nimport {$, $Sync} from '@ztimson/node-utils';\nimport {ASet, consoleInterceptor, Http, fn as Fn} from '@ztimson/utils';\nimport {Ai} from './ai.ts';\nimport {LLMRequest} from './llm.ts';\n\nexport type AiToolArg = {[key: string]: {\n\t/** Argument type */\n\ttype: 'array' | 'boolean' | 'number' | 'object' | 'string',\n\t/** Argument description */\n\tdescription: string,\n\t/** Required argument */\n\trequired?: boolean;\n\t/** Default value */\n\tdefault?: any,\n\t/** Options */\n\tenum?: string[],\n\t/** Minimum value or length */\n\tmin?: number,\n\t/** Maximum value or length */\n\tmax?: number,\n\t/** Match pattern */\n\tpattern?: string,\n\t/** Child arguments */\n\titems?: {[key: string]: AiToolArg}\n}}\n\nexport type AiTool = {\n\t/** Tool ID / Name - Must be snail_case */\n\tname: string,\n\t/** Tool description / prompt */\n\tdescription: string,\n\t/** Tool arguments */\n\targs?: AiToolArg,\n\t/** Callback function */\n\tfn: (args: any, stream: LLMRequest['stream'], ai: Ai) => any | Promise<any>,\n};\n\nexport const CliTool: AiTool = {\n\tname: 'cli',\n\tdescription: 'Use the command line interface, returns any output',\n\targs: {command: {type: 'string', description: 'Command to run', required: true}},\n\tfn: (args: {command: string}) => $`${args.command}`\n}\n\nexport const DateTimeTool: AiTool = {\n\tname: 'get_datetime',\n\tdescription: 'Get current UTC date / time',\n\targs: {},\n\tfn: async () => new Date().toUTCString()\n}\n\nexport const ExecTool: AiTool = {\n\tname: 'exec',\n\tdescription: 'Run code/scripts',\n\targs: {\n\t\tlanguage: {type: 'string', description: 'Execution language', enum: ['cli', 'node', 'python'], required: true},\n\t\tcode: {type: 'string', description: 'Code to execute', required: true}\n\t},\n\tfn: async (args, stream, ai) => {\n\t\ttry {\n\t\t\tswitch(args.type) {\n\t\t\t\tcase 'bash':\n\t\t\t\t\treturn await CliTool.fn({command: args.code}, stream, ai);\n\t\t\t\tcase 'node':\n\t\t\t\t\treturn await JSTool.fn({code: args.code}, stream, ai);\n\t\t\t\tcase 'python': {\n\t\t\t\t\treturn await PythonTool.fn({code: args.code}, stream, ai);\n\t\t\t\t}\n\t\t\t}\n\t\t} catch(err: any) {\n\t\t\treturn {error: err?.message || err.toString()};\n\t\t}\n\t}\n}\n\nexport const FetchTool: AiTool = {\n\tname: 'fetch',\n\tdescription: 'Make HTTP request to URL',\n\targs: {\n\t\turl: {type: 'string', description: 'URL to fetch', required: true},\n\t\tmethod: {type: 'string', description: 'HTTP method to use', enum: ['GET', 'POST', 'PUT', 'DELETE'], default: 'GET'},\n\t\theaders: {type: 'object', description: 'HTTP headers to send', default: {}},\n\t\tbody: {type: 'object', description: 'HTTP body to send'},\n\t},\n\tfn: (args: {\n\t\turl: string;\n\t\tmethod: 'GET' | 'POST' | 'PUT' | 'DELETE';\n\t\theaders: {[key: string]: string};\n\t\tbody: any;\n\t}) => new Http({url: args.url, headers: args.headers}).request({method: args.method || 'GET', body: args.body})\n}\n\nexport const JSTool: AiTool = {\n\tname: 'exec_javascript',\n\tdescription: 'Execute commonjs javascript',\n\targs: {\n\t\tcode: {type: 'string', description: 'CommonJS javascript', required: true}\n\t},\n\tfn: async (args: {code: string}) => {\n\t\tconst console = consoleInterceptor(null);\n\t\tconst resp = await Fn<any>({console}, args.code, true).catch((err: any) => console.output.error.push(err));\n\t\treturn {...console.output, return: resp, stdout: undefined, stderr: undefined};\n\t}\n}\n\nexport const PythonTool: AiTool = {\n\tname: 'exec_javascript',\n\tdescription: 'Execute commonjs javascript',\n\targs: {\n\t\tcode: {type: 'string', description: 'CommonJS javascript', required: true}\n\t},\n\tfn: async (args: {code: string}) => ({result: $Sync`python -c \"${args.code}\"`})\n}\n\nexport const ReadWebpageTool: AiTool = {\n\tname: 'read_webpage',\n\tdescription: 'Extract clean, structured content from a webpage. Use after web_search to read specific URLs',\n\targs: {\n\t\turl: {type: 'string', description: 'URL to extract content from', required: true},\n\t\tfocus: {type: 'string', description: 'Optional: What aspect to focus on (e.g., \"pricing\", \"features\", \"contact info\")'}\n\t},\n\tfn: async (args: {url: string; focus?: string}) => {\n\t\tconst html = await fetch(args.url, {headers: {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64)\"}})\n\t\t\t.then(r => r.text()).catch(err => {throw new Error(`Failed to fetch: ${err.message}`)});\n\n\t\tconst $ = cheerio.load(html);\n\t\t$('script, style, nav, footer, header, aside, iframe, noscript, [role=\"navigation\"], [role=\"banner\"], .ad, .ads, .cookie, .popup').remove();\n\t\tconst metadata = {\n\t\t\ttitle: $('meta[property=\"og:title\"]').attr('content') || $('title').text() || '',\n\t\t\tdescription: $('meta[name=\"description\"]').attr('content') || $('meta[property=\"og:description\"]').attr('content') || '',\n\t\t};\n\n\t\tlet content = '';\n\t\tconst contentSelectors = ['article', 'main', '[role=\"main\"]', '.content', '.post', '.entry', 'body'];\n\t\tfor (const selector of contentSelectors) {\n\t\t\tconst el = $(selector).first();\n\t\t\tif (el.length && el.text().trim().length > 200) {\n\t\t\t\tcontent = el.text();\n\t\t\t\tbreak;\n\t\t\t}\n\t\t}\n\t\tif (!content) content = $('body').text();\n\t\tcontent = content.replace(/\\s+/g, ' ').trim().slice(0, 8000);\n\n\t\treturn {url: args.url, title: metadata.title.trim(), description: metadata.description.trim(), content, focus: args.focus};\n\t}\n}\n\nexport const WebSearchTool: AiTool = {\n\tname: 'web_search',\n\tdescription: 'Use duckduckgo (anonymous) to find find relevant online resources. Returns a list of URLs that works great with the `read_webpage` tool',\n\targs: {\n\t\tquery: {type: 'string', description: 'Search string', required: true},\n\t\tlength: {type: 'string', description: 'Number of results to return', default: 5},\n\t},\n\tfn: async (args: {\n\t\tquery: string;\n\t\tlength: number;\n\t}) => {\n\t\tconst html = await fetch(`https://html.duckduckgo.com/html/?q=${encodeURIComponent(args.query)}`, {\n\t\t\theaders: {\"User-Agent\": \"Mozilla/5.0 (Windows NT 10.0; Win64; x64)\", \"Accept-Language\": \"en-US,en;q=0.9\"}\n\t\t}).then(resp => resp.text());\n\t\tlet match, regex = /<a .*?href=\"(.+?)\".+?<\\/a>/g;\n\t\tconst results = new ASet<string>();\n\t\twhile((match = regex.exec(html)) !== null) {\n\t\t\tlet url = /uddg=(.+)&?/.exec(decodeURIComponent(match[1]))?.[1];\n\t\t\tif(url) url = decodeURIComponent(url);\n\t\t\tif(url) results.add(url);\n\t\t\tif(results.size >= (args.length || 5)) break;\n\t\t}\n\t\treturn results;\n\t}\n}\n"],"names":["LLMProvider","Anthropic","ai","apiToken","model","anthropic","history","timestamp","messages","h","textContent","c","m","i","message","options","controller","res","tools","requestParams","t","objectMap","key","value","resp","isFirstMessage","err","chunk","text","last","JSONAttemptParse","toolCalls","results","toolCall","tool","findByProp","result","JSONSanitize","OpenAi","host","token","openAI","clean","tc","record","h2","rest","rej","args","LLM","config","abort","search","query","subject","limit","o","q","a","b","relevant","compressed","updated","m2","max","min","keep","tokens","system","recent","process","summary","memory","owner","fact","e","v1","v2","dotProduct","normA","normB","denominator","target","maxTokens","overlapTokens","objString","obj","path","p","l","chunks","j","next","embed","resolve","reject","worker","Worker","join","dirname","fileURLToPath","_documentCurrentScript","handleMessage","embedding","handleError","code","index","searchTerms","vector","dimensions","char","v","similarities","refVector","acc","s","codeBlock","jsonStr","Audio","file","speaker","aborted","warning","error","canDiarization","Vision","createWorker","data","Ai","os","CliTool","$","DateTimeTool","ExecTool","stream","JSTool","PythonTool","FetchTool","Http","console","consoleInterceptor","Fn","$Sync","ReadWebpageTool","html","r","cheerio","metadata","content","contentSelectors","selector","el","WebSearchTool","match","regex","ASet","url"],"mappings":"qtBAGO,MAAeA,CAAY,CAElC,CCCO,MAAMC,UAAkBD,CAAY,CAG1C,YAA4BE,EAAwBC,EAAyBC,EAAe,CAC3F,MAAA,EAD2B,KAAA,GAAAF,EAAwB,KAAA,SAAAC,EAAyB,KAAA,MAAAC,EAE5E,KAAK,OAAS,IAAIC,EAAAA,UAAU,CAAC,OAAQF,EAAS,CAC/C,CALA,OAOQ,WAAWG,EAA8B,CAChD,MAAMC,EAAY,KAAK,IAAA,EACjBC,EAAyB,CAAA,EAC/B,QAAQC,KAAKH,EACZ,GAAG,OAAOG,EAAE,SAAW,SACtBD,EAAS,KAAU,CAAC,UAAAD,EAAW,GAAGE,EAAE,MAC9B,CACN,MAAMC,EAAcD,EAAE,SAAS,OAAQE,GAAWA,EAAE,MAAQ,MAAM,EAAE,IAAKA,GAAWA,EAAE,IAAI,EAAE,KAAK;AAAA;AAAA,CAAM,EACpGD,GAAaF,EAAS,KAAK,CAAC,UAAAD,EAAW,KAAME,EAAE,KAAM,QAASC,EAAY,EAC7ED,EAAE,QAAQ,QAASE,GAAW,CAC7B,GAAGA,EAAE,MAAQ,WACZH,EAAS,KAAK,CAAC,UAAAD,EAAW,KAAM,OAAQ,GAAII,EAAE,GAAI,KAAMA,EAAE,KAAM,KAAMA,EAAE,MAAO,QAAS,OAAU,UACzFA,EAAE,MAAQ,cAAe,CAClC,MAAMC,EAASJ,EAAS,SAASI,GAAWA,EAAG,IAAMD,EAAE,WAAW,EAC/DC,IAAGA,EAAED,EAAE,SAAW,QAAU,SAAS,EAAIA,EAAE,QAC/C,CACD,CAAC,CACF,CAED,OAAOH,CACR,CAEQ,aAAaF,EAA8B,CAClD,QAAQO,EAAI,EAAGA,EAAIP,EAAQ,OAAQO,IAClC,GAAGP,EAAQO,CAAC,EAAE,MAAQ,OAAQ,CAC7B,MAAMJ,EAASH,EAAQO,CAAC,EACxBP,EAAQ,OAAOO,EAAG,EACjB,CAAC,KAAM,YAAa,QAAS,CAAC,CAAC,KAAM,WAAY,GAAIJ,EAAE,GAAI,KAAMA,EAAE,KAAM,MAAOA,EAAE,IAAA,CAAK,CAAA,EACvF,CAAC,KAAM,OAAQ,QAAS,CAAC,CAAC,KAAM,cAAe,YAAaA,EAAE,GAAI,SAAU,CAAC,CAACA,EAAE,MAAO,QAAUA,EAAE,OAASA,EAAE,QAAQ,CAAA,CAAC,EAExHI,GACD,CAED,OAAOP,EAAQ,IAAI,CAAC,CAAC,UAAAC,EAAW,GAAGE,CAAA,IAAOA,CAAC,CAC5C,CAEA,IAAIK,EAAiBC,EAAsB,GAA8B,CACxE,MAAMC,EAAa,IAAI,gBACvB,OAAO,OAAO,OAAO,IAAI,QAAa,MAAOC,GAAQ,CACpD,IAAIX,EAAU,KAAK,aAAa,CAAC,GAAGS,EAAQ,SAAW,GAAI,CAAC,KAAM,OAAQ,QAASD,EAAS,UAAW,KAAK,IAAA,CAAI,CAAE,CAAC,EACnH,MAAMI,EAAQH,EAAQ,OAAS,KAAK,GAAG,QAAQ,KAAK,OAAS,CAAA,EACvDI,EAAqB,CAC1B,MAAOJ,EAAQ,OAAS,KAAK,MAC7B,WAAYA,EAAQ,YAAc,KAAK,GAAG,QAAQ,KAAK,YAAc,KACrE,OAAQA,EAAQ,QAAU,KAAK,GAAG,QAAQ,KAAK,QAAU,GACzD,YAAaA,EAAQ,aAAe,KAAK,GAAG,QAAQ,KAAK,aAAe,GACxE,MAAOG,EAAM,IAAIE,IAAM,CACtB,KAAMA,EAAE,KACR,YAAaA,EAAE,YACf,aAAc,CACb,KAAM,SACN,WAAYA,EAAE,KAAOC,EAAAA,UAAUD,EAAE,KAAM,CAACE,EAAKC,KAAW,CAAC,GAAGA,EAAO,SAAU,MAAA,EAAW,EAAI,CAAA,EAC5F,SAAUH,EAAE,KAAO,OAAO,QAAQA,EAAE,IAAI,EAAE,OAAOA,GAAKA,EAAE,CAAC,EAAE,QAAQ,EAAE,IAAIA,GAAKA,EAAE,CAAC,CAAC,EAAI,CAAA,CAAC,EAExF,GAAI,MAAA,EACH,EACF,SAAUd,EACV,OAAQ,CAAC,CAACS,EAAQ,MAAA,EAGnB,IAAIS,EAAWC,EAAiB,GAChC,EAAG,CAOF,GANAD,EAAO,MAAM,KAAK,OAAO,SAAS,OAAOL,CAAa,EAAE,MAAMO,GAAO,CACpE,MAAAA,EAAI,SAAW;AAAA;AAAA;AAAA,EAAkB,KAAK,UAAUpB,EAAS,KAAM,CAAC,CAAC,GAC3DoB,CACP,CAAC,EAGEX,EAAQ,OAAQ,CACdU,EACCA,EAAiB,GADFV,EAAQ,OAAO,CAAC,KAAM;AAAA;AAAA,EAAO,EAEjDS,EAAK,QAAU,CAAA,EACf,gBAAiBG,KAASH,EAAM,CAC/B,GAAGR,EAAW,OAAO,QAAS,MAC9B,GAAGW,EAAM,OAAS,sBACdA,EAAM,cAAc,OAAS,OAC/BH,EAAK,QAAQ,KAAK,CAAC,KAAM,OAAQ,KAAM,GAAG,EACjCG,EAAM,cAAc,OAAS,YACtCH,EAAK,QAAQ,KAAK,CAAC,KAAM,WAAY,GAAIG,EAAM,cAAc,GAAI,KAAMA,EAAM,cAAc,KAAM,MAAY,GAAG,UAExGA,EAAM,OAAS,sBACxB,GAAGA,EAAM,MAAM,OAAS,aAAc,CACrC,MAAMC,EAAOD,EAAM,MAAM,KACzBH,EAAK,QAAQ,GAAG,EAAE,EAAE,MAAQI,EAC5Bb,EAAQ,OAAO,CAAC,KAAAa,EAAK,CACtB,MAAUD,EAAM,MAAM,OAAS,qBAC9BH,EAAK,QAAQ,GAAG,EAAE,EAAE,OAASG,EAAM,MAAM,sBAEjCA,EAAM,OAAS,qBAAsB,CAC9C,MAAME,EAAOL,EAAK,QAAQ,GAAG,EAAE,EAC5BK,EAAK,OAAS,OAAMA,EAAK,MAAQA,EAAK,MAAQC,EAAAA,iBAAiBD,EAAK,MAAO,CAAA,CAAE,EAAI,CAAA,EACrF,SAAUF,EAAM,OAAS,eACxB,KAEF,CACD,CAGA,MAAMI,EAAYP,EAAK,QAAQ,OAAQ,GAAW,EAAE,OAAS,UAAU,EACvE,GAAGO,EAAU,QAAU,CAACf,EAAW,OAAO,QAAS,CAClDV,EAAQ,KAAK,CAAC,KAAM,YAAa,QAASkB,EAAK,QAAQ,EACvD,MAAMQ,EAAU,MAAM,QAAQ,IAAID,EAAU,IAAI,MAAOE,GAAkB,CACxE,MAAMC,EAAOhB,EAAM,KAAKiB,EAAAA,WAAW,OAAQF,EAAS,IAAI,CAAC,EAEzD,GADGlB,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAMkB,EAAS,KAAK,EACpD,CAACC,EAAM,MAAO,CAAC,YAAaD,EAAS,GAAI,SAAU,GAAM,QAAS,gBAAA,EACrE,GAAI,CACH,MAAMG,EAAS,MAAMF,EAAK,GAAGD,EAAS,MAAOlB,GAAS,OAAQ,KAAK,EAAE,EACrE,MAAO,CAAC,KAAM,cAAe,YAAakB,EAAS,GAAI,QAASI,eAAaD,CAAM,CAAA,CACpF,OAASV,EAAU,CAClB,MAAO,CAAC,KAAM,cAAe,YAAaO,EAAS,GAAI,SAAU,GAAM,QAASP,GAAK,SAAWA,GAAK,SAAA,GAAc,SAAA,CACpH,CACD,CAAC,CAAC,EACFpB,EAAQ,KAAK,CAAC,KAAM,OAAQ,QAAS0B,EAAQ,EAC7Cb,EAAc,SAAWb,CAC1B,CACD,OAAS,CAACU,EAAW,OAAO,SAAWQ,EAAK,QAAQ,KAAMb,GAAWA,EAAE,OAAS,UAAU,GAC1FL,EAAQ,KAAK,CAAC,KAAM,YAAa,QAASkB,EAAK,QAAQ,OAAQb,GAAWA,EAAE,MAAQ,MAAM,EAAE,IAAKA,GAAWA,EAAE,IAAI,EAAE,KAAK;AAAA;AAAA,CAAM,EAAE,EACjIL,EAAU,KAAK,WAAWA,CAAO,EAE9BS,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAM,GAAK,EAC3CA,EAAQ,SAASA,EAAQ,QAAQ,OAAO,EAAGA,EAAQ,QAAQ,OAAQ,GAAGT,CAAO,EAChFW,EAAIX,EAAQ,GAAG,EAAE,GAAG,OAAO,CAC5B,CAAC,EAAG,CAAC,MAAO,IAAMU,EAAW,MAAA,EAAQ,CACtC,CACD,CCpIO,MAAMsB,UAAetC,CAAY,CAGvC,YAA4BE,EAAwBqC,EAAqCC,EAAsBpC,EAAe,CAC7H,MAAA,EAD2B,KAAA,GAAAF,EAAwB,KAAA,KAAAqC,EAAqC,KAAA,MAAAC,EAAsB,KAAA,MAAApC,EAE9G,KAAK,OAAS,IAAIqC,EAAAA,OAAOC,QAAM,CAC9B,QAASH,EACT,OAAQC,CAAA,CACR,CAAC,CACH,CARA,OAUQ,WAAWlC,EAA8B,CAChD,QAAQO,EAAI,EAAGA,EAAIP,EAAQ,OAAQO,IAAK,CACvC,MAAMJ,EAAIH,EAAQO,CAAC,EACnB,GAAGJ,EAAE,OAAS,aAAeA,EAAE,WAAY,CAC1C,MAAMS,EAAQT,EAAE,WAAW,IAAKkC,IAAa,CAC5C,KAAM,OACN,GAAIA,EAAG,GACP,KAAMA,EAAG,SAAS,KAClB,KAAMb,EAAAA,iBAAiBa,EAAG,SAAS,UAAW,CAAA,CAAE,EAChD,UAAWlC,EAAE,SAAA,EACZ,EACFH,EAAQ,OAAOO,EAAG,EAAG,GAAGK,CAAK,EAC7BL,GAAKK,EAAM,OAAS,CACrB,SAAUT,EAAE,OAAS,QAAUA,EAAE,QAAS,CACzC,MAAMmC,EAAStC,EAAQ,QAAWG,EAAE,cAAgBoC,EAAG,EAAE,EACtDD,IACCnC,EAAE,QAAQ,SAAS,UAAU,EAAGmC,EAAO,MAAQnC,EAAE,QAC/CmC,EAAO,QAAUnC,EAAE,SAEzBH,EAAQ,OAAOO,EAAG,CAAC,EACnBA,GACD,CACIP,EAAQO,CAAC,GAAG,cAAmBA,CAAC,EAAE,UAAY,KAAK,IAAA,EACxD,CACA,OAAOP,CACR,CAEQ,aAAaA,EAA8B,CAClD,OAAOA,EAAQ,OAAO,CAAC8B,EAAQ3B,IAAM,CACpC,GAAGA,EAAE,OAAS,OACb2B,EAAO,KAAK,CACX,KAAM,YACN,QAAS,KACT,WAAY,CAAC,CAAE,GAAI3B,EAAE,GAAI,KAAM,WAAY,SAAU,CAAE,KAAMA,EAAE,KAAM,UAAW,KAAK,UAAUA,EAAE,IAAI,CAAA,EAAK,EAC1G,QAAS,KACT,YAAa,CAAA,CAAC,EACZ,CACF,KAAM,OACN,aAAcA,EAAE,GAChB,QAASA,EAAE,OAASA,EAAE,OAAA,CACtB,MACK,CACN,KAAM,CAAC,UAAAF,EAAW,GAAGuC,CAAA,EAAQrC,EAC7B2B,EAAO,KAAKU,CAAI,CACjB,CACA,OAAOV,CACR,EAAG,CAAA,CAAW,CACf,CAEA,IAAItB,EAAiBC,EAAsB,GAA8B,CACxE,MAAMC,EAAa,IAAI,gBACvB,OAAO,OAAO,OAAO,IAAI,QAAa,MAAOC,EAAK8B,IAAQ,CACtDhC,EAAQ,QAAUA,EAAQ,UAAU,CAAC,GAAG,MAAQ,UAAUA,EAAQ,SAAS,OAAO,EAAG,EAAG,CAAC,KAAM,SAAU,QAASA,EAAQ,OAAQ,UAAW,KAAK,IAAA,EAAM,EAC3J,IAAIT,EAAU,KAAK,aAAa,CAAC,GAAGS,EAAQ,SAAW,GAAI,CAAC,KAAM,OAAQ,QAASD,EAAS,UAAW,KAAK,IAAA,CAAI,CAAE,CAAC,EACnH,MAAMI,EAAQH,EAAQ,OAAS,KAAK,GAAG,QAAQ,KAAK,OAAS,CAAA,EACvDI,EAAqB,CAC1B,MAAOJ,EAAQ,OAAS,KAAK,MAC7B,SAAUT,EACV,OAAQ,CAAC,CAACS,EAAQ,OAClB,WAAYA,EAAQ,YAAc,KAAK,GAAG,QAAQ,KAAK,YAAc,KACrE,YAAaA,EAAQ,aAAe,KAAK,GAAG,QAAQ,KAAK,aAAe,GACxE,MAAOG,EAAM,IAAIE,IAAM,CACtB,KAAM,WACN,SAAU,CACT,KAAMA,EAAE,KACR,YAAaA,EAAE,YACf,WAAY,CACX,KAAM,SACN,WAAYA,EAAE,KAAOC,EAAAA,UAAUD,EAAE,KAAM,CAACE,EAAKC,KAAW,CAAC,GAAGA,EAAO,SAAU,MAAA,EAAW,EAAI,CAAA,EAC5F,SAAUH,EAAE,KAAO,OAAO,QAAQA,EAAE,IAAI,EAAE,OAAOA,GAAKA,EAAE,CAAC,EAAE,QAAQ,EAAE,IAAIA,GAAKA,EAAE,CAAC,CAAC,EAAI,CAAA,CAAC,CACxF,CACD,EACC,CAAA,EAGH,IAAII,EAAWC,EAAiB,GAChC,EAAG,CAMF,GALAD,EAAO,MAAM,KAAK,OAAO,KAAK,YAAY,OAAOL,CAAa,EAAE,MAAMO,GAAO,CAC5E,MAAAA,EAAI,SAAW;AAAA;AAAA;AAAA,EAAkB,KAAK,UAAUpB,EAAS,KAAM,CAAC,CAAC,GAC3DoB,CACP,CAAC,EAEEX,EAAQ,OAAQ,CACdU,EACCA,EAAiB,GADFV,EAAQ,OAAO,CAAC,KAAM;AAAA;AAAA,EAAO,EAEjDS,EAAK,QAAU,CAAC,CAAC,QAAS,CAAC,QAAS,GAAI,WAAY,CAAA,CAAC,EAAG,EACxD,gBAAiBG,KAASH,EAAM,CAC/B,GAAGR,EAAW,OAAO,QAAS,MAC3BW,EAAM,QAAQ,CAAC,EAAE,MAAM,UACzBH,EAAK,QAAQ,CAAC,EAAE,QAAQ,SAAWG,EAAM,QAAQ,CAAC,EAAE,MAAM,QAC1DZ,EAAQ,OAAO,CAAC,KAAMY,EAAM,QAAQ,CAAC,EAAE,MAAM,QAAQ,GAEnDA,EAAM,QAAQ,CAAC,EAAE,MAAM,aACzBH,EAAK,QAAQ,CAAC,EAAE,QAAQ,WAAaG,EAAM,QAAQ,CAAC,EAAE,MAAM,WAE9D,CACD,CAEA,MAAMI,EAAYP,EAAK,QAAQ,CAAC,EAAE,QAAQ,YAAc,CAAA,EACxD,GAAGO,EAAU,QAAU,CAACf,EAAW,OAAO,QAAS,CAClDV,EAAQ,KAAKkB,EAAK,QAAQ,CAAC,EAAE,OAAO,EACpC,MAAMQ,EAAU,MAAM,QAAQ,IAAID,EAAU,IAAI,MAAOE,GAAkB,CACxE,MAAMC,EAAOhB,GAAO,KAAKiB,EAAAA,WAAW,OAAQF,EAAS,SAAS,IAAI,CAAC,EAEnE,GADGlB,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAMkB,EAAS,SAAS,KAAK,EAC7D,CAACC,EAAM,MAAO,CAAC,KAAM,OAAQ,aAAcD,EAAS,GAAI,QAAS,6BAAA,EACpE,GAAI,CACH,MAAMe,EAAOlB,EAAAA,iBAAiBG,EAAS,SAAS,UAAW,CAAA,CAAE,EACvDG,EAAS,MAAMF,EAAK,GAAGc,EAAMjC,EAAQ,OAAQ,KAAK,EAAE,EAC1D,MAAO,CAAC,KAAM,OAAQ,aAAckB,EAAS,GAAI,QAASI,eAAaD,CAAM,CAAA,CAC9E,OAASV,EAAU,CAClB,MAAO,CAAC,KAAM,OAAQ,aAAcO,EAAS,GAAI,QAASI,EAAAA,aAAa,CAAC,MAAOX,GAAK,SAAWA,GAAK,YAAc,SAAA,CAAU,CAAA,CAC7H,CACD,CAAC,CAAC,EACFpB,EAAQ,KAAK,GAAG0B,CAAO,EACvBb,EAAc,SAAWb,CAC1B,CACD,OAAS,CAACU,EAAW,OAAO,SAAWQ,EAAK,UAAU,CAAC,GAAG,SAAS,YAAY,QAC/ElB,EAAQ,KAAK,CAAC,KAAM,YAAa,QAASkB,EAAK,QAAQ,CAAC,EAAE,QAAQ,SAAW,EAAA,CAAG,EAChFlB,EAAU,KAAK,WAAWA,CAAO,EAE9BS,EAAQ,QAAQA,EAAQ,OAAO,CAAC,KAAM,GAAK,EAC3CA,EAAQ,SAASA,EAAQ,QAAQ,OAAO,EAAGA,EAAQ,QAAQ,OAAQ,GAAGT,CAAO,EAChFW,EAAIX,EAAQ,GAAG,EAAE,GAAG,OAAO,CAC5B,CAAC,EAAG,CAAC,MAAO,IAAMU,EAAW,MAAA,EAAQ,CACtC,CACD,CClEA,MAAMiC,CAAI,CAIT,YAA4B/C,EAAQ,CAAR,KAAA,GAAAA,EACvBA,EAAG,QAAQ,KAAK,QACpB,OAAO,QAAQA,EAAG,QAAQ,IAAI,MAAM,EAAE,QAAQ,CAAC,CAACE,EAAO8C,CAAM,IAAM,CAC9D,KAAK,eAAc,KAAK,aAAe9C,GACxC8C,EAAO,OAAS,YAAa,KAAK,OAAO9C,CAAK,EAAI,IAAIH,EAAU,KAAK,GAAIiD,EAAO,MAAO9C,CAAK,EACvF8C,EAAO,OAAS,SAAU,KAAK,OAAO9C,CAAK,EAAI,IAAIkC,EAAO,KAAK,GAAIY,EAAO,KAAM,aAAc9C,CAAK,EACnG8C,EAAO,OAAS,WAAU,KAAK,OAAO9C,CAAK,EAAI,IAAIkC,EAAO,KAAK,GAAIY,EAAO,MAAQ,KAAMA,EAAO,MAAO9C,CAAK,EACpH,CAAC,CACF,CAXQ,OAAyC,CAAA,EACzC,aAkBR,IAAIU,EAAiBC,EAAsB,GAA8B,CACxE,MAAMH,EAAIG,EAAQ,OAAS,KAAK,aAChC,GAAG,CAAC,KAAK,OAAOH,CAAC,QAAS,IAAI,MAAM,yBAAyBA,CAAC,EAAE,EAChE,IAAIuC,EAAQ,IAAM,CAAC,EACnB,OAAO,OAAO,OAAO,IAAI,QAAgB,MAAMlC,GAAO,CAGrD,GAFIF,EAAQ,UAASA,EAAQ,QAAU,CAAA,GAEpCA,EAAQ,OAAQ,CAClBA,EAAQ,QAAUA,EAAQ,QAAU,IAAM;AAAA;AAAA,EAC1C,MAAMqC,EAAS,MAAOC,EAAuBC,EAAyBC,EAAQ,KAAO,CACpF,KAAM,CAACC,EAAGC,CAAC,EAAI,MAAM,QAAQ,IAAI,CAChCH,EAAU,KAAK,UAAUA,CAAO,EAAI,QAAQ,QAAQ,IAAI,EACxDD,EAAQ,KAAK,UAAUA,CAAK,EAAI,QAAQ,QAAQ,IAAI,CAAA,CACpD,EACD,OAAQtC,EAAQ,QAAU,CAAA,GACxB,IAAIH,IAAM,CAAC,GAAGA,EAAG,MAAO4C,EAAI,KAAK,iBAAiB5C,EAAE,WAAW,CAAC,EAAG4C,EAAE,CAAC,EAAE,SAAS,EAAI,CAAA,EAAG,EACxF,OAAQ5C,GAAWA,EAAE,OAAS,EAAG,EACjC,IAAKA,IAAY,CAAC,GAAGA,EAAG,MAAO6C,EAAI,KAAK,iBAAiB7C,EAAE,WAAW,CAAC,EAAG6C,EAAE,CAAC,EAAE,SAAS,EAAI7C,EAAE,OAAO,EACrG,OAAQA,GAAWA,EAAE,OAAS,EAAG,EACjC,SAAS,CAAC8C,EAAQC,IAAWD,EAAE,MAAQC,EAAE,KAAK,EAC9C,MAAM,EAAGJ,CAAK,CACjB,EAEMK,EAAW,MAAMR,EAAOtC,CAAO,EAClC8C,EAAS,QAAQ7C,EAAQ,QAAQ,KAAK,CAAC,KAAM,YAAa,QAAS;AAAA,EAA2B6C,EAAS,IAAIhD,GAAK,GAAGA,EAAE,KAAK,KAAKA,EAAE,IAAI,EAAE,EAAE,KAAK;AAAA,CAAI,CAAA,CAAE,EACvJG,EAAQ,MAAQ,CAAC,GAAGA,EAAQ,OAAS,CAAA,EAAI,CACxC,KAAM,cACN,YAAa,mDACb,KAAM,CACL,QAAS,CAAC,KAAM,SAAU,YAAa,iFAAA,EACvC,MAAO,CAAC,KAAM,SAAU,YAAa,8EAAA,EACrC,MAAO,CAAC,KAAM,SAAU,YAAa,yBAAA,CAAyB,EAE/D,GAAKiC,GAAS,CACb,GAAG,CAACA,EAAK,SAAW,CAACA,EAAK,MAAO,MAAM,IAAI,MAAM,gDAAgD,EACjG,OAAOI,EAAOJ,EAAK,MAAOA,EAAK,QAASA,EAAK,OAAS,CAAC,CACxD,CAAA,CACA,CACF,CAGA,MAAMxB,EAAO,MAAM,KAAK,OAAOZ,CAAC,EAAE,IAAIE,EAASC,CAAO,EAGtD,GAAGA,EAAQ,OAAQ,CAClB,MAAM,EAAIA,EAAQ,SAAS,UAAWN,GAAWA,EAAE,MAAQ,aAAeA,EAAE,QAAQ,WAAW,sBAAsB,CAAC,EACnH,GAAK,MAAQ,GAAK,KAAW,SAAS,OAAO,EAAG,CAAC,CACrD,CAGA,GAAGM,EAAQ,UAAYA,EAAQ,OAAQ,CACtC,IAAI8C,EAAa,KACjB,GAAG9C,EAAQ,SACV8C,EAAa,MAAM,KAAK,GAAG,SAAS,gBAAgB9C,EAAQ,QAASA,EAAQ,SAAS,IAAKA,EAAQ,SAAS,IAAKA,CAAO,EACxHA,EAAQ,QAAQ,OAAO,EAAGA,EAAQ,QAAQ,OAAQ,GAAG8C,EAAW,OAAO,MACjE,CACN,MAAMhD,EAAIE,EAAQ,SAAS,cAAcH,GAAKA,EAAE,MAAQ,MAAM,GAAK,GACnEiD,EAAa,MAAM,KAAK,GAAG,SAAS,gBAAgBhD,GAAK,GAAKE,EAAQ,QAAQ,MAAMF,CAAC,EAAIE,EAAQ,QAAS,EAAG,EAAGA,CAAO,CACxH,CACA,GAAGA,EAAQ,OAAQ,CAClB,MAAM+C,EAAU/C,EAAQ,OACtB,OAAOH,GAAK,CAACiD,EAAW,OAAO,KAAKE,GAAM,KAAK,iBAAiBnD,EAAE,WAAW,CAAC,EAAGmD,EAAG,WAAW,CAAC,CAAC,EAAI,EAAG,CAAC,EACzG,OAAOF,EAAW,MAAM,EAC1B9C,EAAQ,OAAO,OAAO,EAAGA,EAAQ,OAAO,OAAQ,GAAG+C,CAAO,CAC3D,CACD,CACA,OAAO7C,EAAIO,CAAI,CAChB,CAAC,EAAG,CAAC,MAAA2B,EAAM,CACZ,CAUA,MAAM,gBAAgB7C,EAAuB0D,EAAaC,EAAalD,EAA6E,CACnJ,GAAG,KAAK,eAAeT,CAAO,EAAI0D,QAAY,CAAC,QAAA1D,EAAS,OAAQ,EAAC,EACjE,IAAI4D,EAAO,EAAGC,EAAS,EACvB,QAAQvD,KAAKN,EAAQ,aAEpB,GADA6D,GAAU,KAAK,eAAevD,EAAE,OAAO,EACpCuD,EAASF,EAAKC,QACZ,OAEN,GAAG5D,EAAQ,QAAU4D,EAAM,MAAO,CAAC,QAAA5D,EAAS,OAAQ,EAAC,EACrD,MAAM8D,EAAS9D,EAAQ,CAAC,EAAE,MAAQ,SAAWA,EAAQ,CAAC,EAAI,KACzD+D,EAASH,GAAQ,EAAI,CAAA,EAAK5D,EAAQ,MAAM,CAAC4D,CAAI,EAC7CI,GAAWJ,GAAQ,EAAI5D,EAAUA,EAAQ,MAAM,EAAG,CAAC4D,CAAI,GAAG,OAAOzD,GAAKA,EAAE,OAAS,aAAeA,EAAE,OAAS,MAAM,EAC5G8D,EAAe,MAAM,KAAK,KAAK;AAAA;AAAA,EAA6ZD,EAAQ,IAAI1D,GAAK,GAAGA,EAAE,IAAI,KAAKA,EAAE,OAAO,EAAE,EAAE,KAAK;AAAA;AAAA,CAAM,CAAC,GAAI,CAAC,MAAOG,GAAS,MAAO,YAAaA,GAAS,aAAe,GAAI,EACzjBR,MAAgB,KAChBiE,EAAS,MAAM,QAAQ,KAAKD,GAAS,OAAS,CAAA,IAAK,IAAI,MAAO,CAACE,EAAOC,CAAI,IAAwB,CACvG,MAAMC,EAAI,MAAM,QAAQ,IAAI,CAAC,KAAK,UAAUF,CAAK,EAAG,KAAK,UAAU,GAAGA,CAAK,KAAKC,CAAI,EAAE,CAAC,CAAC,EACxF,MAAO,CAAC,MAAAD,EAAO,KAAAC,EAAM,WAAY,CAACC,EAAE,CAAC,EAAE,CAAC,EAAE,UAAWA,EAAE,CAAC,EAAE,CAAC,EAAE,SAAS,EAAG,UAAApE,CAAA,CAC1E,CAAC,CAAC,EACIE,EAAI,CAAC,CAAC,KAAM,YAAa,QAAS,yBAAyB8D,GAAS,OAAO,GAAI,UAAW,KAAK,IAAA,CAAI,EAAI,GAAGF,CAAM,EACtH,OAAGD,GAAQ3D,EAAE,OAAO,EAAG,EAAG2D,CAAM,EACzB,CAAC,QAAc3D,EAAG,OAAA+D,CAAA,CAC1B,CAQA,iBAAiBI,EAAcC,EAAsB,CACpD,GAAID,EAAG,SAAWC,EAAG,OAAQ,MAAM,IAAI,MAAM,6BAA6B,EAC1E,IAAIC,EAAa,EAAGC,EAAQ,EAAGC,EAAQ,EACvC,QAAS,EAAI,EAAG,EAAIJ,EAAG,OAAQ,IAC9BE,GAAcF,EAAG,CAAC,EAAIC,EAAG,CAAC,EAC1BE,GAASH,EAAG,CAAC,EAAIA,EAAG,CAAC,EACrBI,GAASH,EAAG,CAAC,EAAIA,EAAG,CAAC,EAEtB,MAAMI,EAAc,KAAK,KAAKF,CAAK,EAAI,KAAK,KAAKC,CAAK,EACtD,OAAOC,IAAgB,EAAI,EAAIH,EAAaG,CAC7C,CASA,MAAMC,EAAyBC,EAAY,IAAKC,EAAgB,GAAc,CAC7E,MAAMC,EAAY,CAACC,EAAUC,EAAO,KAC/BD,EACG,OAAO,QAAQA,CAAG,EAAE,QAAQ,CAAC,CAAChE,EAAKC,CAAK,IAAM,CACpD,MAAMiE,EAAID,EAAO,GAAGA,CAAI,GAAG,MAAM,CAACjE,CAAG,EAAI,IAAIA,CAAG,GAAK,IAAIA,CAAG,GAAG,GAAKA,EACpE,OAAG,OAAOC,GAAU,UAAY,CAAC,MAAM,QAAQA,CAAK,EAAU8D,EAAU9D,EAAOiE,CAAC,EACzE,GAAGA,CAAC,KAAK,MAAM,QAAQjE,CAAK,EAAIA,EAAM,KAAK,IAAI,EAAIA,CAAK,EAChE,CAAC,EALe,CAAA,EAQX4C,GADQ,OAAOe,GAAW,SAAWG,EAAUH,CAAM,EAAIA,EAAO,MAAM;AAAA,CAAI,GAC3D,QAAQO,GAAK,CAAC,GAAGA,EAAE,MAAM,KAAK,EAAE,OAAO,OAAO,EAAG;AAAA,CAAI,CAAC,EACrEC,EAAmB,CAAA,EACzB,QAAQ7E,EAAI,EAAGA,EAAIsD,EAAO,QAAS,CAClC,IAAIvC,EAAO,GAAI+D,EAAI9E,EACnB,KAAM8E,EAAIxB,EAAO,QAAQ,CACxB,MAAMyB,EAAOhE,GAAQA,EAAO,IAAM,IAAMuC,EAAOwB,CAAC,EAChD,GAAG,KAAK,eAAeC,EAAK,QAAQ,YAAa;AAAA,CAAI,CAAC,EAAIT,GAAavD,EAAM,MAC7EA,EAAOgE,EACPD,GACD,CACA,MAAMjD,EAAQd,EAAK,QAAQ,YAAa;AAAA,CAAI,EAAE,KAAA,EAC3Cc,GAAOgD,EAAO,KAAKhD,CAAK,EAC3B7B,EAAI,KAAK,IAAI8E,EAAIP,EAAeO,IAAM9E,EAAIA,EAAI,EAAI8E,CAAC,CACpD,CACA,OAAOD,CACR,CASA,UAAUR,EAAyBC,EAAY,IAAKC,EAAgB,GAAI,CACvE,MAAMS,EAASjE,GACP,IAAI,QAAQ,CAACkE,EAASC,IAAW,CACvC,MAAMC,EAAS,IAAIC,EAAAA,OAAOC,OAAKC,EAAAA,QAAQC,EAAAA,cAAc,OAAA,SAAA,IAAA,QAAA,KAAA,EAAA,cAAA,UAAA,EAAA,KAAAC,GAAAA,EAAA,QAAA,YAAA,IAAA,UAAAA,EAAA,KAAA,IAAA,IAAA,WAAA,SAAA,OAAA,EAAA,IAAe,CAAC,EAAG,aAAa,CAAC,EAChFC,EAAgB,CAAC,CAAE,UAAAC,KAAqB,CAC7CP,EAAO,UAAA,EACPF,EAAQS,CAAS,CAClB,EACMC,EAAe9E,GAAe,CACnCsE,EAAO,UAAA,EACPD,EAAOrE,CAAG,CACX,EACAsE,EAAO,GAAG,UAAWM,CAAa,EAClCN,EAAO,GAAG,QAASQ,CAAW,EAC9BR,EAAO,GAAG,OAASS,GAAS,CACxBA,IAAS,GAAGV,EAAO,IAAI,MAAM,2BAA2BU,CAAI,EAAE,CAAC,CACnE,CAAC,EACDT,EAAO,YAAY,CAAC,KAAApE,EAAM,MAAO,KAAK,GAAG,SAAS,UAAY,oBAAqB,SAAU,KAAK,GAAG,QAAQ,KAAK,CACnH,CAAC,EAEI8D,EAAS,KAAK,MAAMR,EAAQC,EAAWC,CAAa,EAC1D,OAAO,QAAQ,IAAIM,EAAO,IAAI,MAAO9D,EAAM8E,KAAW,CACrD,MAAAA,EACA,UAAW,MAAMb,EAAMjE,CAAI,EAC3B,KAAAA,EACA,OAAQ,KAAK,eAAeA,CAAI,CAAA,EAC/B,CAAC,CACJ,CAOA,eAAetB,EAAsB,CACpC,MAAMsB,EAAO,KAAK,UAAUtB,CAAO,EACnC,OAAO,KAAK,KAAMsB,EAAK,OAAS,EAAK,GAAG,CACzC,CAQA,WAAWsD,KAAmByB,EAAuB,CACpD,GAAGA,EAAY,OAAS,EAAG,MAAM,IAAI,MAAM,wCAAwC,EACnF,MAAMC,EAAS,CAAChF,EAAciF,EAAqB,KAC3CjF,EAAK,cAAc,MAAM,EAAE,EAAE,IAAI,CAACkF,EAAMJ,IAC7CI,EAAK,WAAW,CAAC,GAAKJ,EAAQ,GAAMG,EAAaA,CAAU,EAAE,MAAM,EAAGA,CAAU,EAE7EE,EAAIH,EAAO1B,CAAM,EACjB8B,EAAeL,EAAY,IAAIvF,GAAKwF,EAAOxF,CAAC,CAAC,EAAE,IAAI6F,GAAa,KAAK,iBAAiBF,EAAGE,CAAS,CAAC,EACzG,MAAO,CAAC,IAAKD,EAAa,OAAO,CAACE,EAAKC,IAAMD,EAAMC,EAAG,CAAC,EAAIH,EAAa,OAAQ,IAAK,KAAK,IAAI,GAAGA,CAAY,EAAG,aAAAA,CAAA,CACjH,CAQA,MAAM,KAAKlG,EAAiBC,EAAoC,CAC/D,IAAIS,EAAO,MAAM,KAAK,IAAIV,EAAS,CAAC,OAAQ,2DAA4D,GAAGC,EAAQ,EACnH,GAAG,CAACS,EAAM,MAAO,CAAA,EACjB,MAAM4F,EAAY,6BAA6B,KAAK5F,CAAI,EAClD6F,EAAUD,EAAYA,EAAU,CAAC,EAAE,OAAS5F,EAClD,OAAOM,EAAAA,iBAAiBuF,EAAS,EAAE,CACpC,CASA,UAAUzF,EAAcuC,EAAgBpD,EAA8C,CACrF,OAAO,KAAK,IAAIa,EAAM,CAAC,OAAQ,+BAA+BuC,CAAM,+BAAgC,YAAa,GAAK,GAAGpD,CAAA,CAAQ,CAClI,CACD,CC1UO,MAAMuG,CAAM,CAClB,YAAoBpH,EAAQ,CAAR,KAAA,GAAAA,CAAS,CAE7B,IAAIqH,EAAcxG,EAAiD,GAAqC,CACvG,KAAM,CAAE,MAAAX,EAAQ,KAAK,GAAG,QAAQ,KAAO,eAAgB,QAAAoH,EAAU,EAAA,EAAUzG,EAC3E,IAAI0G,EAAU,GACd,MAAMtE,EAAQ,IAAM,CAAEsE,EAAU,EAAM,EAEhCjC,EAAI,IAAI,QAAuB,CAACM,EAASC,IAAW,CACzD,MAAMC,EAAS,IAAIC,EAAAA,OAAOC,OAAKC,EAAAA,QAAQC,EAAAA,cAAc,OAAA,SAAA,IAAA,QAAA,KAAA,EAAA,cAAA,UAAA,EAAA,KAAAC,GAAAA,EAAA,QAAA,YAAA,IAAA,UAAAA,EAAA,KAAA,IAAA,IAAA,WAAA,SAAA,OAAA,EAAA,IAAe,CAAC,EAAG,QAAQ,CAAC,EAC3EC,EAAgB,CAAC,CAAE,KAAA1E,EAAM,QAAA8F,EAAS,MAAAC,KAAiB,CACxD3B,EAAO,UAAA,EACJ,CAAAyB,IACAE,EAAO5B,EAAO,IAAI,MAAM4B,CAAK,CAAC,GAE7BD,GAAS,QAAQ,KAAKA,CAAO,EAChC5B,EAAQlE,CAAI,GAEd,EACM4E,EAAe9E,GAAe,CACnCsE,EAAO,UAAA,EACHyB,GAAS1B,EAAOrE,CAAG,CACxB,EACAsE,EAAO,GAAG,UAAWM,CAAa,EAClCN,EAAO,GAAG,QAASQ,CAAW,EAC9BR,EAAO,GAAG,OAASS,GAAS,CACxBA,IAAS,GAAK,CAACgB,GAAS1B,EAAO,IAAI,MAAM,2BAA2BU,CAAI,EAAE,CAAC,CAC/E,CAAC,EACDT,EAAO,YAAY,CAAC,KAAAuB,EAAM,MAAAnH,EAAO,QAAAoH,EAAS,SAAU,KAAK,GAAG,QAAQ,KAAM,MAAO,KAAK,GAAG,QAAQ,QAAQ,CAC1G,CAAC,EACD,OAAO,OAAO,OAAOhC,EAAG,CAAE,MAAArC,EAAO,CAClC,CAEA,eAAiByE,EAAAA,cAClB,CCrCO,MAAMC,CAAO,CAEnB,YAAoB3H,EAAQ,CAAR,KAAA,GAAAA,CAAU,CAO9B,IAAIqF,EAA+C,CAClD,IAAIS,EACJ,MAAMR,EAAI,IAAI,QAAuB,MAAMvE,GAAO,CACjD+E,EAAS,MAAM8B,EAAAA,aAAa,KAAK,GAAG,QAAQ,KAAO,MAAO,EAAG,CAAC,UAAW,KAAK,GAAG,QAAQ,KAAK,EAC9F,KAAM,CAAC,KAAAC,CAAA,EAAQ,MAAM/B,EAAO,UAAUT,CAAI,EAC1C,MAAMS,EAAO,UAAA,EACb/E,EAAI8G,EAAK,KAAK,KAAA,GAAU,IAAI,CAC7B,CAAC,EACD,OAAO,OAAO,OAAOvC,EAAG,CAAC,MAAO,IAAMQ,GAAQ,UAAA,EAAY,CAC3D,CACD,CCIO,MAAMgC,CAAG,CAQf,YAA4BjH,EAAoB,CAApB,KAAA,QAAAA,EACvBA,EAAQ,OAAMA,EAAQ,KAAOkH,EAAG,OAAA,GACpC,QAAQ,IAAI,mBAAqBlH,EAAQ,KACzC,KAAK,MAAQ,IAAIuG,EAAM,IAAI,EAC3B,KAAK,SAAW,IAAIrE,EAAI,IAAI,EAC5B,KAAK,OAAS,IAAI4E,EAAO,IAAI,CAC9B,CAZA,MAEA,SAEA,MASD,CCHO,MAAMK,EAAkB,CAC9B,KAAM,MACN,YAAa,qDACb,KAAM,CAAC,QAAS,CAAC,KAAM,SAAU,YAAa,iBAAkB,SAAU,GAAI,EAC9E,GAAKlF,GAA4BmF,EAAAA,IAAInF,EAAK,OAAO,EAClD,EAEaoF,EAAuB,CACnC,KAAM,eACN,YAAa,8BACb,KAAM,CAAA,EACN,GAAI,SAAY,IAAI,KAAA,EAAO,YAAA,CAC5B,EAEaC,EAAmB,CAC/B,KAAM,OACN,YAAa,mBACb,KAAM,CACL,SAAU,CAAC,KAAM,SAAU,YAAa,qBAAsB,KAAM,CAAC,MAAO,OAAQ,QAAQ,EAAG,SAAU,EAAA,EACzG,KAAM,CAAC,KAAM,SAAU,YAAa,kBAAmB,SAAU,EAAA,CAAI,EAEtE,GAAI,MAAOrF,EAAMsF,EAAQpI,IAAO,CAC/B,GAAI,CACH,OAAO8C,EAAK,KAAA,CACX,IAAK,OACJ,OAAO,MAAMkF,EAAQ,GAAG,CAAC,QAASlF,EAAK,IAAA,EAAOsF,EAAQpI,CAAE,EACzD,IAAK,OACJ,OAAO,MAAMqI,EAAO,GAAG,CAAC,KAAMvF,EAAK,IAAA,EAAOsF,EAAQpI,CAAE,EACrD,IAAK,SACJ,OAAO,MAAMsI,EAAW,GAAG,CAAC,KAAMxF,EAAK,IAAA,EAAOsF,EAAQpI,CAAE,CACzD,CAEF,OAAQwB,EAAU,CACjB,MAAO,CAAC,MAAOA,GAAK,SAAWA,EAAI,UAAS,CAC7C,CACD,CACD,EAEa+G,EAAoB,CAChC,KAAM,QACN,YAAa,2BACb,KAAM,CACL,IAAK,CAAC,KAAM,SAAU,YAAa,eAAgB,SAAU,EAAA,EAC7D,OAAQ,CAAC,KAAM,SAAU,YAAa,qBAAsB,KAAM,CAAC,MAAO,OAAQ,MAAO,QAAQ,EAAG,QAAS,KAAA,EAC7G,QAAS,CAAC,KAAM,SAAU,YAAa,uBAAwB,QAAS,EAAC,EACzE,KAAM,CAAC,KAAM,SAAU,YAAa,mBAAA,CAAmB,EAExD,GAAKzF,GAKC,IAAI0F,EAAAA,KAAK,CAAC,IAAK1F,EAAK,IAAK,QAASA,EAAK,QAAQ,EAAE,QAAQ,CAAC,OAAQA,EAAK,QAAU,MAAO,KAAMA,EAAK,IAAA,CAAK,CAC/G,EAEauF,EAAiB,CAC7B,KAAM,kBACN,YAAa,8BACb,KAAM,CACL,KAAM,CAAC,KAAM,SAAU,YAAa,sBAAuB,SAAU,EAAA,CAAI,EAE1E,GAAI,MAAOvF,GAAyB,CACnC,MAAM2F,EAAUC,EAAAA,mBAAmB,IAAI,EACjCpH,EAAO,MAAMqH,KAAQ,CAAC,QAAAF,CAAA,EAAU3F,EAAK,KAAM,EAAI,EAAE,MAAOtB,GAAaiH,EAAQ,OAAO,MAAM,KAAKjH,CAAG,CAAC,EACzG,MAAO,CAAC,GAAGiH,EAAQ,OAAQ,OAAQnH,EAAM,OAAQ,OAAW,OAAQ,MAAA,CACrE,CACD,EAEagH,EAAqB,CACjC,KAAM,kBACN,YAAa,8BACb,KAAM,CACL,KAAM,CAAC,KAAM,SAAU,YAAa,sBAAuB,SAAU,EAAA,CAAI,EAE1E,GAAI,MAAOxF,IAA0B,CAAC,OAAQ8F,EAAAA,mBAAmB9F,EAAK,IAAI,GAAA,EAC3E,EAEa+F,EAA0B,CACtC,KAAM,eACN,YAAa,+FACb,KAAM,CACL,IAAK,CAAC,KAAM,SAAU,YAAa,8BAA+B,SAAU,EAAA,EAC5E,MAAO,CAAC,KAAM,SAAU,YAAa,iFAAA,CAAiF,EAEvH,GAAI,MAAO/F,GAAwC,CAClD,MAAMgG,EAAO,MAAM,MAAMhG,EAAK,IAAK,CAAC,QAAS,CAAC,aAAc,2CAAA,EAA6C,EACvG,KAAKiG,GAAKA,EAAE,MAAM,EAAE,MAAMvH,GAAO,CAAC,MAAM,IAAI,MAAM,oBAAoBA,EAAI,OAAO,EAAE,CAAC,CAAC,EAEjFyG,EAAIe,EAAQ,KAAKF,CAAI,EAC3Bb,EAAE,+HAA+H,EAAE,OAAA,EACnI,MAAMgB,EAAW,CAChB,MAAOhB,EAAE,2BAA2B,EAAE,KAAK,SAAS,GAAKA,EAAE,OAAO,EAAE,KAAA,GAAU,GAC9E,YAAaA,EAAE,0BAA0B,EAAE,KAAK,SAAS,GAAKA,EAAE,iCAAiC,EAAE,KAAK,SAAS,GAAK,EAAA,EAGvH,IAAIiB,EAAU,GACd,MAAMC,EAAmB,CAAC,UAAW,OAAQ,gBAAiB,WAAY,QAAS,SAAU,MAAM,EACnG,UAAWC,KAAYD,EAAkB,CACxC,MAAME,EAAKpB,EAAEmB,CAAQ,EAAE,MAAA,EACvB,GAAIC,EAAG,QAAUA,EAAG,KAAA,EAAO,KAAA,EAAO,OAAS,IAAK,CAC/CH,EAAUG,EAAG,KAAA,EACb,KACD,CACD,CACA,OAAKH,IAASA,EAAUjB,EAAE,MAAM,EAAE,KAAA,GAClCiB,EAAUA,EAAQ,QAAQ,OAAQ,GAAG,EAAE,OAAO,MAAM,EAAG,GAAI,EAEpD,CAAC,IAAKpG,EAAK,IAAK,MAAOmG,EAAS,MAAM,KAAA,EAAQ,YAAaA,EAAS,YAAY,KAAA,EAAQ,QAAAC,EAAS,MAAOpG,EAAK,KAAA,CACrH,CACD,EAEawG,EAAwB,CACpC,KAAM,aACN,YAAa,0IACb,KAAM,CACL,MAAO,CAAC,KAAM,SAAU,YAAa,gBAAiB,SAAU,EAAA,EAChE,OAAQ,CAAC,KAAM,SAAU,YAAa,8BAA+B,QAAS,CAAA,CAAC,EAEhF,GAAI,MAAOxG,GAGL,CACL,MAAMgG,EAAO,MAAM,MAAM,uCAAuC,mBAAmBhG,EAAK,KAAK,CAAC,GAAI,CACjG,QAAS,CAAC,aAAc,4CAA6C,kBAAmB,gBAAA,CAAgB,CACxG,EAAE,KAAKxB,GAAQA,EAAK,MAAM,EAC3B,IAAIiI,EAAOC,EAAQ,8BACnB,MAAM1H,EAAU,IAAI2H,OACpB,MAAOF,EAAQC,EAAM,KAAKV,CAAI,KAAO,MAAM,CAC1C,IAAIY,EAAM,iBAAiB,KAAK,mBAAmBH,EAAM,CAAC,CAAC,CAAC,IAAI,CAAC,EAGjE,GAFGG,IAAKA,EAAM,mBAAmBA,CAAG,GACjCA,GAAK5H,EAAQ,IAAI4H,CAAG,EACpB5H,EAAQ,OAASgB,EAAK,QAAU,GAAI,KACxC,CACA,OAAOhB,CACR,CACD"}
|
package/dist/index.mjs
CHANGED
|
@@ -1,21 +1,20 @@
|
|
|
1
|
-
import * as
|
|
2
|
-
import { objectMap as _, JSONAttemptParse as
|
|
3
|
-
import { Anthropic as
|
|
4
|
-
import { OpenAI as
|
|
1
|
+
import * as M from "node:os";
|
|
2
|
+
import { objectMap as _, JSONAttemptParse as g, findByProp as k, JSONSanitize as b, clean as P, Http as $, consoleInterceptor as E, fn as A, ASet as v } from "@ztimson/utils";
|
|
3
|
+
import { Anthropic as O } from "@anthropic-ai/sdk";
|
|
4
|
+
import { OpenAI as U } from "openai";
|
|
5
5
|
import { Worker as x } from "worker_threads";
|
|
6
|
-
import { fileURLToPath as
|
|
7
|
-
import { join as
|
|
8
|
-
import R from "
|
|
9
|
-
import {
|
|
10
|
-
import { createWorker as D } from "tesseract.js";
|
|
6
|
+
import { fileURLToPath as j } from "url";
|
|
7
|
+
import { join as S, dirname as T } from "path";
|
|
8
|
+
import { canDiarization as R } from "./asr.mjs";
|
|
9
|
+
import { createWorker as L } from "tesseract.js";
|
|
11
10
|
import "./embedder.mjs";
|
|
12
|
-
import * as
|
|
13
|
-
import { $ as
|
|
14
|
-
class
|
|
11
|
+
import * as D from "cheerio";
|
|
12
|
+
import { $ as N, $Sync as C } from "@ztimson/node-utils";
|
|
13
|
+
class q {
|
|
15
14
|
}
|
|
16
|
-
class
|
|
15
|
+
class W extends q {
|
|
17
16
|
constructor(r, e, t) {
|
|
18
|
-
super(), this.ai = r, this.apiToken = e, this.model = t, this.client = new
|
|
17
|
+
super(), this.ai = r, this.apiToken = e, this.model = t, this.client = new O({ apiKey: e });
|
|
19
18
|
}
|
|
20
19
|
client;
|
|
21
20
|
toStandard(r) {
|
|
@@ -95,7 +94,7 @@ ${JSON.stringify(s, null, 2)}`, c;
|
|
|
95
94
|
} else c.delta.type === "input_json_delta" && (m.content.at(-1).input += c.delta.partial_json);
|
|
96
95
|
else if (c.type === "content_block_stop") {
|
|
97
96
|
const l = m.content.at(-1);
|
|
98
|
-
l.input != null && (l.input = l.input ?
|
|
97
|
+
l.input != null && (l.input = l.input ? g(l.input, {}) : {});
|
|
99
98
|
} else if (c.type === "message_stop")
|
|
100
99
|
break;
|
|
101
100
|
}
|
|
@@ -122,9 +121,9 @@ ${JSON.stringify(s, null, 2)}`, c;
|
|
|
122
121
|
}), { abort: () => t.abort() });
|
|
123
122
|
}
|
|
124
123
|
}
|
|
125
|
-
class w extends
|
|
124
|
+
class w extends q {
|
|
126
125
|
constructor(r, e, t, o) {
|
|
127
|
-
super(), this.ai = r, this.host = e, this.token = t, this.model = o, this.client = new
|
|
126
|
+
super(), this.ai = r, this.host = e, this.token = t, this.model = o, this.client = new U(P({
|
|
128
127
|
baseURL: e,
|
|
129
128
|
apiKey: t
|
|
130
129
|
}));
|
|
@@ -138,7 +137,7 @@ class w extends j {
|
|
|
138
137
|
role: "tool",
|
|
139
138
|
id: s.id,
|
|
140
139
|
name: s.function.name,
|
|
141
|
-
args:
|
|
140
|
+
args: g(s.function.arguments, {}),
|
|
142
141
|
timestamp: t.timestamp
|
|
143
142
|
}));
|
|
144
143
|
r.splice(e, 1, ...o), e += o.length - 1;
|
|
@@ -218,8 +217,8 @@ ${JSON.stringify(n, null, 2)}`, l;
|
|
|
218
217
|
const u = i?.find(k("name", p.function.name));
|
|
219
218
|
if (e.stream && e.stream({ tool: p.function.name }), !u) return { role: "tool", tool_call_id: p.id, content: '{"error": "Tool not found"}' };
|
|
220
219
|
try {
|
|
221
|
-
const f =
|
|
222
|
-
return { role: "tool", tool_call_id: p.id, content: b(
|
|
220
|
+
const f = g(p.function.arguments, {}), y = await u.fn(f, e.stream, this.ai);
|
|
221
|
+
return { role: "tool", tool_call_id: p.id, content: b(y) };
|
|
223
222
|
} catch (f) {
|
|
224
223
|
return { role: "tool", tool_call_id: p.id, content: b({ error: f?.message || f?.toString() || "Unknown" }) };
|
|
225
224
|
}
|
|
@@ -231,10 +230,10 @@ ${JSON.stringify(n, null, 2)}`, l;
|
|
|
231
230
|
}), { abort: () => t.abort() });
|
|
232
231
|
}
|
|
233
232
|
}
|
|
234
|
-
class
|
|
233
|
+
class z {
|
|
235
234
|
constructor(r) {
|
|
236
235
|
this.ai = r, r.options.llm?.models && Object.entries(r.options.llm.models).forEach(([e, t]) => {
|
|
237
|
-
this.defaultModel || (this.defaultModel = e), t.proto == "anthropic" ? this.models[e] = new
|
|
236
|
+
this.defaultModel || (this.defaultModel = e), t.proto == "anthropic" ? this.models[e] = new W(this.ai, t.token, e) : t.proto == "ollama" ? this.models[e] = new w(this.ai, t.host, "not-needed", e) : t.proto == "openai" && (this.models[e] = new w(this.ai, t.host || null, t.token, e));
|
|
238
237
|
});
|
|
239
238
|
}
|
|
240
239
|
models = {};
|
|
@@ -319,8 +318,8 @@ You have passive persistent memory never make any mention of your memory capabil
|
|
|
319
318
|
${a.map((u) => `${u.role}: ${u.content}`).join(`
|
|
320
319
|
|
|
321
320
|
`)}`, { model: o?.model, temperature: o?.temperature || 0.3 }), c = /* @__PURE__ */ new Date(), l = await Promise.all((d?.facts || [])?.map(async ([u, f]) => {
|
|
322
|
-
const
|
|
323
|
-
return { owner: u, fact: f, embeddings: [
|
|
321
|
+
const y = await Promise.all([this.embedding(u), this.embedding(`${u}: ${f}`)]);
|
|
322
|
+
return { owner: u, fact: f, embeddings: [y[0][0].embedding, y[1][0].embedding], timestamp: c };
|
|
324
323
|
})), p = [{ role: "assistant", content: `Conversation Summary: ${d?.summary}`, timestamp: Date.now() }, ...m];
|
|
325
324
|
return i && p.splice(0, 0, i), { history: p, memory: l };
|
|
326
325
|
}
|
|
@@ -375,7 +374,7 @@ ${a.map((u) => `${u.role}: ${u.content}`).join(`
|
|
|
375
374
|
*/
|
|
376
375
|
embedding(r, e = 500, t = 50) {
|
|
377
376
|
const o = (n) => new Promise((i, m) => {
|
|
378
|
-
const a = new x(
|
|
377
|
+
const a = new x(S(T(j(import.meta.url)), "embedder.js")), d = ({ embedding: l }) => {
|
|
379
378
|
a.terminate(), i(l);
|
|
380
379
|
}, c = (l) => {
|
|
381
380
|
a.terminate(), m(l);
|
|
@@ -421,7 +420,7 @@ ${a.map((u) => `${u.role}: ${u.content}`).join(`
|
|
|
421
420
|
let t = await this.ask(r, { system: "Respond using a JSON blob matching any provided examples", ...e });
|
|
422
421
|
if (!t) return {};
|
|
423
422
|
const o = /```(?:.+)?\s*([\s\S]*?)```/.exec(t), s = o ? o[1].trim() : t;
|
|
424
|
-
return
|
|
423
|
+
return g(s, {});
|
|
425
424
|
}
|
|
426
425
|
/**
|
|
427
426
|
* Create a summary of some text
|
|
@@ -434,31 +433,30 @@ ${a.map((u) => `${u.role}: ${u.content}`).join(`
|
|
|
434
433
|
return this.ask(r, { system: `Generate a brief summary <= ${e} tokens. Output nothing else`, temperature: 0.3, ...t });
|
|
435
434
|
}
|
|
436
435
|
}
|
|
437
|
-
class
|
|
436
|
+
class I {
|
|
438
437
|
constructor(r) {
|
|
439
438
|
this.ai = r;
|
|
440
439
|
}
|
|
441
440
|
asr(r, e = {}) {
|
|
442
|
-
console.log("audio", r);
|
|
443
441
|
const { model: t = this.ai.options.asr || "whisper-base", speaker: o = !1 } = e;
|
|
444
442
|
let s = !1;
|
|
445
443
|
const n = () => {
|
|
446
444
|
s = !0;
|
|
447
445
|
}, i = new Promise((m, a) => {
|
|
448
|
-
const d = new x(
|
|
446
|
+
const d = new x(S(T(j(import.meta.url)), "asr.js")), c = ({ text: p, warning: u, error: f }) => {
|
|
449
447
|
d.terminate(), !s && (f ? a(new Error(f)) : (u && console.warn(u), m(p)));
|
|
450
448
|
}, l = (p) => {
|
|
451
449
|
d.terminate(), s || a(p);
|
|
452
450
|
};
|
|
453
451
|
d.on("message", c), d.on("error", l), d.on("exit", (p) => {
|
|
454
452
|
p !== 0 && !s && a(new Error(`Worker exited with code ${p}`));
|
|
455
|
-
}), d.postMessage({ file: r, model: t, speaker: o, modelDir: this.ai.options.path });
|
|
453
|
+
}), d.postMessage({ file: r, model: t, speaker: o, modelDir: this.ai.options.path, token: this.ai.options.hfToken });
|
|
456
454
|
});
|
|
457
455
|
return Object.assign(i, { abort: n });
|
|
458
456
|
}
|
|
459
|
-
canDiarization =
|
|
457
|
+
canDiarization = R;
|
|
460
458
|
}
|
|
461
|
-
class
|
|
459
|
+
class J {
|
|
462
460
|
constructor(r) {
|
|
463
461
|
this.ai = r;
|
|
464
462
|
}
|
|
@@ -470,16 +468,16 @@ class H {
|
|
|
470
468
|
ocr(r) {
|
|
471
469
|
let e;
|
|
472
470
|
const t = new Promise(async (o) => {
|
|
473
|
-
e = await
|
|
471
|
+
e = await L(this.ai.options.ocr || "eng", 2, { cachePath: this.ai.options.path });
|
|
474
472
|
const { data: s } = await e.recognize(r);
|
|
475
473
|
await e.terminate(), o(s.text.trim() || null);
|
|
476
474
|
});
|
|
477
475
|
return Object.assign(t, { abort: () => e?.terminate() });
|
|
478
476
|
}
|
|
479
477
|
}
|
|
480
|
-
class
|
|
478
|
+
class se {
|
|
481
479
|
constructor(r) {
|
|
482
|
-
this.options = r, r.path || (r.path =
|
|
480
|
+
this.options = r, r.path || (r.path = M.tmpdir()), process.env.TRANSFORMERS_CACHE = r.path, this.audio = new I(this), this.language = new z(this), this.vision = new J(this);
|
|
483
481
|
}
|
|
484
482
|
/** Audio processing AI */
|
|
485
483
|
audio;
|
|
@@ -488,17 +486,17 @@ class oe {
|
|
|
488
486
|
/** Vision processing AI */
|
|
489
487
|
vision;
|
|
490
488
|
}
|
|
491
|
-
const
|
|
489
|
+
const H = {
|
|
492
490
|
name: "cli",
|
|
493
491
|
description: "Use the command line interface, returns any output",
|
|
494
492
|
args: { command: { type: "string", description: "Command to run", required: !0 } },
|
|
495
|
-
fn: (h) =>
|
|
496
|
-
},
|
|
493
|
+
fn: (h) => N`${h.command}`
|
|
494
|
+
}, ne = {
|
|
497
495
|
name: "get_datetime",
|
|
498
496
|
description: "Get current UTC date / time",
|
|
499
497
|
args: {},
|
|
500
498
|
fn: async () => (/* @__PURE__ */ new Date()).toUTCString()
|
|
501
|
-
},
|
|
499
|
+
}, oe = {
|
|
502
500
|
name: "exec",
|
|
503
501
|
description: "Run code/scripts",
|
|
504
502
|
args: {
|
|
@@ -509,17 +507,17 @@ const F = {
|
|
|
509
507
|
try {
|
|
510
508
|
switch (h.type) {
|
|
511
509
|
case "bash":
|
|
512
|
-
return await
|
|
510
|
+
return await H.fn({ command: h.code }, r, e);
|
|
513
511
|
case "node":
|
|
514
|
-
return await
|
|
512
|
+
return await F.fn({ code: h.code }, r, e);
|
|
515
513
|
case "python":
|
|
516
|
-
return await
|
|
514
|
+
return await G.fn({ code: h.code }, r, e);
|
|
517
515
|
}
|
|
518
516
|
} catch (t) {
|
|
519
517
|
return { error: t?.message || t.toString() };
|
|
520
518
|
}
|
|
521
519
|
}
|
|
522
|
-
},
|
|
520
|
+
}, ie = {
|
|
523
521
|
name: "fetch",
|
|
524
522
|
description: "Make HTTP request to URL",
|
|
525
523
|
args: {
|
|
@@ -528,25 +526,25 @@ const F = {
|
|
|
528
526
|
headers: { type: "object", description: "HTTP headers to send", default: {} },
|
|
529
527
|
body: { type: "object", description: "HTTP body to send" }
|
|
530
528
|
},
|
|
531
|
-
fn: (h) => new
|
|
532
|
-
},
|
|
529
|
+
fn: (h) => new $({ url: h.url, headers: h.headers }).request({ method: h.method || "GET", body: h.body })
|
|
530
|
+
}, F = {
|
|
533
531
|
name: "exec_javascript",
|
|
534
532
|
description: "Execute commonjs javascript",
|
|
535
533
|
args: {
|
|
536
534
|
code: { type: "string", description: "CommonJS javascript", required: !0 }
|
|
537
535
|
},
|
|
538
536
|
fn: async (h) => {
|
|
539
|
-
const r =
|
|
537
|
+
const r = E(null), e = await A({ console: r }, h.code, !0).catch((t) => r.output.error.push(t));
|
|
540
538
|
return { ...r.output, return: e, stdout: void 0, stderr: void 0 };
|
|
541
539
|
}
|
|
542
|
-
},
|
|
540
|
+
}, G = {
|
|
543
541
|
name: "exec_javascript",
|
|
544
542
|
description: "Execute commonjs javascript",
|
|
545
543
|
args: {
|
|
546
544
|
code: { type: "string", description: "CommonJS javascript", required: !0 }
|
|
547
545
|
},
|
|
548
|
-
fn: async (h) => ({ result:
|
|
549
|
-
},
|
|
546
|
+
fn: async (h) => ({ result: C`python -c "${h.code}"` })
|
|
547
|
+
}, ae = {
|
|
550
548
|
name: "read_webpage",
|
|
551
549
|
description: "Extract clean, structured content from a webpage. Use after web_search to read specific URLs",
|
|
552
550
|
args: {
|
|
@@ -556,7 +554,7 @@ const F = {
|
|
|
556
554
|
fn: async (h) => {
|
|
557
555
|
const r = await fetch(h.url, { headers: { "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64)" } }).then((n) => n.text()).catch((n) => {
|
|
558
556
|
throw new Error(`Failed to fetch: ${n.message}`);
|
|
559
|
-
}), e =
|
|
557
|
+
}), e = D.load(r);
|
|
560
558
|
e('script, style, nav, footer, header, aside, iframe, noscript, [role="navigation"], [role="banner"], .ad, .ads, .cookie, .popup').remove();
|
|
561
559
|
const t = {
|
|
562
560
|
title: e('meta[property="og:title"]').attr("content") || e("title").text() || "",
|
|
@@ -573,7 +571,7 @@ const F = {
|
|
|
573
571
|
}
|
|
574
572
|
return o || (o = e("body").text()), o = o.replace(/\s+/g, " ").trim().slice(0, 8e3), { url: h.url, title: t.title.trim(), description: t.description.trim(), content: o, focus: h.focus };
|
|
575
573
|
}
|
|
576
|
-
},
|
|
574
|
+
}, ce = {
|
|
577
575
|
name: "web_search",
|
|
578
576
|
description: "Use duckduckgo (anonymous) to find find relevant online resources. Returns a list of URLs that works great with the `read_webpage` tool",
|
|
579
577
|
args: {
|
|
@@ -585,7 +583,7 @@ const F = {
|
|
|
585
583
|
headers: { "User-Agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64)", "Accept-Language": "en-US,en;q=0.9" }
|
|
586
584
|
}).then((s) => s.text());
|
|
587
585
|
let e, t = /<a .*?href="(.+?)".+?<\/a>/g;
|
|
588
|
-
const o = new
|
|
586
|
+
const o = new v();
|
|
589
587
|
for (; (e = t.exec(r)) !== null; ) {
|
|
590
588
|
let s = /uddg=(.+)&?/.exec(decodeURIComponent(e[1]))?.[1];
|
|
591
589
|
if (s && (s = decodeURIComponent(s)), s && o.add(s), o.size >= (h.length || 5)) break;
|
|
@@ -594,20 +592,20 @@ const F = {
|
|
|
594
592
|
}
|
|
595
593
|
};
|
|
596
594
|
export {
|
|
597
|
-
|
|
598
|
-
|
|
599
|
-
|
|
600
|
-
|
|
601
|
-
|
|
602
|
-
|
|
603
|
-
|
|
604
|
-
|
|
605
|
-
|
|
595
|
+
se as Ai,
|
|
596
|
+
W as Anthropic,
|
|
597
|
+
I as Audio,
|
|
598
|
+
H as CliTool,
|
|
599
|
+
ne as DateTimeTool,
|
|
600
|
+
oe as ExecTool,
|
|
601
|
+
ie as FetchTool,
|
|
602
|
+
F as JSTool,
|
|
603
|
+
q as LLMProvider,
|
|
606
604
|
w as OpenAi,
|
|
607
|
-
|
|
608
|
-
|
|
609
|
-
|
|
610
|
-
|
|
611
|
-
|
|
605
|
+
G as PythonTool,
|
|
606
|
+
ae as ReadWebpageTool,
|
|
607
|
+
J as Vision,
|
|
608
|
+
ce as WebSearchTool,
|
|
609
|
+
R as canDiarization
|
|
612
610
|
};
|
|
613
611
|
//# sourceMappingURL=index.mjs.map
|