@elizaos/plugin-openai 2.0.0-alpha.14 → 2.0.0-alpha.16
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/browser/index.browser.js +2 -2
- package/dist/browser/index.browser.js.map +4 -4
- package/dist/cjs/index.node.cjs +18 -3
- package/dist/cjs/index.node.js.map +4 -4
- package/dist/models/text.d.ts.map +1 -1
- package/dist/node/index.node.js +18 -3
- package/dist/node/index.node.js.map +4 -4
- package/package.json +3 -3
|
@@ -1,3 +1,3 @@
|
|
|
1
|
-
import{logger as l,ModelType as P}from"@elizaos/core";import{logger as L}from"@elizaos/core";import{logger as _}from"@elizaos/core";function Ko(o){if(typeof process>"u"||!process.env)return;let n=process.env[o];return n===void 0?void 0:String(n)}function w(o,n,c){let r=o.getSetting(n);if(r!==void 0&&r!==null)return String(r);return Ko(n)??c}function x(o,n,c){let r=w(o,n);if(r===void 0)return c;let f=Number.parseInt(r,10);if(!Number.isFinite(f))throw Error(`Setting '${n}' must be a valid integer, got: ${r}`);return f}function v(o,n,c){let r=w(o,n);if(r===void 0)return c;let f=r.toLowerCase();return f==="true"||f==="1"||f==="yes"}function j(){return typeof globalThis<"u"&&typeof globalThis.document<"u"}function g(o){return j()&&!!w(o,"OPENAI_BROWSER_BASE_URL")}function y(o){return w(o,"OPENAI_API_KEY")}function xo(o){let n=w(o,"OPENAI_EMBEDDING_API_KEY");if(n)return _.debug("[OpenAI] Using specific embedding API key"),n;return _.debug("[OpenAI] Falling back to general API key for embeddings"),y(o)}function R(o,n=!1){if(j()&&!v(o,"OPENAI_ALLOW_BROWSER_API_KEY",!1))return{};let c=n?xo(o):y(o);return c?{Authorization:`Bearer ${c}`}:{}}function O(o){let n=w(o,"OPENAI_BROWSER_BASE_URL"),c=j()&&n?n:w(o,"OPENAI_BASE_URL")??"https://api.openai.com/v1";return _.debug(`[OpenAI] Base URL: ${c}`),c}function u(o){let n=j()?w(o,"OPENAI_BROWSER_EMBEDDING_URL")??w(o,"OPENAI_BROWSER_BASE_URL"):w(o,"OPENAI_EMBEDDING_URL");if(n)return _.debug(`[OpenAI] Using embedding base URL: ${n}`),n;return _.debug("[OpenAI] Falling back to general base URL for embeddings"),O(o)}function F(o){return w(o,"OPENAI_SMALL_MODEL")??w(o,"SMALL_MODEL")??"gpt-5-mini"}function d(o){return w(o,"OPENAI_LARGE_MODEL")??w(o,"LARGE_MODEL")??"gpt-5"}function m(o){return w(o,"OPENAI_EMBEDDING_MODEL")??"text-embedding-3-small"}function oo(o){return w(o,"OPENAI_IMAGE_DESCRIPTION_MODEL")??"gpt-5-mini"}function no(o){return w(o,"OPENAI_TRANSCRIPTION_MODEL")??"gpt-5-mini-transcribe"}function co(o){return w(o,"OPENAI_TTS_MODEL")??"tts-1"}function ro(o){return w(o,"OPENAI_TTS_VOICE")??"nova"}function fo(o){return w(o,"OPENAI_TTS_INSTRUCTIONS")??""}function To(o){return w(o,"OPENAI_IMAGE_MODEL")??"dall-e-3"}function io(o){return v(o,"OPENAI_EXPERIMENTAL_TELEMETRY",!1)}function ko(o){return x(o,"OPENAI_EMBEDDING_DIMENSIONS",1536)}function Ao(o){return x(o,"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS",8192)}function to(o){return w(o,"OPENAI_RESEARCH_MODEL")??"o3-deep-research"}function Io(o){return x(o,"OPENAI_RESEARCH_TIMEOUT",3600000)}globalThis.AI_SDK_LOG_WARNINGS??=!1;function Po(o,n){Xo(n)}async function Xo(o){if(j()){L.debug("[OpenAI] Skipping API validation in browser environment");return}if(!y(o)){L.warn("[OpenAI] OPENAI_API_KEY is not configured. OpenAI functionality will fail until a valid API key is provided.");return}try{let c=O(o),r=await fetch(`${c}/models`,{headers:R(o)});if(!r.ok){L.warn(`[OpenAI] API key validation failed: ${r.status} ${r.statusText}. Please verify your OPENAI_API_KEY is correct.`);return}}catch(c){let r=c instanceof Error?c.message:String(c);L.warn(`[OpenAI] API validation error: ${r}. OpenAI functionality may be limited.`)}}import{logger as U}from"@elizaos/core";import{logger as Do}from"@elizaos/core";var $={WAV:{HEADER:[82,73,70,70],IDENTIFIER:[87,65,86,69]},MP3_ID3:[73,68,51],OGG:[79,103,103,83],FLAC:[102,76,97,67],FTYP:[102,116,121,112],WEBM_EBML:[26,69,223,163]},Wo=12;function b(o,n,c){for(let r=0;r<c.length;r++){let f=c[r];if(f===void 0||o[n+r]!==f)return!1}return!0}function X(o){if(o.length<Wo)return"application/octet-stream";if(b(o,0,$.WAV.HEADER)&&b(o,8,$.WAV.IDENTIFIER))return"audio/wav";let n=o[0],c=o[1];if(b(o,0,$.MP3_ID3)||n===255&&c!==void 0&&(c&224)===224)return"audio/mpeg";if(b(o,0,$.OGG))return"audio/ogg";if(b(o,0,$.FLAC))return"audio/flac";if(b(o,4,$.FTYP))return"audio/mp4";if(b(o,0,$.WEBM_EBML))return"audio/webm";return Do.warn("Could not detect audio format from buffer, using generic binary type"),"application/octet-stream"}function Co(o){switch(o){case"audio/wav":return"wav";case"audio/mpeg":return"mp3";case"audio/ogg":return"ogg";case"audio/flac":return"flac";case"audio/mp4":return"m4a";case"audio/webm":return"webm";case"application/octet-stream":return"bin"}}function po(o){return`recording.${Co(o)}`}function so(o){return o instanceof Blob||o instanceof File}function D(o){return Buffer.isBuffer(o)}function Ho(o){return typeof o==="object"&&o!==null&&"audio"in o&&(so(o.audio)||D(o.audio))}function Vo(o){return typeof o==="object"&&o!==null&&"audioUrl"in o&&typeof o.audioUrl==="string"}async function wo(o){let n=await fetch(o);if(!n.ok)throw Error(`Failed to fetch audio from URL: ${n.status}`);return n.blob()}async function W(o,n){let c=no(o),r,f={};if(typeof n==="string")U.debug(`[OpenAI] Fetching audio from URL: ${n}`),r=await wo(n);else if(so(n))r=n;else if(D(n)){let I=X(n);U.debug(`[OpenAI] Auto-detected audio MIME type: ${I}`),r=new Blob([new Uint8Array(n)],{type:I})}else if(Ho(n)){if(f=n,n.model)c=n.model;if(D(n.audio)){let I=n.mimeType??X(n.audio);U.debug(`[OpenAI] Using MIME type: ${I}`),r=new Blob([new Uint8Array(n.audio)],{type:I})}else r=n.audio}else if(Vo(n))U.debug(`[OpenAI] Fetching audio from URL: ${n.audioUrl}`),r=await wo(n.audioUrl),f={prompt:n.prompt};else throw Error("TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object");U.debug(`[OpenAI] Using TRANSCRIPTION model: ${c}`);let T=r.type||"audio/webm",A=r.name||po(T.startsWith("audio/")?T:"audio/webm"),k=new FormData;if(k.append("file",r,A),k.append("model",c),f.language)k.append("language",f.language);if(f.responseFormat)k.append("response_format",f.responseFormat);if(f.prompt)k.append("prompt",f.prompt);if(f.temperature!==void 0)k.append("temperature",String(f.temperature));if(f.timestampGranularities)for(let I of f.timestampGranularities)k.append("timestamp_granularities[]",I);let t=O(o),i=await fetch(`${t}/audio/transcriptions`,{method:"POST",headers:R(o),body:k});if(!i.ok){let I=await i.text().catch(()=>"Unknown error");throw Error(`OpenAI transcription failed: ${i.status} ${i.statusText} - ${I}`)}return(await i.json()).text}async function C(o,n){let c,r,f="mp3",T,A;if(typeof n==="string")c=n,r=void 0;else{if(c=n.text,r=n.voice,"format"in n&&n.format)f=n.format;if("model"in n&&n.model)T=n.model;if("instructions"in n&&n.instructions)A=n.instructions}if(T=T??co(o),r=r??ro(o),A=A??fo(o),U.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${T}`),!c||c.trim().length===0)throw Error("TEXT_TO_SPEECH requires non-empty text");if(c.length>4096)throw Error("TEXT_TO_SPEECH text exceeds 4096 character limit");let k=["alloy","echo","fable","onyx","nova","shimmer"];if(r&&!k.includes(r))throw Error(`Invalid voice: ${r}. Must be one of: ${k.join(", ")}`);let t=O(o),i={model:T,voice:r,input:c,response_format:f};if(A&&A.length>0)i.instructions=A;let p=await fetch(`${t}/audio/speech`,{method:"POST",headers:{...R(o),"Content-Type":"application/json",...f==="mp3"?{Accept:"audio/mpeg"}:{}},body:JSON.stringify(i)});if(!p.ok){let I=await p.text().catch(()=>"Unknown error");throw Error(`OpenAI TTS failed: ${p.status} ${p.statusText} - ${I}`)}return p.arrayBuffer()}import{logger as K,ModelType as Yo,VECTOR_DIMS as qo}from"@elizaos/core";import{EventType as Qo}from"@elizaos/core";var Oo=200;function Go(o){if(o.length<=Oo)return o;return`${o.slice(0,Oo)}…`}function Zo(o){if("promptTokens"in o){let n="promptTokensDetails"in o?o.promptTokensDetails:void 0,c=o.cachedPromptTokens??n?.cachedTokens;return{promptTokens:o.promptTokens??0,completionTokens:o.completionTokens??0,totalTokens:o.totalTokens??(o.promptTokens??0)+(o.completionTokens??0),cachedPromptTokens:c}}if("inputTokens"in o||"outputTokens"in o){let n=o.inputTokens??0,c=o.outputTokens??0,r=o.totalTokens??n+c;return{promptTokens:n,completionTokens:c,totalTokens:r,cachedPromptTokens:o.cachedInputTokens}}return{promptTokens:0,completionTokens:0,totalTokens:0}}function J(o,n,c,r){let f=Zo(r),T={runtime:o,source:"openai",provider:"openai",type:n,prompt:Go(c),tokens:{prompt:f.promptTokens,completion:f.completionTokens,total:f.totalTokens,...f.cachedPromptTokens!==void 0?{cached:f.cachedPromptTokens}:{}}};o.emitEvent(Qo.MODEL_USED,T)}function Mo(o){let n=Object.values(qo);if(!n.includes(o))throw Error(`Invalid embedding dimension: ${o}. Must be one of: ${n.join(", ")}`);return o}function eo(o){if(o===null)return null;if(typeof o==="string")return o;if(typeof o==="object"&&typeof o.text==="string")return o.text;throw Error("Invalid embedding params: expected string, { text: string }, or null")}async function H(o,n){let c=m(o),r=Mo(ko(o)),f=eo(n);if(f===null){K.debug("[OpenAI] Creating test embedding for initialization");let N=Array(r).fill(0);return N[0]=0.1,N}let T=f.trim();if(T.length===0)throw Error("Cannot generate embedding for empty text");let A=32000;if(T.length>A)K.warn(`[OpenAI] Embedding input too long (~${Math.ceil(T.length/4)} tokens), truncating to ~8000 tokens`),T=T.slice(0,A);let t=`${u(o)}/embeddings`;K.debug(`[OpenAI] Generating embedding with model: ${c}`);let i=await fetch(t,{method:"POST",headers:{...R(o,!0),"Content-Type":"application/json"},body:JSON.stringify({model:c,input:T})});if(!i.ok){let N=await i.text().catch(()=>"Unknown error");throw Error(`OpenAI embedding API error: ${i.status} ${i.statusText} - ${N}`)}let p=await i.json(),I=p?.data?.[0];if(!I||!I.embedding)throw Error("OpenAI API returned invalid embedding response structure");let s=I.embedding;if(s.length!==r)throw Error(`Embedding dimension mismatch: got ${s.length}, expected ${r}. Check OPENAI_EMBEDDING_DIMENSIONS setting.`);if(p.usage)J(o,Yo.TEXT_EMBEDDING,T,{promptTokens:p.usage.prompt_tokens,completionTokens:0,totalTokens:p.usage.total_tokens});return K.debug(`[OpenAI] Generated embedding with ${s.length} dimensions`),s}import{logger as Ro,ModelType as Bo}from"@elizaos/core";var ho="Please analyze this image and provide a title and detailed description.";async function V(o,n){let c=To(o),r=n.count??1,f=n.size??"1024x1024",T=n;if(Ro.debug(`[OpenAI] Using IMAGE model: ${c}`),!n.prompt||n.prompt.trim().length===0)throw Error("IMAGE generation requires a non-empty prompt");if(r<1||r>10)throw Error("IMAGE count must be between 1 and 10");let A=O(o),k={model:c,prompt:n.prompt,n:r,size:f};if(T.quality)k.quality=T.quality;if(T.style)k.style=T.style;let t=await fetch(`${A}/images/generations`,{method:"POST",headers:{...R(o),"Content-Type":"application/json"},body:JSON.stringify(k)});if(!t.ok){let p=await t.text().catch(()=>"Unknown error");throw Error(`OpenAI image generation failed: ${t.status} ${t.statusText} - ${p}`)}let i=await t.json();if(!i.data||i.data.length===0)throw Error("OpenAI API returned no images");return i.data.map((p)=>({url:p.url,revisedPrompt:p.revised_prompt}))}function ao(o){return o.match(/title[:\s]+(.+?)(?:\n|$)/i)?.[1]?.trim()??"Image Analysis"}function vo(o){return o.replace(/title[:\s]+(.+?)(?:\n|$)/i,"").trim()}async function Q(o,n){let c=oo(o),r=Ao(o);Ro.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${c}`);let f,T;if(typeof n==="string")f=n,T=ho;else f=n.imageUrl,T=n.prompt??ho;if(!f||f.trim().length===0)throw Error("IMAGE_DESCRIPTION requires a valid image URL");let A=O(o),k={model:c,messages:[{role:"user",content:[{type:"text",text:T},{type:"image_url",image_url:{url:f}}]}],max_tokens:r},t=await fetch(`${A}/chat/completions`,{method:"POST",headers:{...R(o),"Content-Type":"application/json"},body:JSON.stringify(k)});if(!t.ok){let s=await t.text().catch(()=>"Unknown error");throw Error(`OpenAI image description failed: ${t.status} ${t.statusText} - ${s}`)}let i=await t.json();if(i.usage)J(o,Bo.IMAGE_DESCRIPTION,typeof n==="string"?n:n.prompt??"",{promptTokens:i.usage.prompt_tokens,completionTokens:i.usage.completion_tokens,totalTokens:i.usage.total_tokens});let I=i.choices?.[0]?.message?.content;if(!I)throw Error("OpenAI API returned empty image description");return{title:ao(I),description:vo(I)}}import{logger as Jo,ModelType as Eo}from"@elizaos/core";import{generateObject as on}from"ai";import{createOpenAI as lo}from"@ai-sdk/openai";var go="sk-proxy";function z(o){let n=O(o),c=y(o);if(!c&&g(o))return lo({apiKey:go,baseURL:n});if(!c)throw Error("OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.");return lo({apiKey:c,baseURL:n})}import{logger as No}from"@elizaos/core";import{JSONParseError as uo}from"ai";var mo={MARKDOWN_JSON:/```json\n|\n```|```/g,WHITESPACE:/^\s+|\s+$/g};function yo(){return async({text:o,error:n})=>{if(!(n instanceof uo))return null;try{let c=o.replace(mo.MARKDOWN_JSON,"");return JSON.parse(c),No.debug("[JSON Repair] Successfully repaired JSON by removing markdown wrappers"),c}catch{return No.warn("[JSON Repair] Unable to repair JSON text"),null}}}async function So(o,n,c,r){let f=z(o),T=r(o);if(Jo.debug(`[OpenAI] Using ${c} model: ${T}`),!n.prompt||n.prompt.trim().length===0)throw Error("Object generation requires a non-empty prompt");if(n.schema)Jo.debug("[OpenAI] Schema provided but using no-schema mode. Structure is determined by prompt instructions.");let A=f.chat(T),{object:k,usage:t}=await on({model:A,output:"no-schema",prompt:n.prompt,experimental_repairText:yo()});if(t)J(o,c,n.prompt,t);if(typeof k!=="object"||k===null)throw Error(`Object generation returned ${typeof k}, expected object`);return k}async function G(o,n){return So(o,n,Eo.OBJECT_SMALL,F)}async function Z(o,n){return So(o,n,Eo.OBJECT_LARGE,d)}import{logger as E}from"@elizaos/core";function nn(o){switch(o.type){case"web_search_preview":return{type:"web_search_preview"};case"file_search":return{type:"file_search",vector_store_ids:o.vectorStoreIds};case"code_interpreter":return{type:"code_interpreter",container:o.container??{type:"auto"}};case"mcp":return{type:"mcp",server_label:o.serverLabel,server_url:o.serverUrl,require_approval:o.requireApproval??"never"};default:throw Error(`Unknown research tool type: ${o.type}`)}}function cn(o){switch(o.type){case"web_search_call":return{id:o.id??"",type:"web_search_call",status:o.status??"completed",action:{type:o.action?.type??"search",query:o.action?.query,url:o.action?.url}};case"file_search_call":return{id:o.id??"",type:"file_search_call",status:o.status??"completed",query:o.query??"",results:o.results?.map((n)=>({fileId:n.file_id,fileName:n.file_name,score:n.score}))};case"code_interpreter_call":return{id:o.id??"",type:"code_interpreter_call",status:o.status??"completed",code:o.code??"",output:o.output};case"mcp_tool_call":return{id:o.id??"",type:"mcp_tool_call",status:o.status??"completed",serverLabel:o.server_label??"",toolName:o.tool_name??"",arguments:o.arguments??{},result:o.result};case"message":return{type:"message",content:o.content?.map((n)=>({type:"output_text",text:n.text,annotations:n.annotations?.map((c)=>({url:c.url,title:c.title,startIndex:c.start_index,endIndex:c.end_index}))??[]}))??[]};default:return null}}function rn(o){if(o.output_text){let r=[];if(o.output){for(let f of o.output)if(f.type==="message"&&f.content){for(let T of f.content)if(T.annotations)for(let A of T.annotations)r.push({url:A.url,title:A.title,startIndex:A.start_index,endIndex:A.end_index})}}return{text:o.output_text,annotations:r}}let n="",c=[];if(o.output){for(let r of o.output)if(r.type==="message"&&r.content){for(let f of r.content)if(n+=f.text,f.annotations)for(let T of f.annotations)c.push({url:T.url,title:T.title,startIndex:T.start_index,endIndex:T.end_index})}}return{text:n,annotations:c}}async function Y(o,n){let c=y(o);if(!c)throw Error("OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.");let r=O(o),f=n.model??to(o),T=Io(o);E.debug(`[OpenAI] Starting deep research with model: ${f}`),E.debug(`[OpenAI] Research input: ${n.input.substring(0,100)}...`);let A=n.tools?.filter((S)=>S.type==="web_search_preview"||S.type==="file_search"||S.type==="mcp");if(!A||A.length===0)E.debug("[OpenAI] No data source tools specified, defaulting to web_search_preview"),n.tools=[{type:"web_search_preview"},...n.tools??[]];let k={model:f,input:n.input};if(n.instructions)k.instructions=n.instructions;if(n.background!==void 0)k.background=n.background;if(n.tools&&n.tools.length>0)k.tools=n.tools.map(nn);if(n.maxToolCalls!==void 0)k.max_tool_calls=n.maxToolCalls;if(n.reasoningSummary)k.reasoning={summary:n.reasoningSummary};E.debug(`[OpenAI] Research request body: ${JSON.stringify(k,null,2)}`);let t=await fetch(`${r}/responses`,{method:"POST",headers:{Authorization:`Bearer ${c}`,"Content-Type":"application/json"},body:JSON.stringify(k),signal:AbortSignal.timeout(T)});if(!t.ok){let S=await t.text();throw E.error(`[OpenAI] Research request failed: ${t.status} ${S}`),Error(`Deep research request failed: ${t.status} ${t.statusText}`)}let i=await t.json();if(i.error)throw E.error(`[OpenAI] Research API error: ${i.error.message}`),Error(`Deep research error: ${i.error.message}`);E.debug(`[OpenAI] Research response received. Status: ${i.status??"completed"}`);let{text:p,annotations:I}=rn(i),s=[];if(i.output)for(let S of i.output){let a=cn(S);if(a)s.push(a)}let N={id:i.id,text:p,annotations:I,outputItems:s,status:i.status};return E.info(`[OpenAI] Research completed. Text length: ${p.length}, Annotations: ${I.length}, Output items: ${s.length}`),N}import{logger as fn,ModelType as $o}from"@elizaos/core";import{generateText as Tn,streamText as kn}from"ai";function An(o){if(!o)return;let n=o.inputTokens??0,c=o.outputTokens??0,r=o;return{promptTokens:n,completionTokens:c,totalTokens:n+c,cachedPromptTokens:r.cachedInputTokens}}function tn(o){let n=o;return{promptCacheKey:n.providerOptions?.openai?.promptCacheKey,promptCacheRetention:n.providerOptions?.openai?.promptCacheRetention}}async function bo(o,n,c,r){let f=z(o),T=r(o);fn.debug(`[OpenAI] Using ${c} model: ${T}`);let A=tn(n),k=o.character.system??void 0,i={model:f.chat(T),prompt:n.prompt,system:k,maxOutputTokens:n.maxTokens??8192,experimental_telemetry:{isEnabled:io(o)},...A.promptCacheKey||A.promptCacheRetention?{providerOptions:{openai:{...A.promptCacheKey?{promptCacheKey:A.promptCacheKey}:{},...A.promptCacheRetention?{promptCacheRetention:A.promptCacheRetention}:{}}}}:{}};if(n.stream){let s=kn(i);return{textStream:s.textStream,text:Promise.resolve(s.text),usage:Promise.resolve(s.usage).then(An),finishReason:Promise.resolve(s.finishReason).then((N)=>N)}}let{text:p,usage:I}=await Tn(i);if(I)J(o,c,n.prompt,I);return p}async function q(o,n){return bo(o,n,$o.TEXT_SMALL,F)}async function M(o,n){return bo(o,n,$o.TEXT_LARGE,d)}import{ModelType as zo}from"@elizaos/core";import{ModelType as In}from"@elizaos/core";import{encodingForModel as Pn,getEncoding as pn}from"js-tiktoken";function jo(o){let c=o.toLowerCase().includes("4o")?"o200k_base":"cl100k_base";try{return Pn(o)}catch{return pn(c)}}function Fo(o,n){if(n===In.TEXT_SMALL)return F(o);return d(o)}function Uo(o,n,c){let r=Fo(o,n);return jo(r).encode(c)}function _o(o,n,c){let r=Fo(o,n);return jo(r).decode(c)}async function e(o,n){if(!n.prompt)throw Error("Tokenization requires a non-empty prompt");let c=n.modelType??zo.TEXT_LARGE;return Uo(o,c,n.prompt)}async function B(o,n){if(!n.tokens||!Array.isArray(n.tokens))throw Error("Detokenization requires a valid tokens array");if(n.tokens.length===0)return"";for(let r=0;r<n.tokens.length;r++){let f=n.tokens[r];if(typeof f!=="number"||!Number.isFinite(f))throw Error(`Invalid token at index ${r}: expected number`)}let c=n.modelType??zo.TEXT_LARGE;return _o(o,c,n.tokens)}function wn(){if(typeof process>"u")return{};return process.env}var h=wn(),sn={name:"openai",description:"OpenAI API integration for text, image, audio, and embedding models",config:{OPENAI_API_KEY:h.OPENAI_API_KEY??null,OPENAI_BASE_URL:h.OPENAI_BASE_URL??null,OPENAI_SMALL_MODEL:h.OPENAI_SMALL_MODEL??null,OPENAI_LARGE_MODEL:h.OPENAI_LARGE_MODEL??null,SMALL_MODEL:h.SMALL_MODEL??null,LARGE_MODEL:h.LARGE_MODEL??null,OPENAI_EMBEDDING_MODEL:h.OPENAI_EMBEDDING_MODEL??null,OPENAI_EMBEDDING_API_KEY:h.OPENAI_EMBEDDING_API_KEY??null,OPENAI_EMBEDDING_URL:h.OPENAI_EMBEDDING_URL??null,OPENAI_EMBEDDING_DIMENSIONS:h.OPENAI_EMBEDDING_DIMENSIONS??null,OPENAI_IMAGE_DESCRIPTION_MODEL:h.OPENAI_IMAGE_DESCRIPTION_MODEL??null,OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS:h.OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS??null,OPENAI_EXPERIMENTAL_TELEMETRY:h.OPENAI_EXPERIMENTAL_TELEMETRY??null,OPENAI_RESEARCH_MODEL:h.OPENAI_RESEARCH_MODEL??null,OPENAI_RESEARCH_TIMEOUT:h.OPENAI_RESEARCH_TIMEOUT??null},async init(o,n){Po(o,n)},models:{[P.TEXT_EMBEDDING]:async(o,n)=>{return H(o,n)},[P.TEXT_TOKENIZER_ENCODE]:async(o,n)=>{return e(o,n)},[P.TEXT_TOKENIZER_DECODE]:async(o,n)=>{return B(o,n)},[P.TEXT_SMALL]:async(o,n)=>{return q(o,n)},[P.TEXT_LARGE]:async(o,n)=>{return M(o,n)},[P.IMAGE]:async(o,n)=>{return V(o,n)},[P.IMAGE_DESCRIPTION]:async(o,n)=>{return Q(o,n)},[P.TRANSCRIPTION]:async(o,n)=>{return W(o,n)},[P.TEXT_TO_SPEECH]:async(o,n)=>{return C(o,n)},[P.OBJECT_SMALL]:async(o,n)=>{return G(o,n)},[P.OBJECT_LARGE]:async(o,n)=>{return Z(o,n)},[P.RESEARCH]:async(o,n)=>{return Y(o,n)}},tests:[{name:"openai_plugin_tests",tests:[{name:"openai_test_api_connectivity",fn:async(o)=>{let n=O(o),c=await fetch(`${n}/models`,{headers:R(o)});if(!c.ok)throw Error(`API connectivity test failed: ${c.status} ${c.statusText}`);let r=await c.json();l.info(`[OpenAI Test] API connected. ${r.data?.length??0} models available.`)}},{name:"openai_test_text_embedding",fn:async(o)=>{let n=await o.useModel(P.TEXT_EMBEDDING,{text:"Hello, world!"});if(!Array.isArray(n)||n.length===0)throw Error("Embedding should return a non-empty array");l.info(`[OpenAI Test] Generated embedding with ${n.length} dimensions`)}},{name:"openai_test_text_small",fn:async(o)=>{let n=await o.useModel(P.TEXT_SMALL,{prompt:"Say hello in exactly 5 words."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_SMALL should return non-empty string");l.info(`[OpenAI Test] TEXT_SMALL generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_text_large",fn:async(o)=>{let n=await o.useModel(P.TEXT_LARGE,{prompt:"Explain quantum computing in 2 sentences."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_LARGE should return non-empty string");l.info(`[OpenAI Test] TEXT_LARGE generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_tokenizer_roundtrip",fn:async(o)=>{let c=await o.useModel(P.TEXT_TOKENIZER_ENCODE,{prompt:"Hello, tokenizer test!",modelType:P.TEXT_SMALL});if(!Array.isArray(c)||c.length===0)throw Error("Tokenization should return non-empty token array");let r=await o.useModel(P.TEXT_TOKENIZER_DECODE,{tokens:c,modelType:P.TEXT_SMALL});if(r!=="Hello, tokenizer test!")throw Error(`Tokenizer roundtrip failed: expected "Hello, tokenizer test!", got "${r}"`);l.info(`[OpenAI Test] Tokenizer roundtrip successful (${c.length} tokens)`)}},{name:"openai_test_streaming",fn:async(o)=>{let n=[],c=await o.useModel(P.TEXT_LARGE,{prompt:"Count from 1 to 5, one number per line.",stream:!0,onStreamChunk:(r)=>{n.push(r)}});if(typeof c!=="string"||c.length===0)throw Error("Streaming should return non-empty result");if(n.length===0)throw Error("No streaming chunks received");l.info(`[OpenAI Test] Streaming test: ${n.length} chunks received`)}},{name:"openai_test_image_description",fn:async(o)=>{let c=await o.useModel(P.IMAGE_DESCRIPTION,"https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Camponotus_flavomarginatus_ant.jpg/440px-Camponotus_flavomarginatus_ant.jpg");if(!c||typeof c!=="object"||!("title"in c)||!("description"in c))throw Error("Image description should return { title, description }");l.info(`[OpenAI Test] Image described: "${c.title}"`)}},{name:"openai_test_transcription",fn:async(o)=>{let r=await(await fetch("https://upload.wikimedia.org/wikipedia/commons/2/25/En-Open_Source.ogg")).arrayBuffer(),f=Buffer.from(new Uint8Array(r)),T=await o.useModel(P.TRANSCRIPTION,f);if(typeof T!=="string")throw Error("Transcription should return a string");l.info(`[OpenAI Test] Transcription: "${T.substring(0,50)}..."`)}},{name:"openai_test_text_to_speech",fn:async(o)=>{let n=await o.useModel(P.TEXT_TO_SPEECH,{text:"Hello, this is a text-to-speech test."});if(!(n instanceof ArrayBuffer)||n.byteLength===0)throw Error("TTS should return non-empty ArrayBuffer");l.info(`[OpenAI Test] TTS generated ${n.byteLength} bytes of audio`)}},{name:"openai_test_object_generation",fn:async(o)=>{let n=await o.useModel(P.OBJECT_SMALL,{prompt:"Return a JSON object with exactly these fields: name (string), age (number), active (boolean)"});if(!n||typeof n!=="object")throw Error("Object generation should return an object");l.info(`[OpenAI Test] Object generated: ${JSON.stringify(n).substring(0,100)}`)}},{name:"openai_test_research",fn:async(o)=>{let n=await o.useModel(P.RESEARCH,{input:"What is the current date and time?",tools:[{type:"web_search_preview"}],maxToolCalls:3});if(!n||typeof n!=="object"||!("text"in n))throw Error("Research should return an object with text property");if(typeof n.text!=="string"||n.text.length===0)throw Error("Research result text should be a non-empty string");l.info(`[OpenAI Test] Research completed. Text length: ${n.text.length}, Annotations: ${n.annotations?.length??0}`)}}]}]},Lo=sn;var Uc=Lo;export{sn as openaiPlugin,Uc as default};
|
|
1
|
+
import{logger as h,ModelType as P}from"@elizaos/core";import{logger as K}from"@elizaos/core";import{logger as z}from"@elizaos/core";function xo(o){if(typeof process>"u"||!process.env)return;let n=process.env[o];return n===void 0?void 0:String(n)}function p(o,n,c){let r=o.getSetting(n);if(r!==void 0&&r!==null)return String(r);return xo(n)??c}function X(o,n,c){let r=p(o,n);if(r===void 0)return c;let f=Number.parseInt(r,10);if(!Number.isFinite(f))throw Error(`Setting '${n}' must be a valid integer, got: ${r}`);return f}function e(o,n,c){let r=p(o,n);if(r===void 0)return c;let f=r.toLowerCase();return f==="true"||f==="1"||f==="yes"}function F(){return typeof globalThis<"u"&&typeof globalThis.document<"u"}function a(o){return F()&&!!p(o,"OPENAI_BROWSER_BASE_URL")}function J(o){return p(o,"OPENAI_API_KEY")}function Xo(o){let n=p(o,"OPENAI_EMBEDDING_API_KEY");if(n)return z.debug("[OpenAI] Using specific embedding API key"),n;return z.debug("[OpenAI] Falling back to general API key for embeddings"),J(o)}function t(o,n=!1){if(F()&&!e(o,"OPENAI_ALLOW_BROWSER_API_KEY",!1))return{};let c=n?Xo(o):J(o);return c?{Authorization:`Bearer ${c}`}:{}}function s(o){let n=p(o,"OPENAI_BROWSER_BASE_URL"),c=F()&&n?n:p(o,"OPENAI_BASE_URL")??"https://api.openai.com/v1";return z.debug(`[OpenAI] Base URL: ${c}`),c}function u(o){let n=F()?p(o,"OPENAI_BROWSER_EMBEDDING_URL")??p(o,"OPENAI_BROWSER_BASE_URL"):p(o,"OPENAI_EMBEDDING_URL");if(n)return z.debug(`[OpenAI] Using embedding base URL: ${n}`),n;return z.debug("[OpenAI] Falling back to general base URL for embeddings"),s(o)}function _(o){return p(o,"OPENAI_SMALL_MODEL")??p(o,"SMALL_MODEL")??"gpt-5-mini"}function d(o){return p(o,"OPENAI_LARGE_MODEL")??p(o,"LARGE_MODEL")??"gpt-5"}function m(o){return p(o,"OPENAI_EMBEDDING_MODEL")??"text-embedding-3-small"}function oo(o){return p(o,"OPENAI_IMAGE_DESCRIPTION_MODEL")??"gpt-5-mini"}function no(o){return p(o,"OPENAI_TRANSCRIPTION_MODEL")??"gpt-5-mini-transcribe"}function co(o){return p(o,"OPENAI_TTS_MODEL")??"tts-1"}function ro(o){return p(o,"OPENAI_TTS_VOICE")??"nova"}function fo(o){return p(o,"OPENAI_TTS_INSTRUCTIONS")??""}function To(o){return p(o,"OPENAI_IMAGE_MODEL")??"dall-e-3"}function io(o){return e(o,"OPENAI_EXPERIMENTAL_TELEMETRY",!1)}function ko(o){return X(o,"OPENAI_EMBEDDING_DIMENSIONS",1536)}function Ao(o){return X(o,"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS",8192)}function Io(o){return p(o,"OPENAI_RESEARCH_MODEL")??"o3-deep-research"}function Po(o){return X(o,"OPENAI_RESEARCH_TIMEOUT",3600000)}globalThis.AI_SDK_LOG_WARNINGS??=!1;function wo(o,n){Do(n)}async function Do(o){if(F()){K.debug("[OpenAI] Skipping API validation in browser environment");return}if(!J(o)){K.warn("[OpenAI] OPENAI_API_KEY is not configured. OpenAI functionality will fail until a valid API key is provided.");return}try{let c=s(o),r=await fetch(`${c}/models`,{headers:t(o)});if(!r.ok){K.warn(`[OpenAI] API key validation failed: ${r.status} ${r.statusText}. Please verify your OPENAI_API_KEY is correct.`);return}}catch(c){let r=c instanceof Error?c.message:String(c);K.warn(`[OpenAI] API validation error: ${r}. OpenAI functionality may be limited.`)}}import{logger as U}from"@elizaos/core";import{logger as Ho}from"@elizaos/core";var b={WAV:{HEADER:[82,73,70,70],IDENTIFIER:[87,65,86,69]},MP3_ID3:[73,68,51],OGG:[79,103,103,83],FLAC:[102,76,97,67],FTYP:[102,116,121,112],WEBM_EBML:[26,69,223,163]},Wo=12;function j(o,n,c){for(let r=0;r<c.length;r++){let f=c[r];if(f===void 0||o[n+r]!==f)return!1}return!0}function D(o){if(o.length<Wo)return"application/octet-stream";if(j(o,0,b.WAV.HEADER)&&j(o,8,b.WAV.IDENTIFIER))return"audio/wav";let n=o[0],c=o[1];if(j(o,0,b.MP3_ID3)||n===255&&c!==void 0&&(c&224)===224)return"audio/mpeg";if(j(o,0,b.OGG))return"audio/ogg";if(j(o,0,b.FLAC))return"audio/flac";if(j(o,4,b.FTYP))return"audio/mp4";if(j(o,0,b.WEBM_EBML))return"audio/webm";return Ho.warn("Could not detect audio format from buffer, using generic binary type"),"application/octet-stream"}function Vo(o){switch(o){case"audio/wav":return"wav";case"audio/mpeg":return"mp3";case"audio/ogg":return"ogg";case"audio/flac":return"flac";case"audio/mp4":return"m4a";case"audio/webm":return"webm";case"application/octet-stream":return"bin"}}function Oo(o){return`recording.${Vo(o)}`}function so(o){return o instanceof Blob||o instanceof File}function H(o){return Buffer.isBuffer(o)}function Qo(o){return typeof o==="object"&&o!==null&&"audio"in o&&(so(o.audio)||H(o.audio))}function Co(o){return typeof o==="object"&&o!==null&&"audioUrl"in o&&typeof o.audioUrl==="string"}async function po(o){let n=await fetch(o);if(!n.ok)throw Error(`Failed to fetch audio from URL: ${n.status}`);return n.blob()}async function W(o,n){let c=no(o),r,f={};if(typeof n==="string")U.debug(`[OpenAI] Fetching audio from URL: ${n}`),r=await po(n);else if(so(n))r=n;else if(H(n)){let O=D(n);U.debug(`[OpenAI] Auto-detected audio MIME type: ${O}`),r=new Blob([new Uint8Array(n)],{type:O})}else if(Qo(n)){if(f=n,n.model)c=n.model;if(H(n.audio)){let O=n.mimeType??D(n.audio);U.debug(`[OpenAI] Using MIME type: ${O}`),r=new Blob([new Uint8Array(n.audio)],{type:O})}else r=n.audio}else if(Co(n))U.debug(`[OpenAI] Fetching audio from URL: ${n.audioUrl}`),r=await po(n.audioUrl),f={prompt:n.prompt};else throw Error("TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object");U.debug(`[OpenAI] Using TRANSCRIPTION model: ${c}`);let T=r.type||"audio/webm",A=r.name||Oo(T.startsWith("audio/")?T:"audio/webm"),i=new FormData;if(i.append("file",r,A),i.append("model",c),f.language)i.append("language",f.language);if(f.responseFormat)i.append("response_format",f.responseFormat);if(f.prompt)i.append("prompt",f.prompt);if(f.temperature!==void 0)i.append("temperature",String(f.temperature));if(f.timestampGranularities)for(let O of f.timestampGranularities)i.append("timestamp_granularities[]",O);let I=s(o),k=await fetch(`${I}/audio/transcriptions`,{method:"POST",headers:t(o),body:i});if(!k.ok){let O=await k.text().catch(()=>"Unknown error");throw Error(`OpenAI transcription failed: ${k.status} ${k.statusText} - ${O}`)}return(await k.json()).text}async function V(o,n){let c,r,f="mp3",T,A;if(typeof n==="string")c=n,r=void 0;else{if(c=n.text,r=n.voice,"format"in n&&n.format)f=n.format;if("model"in n&&n.model)T=n.model;if("instructions"in n&&n.instructions)A=n.instructions}if(T=T??co(o),r=r??ro(o),A=A??fo(o),U.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${T}`),!c||c.trim().length===0)throw Error("TEXT_TO_SPEECH requires non-empty text");if(c.length>4096)throw Error("TEXT_TO_SPEECH text exceeds 4096 character limit");let i=["alloy","echo","fable","onyx","nova","shimmer"];if(r&&!i.includes(r))throw Error(`Invalid voice: ${r}. Must be one of: ${i.join(", ")}`);let I=s(o),k={model:T,voice:r,input:c,response_format:f};if(A&&A.length>0)k.instructions=A;let w=await fetch(`${I}/audio/speech`,{method:"POST",headers:{...t(o),"Content-Type":"application/json",...f==="mp3"?{Accept:"audio/mpeg"}:{}},body:JSON.stringify(k)});if(!w.ok){let O=await w.text().catch(()=>"Unknown error");throw Error(`OpenAI TTS failed: ${w.status} ${w.statusText} - ${O}`)}return w.arrayBuffer()}import{logger as x,ModelType as qo,VECTOR_DIMS as Mo}from"@elizaos/core";import{EventType as Go}from"@elizaos/core";var Ro=200;function Zo(o){if(o.length<=Ro)return o;return`${o.slice(0,Ro)}…`}function Yo(o){if("promptTokens"in o){let n="promptTokensDetails"in o?o.promptTokensDetails:void 0,c=o.cachedPromptTokens??n?.cachedTokens;return{promptTokens:o.promptTokens??0,completionTokens:o.completionTokens??0,totalTokens:o.totalTokens??(o.promptTokens??0)+(o.completionTokens??0),cachedPromptTokens:c}}if("inputTokens"in o||"outputTokens"in o){let n=o.inputTokens??0,c=o.outputTokens??0,r=o.totalTokens??n+c;return{promptTokens:n,completionTokens:c,totalTokens:r,cachedPromptTokens:o.cachedInputTokens}}return{promptTokens:0,completionTokens:0,totalTokens:0}}function E(o,n,c,r){let f=Yo(r),T={runtime:o,source:"openai",provider:"openai",type:n,prompt:Zo(c),tokens:{prompt:f.promptTokens,completion:f.completionTokens,total:f.totalTokens,...f.cachedPromptTokens!==void 0?{cached:f.cachedPromptTokens}:{}}};o.emitEvent(Go.MODEL_USED,T)}function Bo(o){let n=Object.values(Mo);if(!n.includes(o))throw Error(`Invalid embedding dimension: ${o}. Must be one of: ${n.join(", ")}`);return o}function vo(o){if(o===null)return null;if(typeof o==="string")return o;if(typeof o==="object"&&typeof o.text==="string")return o.text;throw Error("Invalid embedding params: expected string, { text: string }, or null")}async function Q(o,n){let c=m(o),r=Bo(ko(o)),f=vo(n);if(f===null){x.debug("[OpenAI] Creating test embedding for initialization");let y=Array(r).fill(0);return y[0]=0.1,y}let T=f.trim();if(T.length===0)throw Error("Cannot generate embedding for empty text");let A=32000;if(T.length>A)x.warn(`[OpenAI] Embedding input too long (~${Math.ceil(T.length/4)} tokens), truncating to ~8000 tokens`),T=T.slice(0,A);let I=`${u(o)}/embeddings`;x.debug(`[OpenAI] Generating embedding with model: ${c}`);let k=await fetch(I,{method:"POST",headers:{...t(o,!0),"Content-Type":"application/json"},body:JSON.stringify({model:c,input:T})});if(!k.ok){let y=await k.text().catch(()=>"Unknown error");throw Error(`OpenAI embedding API error: ${k.status} ${k.statusText} - ${y}`)}let w=await k.json(),O=w?.data?.[0];if(!O?.embedding)throw Error("OpenAI API returned invalid embedding response structure");let l=O.embedding;if(l.length!==r)throw Error(`Embedding dimension mismatch: got ${l.length}, expected ${r}. Check OPENAI_EMBEDDING_DIMENSIONS setting.`);if(w.usage)E(o,qo.TEXT_EMBEDDING,T,{promptTokens:w.usage.prompt_tokens,completionTokens:0,totalTokens:w.usage.total_tokens});return x.debug(`[OpenAI] Generated embedding with ${l.length} dimensions`),l}import{logger as to,ModelType as go}from"@elizaos/core";var lo="Please analyze this image and provide a title and detailed description.";async function C(o,n){let c=To(o),r=n.count??1,f=n.size??"1024x1024",T=n;if(to.debug(`[OpenAI] Using IMAGE model: ${c}`),!n.prompt||n.prompt.trim().length===0)throw Error("IMAGE generation requires a non-empty prompt");if(r<1||r>10)throw Error("IMAGE count must be between 1 and 10");let A=s(o),i={model:c,prompt:n.prompt,n:r,size:f};if(T.quality)i.quality=T.quality;if(T.style)i.style=T.style;let I=await fetch(`${A}/images/generations`,{method:"POST",headers:{...t(o),"Content-Type":"application/json"},body:JSON.stringify(i)});if(!I.ok){let w=await I.text().catch(()=>"Unknown error");throw Error(`OpenAI image generation failed: ${I.status} ${I.statusText} - ${w}`)}let k=await I.json();if(!k.data||k.data.length===0)throw Error("OpenAI API returned no images");return k.data.map((w)=>({url:w.url,revisedPrompt:w.revised_prompt}))}function eo(o){return o.match(/title[:\s]+(.+?)(?:\n|$)/i)?.[1]?.trim()??"Image Analysis"}function ao(o){return o.replace(/title[:\s]+(.+?)(?:\n|$)/i,"").trim()}async function G(o,n){let c=oo(o),r=Ao(o);to.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${c}`);let f,T;if(typeof n==="string")f=n,T=lo;else f=n.imageUrl,T=n.prompt??lo;if(!f||f.trim().length===0)throw Error("IMAGE_DESCRIPTION requires a valid image URL");let A=s(o),i={model:c,messages:[{role:"user",content:[{type:"text",text:T},{type:"image_url",image_url:{url:f}}]}],max_tokens:r},I=await fetch(`${A}/chat/completions`,{method:"POST",headers:{...t(o),"Content-Type":"application/json"},body:JSON.stringify(i)});if(!I.ok){let l=await I.text().catch(()=>"Unknown error");throw Error(`OpenAI image description failed: ${I.status} ${I.statusText} - ${l}`)}let k=await I.json();if(k.usage)E(o,go.IMAGE_DESCRIPTION,typeof n==="string"?n:n.prompt??"",{promptTokens:k.usage.prompt_tokens,completionTokens:k.usage.completion_tokens,totalTokens:k.usage.total_tokens});let O=k.choices?.[0]?.message?.content;if(!O)throw Error("OpenAI API returned empty image description");return{title:eo(O),description:ao(O)}}import{logger as Jo,ModelType as Eo}from"@elizaos/core";import{generateObject as nn}from"ai";import{createOpenAI as ho}from"@ai-sdk/openai";var uo="sk-proxy";function L(o){let n=s(o),c=J(o);if(!c&&a(o))return ho({apiKey:uo,baseURL:n});if(!c)throw Error("OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.");return ho({apiKey:c,baseURL:n})}import{logger as No}from"@elizaos/core";import{JSONParseError as mo}from"ai";var on={MARKDOWN_JSON:/```json\n|\n```|```/g,WHITESPACE:/^\s+|\s+$/g};function yo(){return async({text:o,error:n})=>{if(!(n instanceof mo))return null;try{let c=o.replace(on.MARKDOWN_JSON,"");return JSON.parse(c),No.debug("[JSON Repair] Successfully repaired JSON by removing markdown wrappers"),c}catch{return No.warn("[JSON Repair] Unable to repair JSON text"),null}}}async function So(o,n,c,r){let f=L(o),T=r(o);if(Jo.debug(`[OpenAI] Using ${c} model: ${T}`),!n.prompt||n.prompt.trim().length===0)throw Error("Object generation requires a non-empty prompt");if(n.schema)Jo.debug("[OpenAI] Schema provided but using no-schema mode. Structure is determined by prompt instructions.");let A=f.chat(T),{object:i,usage:I}=await nn({model:A,output:"no-schema",prompt:n.prompt,experimental_repairText:yo()});if(I)E(o,c,n.prompt,I);if(typeof i!=="object"||i===null)throw Error(`Object generation returned ${typeof i}, expected object`);return i}async function Z(o,n){return So(o,n,Eo.OBJECT_SMALL,_)}async function Y(o,n){return So(o,n,Eo.OBJECT_LARGE,d)}import{logger as S}from"@elizaos/core";function cn(o){switch(o.type){case"web_search_preview":return{type:"web_search_preview"};case"file_search":return{type:"file_search",vector_store_ids:o.vectorStoreIds};case"code_interpreter":return{type:"code_interpreter",container:o.container??{type:"auto"}};case"mcp":return{type:"mcp",server_label:o.serverLabel,server_url:o.serverUrl,require_approval:o.requireApproval??"never"};default:throw Error(`Unknown research tool type: ${o.type}`)}}function rn(o){switch(o.type){case"web_search_call":return{id:o.id??"",type:"web_search_call",status:o.status??"completed",action:{type:o.action?.type??"search",query:o.action?.query,url:o.action?.url}};case"file_search_call":return{id:o.id??"",type:"file_search_call",status:o.status??"completed",query:o.query??"",results:o.results?.map((n)=>({fileId:n.file_id,fileName:n.file_name,score:n.score}))};case"code_interpreter_call":return{id:o.id??"",type:"code_interpreter_call",status:o.status??"completed",code:o.code??"",output:o.output};case"mcp_tool_call":return{id:o.id??"",type:"mcp_tool_call",status:o.status??"completed",serverLabel:o.server_label??"",toolName:o.tool_name??"",arguments:o.arguments??{},result:o.result};case"message":return{type:"message",content:o.content?.map((n)=>({type:"output_text",text:n.text,annotations:n.annotations?.map((c)=>({url:c.url,title:c.title,startIndex:c.start_index,endIndex:c.end_index}))??[]}))??[]};default:return null}}function fn(o){if(o.output_text){let r=[];if(o.output){for(let f of o.output)if(f.type==="message"&&f.content){for(let T of f.content)if(T.annotations)for(let A of T.annotations)r.push({url:A.url,title:A.title,startIndex:A.start_index,endIndex:A.end_index})}}return{text:o.output_text,annotations:r}}let n="",c=[];if(o.output){for(let r of o.output)if(r.type==="message"&&r.content){for(let f of r.content)if(n+=f.text,f.annotations)for(let T of f.annotations)c.push({url:T.url,title:T.title,startIndex:T.start_index,endIndex:T.end_index})}}return{text:n,annotations:c}}async function q(o,n){let c=J(o);if(!c)throw Error("OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.");let r=s(o),f=n.model??Io(o),T=Po(o);S.debug(`[OpenAI] Starting deep research with model: ${f}`),S.debug(`[OpenAI] Research input: ${n.input.substring(0,100)}...`);let A=n.tools?.filter((N)=>N.type==="web_search_preview"||N.type==="file_search"||N.type==="mcp");if(!A||A.length===0)S.debug("[OpenAI] No data source tools specified, defaulting to web_search_preview"),n.tools=[{type:"web_search_preview"},...n.tools??[]];let i={model:f,input:n.input};if(n.instructions)i.instructions=n.instructions;if(n.background!==void 0)i.background=n.background;if(n.tools&&n.tools.length>0)i.tools=n.tools.map(cn);if(n.maxToolCalls!==void 0)i.max_tool_calls=n.maxToolCalls;if(n.reasoningSummary)i.reasoning={summary:n.reasoningSummary};S.debug(`[OpenAI] Research request body: ${JSON.stringify(i,null,2)}`);let I=await fetch(`${r}/responses`,{method:"POST",headers:{Authorization:`Bearer ${c}`,"Content-Type":"application/json"},body:JSON.stringify(i),signal:AbortSignal.timeout(T)});if(!I.ok){let N=await I.text();throw S.error(`[OpenAI] Research request failed: ${I.status} ${N}`),Error(`Deep research request failed: ${I.status} ${I.statusText}`)}let k=await I.json();if(k.error)throw S.error(`[OpenAI] Research API error: ${k.error.message}`),Error(`Deep research error: ${k.error.message}`);S.debug(`[OpenAI] Research response received. Status: ${k.status??"completed"}`);let{text:w,annotations:O}=fn(k),l=[];if(k.output)for(let N of k.output){let $=rn(N);if($)l.push($)}let y={id:k.id,text:w,annotations:O,outputItems:l,status:k.status};return S.info(`[OpenAI] Research completed. Text length: ${w.length}, Annotations: ${O.length}, Output items: ${l.length}`),y}import{logger as Tn,ModelType as $o}from"@elizaos/core";import{generateText as kn,streamText as An}from"ai";function In(o){let n=[{type:"text",text:o.prompt}];for(let c of o.attachments??[])n.push({type:"file",data:c.data,mediaType:c.mediaType,...c.filename?{filename:c.filename}:{}});return n}function Pn(o){if(!o)return;let n=o.inputTokens??0,c=o.outputTokens??0,r=o;return{promptTokens:n,completionTokens:c,totalTokens:n+c,cachedPromptTokens:r.cachedInputTokens}}function wn(o){let n=o;return{promptCacheKey:n.providerOptions?.openai?.promptCacheKey,promptCacheRetention:n.providerOptions?.openai?.promptCacheRetention}}async function bo(o,n,c,r){let f=n,T=L(o),A=r(o);Tn.debug(`[OpenAI] Using ${c} model: ${A}`);let i=wn(n),k=(f.attachments?.length??0)>0?In(f):void 0,w=o.character.system??void 0,l={model:T.chat(A),...k?{messages:[{role:"user",content:k}]}:{prompt:n.prompt},system:w,maxOutputTokens:n.maxTokens??8192,experimental_telemetry:{isEnabled:io(o)},...i.promptCacheKey||i.promptCacheRetention?{providerOptions:{openai:{...i.promptCacheKey?{promptCacheKey:i.promptCacheKey}:{},...i.promptCacheRetention?{promptCacheRetention:i.promptCacheRetention}:{}}}}:{}};if(n.stream){let $=An(l);return{textStream:$.textStream,text:Promise.resolve($.text),usage:Promise.resolve($.usage).then(Pn),finishReason:Promise.resolve($.finishReason).then((Ko)=>Ko)}}let{text:y,usage:N}=await kn(l);if(N)E(o,c,n.prompt,N);return y}async function M(o,n){return bo(o,n,$o.TEXT_SMALL,_)}async function B(o,n){return bo(o,n,$o.TEXT_LARGE,d)}import{ModelType as zo}from"@elizaos/core";import{ModelType as On}from"@elizaos/core";import{encodingForModel as pn,getEncoding as sn}from"js-tiktoken";function jo(o){let c=o.toLowerCase().includes("4o")?"o200k_base":"cl100k_base";try{return pn(o)}catch{return sn(c)}}function Fo(o,n){if(n===On.TEXT_SMALL)return _(o);return d(o)}function _o(o,n,c){let r=Fo(o,n);return jo(r).encode(c)}function Uo(o,n,c){let r=Fo(o,n);return jo(r).decode(c)}async function v(o,n){if(!n.prompt)throw Error("Tokenization requires a non-empty prompt");let c=n.modelType??zo.TEXT_LARGE;return _o(o,c,n.prompt)}async function g(o,n){if(!n.tokens||!Array.isArray(n.tokens))throw Error("Detokenization requires a valid tokens array");if(n.tokens.length===0)return"";for(let r=0;r<n.tokens.length;r++){let f=n.tokens[r];if(typeof f!=="number"||!Number.isFinite(f))throw Error(`Invalid token at index ${r}: expected number`)}let c=n.modelType??zo.TEXT_LARGE;return Uo(o,c,n.tokens)}function Rn(){if(typeof process>"u")return{};return process.env}var R=Rn(),ln={name:"openai",description:"OpenAI API integration for text, image, audio, and embedding models",config:{OPENAI_API_KEY:R.OPENAI_API_KEY??null,OPENAI_BASE_URL:R.OPENAI_BASE_URL??null,OPENAI_SMALL_MODEL:R.OPENAI_SMALL_MODEL??null,OPENAI_LARGE_MODEL:R.OPENAI_LARGE_MODEL??null,SMALL_MODEL:R.SMALL_MODEL??null,LARGE_MODEL:R.LARGE_MODEL??null,OPENAI_EMBEDDING_MODEL:R.OPENAI_EMBEDDING_MODEL??null,OPENAI_EMBEDDING_API_KEY:R.OPENAI_EMBEDDING_API_KEY??null,OPENAI_EMBEDDING_URL:R.OPENAI_EMBEDDING_URL??null,OPENAI_EMBEDDING_DIMENSIONS:R.OPENAI_EMBEDDING_DIMENSIONS??null,OPENAI_IMAGE_DESCRIPTION_MODEL:R.OPENAI_IMAGE_DESCRIPTION_MODEL??null,OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS:R.OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS??null,OPENAI_EXPERIMENTAL_TELEMETRY:R.OPENAI_EXPERIMENTAL_TELEMETRY??null,OPENAI_RESEARCH_MODEL:R.OPENAI_RESEARCH_MODEL??null,OPENAI_RESEARCH_TIMEOUT:R.OPENAI_RESEARCH_TIMEOUT??null},async init(o,n){wo(o,n)},models:{[P.TEXT_EMBEDDING]:async(o,n)=>{return Q(o,n)},[P.TEXT_TOKENIZER_ENCODE]:async(o,n)=>{return v(o,n)},[P.TEXT_TOKENIZER_DECODE]:async(o,n)=>{return g(o,n)},[P.TEXT_SMALL]:async(o,n)=>{return M(o,n)},[P.TEXT_LARGE]:async(o,n)=>{return B(o,n)},[P.IMAGE]:async(o,n)=>{return C(o,n)},[P.IMAGE_DESCRIPTION]:async(o,n)=>{return G(o,n)},[P.TRANSCRIPTION]:async(o,n)=>{return W(o,n)},[P.TEXT_TO_SPEECH]:async(o,n)=>{return V(o,n)},[P.OBJECT_SMALL]:async(o,n)=>{return Z(o,n)},[P.OBJECT_LARGE]:async(o,n)=>{return Y(o,n)},[P.RESEARCH]:async(o,n)=>{return q(o,n)}},tests:[{name:"openai_plugin_tests",tests:[{name:"openai_test_api_connectivity",fn:async(o)=>{let n=s(o),c=await fetch(`${n}/models`,{headers:t(o)});if(!c.ok)throw Error(`API connectivity test failed: ${c.status} ${c.statusText}`);let r=await c.json();h.info(`[OpenAI Test] API connected. ${r.data?.length??0} models available.`)}},{name:"openai_test_text_embedding",fn:async(o)=>{let n=await o.useModel(P.TEXT_EMBEDDING,{text:"Hello, world!"});if(!Array.isArray(n)||n.length===0)throw Error("Embedding should return a non-empty array");h.info(`[OpenAI Test] Generated embedding with ${n.length} dimensions`)}},{name:"openai_test_text_small",fn:async(o)=>{let n=await o.useModel(P.TEXT_SMALL,{prompt:"Say hello in exactly 5 words."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_SMALL should return non-empty string");h.info(`[OpenAI Test] TEXT_SMALL generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_text_large",fn:async(o)=>{let n=await o.useModel(P.TEXT_LARGE,{prompt:"Explain quantum computing in 2 sentences."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_LARGE should return non-empty string");h.info(`[OpenAI Test] TEXT_LARGE generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_tokenizer_roundtrip",fn:async(o)=>{let c=await o.useModel(P.TEXT_TOKENIZER_ENCODE,{prompt:"Hello, tokenizer test!",modelType:P.TEXT_SMALL});if(!Array.isArray(c)||c.length===0)throw Error("Tokenization should return non-empty token array");let r=await o.useModel(P.TEXT_TOKENIZER_DECODE,{tokens:c,modelType:P.TEXT_SMALL});if(r!=="Hello, tokenizer test!")throw Error(`Tokenizer roundtrip failed: expected "Hello, tokenizer test!", got "${r}"`);h.info(`[OpenAI Test] Tokenizer roundtrip successful (${c.length} tokens)`)}},{name:"openai_test_streaming",fn:async(o)=>{let n=[],c=await o.useModel(P.TEXT_LARGE,{prompt:"Count from 1 to 5, one number per line.",stream:!0,onStreamChunk:(r)=>{n.push(r)}});if(typeof c!=="string"||c.length===0)throw Error("Streaming should return non-empty result");if(n.length===0)throw Error("No streaming chunks received");h.info(`[OpenAI Test] Streaming test: ${n.length} chunks received`)}},{name:"openai_test_image_description",fn:async(o)=>{let c=await o.useModel(P.IMAGE_DESCRIPTION,"https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Camponotus_flavomarginatus_ant.jpg/440px-Camponotus_flavomarginatus_ant.jpg");if(!c||typeof c!=="object"||!("title"in c)||!("description"in c))throw Error("Image description should return { title, description }");h.info(`[OpenAI Test] Image described: "${c.title}"`)}},{name:"openai_test_transcription",fn:async(o)=>{let r=await(await fetch("https://upload.wikimedia.org/wikipedia/commons/2/25/En-Open_Source.ogg")).arrayBuffer(),f=Buffer.from(new Uint8Array(r)),T=await o.useModel(P.TRANSCRIPTION,f);if(typeof T!=="string")throw Error("Transcription should return a string");h.info(`[OpenAI Test] Transcription: "${T.substring(0,50)}..."`)}},{name:"openai_test_text_to_speech",fn:async(o)=>{let n=await o.useModel(P.TEXT_TO_SPEECH,{text:"Hello, this is a text-to-speech test."});if(!(n instanceof ArrayBuffer)||n.byteLength===0)throw Error("TTS should return non-empty ArrayBuffer");h.info(`[OpenAI Test] TTS generated ${n.byteLength} bytes of audio`)}},{name:"openai_test_object_generation",fn:async(o)=>{let n=await o.useModel(P.OBJECT_SMALL,{prompt:"Return a JSON object with exactly these fields: name (string), age (number), active (boolean)"});if(!n||typeof n!=="object")throw Error("Object generation should return an object");h.info(`[OpenAI Test] Object generated: ${JSON.stringify(n).substring(0,100)}`)}},{name:"openai_test_research",fn:async(o)=>{let n=await o.useModel(P.RESEARCH,{input:"What is the current date and time?",tools:[{type:"web_search_preview"}],maxToolCalls:3});if(!n||typeof n!=="object"||!("text"in n))throw Error("Research should return an object with text property");if(typeof n.text!=="string"||n.text.length===0)throw Error("Research result text should be a non-empty string");h.info(`[OpenAI Test] Research completed. Text length: ${n.text.length}, Annotations: ${n.annotations?.length??0}`)}}]}]},Lo=ln;var zc=Lo;export{ln as openaiPlugin,zc as default};
|
|
2
2
|
|
|
3
|
-
//# debugId=
|
|
3
|
+
//# debugId=52010ABAEE8E9A2864756E2164756E21
|
|
@@ -7,19 +7,19 @@
|
|
|
7
7
|
"import type { IAgentRuntime } from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\n\nfunction getEnvValue(key: string): string | undefined {\n if (typeof process === \"undefined\" || !process.env) {\n return undefined;\n }\n const value = process.env[key];\n return value === undefined ? undefined : String(value);\n}\n\nexport function getSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue?: string\n): string | undefined {\n const value = runtime.getSetting(key);\n if (value !== undefined && value !== null) {\n return String(value);\n }\n return getEnvValue(key) ?? defaultValue;\n}\nexport function getRequiredSetting(\n runtime: IAgentRuntime,\n key: string,\n errorMessage?: string\n): string {\n const value = getSetting(runtime, key);\n if (value === undefined || value.trim() === \"\") {\n throw new Error(errorMessage ?? `Required setting '${key}' is not configured`);\n }\n return value;\n}\n\nexport function getNumericSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue: number\n): number {\n const value = getSetting(runtime, key);\n if (value === undefined) {\n return defaultValue;\n }\n const parsed = Number.parseInt(value, 10);\n if (!Number.isFinite(parsed)) {\n throw new Error(`Setting '${key}' must be a valid integer, got: ${value}`);\n }\n return parsed;\n}\n\nexport function getBooleanSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue: boolean\n): boolean {\n const value = getSetting(runtime, key);\n if (value === undefined) {\n return defaultValue;\n }\n const normalized = value.toLowerCase();\n return normalized === \"true\" || normalized === \"1\" || normalized === \"yes\";\n}\n\nexport function isBrowser(): boolean {\n return (\n typeof globalThis !== \"undefined\" &&\n typeof (globalThis as { document?: Document }).document !== \"undefined\"\n );\n}\n\nexport function isProxyMode(runtime: IAgentRuntime): boolean {\n return isBrowser() && !!getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n}\n\nexport function getApiKey(runtime: IAgentRuntime): string | undefined {\n return getSetting(runtime, \"OPENAI_API_KEY\");\n}\n\nexport function getEmbeddingApiKey(runtime: IAgentRuntime): string | undefined {\n const embeddingApiKey = getSetting(runtime, \"OPENAI_EMBEDDING_API_KEY\");\n if (embeddingApiKey) {\n logger.debug(\"[OpenAI] Using specific embedding API key\");\n return embeddingApiKey;\n }\n logger.debug(\"[OpenAI] Falling back to general API key for embeddings\");\n return getApiKey(runtime);\n}\n\nexport function getAuthHeader(\n runtime: IAgentRuntime,\n forEmbedding = false\n): Record<string, string> {\n // By default this plugin does NOT send auth headers in the browser. This is safer because\n // frontend builds would otherwise expose secrets. For local demos, you can explicitly\n // opt-in to sending the Authorization header by setting OPENAI_ALLOW_BROWSER_API_KEY=true.\n if (isBrowser() && !getBooleanSetting(runtime, \"OPENAI_ALLOW_BROWSER_API_KEY\", false)) {\n return {};\n }\n const key = forEmbedding ? getEmbeddingApiKey(runtime) : getApiKey(runtime);\n return key ? { Authorization: `Bearer ${key}` } : {};\n}\n\nexport function getBaseURL(runtime: IAgentRuntime): string {\n const browserURL = getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n const baseURL =\n isBrowser() && browserURL\n ? browserURL\n : (getSetting(runtime, \"OPENAI_BASE_URL\") ?? \"https://api.openai.com/v1\");\n logger.debug(`[OpenAI] Base URL: ${baseURL}`);\n return baseURL;\n}\n\nexport function getEmbeddingBaseURL(runtime: IAgentRuntime): string {\n const embeddingURL = isBrowser()\n ? (getSetting(runtime, \"OPENAI_BROWSER_EMBEDDING_URL\") ??\n getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\"))\n : getSetting(runtime, \"OPENAI_EMBEDDING_URL\");\n\n if (embeddingURL) {\n logger.debug(`[OpenAI] Using embedding base URL: ${embeddingURL}`);\n return embeddingURL;\n }\n\n logger.debug(\"[OpenAI] Falling back to general base URL for embeddings\");\n return getBaseURL(runtime);\n}\n\nexport function getSmallModel(runtime: IAgentRuntime): string {\n return (\n getSetting(runtime, \"OPENAI_SMALL_MODEL\") ?? getSetting(runtime, \"SMALL_MODEL\") ?? \"gpt-5-mini\"\n );\n}\n\nexport function getLargeModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_LARGE_MODEL\") ?? getSetting(runtime, \"LARGE_MODEL\") ?? \"gpt-5\";\n}\n\nexport function getEmbeddingModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_EMBEDDING_MODEL\") ?? \"text-embedding-3-small\";\n}\n\nexport function getImageDescriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MODEL\") ?? \"gpt-5-mini\";\n}\n\nexport function getTranscriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TRANSCRIPTION_MODEL\") ?? \"gpt-5-mini-transcribe\";\n}\n\nexport function getTTSModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_MODEL\") ?? \"tts-1\";\n}\n\nexport function getTTSVoice(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_VOICE\") ?? \"nova\";\n}\n\nexport function getTTSInstructions(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_INSTRUCTIONS\") ?? \"\";\n}\n\nexport function getImageModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_MODEL\") ?? \"dall-e-3\";\n}\n\nexport function getExperimentalTelemetry(runtime: IAgentRuntime): boolean {\n return getBooleanSetting(runtime, \"OPENAI_EXPERIMENTAL_TELEMETRY\", false);\n}\n\nexport function getEmbeddingDimensions(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_EMBEDDING_DIMENSIONS\", 1536);\n}\n\nexport function getImageDescriptionMaxTokens(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS\", 8192);\n}\n\nexport function getResearchModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_RESEARCH_MODEL\") ?? \"o3-deep-research\";\n}\n\nexport function getResearchTimeout(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_RESEARCH_TIMEOUT\", 3600000);\n}\n",
|
|
8
8
|
"import type {\n TextToSpeechParams as CoreTextToSpeechParams,\n TranscriptionParams as CoreTranscriptionParams,\n IAgentRuntime,\n} from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport type {\n TextToSpeechParams as LocalTextToSpeechParams,\n TranscriptionParams as LocalTranscriptionParams,\n OpenAITranscriptionResponse,\n TTSOutputFormat,\n TTSVoice,\n} from \"../types\";\nimport { detectAudioMimeType, getFilenameForMimeType } from \"../utils/audio\";\nimport {\n getAuthHeader,\n getBaseURL,\n getTranscriptionModel,\n getTTSInstructions,\n getTTSModel,\n getTTSVoice,\n} from \"../utils/config\";\n\ntype AudioInput = Blob | File | Buffer;\ntype TranscriptionInput = AudioInput | LocalTranscriptionParams | CoreTranscriptionParams | string;\ntype TTSInput = string | LocalTextToSpeechParams | CoreTextToSpeechParams;\n\nfunction isBlobOrFile(value: unknown): value is Blob | File {\n return value instanceof Blob || value instanceof File;\n}\n\nfunction isBuffer(value: unknown): value is Buffer {\n return Buffer.isBuffer(value);\n}\n\nfunction isLocalTranscriptionParams(value: unknown): value is LocalTranscriptionParams {\n return (\n typeof value === \"object\" &&\n value !== null &&\n \"audio\" in value &&\n (isBlobOrFile((value as LocalTranscriptionParams).audio) ||\n isBuffer((value as LocalTranscriptionParams).audio))\n );\n}\n\nfunction isCoreTranscriptionParams(value: unknown): value is CoreTranscriptionParams {\n return (\n typeof value === \"object\" &&\n value !== null &&\n \"audioUrl\" in value &&\n typeof (value as CoreTranscriptionParams).audioUrl === \"string\"\n );\n}\n\nasync function fetchAudioFromUrl(url: string): Promise<Blob> {\n const response = await fetch(url);\n if (!response.ok) {\n throw new Error(`Failed to fetch audio from URL: ${response.status}`);\n }\n return response.blob();\n}\nexport async function handleTranscription(\n runtime: IAgentRuntime,\n input: TranscriptionInput\n): Promise<string> {\n let modelName = getTranscriptionModel(runtime);\n let blob: Blob;\n let extraParams: Partial<LocalTranscriptionParams> = {};\n\n if (typeof input === \"string\") {\n logger.debug(`[OpenAI] Fetching audio from URL: ${input}`);\n blob = await fetchAudioFromUrl(input);\n } else if (isBlobOrFile(input)) {\n blob = input;\n } else if (isBuffer(input)) {\n const mimeType = detectAudioMimeType(input);\n logger.debug(`[OpenAI] Auto-detected audio MIME type: ${mimeType}`);\n blob = new Blob([new Uint8Array(input)], { type: mimeType });\n } else if (isLocalTranscriptionParams(input)) {\n extraParams = input;\n if (input.model) {\n modelName = input.model;\n }\n if (isBuffer(input.audio)) {\n const mimeType = input.mimeType ?? detectAudioMimeType(input.audio);\n logger.debug(`[OpenAI] Using MIME type: ${mimeType}`);\n blob = new Blob([new Uint8Array(input.audio)], { type: mimeType });\n } else {\n blob = input.audio;\n }\n } else if (isCoreTranscriptionParams(input)) {\n logger.debug(`[OpenAI] Fetching audio from URL: ${input.audioUrl}`);\n blob = await fetchAudioFromUrl(input.audioUrl);\n extraParams = { prompt: input.prompt };\n } else {\n throw new Error(\n \"TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object\"\n );\n }\n\n logger.debug(`[OpenAI] Using TRANSCRIPTION model: ${modelName}`);\n\n const mimeType = (blob as File).type || \"audio/webm\";\n const filename =\n (blob as File).name ||\n getFilenameForMimeType(\n mimeType.startsWith(\"audio/\")\n ? (mimeType as ReturnType<typeof detectAudioMimeType>)\n : \"audio/webm\"\n );\n\n const formData = new FormData();\n formData.append(\"file\", blob, filename);\n formData.append(\"model\", modelName);\n\n if (extraParams.language) {\n formData.append(\"language\", extraParams.language);\n }\n if (extraParams.responseFormat) {\n formData.append(\"response_format\", extraParams.responseFormat);\n }\n if (extraParams.prompt) {\n formData.append(\"prompt\", extraParams.prompt);\n }\n if (extraParams.temperature !== undefined) {\n formData.append(\"temperature\", String(extraParams.temperature));\n }\n if (extraParams.timestampGranularities) {\n for (const granularity of extraParams.timestampGranularities) {\n formData.append(\"timestamp_granularities[]\", granularity);\n }\n }\n\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/audio/transcriptions`, {\n method: \"POST\",\n headers: getAuthHeader(runtime),\n body: formData,\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI transcription failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAITranscriptionResponse;\n return data.text;\n}\n\nexport async function handleTextToSpeech(\n runtime: IAgentRuntime,\n input: TTSInput\n): Promise<ArrayBuffer> {\n let text: string;\n let voice: string | undefined;\n let format: TTSOutputFormat = \"mp3\";\n let model: string | undefined;\n let instructions: string | undefined;\n\n if (typeof input === \"string\") {\n text = input;\n voice = undefined;\n } else {\n text = input.text;\n voice = input.voice;\n if (\"format\" in input && input.format) {\n format = input.format;\n }\n if (\"model\" in input && input.model) {\n model = input.model;\n }\n if (\"instructions\" in input && input.instructions) {\n instructions = input.instructions;\n }\n }\n\n model = model ?? getTTSModel(runtime);\n voice = voice ?? getTTSVoice(runtime);\n instructions = instructions ?? getTTSInstructions(runtime);\n\n logger.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${model}`);\n\n if (!text || text.trim().length === 0) {\n throw new Error(\"TEXT_TO_SPEECH requires non-empty text\");\n }\n\n if (text.length > 4096) {\n throw new Error(\"TEXT_TO_SPEECH text exceeds 4096 character limit\");\n }\n\n const validVoices: TTSVoice[] = [\"alloy\", \"echo\", \"fable\", \"onyx\", \"nova\", \"shimmer\"];\n if (voice && !validVoices.includes(voice as TTSVoice)) {\n throw new Error(`Invalid voice: ${voice}. Must be one of: ${validVoices.join(\", \")}`);\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody: Record<string, string> = {\n model,\n voice: voice as TTSVoice,\n input: text,\n response_format: format,\n };\n\n if (instructions && instructions.length > 0) {\n requestBody.instructions = instructions;\n }\n\n const response = await fetch(`${baseURL}/audio/speech`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n ...(format === \"mp3\" ? { Accept: \"audio/mpeg\" } : {}),\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(`OpenAI TTS failed: ${response.status} ${response.statusText} - ${errorText}`);\n }\n\n return response.arrayBuffer();\n}\n",
|
|
9
9
|
"import { logger } from \"@elizaos/core\";\n\nconst MAGIC_BYTES = {\n WAV: {\n HEADER: [0x52, 0x49, 0x46, 0x46] as const, // \"RIFF\"\n IDENTIFIER: [0x57, 0x41, 0x56, 0x45] as const, // \"WAVE\"\n },\n MP3_ID3: [0x49, 0x44, 0x33] as const, // \"ID3\"\n OGG: [0x4f, 0x67, 0x67, 0x53] as const, // \"OggS\"\n FLAC: [0x66, 0x4c, 0x61, 0x43] as const, // \"fLaC\"\n FTYP: [0x66, 0x74, 0x79, 0x70] as const, // \"ftyp\" at offset 4 for mp4/m4a\n WEBM_EBML: [0x1a, 0x45, 0xdf, 0xa3] as const, // EBML header\n} as const;\n\nconst MIN_DETECTION_BUFFER_SIZE = 12;\n\nexport type AudioMimeType =\n | \"audio/wav\"\n | \"audio/mpeg\"\n | \"audio/ogg\"\n | \"audio/flac\"\n | \"audio/mp4\"\n | \"audio/webm\"\n | \"application/octet-stream\";\n\nfunction matchBytes(buffer: Buffer, offset: number, expected: readonly number[]): boolean {\n for (let i = 0; i < expected.length; i++) {\n const expectedByte = expected[i];\n if (expectedByte === undefined || buffer[offset + i] !== expectedByte) {\n return false;\n }\n }\n return true;\n}\n\nexport function detectAudioMimeType(buffer: Buffer): AudioMimeType {\n if (buffer.length < MIN_DETECTION_BUFFER_SIZE) {\n return \"application/octet-stream\";\n }\n\n // WAV: \"RIFF\" + size + \"WAVE\"\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.WAV.HEADER) &&\n matchBytes(buffer, 8, MAGIC_BYTES.WAV.IDENTIFIER)\n ) {\n return \"audio/wav\";\n }\n\n // MP3: ID3 tag or MPEG frame sync\n const firstByte = buffer[0];\n const secondByte = buffer[1];\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.MP3_ID3) ||\n (firstByte === 0xff && secondByte !== undefined && (secondByte & 0xe0) === 0xe0)\n ) {\n return \"audio/mpeg\";\n }\n\n // OGG: \"OggS\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.OGG)) {\n return \"audio/ogg\";\n }\n\n // FLAC: \"fLaC\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.FLAC)) {\n return \"audio/flac\";\n }\n\n // M4A/MP4: \"ftyp\" at offset 4\n if (matchBytes(buffer, 4, MAGIC_BYTES.FTYP)) {\n return \"audio/mp4\";\n }\n\n // WebM: EBML header\n if (matchBytes(buffer, 0, MAGIC_BYTES.WEBM_EBML)) {\n return \"audio/webm\";\n }\n\n logger.warn(\"Could not detect audio format from buffer, using generic binary type\");\n return \"application/octet-stream\";\n}\n\nexport function getExtensionForMimeType(mimeType: AudioMimeType): string {\n switch (mimeType) {\n case \"audio/wav\":\n return \"wav\";\n case \"audio/mpeg\":\n return \"mp3\";\n case \"audio/ogg\":\n return \"ogg\";\n case \"audio/flac\":\n return \"flac\";\n case \"audio/mp4\":\n return \"m4a\";\n case \"audio/webm\":\n return \"webm\";\n case \"application/octet-stream\":\n return \"bin\";\n }\n}\n\nexport function getFilenameForMimeType(mimeType: AudioMimeType): string {\n const ext = getExtensionForMimeType(mimeType);\n return `recording.${ext}`;\n}\n\nexport function validateAudioFormat(buffer: Buffer): AudioMimeType {\n const mimeType = detectAudioMimeType(buffer);\n if (mimeType === \"application/octet-stream\") {\n throw new Error(\n \"Unable to detect audio format. Supported formats: WAV, MP3, OGG, FLAC, M4A, WebM\"\n );\n }\n return mimeType;\n}\n",
|
|
10
|
-
"import type { IAgentRuntime, TextEmbeddingParams } from \"@elizaos/core\";\nimport { logger, ModelType, VECTOR_DIMS } from \"@elizaos/core\";\n\nimport type { OpenAIEmbeddingResponse } from \"../types\";\nimport {\n getAuthHeader,\n getEmbeddingBaseURL,\n getEmbeddingDimensions,\n getEmbeddingModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ntype VectorDimension = (typeof VECTOR_DIMS)[keyof typeof VECTOR_DIMS];\n\nfunction validateDimension(dimension: number): VectorDimension {\n const validDimensions = Object.values(VECTOR_DIMS) as number[];\n if (!validDimensions.includes(dimension)) {\n throw new Error(\n `Invalid embedding dimension: ${dimension}. Must be one of: ${validDimensions.join(\", \")}`\n );\n }\n return dimension as VectorDimension;\n}\n\nfunction extractText(params: TextEmbeddingParams | string | null): string | null {\n if (params === null) {\n return null;\n }\n if (typeof params === \"string\") {\n return params;\n }\n if (typeof params === \"object\" && typeof params.text === \"string\") {\n return params.text;\n }\n throw new Error(\"Invalid embedding params: expected string, { text: string }, or null\");\n}\n\nexport async function handleTextEmbedding(\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null\n): Promise<number[]> {\n const embeddingModel = getEmbeddingModel(runtime);\n const embeddingDimension = validateDimension(getEmbeddingDimensions(runtime));\n\n const text = extractText(params);\n if (text === null) {\n logger.debug(\"[OpenAI] Creating test embedding for initialization\");\n const testVector = new Array(embeddingDimension).fill(0);\n testVector[0] = 0.1;\n return testVector;\n }\n\n let trimmedText = text.trim();\n if (trimmedText.length === 0) {\n throw new Error(\"Cannot generate embedding for empty text\");\n }\n\n // Truncate to stay within embedding model token limits.\n // OpenAI embedding models support up to 8191 tokens per input;\n // 8000 tokens provides a safe buffer (~4 chars per token).\n const maxChars = 8_000 * 4;\n if (trimmedText.length > maxChars) {\n logger.warn(\n `[OpenAI] Embedding input too long (~${Math.ceil(trimmedText.length / 4)} tokens), truncating to ~8000 tokens`\n );\n trimmedText = trimmedText.slice(0, maxChars);\n }\n\n const baseURL = getEmbeddingBaseURL(runtime);\n const url = `${baseURL}/embeddings`;\n\n logger.debug(`[OpenAI] Generating embedding with model: ${embeddingModel}`);\n\n const response = await fetch(url, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime, true),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: embeddingModel,\n input: trimmedText,\n }),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI embedding API error: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIEmbeddingResponse;\n\n const firstResult = data?.data?.[0];\n if (!firstResult
|
|
10
|
+
"import type { IAgentRuntime, TextEmbeddingParams } from \"@elizaos/core\";\nimport { logger, ModelType, VECTOR_DIMS } from \"@elizaos/core\";\n\nimport type { OpenAIEmbeddingResponse } from \"../types\";\nimport {\n getAuthHeader,\n getEmbeddingBaseURL,\n getEmbeddingDimensions,\n getEmbeddingModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ntype VectorDimension = (typeof VECTOR_DIMS)[keyof typeof VECTOR_DIMS];\n\nfunction validateDimension(dimension: number): VectorDimension {\n const validDimensions = Object.values(VECTOR_DIMS) as number[];\n if (!validDimensions.includes(dimension)) {\n throw new Error(\n `Invalid embedding dimension: ${dimension}. Must be one of: ${validDimensions.join(\", \")}`\n );\n }\n return dimension as VectorDimension;\n}\n\nfunction extractText(params: TextEmbeddingParams | string | null): string | null {\n if (params === null) {\n return null;\n }\n if (typeof params === \"string\") {\n return params;\n }\n if (typeof params === \"object\" && typeof params.text === \"string\") {\n return params.text;\n }\n throw new Error(\"Invalid embedding params: expected string, { text: string }, or null\");\n}\n\nexport async function handleTextEmbedding(\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null\n): Promise<number[]> {\n const embeddingModel = getEmbeddingModel(runtime);\n const embeddingDimension = validateDimension(getEmbeddingDimensions(runtime));\n\n const text = extractText(params);\n if (text === null) {\n logger.debug(\"[OpenAI] Creating test embedding for initialization\");\n const testVector = new Array(embeddingDimension).fill(0);\n testVector[0] = 0.1;\n return testVector;\n }\n\n let trimmedText = text.trim();\n if (trimmedText.length === 0) {\n throw new Error(\"Cannot generate embedding for empty text\");\n }\n\n // Truncate to stay within embedding model token limits.\n // OpenAI embedding models support up to 8191 tokens per input;\n // 8000 tokens provides a safe buffer (~4 chars per token).\n const maxChars = 8_000 * 4;\n if (trimmedText.length > maxChars) {\n logger.warn(\n `[OpenAI] Embedding input too long (~${Math.ceil(trimmedText.length / 4)} tokens), truncating to ~8000 tokens`\n );\n trimmedText = trimmedText.slice(0, maxChars);\n }\n\n const baseURL = getEmbeddingBaseURL(runtime);\n const url = `${baseURL}/embeddings`;\n\n logger.debug(`[OpenAI] Generating embedding with model: ${embeddingModel}`);\n\n const response = await fetch(url, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime, true),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: embeddingModel,\n input: trimmedText,\n }),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI embedding API error: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIEmbeddingResponse;\n\n const firstResult = data?.data?.[0];\n if (!firstResult?.embedding) {\n throw new Error(\"OpenAI API returned invalid embedding response structure\");\n }\n\n const embedding = firstResult.embedding;\n\n if (embedding.length !== embeddingDimension) {\n throw new Error(\n `Embedding dimension mismatch: got ${embedding.length}, expected ${embeddingDimension}. ` +\n `Check OPENAI_EMBEDDING_DIMENSIONS setting.`\n );\n }\n\n if (data.usage) {\n emitModelUsageEvent(runtime, ModelType.TEXT_EMBEDDING, trimmedText, {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: 0,\n totalTokens: data.usage.total_tokens,\n });\n }\n\n logger.debug(`[OpenAI] Generated embedding with ${embedding.length} dimensions`);\n return embedding;\n}\n",
|
|
11
11
|
"import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { EventType } from \"@elizaos/core\";\nimport type { TokenUsage } from \"../types\";\n\nconst MAX_PROMPT_LENGTH = 200;\n\ninterface ModelUsageEventPayload {\n runtime: IAgentRuntime;\n source: \"openai\";\n provider: \"openai\";\n type: ModelTypeName;\n prompt: string;\n tokens: {\n prompt: number;\n completion: number;\n total: number;\n cached?: number;\n };\n}\n\ninterface AISDKUsage {\n inputTokens?: number;\n outputTokens?: number;\n totalTokens?: number;\n cachedInputTokens?: number;\n}\n\ninterface OpenAIAPIUsage {\n promptTokens?: number;\n completionTokens?: number;\n totalTokens?: number;\n cachedPromptTokens?: number;\n promptTokensDetails?: {\n cachedTokens?: number;\n };\n}\n\ntype ModelUsage = TokenUsage | AISDKUsage | OpenAIAPIUsage;\n\nfunction truncatePrompt(prompt: string): string {\n if (prompt.length <= MAX_PROMPT_LENGTH) {\n return prompt;\n }\n return `${prompt.slice(0, MAX_PROMPT_LENGTH)}…`;\n}\n\nfunction normalizeUsage(usage: ModelUsage): TokenUsage {\n if (\"promptTokens\" in usage) {\n const promptTokensDetails =\n \"promptTokensDetails\" in usage ? usage.promptTokensDetails : undefined;\n const cachedPromptTokens = usage.cachedPromptTokens ?? promptTokensDetails?.cachedTokens;\n return {\n promptTokens: usage.promptTokens ?? 0,\n completionTokens: usage.completionTokens ?? 0,\n totalTokens: usage.totalTokens ?? (usage.promptTokens ?? 0) + (usage.completionTokens ?? 0),\n cachedPromptTokens,\n };\n }\n if (\"inputTokens\" in usage || \"outputTokens\" in usage) {\n const input = (usage as AISDKUsage).inputTokens ?? 0;\n const output = (usage as AISDKUsage).outputTokens ?? 0;\n const total = (usage as AISDKUsage).totalTokens ?? input + output;\n return {\n promptTokens: input,\n completionTokens: output,\n totalTokens: total,\n cachedPromptTokens: (usage as AISDKUsage).cachedInputTokens,\n };\n }\n return {\n promptTokens: 0,\n completionTokens: 0,\n totalTokens: 0,\n };\n}\n\nexport function emitModelUsageEvent(\n runtime: IAgentRuntime,\n type: ModelTypeName,\n prompt: string,\n usage: ModelUsage\n): void {\n const normalized = normalizeUsage(usage);\n\n const payload: ModelUsageEventPayload = {\n runtime,\n source: \"openai\",\n provider: \"openai\",\n type,\n prompt: truncatePrompt(prompt),\n tokens: {\n prompt: normalized.promptTokens,\n completion: normalized.completionTokens,\n total: normalized.totalTokens,\n ...(normalized.cachedPromptTokens !== undefined\n ? { cached: normalized.cachedPromptTokens }\n : {}),\n },\n };\n\n runtime.emitEvent(EventType.MODEL_USED, payload);\n}\n",
|
|
12
12
|
"import type { IAgentRuntime, ImageDescriptionParams, ImageGenerationParams } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport type {\n ImageDescriptionResult,\n ImageGenerationResult,\n ImageQuality,\n ImageSize,\n ImageStyle,\n OpenAIChatCompletionResponse,\n OpenAIImageGenerationResponse,\n} from \"../types\";\nimport {\n getAuthHeader,\n getBaseURL,\n getImageDescriptionMaxTokens,\n getImageDescriptionModel,\n getImageModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ninterface ExtendedImageGenerationParams extends ImageGenerationParams {\n quality?: ImageQuality;\n style?: ImageStyle;\n}\n\nconst DEFAULT_IMAGE_DESCRIPTION_PROMPT =\n \"Please analyze this image and provide a title and detailed description.\";\n\nexport async function handleImageGeneration(\n runtime: IAgentRuntime,\n params: ImageGenerationParams\n): Promise<ImageGenerationResult[]> {\n const modelName = getImageModel(runtime);\n const count = params.count ?? 1;\n const size: ImageSize = (params.size as ImageSize) ?? \"1024x1024\";\n const extendedParams = params as ExtendedImageGenerationParams;\n\n logger.debug(`[OpenAI] Using IMAGE model: ${modelName}`);\n\n if (!params.prompt || params.prompt.trim().length === 0) {\n throw new Error(\"IMAGE generation requires a non-empty prompt\");\n }\n\n if (count < 1 || count > 10) {\n throw new Error(\"IMAGE count must be between 1 and 10\");\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody: Record<string, string | number> = {\n model: modelName,\n prompt: params.prompt,\n n: count,\n size,\n };\n\n if (extendedParams.quality) {\n requestBody.quality = extendedParams.quality;\n }\n if (extendedParams.style) {\n requestBody.style = extendedParams.style;\n }\n\n const response = await fetch(`${baseURL}/images/generations`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI image generation failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIImageGenerationResponse;\n\n if (!data.data || data.data.length === 0) {\n throw new Error(\"OpenAI API returned no images\");\n }\n\n return data.data.map((item) => ({\n url: item.url,\n revisedPrompt: item.revised_prompt,\n }));\n}\n\nfunction parseTitleFromResponse(content: string): string {\n const titleMatch = content.match(/title[:\\s]+(.+?)(?:\\n|$)/i);\n return titleMatch?.[1]?.trim() ?? \"Image Analysis\";\n}\n\nfunction parseDescriptionFromResponse(content: string): string {\n return content.replace(/title[:\\s]+(.+?)(?:\\n|$)/i, \"\").trim();\n}\n\nexport async function handleImageDescription(\n runtime: IAgentRuntime,\n params: ImageDescriptionParams | string\n): Promise<ImageDescriptionResult> {\n const modelName = getImageDescriptionModel(runtime);\n const maxTokens = getImageDescriptionMaxTokens(runtime);\n\n logger.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${modelName}`);\n\n let imageUrl: string;\n let promptText: string;\n\n if (typeof params === \"string\") {\n imageUrl = params;\n promptText = DEFAULT_IMAGE_DESCRIPTION_PROMPT;\n } else {\n imageUrl = params.imageUrl;\n promptText = params.prompt ?? DEFAULT_IMAGE_DESCRIPTION_PROMPT;\n }\n\n if (!imageUrl || imageUrl.trim().length === 0) {\n throw new Error(\"IMAGE_DESCRIPTION requires a valid image URL\");\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody = {\n model: modelName,\n messages: [\n {\n role: \"user\",\n content: [\n { type: \"text\", text: promptText },\n { type: \"image_url\", image_url: { url: imageUrl } },\n ],\n },\n ],\n max_tokens: maxTokens,\n };\n\n const response = await fetch(`${baseURL}/chat/completions`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI image description failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIChatCompletionResponse;\n\n if (data.usage) {\n emitModelUsageEvent(\n runtime,\n ModelType.IMAGE_DESCRIPTION,\n typeof params === \"string\" ? params : (params.prompt ?? \"\"),\n {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: data.usage.completion_tokens,\n totalTokens: data.usage.total_tokens,\n }\n );\n }\n\n const firstChoice = data.choices?.[0];\n const content = firstChoice?.message?.content;\n\n if (!content) {\n throw new Error(\"OpenAI API returned empty image description\");\n }\n\n return {\n title: parseTitleFromResponse(content),\n description: parseDescriptionFromResponse(content),\n };\n}\n",
|
|
13
13
|
"import type {\n IAgentRuntime,\n JsonValue,\n ModelTypeName,\n ObjectGenerationParams,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateObject, type LanguageModel } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport { getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\nimport { getJsonRepairFunction } from \"../utils/json\";\n\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\ntype ChatModelFactory = { chat: (modelName: string) => LanguageModel };\n\nasync function generateObjectByModelType(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<Record<string, JsonValue>> {\n const openai = createOpenAIClient(runtime) as ChatModelFactory;\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n\n if (!params.prompt || params.prompt.trim().length === 0) {\n throw new Error(\"Object generation requires a non-empty prompt\");\n }\n\n if (params.schema) {\n logger.debug(\n \"[OpenAI] Schema provided but using no-schema mode. \" +\n \"Structure is determined by prompt instructions.\"\n );\n }\n\n const model = openai.chat(modelName);\n const { object, usage } = await generateObject({\n model,\n output: \"no-schema\",\n prompt: params.prompt,\n experimental_repairText: getJsonRepairFunction(),\n });\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n if (typeof object !== \"object\" || object === null) {\n throw new Error(`Object generation returned ${typeof object}, expected object`);\n }\n\n return object as Record<string, JsonValue>;\n}\n\nexport async function handleObjectSmall(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n): Promise<Record<string, JsonValue>> {\n return generateObjectByModelType(runtime, params, ModelType.OBJECT_SMALL, getSmallModel);\n}\n\nexport async function handleObjectLarge(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n): Promise<Record<string, JsonValue>> {\n return generateObjectByModelType(runtime, params, ModelType.OBJECT_LARGE, getLargeModel);\n}\n",
|
|
14
14
|
"import { createOpenAI, type OpenAIProvider } from \"@ai-sdk/openai\";\nimport type { IAgentRuntime } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, isProxyMode } from \"../utils/config\";\n\nconst PROXY_API_KEY = \"sk-proxy\";\n\nexport function createOpenAIClient(runtime: IAgentRuntime): OpenAIProvider {\n const baseURL = getBaseURL(runtime);\n const apiKey = getApiKey(runtime);\n\n if (!apiKey && isProxyMode(runtime)) {\n return createOpenAI({\n apiKey: PROXY_API_KEY,\n baseURL,\n });\n }\n\n if (!apiKey) {\n throw new Error(\n \"OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.\"\n );\n }\n\n return createOpenAI({\n apiKey,\n baseURL,\n });\n}\n",
|
|
15
15
|
"import { logger } from \"@elizaos/core\";\nimport { JSONParseError } from \"ai\";\n\ninterface JsonRepairParams {\n text: string;\n error: Error;\n}\n\ntype JsonRepairFunction = (params: JsonRepairParams) => Promise<string | null>;\n\nconst JSON_CLEANUP_PATTERNS = {\n MARKDOWN_JSON: /```json\\n|\\n```|```/g,\n WHITESPACE: /^\\s+|\\s+$/g,\n} as const;\n\nexport function getJsonRepairFunction(): JsonRepairFunction {\n return async ({ text, error }: JsonRepairParams): Promise<string | null> => {\n if (!(error instanceof JSONParseError)) {\n return null;\n }\n try {\n const cleanedText = text.replace(JSON_CLEANUP_PATTERNS.MARKDOWN_JSON, \"\");\n JSON.parse(cleanedText);\n logger.debug(\"[JSON Repair] Successfully repaired JSON by removing markdown wrappers\");\n return cleanedText;\n } catch {\n logger.warn(\"[JSON Repair] Unable to repair JSON text\");\n return null;\n }\n };\n}\n\nexport function parseJsonWithRepair<T>(text: string): T {\n try {\n return JSON.parse(text) as T;\n } catch (firstError) {\n const cleanedText = text.replace(JSON_CLEANUP_PATTERNS.MARKDOWN_JSON, \"\");\n try {\n return JSON.parse(cleanedText) as T;\n } catch {\n const message = firstError instanceof Error ? firstError.message : String(firstError);\n throw new Error(`Failed to parse JSON: ${message}`);\n }\n }\n}\n\nexport function safeStringify(value: unknown, indent = 0): string {\n const seen = new WeakSet();\n\n const replacer = (_key: string, val: unknown): unknown => {\n if (typeof val === \"object\" && val !== null) {\n if (seen.has(val)) {\n return \"[Circular]\";\n }\n seen.add(val);\n }\n\n // Handle special types\n if (typeof val === \"bigint\") {\n return val.toString();\n }\n\n if (val instanceof Error) {\n return {\n name: val.name,\n message: val.message,\n stack: val.stack,\n };\n }\n\n if (val instanceof Date) {\n return val.toISOString();\n }\n\n if (val instanceof Map) {\n return Object.fromEntries(val);\n }\n\n if (val instanceof Set) {\n return Array.from(val);\n }\n\n return val;\n };\n\n return JSON.stringify(value, replacer, indent);\n}\n",
|
|
16
16
|
"/**\n * Deep Research model handler\n *\n * Provides deep research capabilities using OpenAI's o3-deep-research and o4-mini-deep-research models.\n * These models can find, analyze, and synthesize hundreds of sources to create comprehensive reports.\n *\n * @see https://platform.openai.com/docs/guides/deep-research\n */\n\nimport type {\n IAgentRuntime,\n JsonValue,\n ResearchAnnotation,\n ResearchCodeInterpreterCall,\n ResearchFileSearchCall,\n ResearchMcpToolCall,\n ResearchMessageOutput,\n ResearchOutputItem,\n ResearchParams,\n ResearchResult,\n ResearchTool,\n ResearchWebSearchCall,\n} from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, getResearchModel, getResearchTimeout } from \"../utils/config\";\n\n// ============================================================================\n// Types for OpenAI Responses API\n// ============================================================================\n\n/**\n * Tool configuration for the Responses API\n */\ninterface ResponsesApiTool {\n type: \"web_search_preview\" | \"file_search\" | \"code_interpreter\" | \"mcp\";\n vector_store_ids?: string[];\n container?: { type: \"auto\" };\n server_label?: string;\n server_url?: string;\n require_approval?: \"never\";\n}\n\n/**\n * Raw response from the OpenAI Responses API\n */\ninterface ResponsesApiResponse {\n id: string;\n object: string;\n status?: \"queued\" | \"in_progress\" | \"completed\" | \"failed\";\n output?: ResponsesApiOutputItem[];\n output_text?: string;\n error?: {\n message: string;\n code: string;\n };\n}\n\n/**\n * Raw output item from the Responses API\n */\ninterface ResponsesApiOutputItem {\n id?: string;\n type: string;\n status?: string;\n action?: {\n type: string;\n query?: string;\n url?: string;\n };\n query?: string;\n results?: Array<{\n file_id: string;\n file_name: string;\n score: number;\n }>;\n code?: string;\n output?: string;\n server_label?: string;\n tool_name?: string;\n arguments?: Record<string, unknown>;\n result?: unknown;\n content?: Array<{\n type: string;\n text: string;\n annotations?: Array<{\n url: string;\n title: string;\n start_index: number;\n end_index: number;\n }>;\n }>;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts ResearchTool params to Responses API tool format\n */\nfunction convertToolToApi(tool: ResearchTool): ResponsesApiTool {\n switch (tool.type) {\n case \"web_search_preview\":\n return { type: \"web_search_preview\" };\n case \"file_search\":\n return {\n type: \"file_search\",\n vector_store_ids: tool.vectorStoreIds,\n };\n case \"code_interpreter\":\n return {\n type: \"code_interpreter\",\n container: tool.container ?? { type: \"auto\" },\n };\n case \"mcp\":\n return {\n type: \"mcp\",\n server_label: tool.serverLabel,\n server_url: tool.serverUrl,\n require_approval: tool.requireApproval ?? \"never\",\n };\n default:\n throw new Error(`Unknown research tool type: ${(tool as ResearchTool).type}`);\n }\n}\n\n/**\n * Converts raw API output items to typed ResearchOutputItem\n */\nfunction convertOutputItem(item: ResponsesApiOutputItem): ResearchOutputItem | null {\n switch (item.type) {\n case \"web_search_call\":\n return {\n id: item.id ?? \"\",\n type: \"web_search_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n action: {\n type: (item.action?.type as \"search\" | \"open_page\" | \"find_in_page\") ?? \"search\",\n query: item.action?.query,\n url: item.action?.url,\n },\n } satisfies ResearchWebSearchCall;\n\n case \"file_search_call\":\n return {\n id: item.id ?? \"\",\n type: \"file_search_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n query: item.query ?? \"\",\n results: item.results?.map((r) => ({\n fileId: r.file_id,\n fileName: r.file_name,\n score: r.score,\n })),\n } satisfies ResearchFileSearchCall;\n\n case \"code_interpreter_call\":\n return {\n id: item.id ?? \"\",\n type: \"code_interpreter_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n code: item.code ?? \"\",\n output: item.output,\n } satisfies ResearchCodeInterpreterCall;\n\n case \"mcp_tool_call\":\n return {\n id: item.id ?? \"\",\n type: \"mcp_tool_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n serverLabel: item.server_label ?? \"\",\n toolName: item.tool_name ?? \"\",\n arguments: (item.arguments ?? {}) as Record<string, JsonValue>,\n result: item.result as JsonValue,\n } satisfies ResearchMcpToolCall;\n\n case \"message\":\n return {\n type: \"message\",\n content:\n item.content?.map((c) => ({\n type: \"output_text\" as const,\n text: c.text,\n annotations:\n c.annotations?.map((a) => ({\n url: a.url,\n title: a.title,\n startIndex: a.start_index,\n endIndex: a.end_index,\n })) ?? [],\n })) ?? [],\n } satisfies ResearchMessageOutput;\n\n default:\n // Unknown output type, skip\n return null;\n }\n}\n\n/**\n * Extracts text and annotations from the response\n */\nfunction extractTextAndAnnotations(response: ResponsesApiResponse): {\n text: string;\n annotations: ResearchAnnotation[];\n} {\n // Try output_text first (convenience field)\n if (response.output_text) {\n // Find annotations from message output items\n const annotations: ResearchAnnotation[] = [];\n if (response.output) {\n for (const item of response.output) {\n if (item.type === \"message\" && item.content) {\n for (const content of item.content) {\n if (content.annotations) {\n for (const ann of content.annotations) {\n annotations.push({\n url: ann.url,\n title: ann.title,\n startIndex: ann.start_index,\n endIndex: ann.end_index,\n });\n }\n }\n }\n }\n }\n }\n return { text: response.output_text, annotations };\n }\n\n // Fall back to extracting from message output items\n let text = \"\";\n const annotations: ResearchAnnotation[] = [];\n\n if (response.output) {\n for (const item of response.output) {\n if (item.type === \"message\" && item.content) {\n for (const content of item.content) {\n text += content.text;\n if (content.annotations) {\n for (const ann of content.annotations) {\n annotations.push({\n url: ann.url,\n title: ann.title,\n startIndex: ann.start_index,\n endIndex: ann.end_index,\n });\n }\n }\n }\n }\n }\n }\n\n return { text, annotations };\n}\n\n// ============================================================================\n// Main Handler\n// ============================================================================\n\n/**\n * Handles RESEARCH model requests using OpenAI's deep research models.\n *\n * Deep research models can take tens of minutes to complete tasks.\n * Use background mode for long-running tasks.\n *\n * @param runtime - The agent runtime\n * @param params - Research parameters\n * @returns Research result with text, annotations, and output items\n *\n * @example\n * ```typescript\n * const result = await handleResearch(runtime, {\n * input: \"Research the economic impact of AI on global labor markets\",\n * tools: [\n * { type: \"web_search_preview\" },\n * { type: \"code_interpreter\", container: { type: \"auto\" } }\n * ],\n * background: true,\n * });\n * console.log(result.text);\n * ```\n */\nexport async function handleResearch(\n runtime: IAgentRuntime,\n params: ResearchParams\n): Promise<ResearchResult> {\n const apiKey = getApiKey(runtime);\n if (!apiKey) {\n throw new Error(\n \"OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.\"\n );\n }\n\n const baseURL = getBaseURL(runtime);\n const modelName = params.model ?? getResearchModel(runtime);\n const timeout = getResearchTimeout(runtime);\n\n logger.debug(`[OpenAI] Starting deep research with model: ${modelName}`);\n logger.debug(`[OpenAI] Research input: ${params.input.substring(0, 100)}...`);\n\n // Validate that at least one data source tool is provided\n const dataSourceTools = params.tools?.filter(\n (t) => t.type === \"web_search_preview\" || t.type === \"file_search\" || t.type === \"mcp\"\n );\n\n if (!dataSourceTools || dataSourceTools.length === 0) {\n // Default to web search if no tools specified\n logger.debug(\"[OpenAI] No data source tools specified, defaulting to web_search_preview\");\n params.tools = [{ type: \"web_search_preview\" }, ...(params.tools ?? [])];\n }\n\n // Build the request body for the Responses API\n const requestBody: Record<string, unknown> = {\n model: modelName,\n input: params.input,\n };\n\n if (params.instructions) {\n requestBody.instructions = params.instructions;\n }\n\n if (params.background !== undefined) {\n requestBody.background = params.background;\n }\n\n if (params.tools && params.tools.length > 0) {\n requestBody.tools = params.tools.map(convertToolToApi);\n }\n\n if (params.maxToolCalls !== undefined) {\n requestBody.max_tool_calls = params.maxToolCalls;\n }\n\n if (params.reasoningSummary) {\n requestBody.reasoning = { summary: params.reasoningSummary };\n }\n\n logger.debug(`[OpenAI] Research request body: ${JSON.stringify(requestBody, null, 2)}`);\n\n // Make the API request\n const response = await fetch(`${baseURL}/responses`, {\n method: \"POST\",\n headers: {\n Authorization: `Bearer ${apiKey}`,\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n signal: AbortSignal.timeout(timeout),\n });\n\n if (!response.ok) {\n const errorText = await response.text();\n logger.error(`[OpenAI] Research request failed: ${response.status} ${errorText}`);\n throw new Error(`Deep research request failed: ${response.status} ${response.statusText}`);\n }\n\n const data = (await response.json()) as ResponsesApiResponse;\n\n if (data.error) {\n logger.error(`[OpenAI] Research API error: ${data.error.message}`);\n throw new Error(`Deep research error: ${data.error.message}`);\n }\n\n logger.debug(`[OpenAI] Research response received. Status: ${data.status ?? \"completed\"}`);\n\n // Extract text and annotations\n const { text, annotations } = extractTextAndAnnotations(data);\n\n // Convert output items\n const outputItems: ResearchOutputItem[] = [];\n if (data.output) {\n for (const item of data.output) {\n const converted = convertOutputItem(item);\n if (converted) {\n outputItems.push(converted);\n }\n }\n }\n\n const result: ResearchResult = {\n id: data.id,\n text,\n annotations,\n outputItems,\n status: data.status,\n };\n\n logger.info(\n `[OpenAI] Research completed. Text length: ${text.length}, Annotations: ${annotations.length}, Output items: ${outputItems.length}`\n );\n\n return result;\n}\n",
|
|
17
|
-
"/**\n * Text generation model handlers\n *\n * Provides text generation using OpenAI's language models.\n */\n\nimport type { GenerateTextParams, IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateText, type LanguageModelUsage, streamText } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport type { TextStreamResult, TokenUsage } from \"../types\";\nimport { getExperimentalTelemetry, getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n// ============================================================================\n// Types\n// ============================================================================\n\n/**\n * Function to get model name from runtime\n */\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\n\ntype PromptCacheRetention = \"in_memory\" | \"24h\";\n\ninterface OpenAIPromptCacheOptions {\n promptCacheKey?: string;\n promptCacheRetention?: PromptCacheRetention;\n}\n\ninterface GenerateTextParamsWithOpenAIOptions extends GenerateTextParams {\n providerOptions?: {\n openai?: OpenAIPromptCacheOptions;\n };\n}\n\ninterface LanguageModelUsageWithCache extends LanguageModelUsage {\n cachedInputTokens?: number;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts AI SDK usage to our token usage format\n */\nfunction convertUsage(usage: LanguageModelUsage | undefined): TokenUsage | undefined {\n if (!usage) {\n return undefined;\n }\n\n // The AI SDK uses inputTokens/outputTokens\n const promptTokens = usage.inputTokens ?? 0;\n const completionTokens = usage.outputTokens ?? 0;\n const usageWithCache = usage as LanguageModelUsageWithCache;\n\n return {\n promptTokens,\n completionTokens,\n totalTokens: promptTokens + completionTokens,\n cachedPromptTokens: usageWithCache.cachedInputTokens,\n };\n}\n\nfunction resolvePromptCacheOptions(params: GenerateTextParams): OpenAIPromptCacheOptions {\n const withOpenAIOptions = params as GenerateTextParamsWithOpenAIOptions;\n return {\n promptCacheKey: withOpenAIOptions.providerOptions?.openai?.promptCacheKey,\n promptCacheRetention: withOpenAIOptions.providerOptions?.openai?.promptCacheRetention,\n };\n}\n\n// ============================================================================\n// Core Generation Function\n// ============================================================================\n\n/**\n * Generates text using the specified model type.\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @param modelType - The type of model (TEXT_SMALL or TEXT_LARGE)\n * @param getModelFn - Function to get the model name\n * @returns Generated text or stream result\n */\nasync function generateTextByModelType(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<string | TextStreamResult> {\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n const promptCacheOptions = resolvePromptCacheOptions(params);\n\n // Get system prompt from character if available\n const systemPrompt = runtime.character.system ?? undefined;\n\n // Use chat() instead of languageModel() to use the Chat Completions API\n // which has better compatibility than the Responses API\n // gpt-5 and gpt-5-mini (reasoning models) don't support temperature,\n // frequencyPenalty, presencePenalty, or stop parameters - use defaults only\n const model = openai.chat(modelName);\n const generateParams = {\n model,\n prompt: params.prompt,\n system: systemPrompt,\n maxOutputTokens: params.maxTokens ?? 8192,\n experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },\n ...(promptCacheOptions.promptCacheKey || promptCacheOptions.promptCacheRetention\n ? {\n providerOptions: {\n openai: {\n ...(promptCacheOptions.promptCacheKey\n ? { promptCacheKey: promptCacheOptions.promptCacheKey }\n : {}),\n ...(promptCacheOptions.promptCacheRetention\n ? { promptCacheRetention: promptCacheOptions.promptCacheRetention }\n : {}),\n },\n },\n }\n : {}),\n };\n\n // Handle streaming mode\n if (params.stream) {\n const result = streamText(generateParams);\n\n return {\n textStream: result.textStream,\n text: Promise.resolve(result.text),\n usage: Promise.resolve(result.usage).then(convertUsage),\n finishReason: Promise.resolve(result.finishReason).then((r) => r as string | undefined),\n };\n }\n\n // Non-streaming mode\n const { text, usage } = await generateText(generateParams);\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n return text;\n}\n\n// ============================================================================\n// Public Handlers\n// ============================================================================\n\n/**\n * Handles TEXT_SMALL model requests.\n *\n * Uses the configured small model (default: gpt-5-mini).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextSmall(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_SMALL, getSmallModel);\n}\n\n/**\n * Handles TEXT_LARGE model requests.\n *\n * Uses the configured large model (default: gpt-5).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextLarge(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_LARGE, getLargeModel);\n}\n",
|
|
17
|
+
"/**\n * Text generation model handlers\n *\n * Provides text generation using OpenAI's language models.\n */\n\nimport type { GenerateTextParams, IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateText, type LanguageModelUsage, streamText } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport type { TextStreamResult, TokenUsage } from \"../types\";\nimport { getExperimentalTelemetry, getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n// ============================================================================\n// Types\n// ============================================================================\n\n/**\n * Function to get model name from runtime\n */\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\n\ntype PromptCacheRetention = \"in_memory\" | \"24h\";\ntype ChatAttachment = {\n data: string | Uint8Array | URL;\n mediaType: string;\n filename?: string;\n};\n\ninterface OpenAIPromptCacheOptions {\n promptCacheKey?: string;\n promptCacheRetention?: PromptCacheRetention;\n}\n\ninterface GenerateTextParamsWithOpenAIOptions extends GenerateTextParams {\n attachments?: ChatAttachment[];\n providerOptions?: {\n openai?: OpenAIPromptCacheOptions;\n };\n}\n\ninterface LanguageModelUsageWithCache extends LanguageModelUsage {\n cachedInputTokens?: number;\n}\n\nfunction buildUserContent(params: GenerateTextParamsWithOpenAIOptions) {\n const content: Array<\n | { type: \"text\"; text: string }\n | {\n type: \"file\";\n data: string | Uint8Array | URL;\n mediaType: string;\n filename?: string;\n }\n > = [{ type: \"text\", text: params.prompt }];\n\n for (const attachment of params.attachments ?? []) {\n content.push({\n type: \"file\",\n data: attachment.data,\n mediaType: attachment.mediaType,\n ...(attachment.filename ? { filename: attachment.filename } : {}),\n });\n }\n\n return content;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts AI SDK usage to our token usage format\n */\nfunction convertUsage(usage: LanguageModelUsage | undefined): TokenUsage | undefined {\n if (!usage) {\n return undefined;\n }\n\n // The AI SDK uses inputTokens/outputTokens\n const promptTokens = usage.inputTokens ?? 0;\n const completionTokens = usage.outputTokens ?? 0;\n const usageWithCache = usage as LanguageModelUsageWithCache;\n\n return {\n promptTokens,\n completionTokens,\n totalTokens: promptTokens + completionTokens,\n cachedPromptTokens: usageWithCache.cachedInputTokens,\n };\n}\n\nfunction resolvePromptCacheOptions(params: GenerateTextParams): OpenAIPromptCacheOptions {\n const withOpenAIOptions = params as GenerateTextParamsWithOpenAIOptions;\n return {\n promptCacheKey: withOpenAIOptions.providerOptions?.openai?.promptCacheKey,\n promptCacheRetention: withOpenAIOptions.providerOptions?.openai?.promptCacheRetention,\n };\n}\n\n// ============================================================================\n// Core Generation Function\n// ============================================================================\n\n/**\n * Generates text using the specified model type.\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @param modelType - The type of model (TEXT_SMALL or TEXT_LARGE)\n * @param getModelFn - Function to get the model name\n * @returns Generated text or stream result\n */\nasync function generateTextByModelType(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<string | TextStreamResult> {\n const paramsWithAttachments = params as GenerateTextParamsWithOpenAIOptions;\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n const promptCacheOptions = resolvePromptCacheOptions(params);\n const hasAttachments = (paramsWithAttachments.attachments?.length ?? 0) > 0;\n const userContent = hasAttachments ? buildUserContent(paramsWithAttachments) : undefined;\n\n // Get system prompt from character if available\n const systemPrompt = runtime.character.system ?? undefined;\n\n // Use chat() instead of languageModel() to use the Chat Completions API\n // which has better compatibility than the Responses API\n // gpt-5 and gpt-5-mini (reasoning models) don't support temperature,\n // frequencyPenalty, presencePenalty, or stop parameters - use defaults only\n const model = openai.chat(modelName);\n const generateParams = {\n model,\n ...(userContent\n ? { messages: [{ role: \"user\" as const, content: userContent }] }\n : { prompt: params.prompt }),\n system: systemPrompt,\n maxOutputTokens: params.maxTokens ?? 8192,\n experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },\n ...(promptCacheOptions.promptCacheKey || promptCacheOptions.promptCacheRetention\n ? {\n providerOptions: {\n openai: {\n ...(promptCacheOptions.promptCacheKey\n ? { promptCacheKey: promptCacheOptions.promptCacheKey }\n : {}),\n ...(promptCacheOptions.promptCacheRetention\n ? { promptCacheRetention: promptCacheOptions.promptCacheRetention }\n : {}),\n },\n },\n }\n : {}),\n };\n\n // Handle streaming mode\n if (params.stream) {\n const result = streamText(generateParams);\n\n return {\n textStream: result.textStream,\n text: Promise.resolve(result.text),\n usage: Promise.resolve(result.usage).then(convertUsage),\n finishReason: Promise.resolve(result.finishReason).then((r) => r as string | undefined),\n };\n }\n\n // Non-streaming mode\n const { text, usage } = await generateText(generateParams);\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n return text;\n}\n\n// ============================================================================\n// Public Handlers\n// ============================================================================\n\n/**\n * Handles TEXT_SMALL model requests.\n *\n * Uses the configured small model (default: gpt-5-mini).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextSmall(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_SMALL, getSmallModel);\n}\n\n/**\n * Handles TEXT_LARGE model requests.\n *\n * Uses the configured large model (default: gpt-5).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextLarge(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_LARGE, getLargeModel);\n}\n",
|
|
18
18
|
"import type { DetokenizeTextParams, IAgentRuntime, TokenizeTextParams } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport { detokenizeText, tokenizeText } from \"../utils/tokenization\";\n\nexport async function handleTokenizerEncode(\n runtime: IAgentRuntime,\n params: TokenizeTextParams\n): Promise<number[]> {\n if (!params.prompt) {\n throw new Error(\"Tokenization requires a non-empty prompt\");\n }\n const modelType = params.modelType ?? ModelType.TEXT_LARGE;\n return tokenizeText(runtime, modelType, params.prompt);\n}\n\nexport async function handleTokenizerDecode(\n runtime: IAgentRuntime,\n params: DetokenizeTextParams\n): Promise<string> {\n if (!params.tokens || !Array.isArray(params.tokens)) {\n throw new Error(\"Detokenization requires a valid tokens array\");\n }\n if (params.tokens.length === 0) {\n return \"\";\n }\n for (let i = 0; i < params.tokens.length; i++) {\n const token = params.tokens[i];\n if (typeof token !== \"number\" || !Number.isFinite(token)) {\n throw new Error(`Invalid token at index ${i}: expected number`);\n }\n }\n const modelType = params.modelType ?? ModelType.TEXT_LARGE;\n return detokenizeText(runtime, modelType, params.tokens);\n}\n",
|
|
19
19
|
"import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport {\n encodingForModel,\n getEncoding,\n type Tiktoken,\n type TiktokenEncoding,\n type TiktokenModel,\n} from \"js-tiktoken\";\nimport { getLargeModel, getSmallModel } from \"./config\";\n\ntype SupportedEncoding = \"cl100k_base\" | \"o200k_base\";\n\nfunction resolveTokenizerEncoding(modelName: string): Tiktoken {\n const normalized = modelName.toLowerCase();\n const fallbackEncoding: SupportedEncoding = normalized.includes(\"4o\")\n ? \"o200k_base\"\n : \"cl100k_base\";\n try {\n return encodingForModel(modelName as TiktokenModel);\n } catch {\n return getEncoding(fallbackEncoding as TiktokenEncoding);\n }\n}\n\nfunction getModelName(runtime: IAgentRuntime, modelType: ModelTypeName): string {\n if (modelType === ModelType.TEXT_SMALL) {\n return getSmallModel(runtime);\n }\n return getLargeModel(runtime);\n}\n\nexport function tokenizeText(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string\n): number[] {\n const modelName = getModelName(runtime, modelType);\n const encoder = resolveTokenizerEncoding(modelName);\n return encoder.encode(text);\n}\n\nexport function detokenizeText(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n tokens: number[]\n): string {\n const modelName = getModelName(runtime, modelType);\n const encoder = resolveTokenizerEncoding(modelName);\n return encoder.decode(tokens);\n}\n\nexport function countTokens(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string\n): number {\n const tokens = tokenizeText(runtime, modelType, text);\n return tokens.length;\n}\n\nexport function truncateToTokenLimit(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string,\n maxTokens: number\n): string {\n const tokens = tokenizeText(runtime, modelType, text);\n if (tokens.length <= maxTokens) {\n return text;\n }\n const truncatedTokens = tokens.slice(0, maxTokens);\n return detokenizeText(runtime, modelType, truncatedTokens);\n}\n",
|
|
20
20
|
"import pluginDefault from \"./index\";\n\nexport * from \"./index\";\nexport default pluginDefault;\n"
|
|
21
21
|
],
|
|
22
|
-
"mappings": "AAgBA,iBAAS,eAAQ,sBCfjB,iBAAS,sBCAT,iBAAS,sBAET,SAAS,EAAW,CAAC,EAAiC,CACpD,GAAI,OAAO,QAAY,KAAe,CAAC,QAAQ,IAC7C,OAEF,IAAM,EAAQ,QAAQ,IAAI,GAC1B,OAAO,IAAU,OAAY,OAAY,OAAO,CAAK,EAGhD,SAAS,CAAU,CACxB,EACA,EACA,EACoB,CACpB,IAAM,EAAQ,EAAQ,WAAW,CAAG,EACpC,GAAI,IAAU,QAAa,IAAU,KACnC,OAAO,OAAO,CAAK,EAErB,OAAO,GAAY,CAAG,GAAK,EActB,SAAS,CAAiB,CAC/B,EACA,EACA,EACQ,CACR,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAS,OAAO,SAAS,EAAO,EAAE,EACxC,GAAI,CAAC,OAAO,SAAS,CAAM,EACzB,MAAU,MAAM,YAAY,oCAAsC,GAAO,EAE3E,OAAO,EAGF,SAAS,CAAiB,CAC/B,EACA,EACA,EACS,CACT,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAa,EAAM,YAAY,EACrC,OAAO,IAAe,QAAU,IAAe,KAAO,IAAe,MAGhE,SAAS,CAAS,EAAY,CACnC,OACE,OAAO,WAAe,KACtB,OAAQ,WAAuC,SAAa,IAIzD,SAAS,CAAW,CAAC,EAAiC,CAC3D,OAAO,EAAU,GAAK,CAAC,CAAC,EAAW,EAAS,yBAAyB,EAGhE,SAAS,CAAS,CAAC,EAA4C,CACpE,OAAO,EAAW,EAAS,gBAAgB,EAGtC,SAAS,EAAkB,CAAC,EAA4C,CAC7E,IAAM,EAAkB,EAAW,EAAS,0BAA0B,EACtE,GAAI,EAEF,OADA,EAAO,MAAM,2CAA2C,EACjD,EAGT,OADA,EAAO,MAAM,yDAAyD,EAC/D,EAAU,CAAO,EAGnB,SAAS,CAAa,CAC3B,EACA,EAAe,GACS,CAIxB,GAAI,EAAU,GAAK,CAAC,EAAkB,EAAS,+BAAgC,EAAK,EAClF,MAAO,CAAC,EAEV,IAAM,EAAM,EAAe,GAAmB,CAAO,EAAI,EAAU,CAAO,EAC1E,OAAO,EAAM,CAAE,cAAe,UAAU,GAAM,EAAI,CAAC,EAG9C,SAAS,CAAU,CAAC,EAAgC,CACzD,IAAM,EAAa,EAAW,EAAS,yBAAyB,EAC1D,EACJ,EAAU,GAAK,EACX,EACC,EAAW,EAAS,iBAAiB,GAAK,4BAEjD,OADA,EAAO,MAAM,sBAAsB,GAAS,EACrC,EAGF,SAAS,CAAmB,CAAC,EAAgC,CAClE,IAAM,EAAe,EAAU,EAC1B,EAAW,EAAS,8BAA8B,GACnD,EAAW,EAAS,yBAAyB,EAC7C,EAAW,EAAS,sBAAsB,EAE9C,GAAI,EAEF,OADA,EAAO,MAAM,sCAAsC,GAAc,EAC1D,EAIT,OADA,EAAO,MAAM,0DAA0D,EAChE,EAAW,CAAO,EAGpB,SAAS,CAAa,CAAC,EAAgC,CAC5D,OACE,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,aAIhF,SAAS,CAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,QAGrF,SAAS,CAAiB,CAAC,EAAgC,CAChE,OAAO,EAAW,EAAS,wBAAwB,GAAK,yBAGnD,SAAS,EAAwB,CAAC,EAAgC,CACvE,OAAO,EAAW,EAAS,gCAAgC,GAAK,aAG3D,SAAS,EAAqB,CAAC,EAAgC,CACpE,OAAO,EAAW,EAAS,4BAA4B,GAAK,wBAGvD,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,QAG7C,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,OAG7C,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAW,EAAS,yBAAyB,GAAK,GAGpD,SAAS,EAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,WAG/C,SAAS,EAAwB,CAAC,EAAiC,CACxE,OAAO,EAAkB,EAAS,gCAAiC,EAAK,EAGnE,SAAS,EAAsB,CAAC,EAAgC,CACrE,OAAO,EAAkB,EAAS,8BAA+B,IAAI,EAGhE,SAAS,EAA4B,CAAC,EAAgC,CAC3E,OAAO,EAAkB,EAAS,sCAAuC,IAAI,EAGxE,SAAS,EAAgB,CAAC,EAAgC,CAC/D,OAAO,EAAW,EAAS,uBAAuB,GAAK,mBAGlD,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAkB,EAAS,0BAA2B,OAAO,EDjLrE,WAAuC,sBAAwB,GAEzD,SAAS,EAAgB,CAC9B,EACA,EACM,CACD,GAA4B,CAAO,EAG1C,eAAe,EAA2B,CAAC,EAAuC,CAChF,GAAI,EAAU,EAAG,CACf,EAAO,MAAM,yDAAyD,EACtE,OAKF,GAAI,CAFW,EAAU,CAAO,EAEnB,CACX,EAAO,KACL,8GAEF,EACA,OAGF,GAAI,CACF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,EAAO,KACL,uCAAuC,EAAS,UAAU,EAAS,2DAErE,EACA,QAEF,MAAO,EAAO,CACd,IAAM,EAAU,aAAiB,MAAQ,EAAM,QAAU,OAAO,CAAK,EACrE,EAAO,KAAK,kCAAkC,yCAA+C,GExCjG,iBAAS,sBCLT,iBAAS,uBAET,IAAM,EAAc,CAClB,IAAK,CACH,OAAQ,CAAC,GAAM,GAAM,GAAM,EAAI,EAC/B,WAAY,CAAC,GAAM,GAAM,GAAM,EAAI,CACrC,EACA,QAAS,CAAC,GAAM,GAAM,EAAI,EAC1B,IAAK,CAAC,GAAM,IAAM,IAAM,EAAI,EAC5B,KAAM,CAAC,IAAM,GAAM,GAAM,EAAI,EAC7B,KAAM,CAAC,IAAM,IAAM,IAAM,GAAI,EAC7B,UAAW,CAAC,GAAM,GAAM,IAAM,GAAI,CACpC,EAEM,GAA4B,GAWlC,SAAS,CAAU,CAAC,EAAgB,EAAgB,EAAsC,CACxF,QAAS,EAAI,EAAG,EAAI,EAAS,OAAQ,IAAK,CACxC,IAAM,EAAe,EAAS,GAC9B,GAAI,IAAiB,QAAa,EAAO,EAAS,KAAO,EACvD,MAAO,GAGX,MAAO,GAGF,SAAS,CAAmB,CAAC,EAA+B,CACjE,GAAI,EAAO,OAAS,GAClB,MAAO,2BAIT,GACE,EAAW,EAAQ,EAAG,EAAY,IAAI,MAAM,GAC5C,EAAW,EAAQ,EAAG,EAAY,IAAI,UAAU,EAEhD,MAAO,YAIT,IAAM,EAAY,EAAO,GACnB,EAAa,EAAO,GAC1B,GACE,EAAW,EAAQ,EAAG,EAAY,OAAO,GACxC,IAAc,KAAQ,IAAe,SAAc,EAAa,OAAU,IAE3E,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,GAAG,EACvC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,SAAS,EAC7C,MAAO,aAIT,OADA,GAAO,KAAK,sEAAsE,EAC3E,2BAGF,SAAS,EAAuB,CAAC,EAAiC,CACvE,OAAQ,OACD,YACH,MAAO,UACJ,aACH,MAAO,UACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,2BACH,MAAO,OAIN,SAAS,EAAsB,CAAC,EAAiC,CAEtE,MAAO,aADK,GAAwB,CAAQ,ID3E9C,SAAS,EAAY,CAAC,EAAsC,CAC1D,OAAO,aAAiB,MAAQ,aAAiB,KAGnD,SAAS,CAAQ,CAAC,EAAiC,CACjD,OAAO,OAAO,SAAS,CAAK,EAG9B,SAAS,EAA0B,CAAC,EAAmD,CACrF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,UAAW,IACV,GAAc,EAAmC,KAAK,GACrD,EAAU,EAAmC,KAAK,GAIxD,SAAS,EAAyB,CAAC,EAAkD,CACnF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,aAAc,GACd,OAAQ,EAAkC,WAAa,SAI3D,eAAe,EAAiB,CAAC,EAA4B,CAC3D,IAAM,EAAW,MAAM,MAAM,CAAG,EAChC,GAAI,CAAC,EAAS,GACZ,MAAU,MAAM,mCAAmC,EAAS,QAAQ,EAEtE,OAAO,EAAS,KAAK,EAEvB,eAAsB,CAAmB,CACvC,EACA,EACiB,CACjB,IAAI,EAAY,GAAsB,CAAO,EACzC,EACA,EAAiD,CAAC,EAEtD,GAAI,OAAO,IAAU,SACnB,EAAO,MAAM,qCAAqC,GAAO,EACzD,EAAO,MAAM,GAAkB,CAAK,EAC/B,QAAI,GAAa,CAAK,EAC3B,EAAO,EACF,QAAI,EAAS,CAAK,EAAG,CAC1B,IAAM,EAAW,EAAoB,CAAK,EAC1C,EAAO,MAAM,2CAA2C,GAAU,EAClE,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,CAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EACtD,QAAI,GAA2B,CAAK,EAAG,CAE5C,GADA,EAAc,EACV,EAAM,MACR,EAAY,EAAM,MAEpB,GAAI,EAAS,EAAM,KAAK,EAAG,CACzB,IAAM,EAAW,EAAM,UAAY,EAAoB,EAAM,KAAK,EAClE,EAAO,MAAM,6BAA6B,GAAU,EACpD,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,EAAM,KAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EAEjE,OAAO,EAAM,MAEV,QAAI,GAA0B,CAAK,EACxC,EAAO,MAAM,qCAAqC,EAAM,UAAU,EAClE,EAAO,MAAM,GAAkB,EAAM,QAAQ,EAC7C,EAAc,CAAE,OAAQ,EAAM,MAAO,EAErC,WAAU,MACR,qFACF,EAGF,EAAO,MAAM,uCAAuC,GAAW,EAE/D,IAAM,EAAY,EAAc,MAAQ,aAClC,EACH,EAAc,MACf,GACE,EAAS,WAAW,QAAQ,EACvB,EACD,YACN,EAEI,EAAW,IAAI,SAIrB,GAHA,EAAS,OAAO,OAAQ,EAAM,CAAQ,EACtC,EAAS,OAAO,QAAS,CAAS,EAE9B,EAAY,SACd,EAAS,OAAO,WAAY,EAAY,QAAQ,EAElD,GAAI,EAAY,eACd,EAAS,OAAO,kBAAmB,EAAY,cAAc,EAE/D,GAAI,EAAY,OACd,EAAS,OAAO,SAAU,EAAY,MAAM,EAE9C,GAAI,EAAY,cAAgB,OAC9B,EAAS,OAAO,cAAe,OAAO,EAAY,WAAW,CAAC,EAEhE,GAAI,EAAY,uBACd,QAAW,KAAe,EAAY,uBACpC,EAAS,OAAO,4BAA6B,CAAW,EAI5D,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,yBAAgC,CAC9D,OAAQ,OACR,QAAS,EAAc,CAAO,EAC9B,KAAM,CACR,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,gCAAgC,EAAS,UAAU,EAAS,gBAAgB,GAC9E,EAIF,OADc,MAAM,EAAS,KAAK,GACtB,KAGd,eAAsB,CAAkB,CACtC,EACA,EACsB,CACtB,IAAI,EACA,EACA,EAA0B,MAC1B,EACA,EAEJ,GAAI,OAAO,IAAU,SACnB,EAAO,EACP,EAAQ,OACH,KAGL,GAFA,EAAO,EAAM,KACb,EAAQ,EAAM,MACV,WAAY,GAAS,EAAM,OAC7B,EAAS,EAAM,OAEjB,GAAI,UAAW,GAAS,EAAM,MAC5B,EAAQ,EAAM,MAEhB,GAAI,iBAAkB,GAAS,EAAM,aACnC,EAAe,EAAM,aAUzB,GANA,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAe,GAAgB,GAAmB,CAAO,EAEzD,EAAO,MAAM,wCAAwC,GAAO,EAExD,CAAC,GAAQ,EAAK,KAAK,EAAE,SAAW,EAClC,MAAU,MAAM,wCAAwC,EAG1D,GAAI,EAAK,OAAS,KAChB,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAA0B,CAAC,QAAS,OAAQ,QAAS,OAAQ,OAAQ,SAAS,EACpF,GAAI,GAAS,CAAC,EAAY,SAAS,CAAiB,EAClD,MAAU,MAAM,kBAAkB,sBAA0B,EAAY,KAAK,IAAI,GAAG,EAGtF,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAsC,CAC1C,QACA,MAAO,EACP,MAAO,EACP,gBAAiB,CACnB,EAEA,GAAI,GAAgB,EAAa,OAAS,EACxC,EAAY,aAAe,EAG7B,IAAM,EAAW,MAAM,MAAM,GAAG,iBAAwB,CACtD,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,sBACZ,IAAW,MAAQ,CAAE,OAAQ,YAAa,EAAI,CAAC,CACrD,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MAAM,sBAAsB,EAAS,UAAU,EAAS,gBAAgB,GAAW,EAG/F,OAAO,EAAS,YAAY,EEhO9B,iBAAS,eAAQ,kBAAW,uBCA5B,oBAAS,uBAGT,IAAM,GAAoB,IAmC1B,SAAS,EAAc,CAAC,EAAwB,CAC9C,GAAI,EAAO,QAAU,GACnB,OAAO,EAET,MAAO,GAAG,EAAO,MAAM,EAAG,EAAiB,KAG7C,SAAS,EAAc,CAAC,EAA+B,CACrD,GAAI,iBAAkB,EAAO,CAC3B,IAAM,EACJ,wBAAyB,EAAQ,EAAM,oBAAsB,OACzD,EAAqB,EAAM,oBAAsB,GAAqB,aAC5E,MAAO,CACL,aAAc,EAAM,cAAgB,EACpC,iBAAkB,EAAM,kBAAoB,EAC5C,YAAa,EAAM,cAAgB,EAAM,cAAgB,IAAM,EAAM,kBAAoB,GACzF,oBACF,EAEF,GAAI,gBAAiB,GAAS,iBAAkB,EAAO,CACrD,IAAM,EAAS,EAAqB,aAAe,EAC7C,EAAU,EAAqB,cAAgB,EAC/C,EAAS,EAAqB,aAAe,EAAQ,EAC3D,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,EACb,mBAAqB,EAAqB,iBAC5C,EAEF,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,CACf,EAGK,SAAS,CAAmB,CACjC,EACA,EACA,EACA,EACM,CACN,IAAM,EAAa,GAAe,CAAK,EAEjC,EAAkC,CACtC,UACA,OAAQ,SACR,SAAU,SACV,OACA,OAAQ,GAAe,CAAM,EAC7B,OAAQ,CACN,OAAQ,EAAW,aACnB,WAAY,EAAW,iBACvB,MAAO,EAAW,eACd,EAAW,qBAAuB,OAClC,CAAE,OAAQ,EAAW,kBAAmB,EACxC,CAAC,CACP,CACF,EAEA,EAAQ,UAAU,GAAU,WAAY,CAAO,EDtFjD,SAAS,EAAiB,CAAC,EAAoC,CAC7D,IAAM,EAAkB,OAAO,OAAO,EAAW,EACjD,GAAI,CAAC,EAAgB,SAAS,CAAS,EACrC,MAAU,MACR,gCAAgC,sBAA8B,EAAgB,KAAK,IAAI,GACzF,EAEF,OAAO,EAGT,SAAS,EAAW,CAAC,EAA4D,CAC/E,GAAI,IAAW,KACb,OAAO,KAET,GAAI,OAAO,IAAW,SACpB,OAAO,EAET,GAAI,OAAO,IAAW,UAAY,OAAO,EAAO,OAAS,SACvD,OAAO,EAAO,KAEhB,MAAU,MAAM,sEAAsE,EAGxF,eAAsB,CAAmB,CACvC,EACA,EACmB,CACnB,IAAM,EAAiB,EAAkB,CAAO,EAC1C,EAAqB,GAAkB,GAAuB,CAAO,CAAC,EAEtE,EAAO,GAAY,CAAM,EAC/B,GAAI,IAAS,KAAM,CACjB,EAAO,MAAM,qDAAqD,EAClE,IAAM,EAAiB,MAAM,CAAkB,EAAE,KAAK,CAAC,EAEvD,OADA,EAAW,GAAK,IACT,EAGT,IAAI,EAAc,EAAK,KAAK,EAC5B,GAAI,EAAY,SAAW,EACzB,MAAU,MAAM,0CAA0C,EAM5D,IAAM,EAAW,MACjB,GAAI,EAAY,OAAS,EACvB,EAAO,KACL,uCAAuC,KAAK,KAAK,EAAY,OAAS,CAAC,uCACzE,EACA,EAAc,EAAY,MAAM,EAAG,CAAQ,EAI7C,IAAM,EAAM,GADI,EAAoB,CAAO,eAG3C,EAAO,MAAM,6CAA6C,GAAgB,EAE1E,IAAM,EAAW,MAAM,MAAM,EAAK,CAChC,OAAQ,OACR,QAAS,IACJ,EAAc,EAAS,EAAI,EAC9B,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CACnB,MAAO,EACP,MAAO,CACT,CAAC,CACH,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,+BAA+B,EAAS,UAAU,EAAS,gBAAgB,GAC7E,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAE5B,EAAc,GAAM,OAAO,GACjC,GAAI,CAAC,GAAe,CAAC,EAAY,UAC/B,MAAU,MAAM,0DAA0D,EAG5E,IAAM,EAAY,EAAY,UAE9B,GAAI,EAAU,SAAW,EACvB,MAAU,MACR,qCAAqC,EAAU,oBAAoB,+CAErE,EAGF,GAAI,EAAK,MACP,EAAoB,EAAS,GAAU,eAAgB,EAAa,CAClE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAClB,YAAa,EAAK,MAAM,YAC1B,CAAC,EAIH,OADA,EAAO,MAAM,qCAAqC,EAAU,mBAAmB,EACxE,EEpHT,iBAAS,gBAAQ,uBAwBjB,IAAM,GACJ,0EAEF,eAAsB,CAAqB,CACzC,EACA,EACkC,CAClC,IAAM,EAAY,GAAc,CAAO,EACjC,EAAQ,EAAO,OAAS,EACxB,EAAmB,EAAO,MAAsB,YAChD,EAAiB,EAIvB,GAFA,GAAO,MAAM,+BAA+B,GAAW,EAEnD,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,8CAA8C,EAGhE,GAAI,EAAQ,GAAK,EAAQ,GACvB,MAAU,MAAM,sCAAsC,EAGxD,IAAM,EAAU,EAAW,CAAO,EAE5B,EAA+C,CACnD,MAAO,EACP,OAAQ,EAAO,OACf,EAAG,EACH,MACF,EAEA,GAAI,EAAe,QACjB,EAAY,QAAU,EAAe,QAEvC,GAAI,EAAe,MACjB,EAAY,MAAQ,EAAe,MAGrC,IAAM,EAAW,MAAM,MAAM,GAAG,uBAA8B,CAC5D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,mCAAmC,EAAS,UAAU,EAAS,gBAAgB,GACjF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,CAAC,EAAK,MAAQ,EAAK,KAAK,SAAW,EACrC,MAAU,MAAM,+BAA+B,EAGjD,OAAO,EAAK,KAAK,IAAI,CAAC,KAAU,CAC9B,IAAK,EAAK,IACV,cAAe,EAAK,cACtB,EAAE,EAGJ,SAAS,EAAsB,CAAC,EAAyB,CAEvD,OADmB,EAAQ,MAAM,2BAA2B,IACxC,IAAI,KAAK,GAAK,iBAGpC,SAAS,EAA4B,CAAC,EAAyB,CAC7D,OAAO,EAAQ,QAAQ,4BAA6B,EAAE,EAAE,KAAK,EAG/D,eAAsB,CAAsB,CAC1C,EACA,EACiC,CACjC,IAAM,EAAY,GAAyB,CAAO,EAC5C,EAAY,GAA6B,CAAO,EAEtD,GAAO,MAAM,2CAA2C,GAAW,EAEnE,IAAI,EACA,EAEJ,GAAI,OAAO,IAAW,SACpB,EAAW,EACX,EAAa,GAEb,OAAW,EAAO,SAClB,EAAa,EAAO,QAAU,GAGhC,GAAI,CAAC,GAAY,EAAS,KAAK,EAAE,SAAW,EAC1C,MAAU,MAAM,8CAA8C,EAGhE,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAc,CAClB,MAAO,EACP,SAAU,CACR,CACE,KAAM,OACN,QAAS,CACP,CAAE,KAAM,OAAQ,KAAM,CAAW,EACjC,CAAE,KAAM,YAAa,UAAW,CAAE,IAAK,CAAS,CAAE,CACpD,CACF,CACF,EACA,WAAY,CACd,EAEM,EAAW,MAAM,MAAM,GAAG,qBAA4B,CAC1D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,oCAAoC,EAAS,UAAU,EAAS,gBAAgB,GAClF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MACP,EACE,EACA,GAAU,kBACV,OAAO,IAAW,SAAW,EAAU,EAAO,QAAU,GACxD,CACE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAAK,MAAM,kBAC7B,YAAa,EAAK,MAAM,YAC1B,CACF,EAIF,IAAM,EADc,EAAK,UAAU,IACN,SAAS,QAEtC,GAAI,CAAC,EACH,MAAU,MAAM,6CAA6C,EAG/D,MAAO,CACL,MAAO,GAAuB,CAAO,EACrC,YAAa,GAA6B,CAAO,CACnD,EC/KF,iBAAS,gBAAQ,uBACjB,yBAAS,YCPT,uBAAS,wBAIT,IAAM,GAAgB,WAEf,SAAS,CAAkB,CAAC,EAAwC,CACzE,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAS,EAAU,CAAO,EAEhC,GAAI,CAAC,GAAU,EAAY,CAAO,EAChC,OAAO,GAAa,CAClB,OAAQ,GACR,SACF,CAAC,EAGH,GAAI,CAAC,EACH,MAAU,MACR,uFACF,EAGF,OAAO,GAAa,CAClB,SACA,SACF,CAAC,EC1BH,iBAAS,uBACT,yBAAS,YAST,IAAM,GAAwB,CAC5B,cAAe,uBACf,WAAY,YACd,EAEO,SAAS,EAAqB,EAAuB,CAC1D,MAAO,QAAS,OAAM,WAAsD,CAC1E,GAAI,EAAE,aAAiB,IACrB,OAAO,KAET,GAAI,CACF,IAAM,EAAc,EAAK,QAAQ,GAAsB,cAAe,EAAE,EAGxE,OAFA,KAAK,MAAM,CAAW,EACtB,GAAO,MAAM,wEAAwE,EAC9E,EACP,KAAM,CAEN,OADA,GAAO,KAAK,0CAA0C,EAC/C,OFXb,eAAe,EAAyB,CACtC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAIpC,GAFA,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAE1D,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,+CAA+C,EAGjE,GAAI,EAAO,OACT,GAAO,MACL,oGAEF,EAGF,IAAM,EAAQ,EAAO,KAAK,CAAS,GAC3B,SAAQ,SAAU,MAAM,GAAe,CAC7C,QACA,OAAQ,YACR,OAAQ,EAAO,OACf,wBAAyB,GAAsB,CACjD,CAAC,EAED,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,GAAI,OAAO,IAAW,UAAY,IAAW,KAC3C,MAAU,MAAM,8BAA8B,OAAO,oBAAyB,EAGhF,OAAO,EAGT,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EAGzF,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EG7CzF,iBAAS,sBA6ET,SAAS,EAAgB,CAAC,EAAsC,CAC9D,OAAQ,EAAK,UACN,qBACH,MAAO,CAAE,KAAM,oBAAqB,MACjC,cACH,MAAO,CACL,KAAM,cACN,iBAAkB,EAAK,cACzB,MACG,mBACH,MAAO,CACL,KAAM,mBACN,UAAW,EAAK,WAAa,CAAE,KAAM,MAAO,CAC9C,MACG,MACH,MAAO,CACL,KAAM,MACN,aAAc,EAAK,YACnB,WAAY,EAAK,UACjB,iBAAkB,EAAK,iBAAmB,OAC5C,UAEA,MAAU,MAAM,+BAAgC,EAAsB,MAAM,GAOlF,SAAS,EAAiB,CAAC,EAAyD,CAClF,OAAQ,EAAK,UACN,kBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,kBACN,OAAS,EAAK,QAAqC,YACnD,OAAQ,CACN,KAAO,EAAK,QAAQ,MAAoD,SACxE,MAAO,EAAK,QAAQ,MACpB,IAAK,EAAK,QAAQ,GACpB,CACF,MAEG,mBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,mBACN,OAAS,EAAK,QAAqC,YACnD,MAAO,EAAK,OAAS,GACrB,QAAS,EAAK,SAAS,IAAI,CAAC,KAAO,CACjC,OAAQ,EAAE,QACV,SAAU,EAAE,UACZ,MAAO,EAAE,KACX,EAAE,CACJ,MAEG,wBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,wBACN,OAAS,EAAK,QAAqC,YACnD,KAAM,EAAK,MAAQ,GACnB,OAAQ,EAAK,MACf,MAEG,gBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,gBACN,OAAS,EAAK,QAAqC,YACnD,YAAa,EAAK,cAAgB,GAClC,SAAU,EAAK,WAAa,GAC5B,UAAY,EAAK,WAAa,CAAC,EAC/B,OAAQ,EAAK,MACf,MAEG,UACH,MAAO,CACL,KAAM,UACN,QACE,EAAK,SAAS,IAAI,CAAC,KAAO,CACxB,KAAM,cACN,KAAM,EAAE,KACR,YACE,EAAE,aAAa,IAAI,CAAC,KAAO,CACzB,IAAK,EAAE,IACP,MAAO,EAAE,MACT,WAAY,EAAE,YACd,SAAU,EAAE,SACd,EAAE,GAAK,CAAC,CACZ,EAAE,GAAK,CAAC,CACZ,UAIA,OAAO,MAOb,SAAS,EAAyB,CAAC,EAGjC,CAEA,GAAI,EAAS,YAAa,CAExB,IAAM,EAAoC,CAAC,EAC3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QACzB,GAAI,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAOb,MAAO,CAAE,KAAM,EAAS,YAAa,aAAY,EAInD,IAAI,EAAO,GACL,EAAoC,CAAC,EAE3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QAEzB,GADA,GAAQ,EAAQ,KACZ,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAQb,MAAO,CAAE,OAAM,aAAY,EA8B7B,eAAsB,CAAc,CAClC,EACA,EACyB,CACzB,IAAM,EAAS,EAAU,CAAO,EAChC,GAAI,CAAC,EACH,MAAU,MACR,yGACF,EAGF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAY,EAAO,OAAS,GAAiB,CAAO,EACpD,EAAU,GAAmB,CAAO,EAE1C,EAAO,MAAM,+CAA+C,GAAW,EACvE,EAAO,MAAM,4BAA4B,EAAO,MAAM,UAAU,EAAG,GAAG,MAAM,EAG5E,IAAM,EAAkB,EAAO,OAAO,OACpC,CAAC,IAAM,EAAE,OAAS,sBAAwB,EAAE,OAAS,eAAiB,EAAE,OAAS,KACnF,EAEA,GAAI,CAAC,GAAmB,EAAgB,SAAW,EAEjD,EAAO,MAAM,2EAA2E,EACxF,EAAO,MAAQ,CAAC,CAAE,KAAM,oBAAqB,EAAG,GAAI,EAAO,OAAS,CAAC,CAAE,EAIzE,IAAM,EAAuC,CAC3C,MAAO,EACP,MAAO,EAAO,KAChB,EAEA,GAAI,EAAO,aACT,EAAY,aAAe,EAAO,aAGpC,GAAI,EAAO,aAAe,OACxB,EAAY,WAAa,EAAO,WAGlC,GAAI,EAAO,OAAS,EAAO,MAAM,OAAS,EACxC,EAAY,MAAQ,EAAO,MAAM,IAAI,EAAgB,EAGvD,GAAI,EAAO,eAAiB,OAC1B,EAAY,eAAiB,EAAO,aAGtC,GAAI,EAAO,iBACT,EAAY,UAAY,CAAE,QAAS,EAAO,gBAAiB,EAG7D,EAAO,MAAM,mCAAmC,KAAK,UAAU,EAAa,KAAM,CAAC,GAAG,EAGtF,IAAM,EAAW,MAAM,MAAM,GAAG,cAAqB,CACnD,OAAQ,OACR,QAAS,CACP,cAAe,UAAU,IACzB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,EAChC,OAAQ,YAAY,QAAQ,CAAO,CACrC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAEtC,MADA,EAAO,MAAM,qCAAqC,EAAS,UAAU,GAAW,EACtE,MAAM,iCAAiC,EAAS,UAAU,EAAS,YAAY,EAG3F,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MAEP,MADA,EAAO,MAAM,gCAAgC,EAAK,MAAM,SAAS,EACvD,MAAM,wBAAwB,EAAK,MAAM,SAAS,EAG9D,EAAO,MAAM,gDAAgD,EAAK,QAAU,aAAa,EAGzF,IAAQ,OAAM,eAAgB,GAA0B,CAAI,EAGtD,EAAoC,CAAC,EAC3C,GAAI,EAAK,OACP,QAAW,KAAQ,EAAK,OAAQ,CAC9B,IAAM,EAAY,GAAkB,CAAI,EACxC,GAAI,EACF,EAAY,KAAK,CAAS,EAKhC,IAAM,EAAyB,CAC7B,GAAI,EAAK,GACT,OACA,cACA,cACA,OAAQ,EAAK,MACf,EAMA,OAJA,EAAO,KACL,6CAA6C,EAAK,wBAAwB,EAAY,yBAAyB,EAAY,QAC7H,EAEO,ECnYT,iBAAS,gBAAQ,uBACjB,uBAAS,iBAAuC,YAuChD,SAAS,EAAY,CAAC,EAA+D,CACnF,GAAI,CAAC,EACH,OAIF,IAAM,EAAe,EAAM,aAAe,EACpC,EAAmB,EAAM,cAAgB,EACzC,EAAiB,EAEvB,MAAO,CACL,eACA,mBACA,YAAa,EAAe,EAC5B,mBAAoB,EAAe,iBACrC,EAGF,SAAS,EAAyB,CAAC,EAAsD,CACvF,IAAM,EAAoB,EAC1B,MAAO,CACL,eAAgB,EAAkB,iBAAiB,QAAQ,eAC3D,qBAAsB,EAAkB,iBAAiB,QAAQ,oBACnE,EAgBF,eAAe,EAAuB,CACpC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAEpC,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAC9D,IAAM,EAAqB,GAA0B,CAAM,EAGrD,EAAe,EAAQ,UAAU,QAAU,OAO3C,EAAiB,CACrB,MAFY,EAAO,KAAK,CAAS,EAGjC,OAAQ,EAAO,OACf,OAAQ,EACR,gBAAiB,EAAO,WAAa,KACrC,uBAAwB,CAAE,UAAW,GAAyB,CAAO,CAAE,KACnE,EAAmB,gBAAkB,EAAmB,qBACxD,CACE,gBAAiB,CACf,OAAQ,IACF,EAAmB,eACnB,CAAE,eAAgB,EAAmB,cAAe,EACpD,CAAC,KACD,EAAmB,qBACnB,CAAE,qBAAsB,EAAmB,oBAAqB,EAChE,CAAC,CACP,CACF,CACF,EACA,CAAC,CACP,EAGA,GAAI,EAAO,OAAQ,CACjB,IAAM,EAAS,GAAW,CAAc,EAExC,MAAO,CACL,WAAY,EAAO,WACnB,KAAM,QAAQ,QAAQ,EAAO,IAAI,EACjC,MAAO,QAAQ,QAAQ,EAAO,KAAK,EAAE,KAAK,EAAY,EACtD,aAAc,QAAQ,QAAQ,EAAO,YAAY,EAAE,KAAK,CAAC,IAAM,CAAuB,CACxF,EAIF,IAAQ,OAAM,SAAU,MAAM,GAAa,CAAc,EAEzD,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,OAAO,EAgBT,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,EAYrF,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,ECtLrF,oBAAS,uBCAT,oBAAS,uBACT,2BACE,kBACA,qBASF,SAAS,EAAwB,CAAC,EAA6B,CAE7D,IAAM,EADa,EAAU,YAAY,EACc,SAAS,IAAI,EAChE,aACA,cACJ,GAAI,CACF,OAAO,GAAiB,CAA0B,EAClD,KAAM,CACN,OAAO,GAAY,CAAoC,GAI3D,SAAS,EAAY,CAAC,EAAwB,EAAkC,CAC9E,GAAI,IAAc,GAAU,WAC1B,OAAO,EAAc,CAAO,EAE9B,OAAO,EAAc,CAAO,EAGvB,SAAS,EAAY,CAC1B,EACA,EACA,EACU,CACV,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAI,EAGrB,SAAS,EAAc,CAC5B,EACA,EACA,EACQ,CACR,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAM,ED7C9B,eAAsB,CAAqB,CACzC,EACA,EACmB,CACnB,GAAI,CAAC,EAAO,OACV,MAAU,MAAM,0CAA0C,EAE5D,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAa,EAAS,EAAW,EAAO,MAAM,EAGvD,eAAsB,CAAqB,CACzC,EACA,EACiB,CACjB,GAAI,CAAC,EAAO,QAAU,CAAC,MAAM,QAAQ,EAAO,MAAM,EAChD,MAAU,MAAM,8CAA8C,EAEhE,GAAI,EAAO,OAAO,SAAW,EAC3B,MAAO,GAET,QAAS,EAAI,EAAG,EAAI,EAAO,OAAO,OAAQ,IAAK,CAC7C,IAAM,EAAQ,EAAO,OAAO,GAC5B,GAAI,OAAO,IAAU,UAAY,CAAC,OAAO,SAAS,CAAK,EACrD,MAAU,MAAM,0BAA0B,oBAAoB,EAGlE,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAe,EAAS,EAAW,EAAO,MAAM,EbKzD,SAAS,EAAa,EAAmB,CACvC,GAAI,OAAO,QAAY,IACrB,MAAO,CAAC,EAEV,OAAO,QAAQ,IAGjB,IAAM,EAAM,GAAc,EAEb,GAAuB,CAClC,KAAM,SACN,YAAa,sEAEb,OAAQ,CACN,eAAgB,EAAI,gBAAkB,KACtC,gBAAiB,EAAI,iBAAmB,KACxC,mBAAoB,EAAI,oBAAsB,KAC9C,mBAAoB,EAAI,oBAAsB,KAC9C,YAAa,EAAI,aAAe,KAChC,YAAa,EAAI,aAAe,KAChC,uBAAwB,EAAI,wBAA0B,KACtD,yBAA0B,EAAI,0BAA4B,KAC1D,qBAAsB,EAAI,sBAAwB,KAClD,4BAA6B,EAAI,6BAA+B,KAChE,+BAAgC,EAAI,gCAAkC,KACtE,oCAAqC,EAAI,qCAAuC,KAChF,8BAA+B,EAAI,+BAAiC,KACpE,sBAAuB,EAAI,uBAAyB,KACpD,wBAAyB,EAAI,yBAA2B,IAC1D,OAEM,KAAI,CAAC,EAAgC,EAAuC,CAChF,GAAiB,EAA0C,CAAO,GAGpE,OAAQ,EACL,EAAU,gBAAiB,MAC1B,EACA,IACsB,CACtB,OAAO,EAAoB,EAAS,CAAM,IAG3C,EAAU,uBAAwB,MACjC,EACA,IACsB,CACtB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,uBAAwB,MACjC,EACA,IACoB,CACpB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,OAAQ,MACjB,EACA,IACqC,CACrC,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,mBAAoB,MAC7B,EACA,IACoD,CACpD,OAAO,EAAuB,EAAS,CAAM,IAG9C,EAAU,eAAgB,MACzB,EACA,IACoB,CACpB,OAAO,EAAoB,EAAS,CAAK,IAG1C,EAAU,gBAAiB,MAC1B,EACA,IACyB,CACzB,OAAO,EAAmB,EAAS,CAAK,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,UAAW,MACpB,EACA,IAC4B,CAC5B,OAAO,EAAe,EAAS,CAAM,EAEzC,EAEA,MAAO,CACL,CACE,KAAM,sBACN,MAAO,CACL,CACE,KAAM,+BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GACZ,MAAU,MACR,iCAAiC,EAAS,UAAU,EAAS,YAC/D,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAClC,EAAO,KAAK,gCAAgC,EAAK,MAAM,QAAU,qBAAqB,EAE1F,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,eACR,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAS,GAAK,EAAU,SAAW,EACpD,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,0CAA0C,EAAU,mBAAmB,EAEvF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,+BACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,2CACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,kCACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CACrE,OAHmB,yBAInB,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAM,GAAK,EAAO,SAAW,EAC9C,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAAc,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CAC1E,SACA,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,IAhBiB,yBAiBnB,MAAU,MACR,uEAAgE,IAClE,EAGF,EAAO,KAAK,iDAAiD,EAAO,gBAAgB,EAExF,EACA,CACE,KAAM,wBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAmB,CAAC,EAEpB,EAAS,MAAM,EAAQ,SAAS,EAAU,WAAY,CAC1D,OAAQ,0CACR,OAAQ,GACR,cAAe,CAAC,IAAkB,CAChC,EAAO,KAAK,CAAK,EAErB,CAAC,EAED,GAAI,OAAO,IAAW,UAAY,EAAO,SAAW,EAClD,MAAU,MAAM,0CAA0C,EAG5D,GAAI,EAAO,SAAW,EACpB,MAAU,MAAM,8BAA8B,EAGhD,EAAO,KAAK,iCAAiC,EAAO,wBAAwB,EAEhF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CAInD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,kBAF9C,uIAE6E,EAE/E,GACE,CAAC,GACD,OAAO,IAAW,UAClB,EAAE,UAAW,IACb,EAAE,gBAAiB,GAEnB,MAAU,MAAM,wDAAwD,EAG1E,EAAO,KAAK,mCAAmC,EAAO,QAAQ,EAElE,EACA,CACE,KAAM,4BACN,GAAI,MAAO,IAA0C,CAMnD,IAAM,EAAc,MADH,MAAM,MAFrB,wEAEmC,GACF,YAAY,EACzC,EAAc,OAAO,KAAK,IAAI,WAAW,CAAW,CAAC,EAErD,EAAgB,MAAM,EAAQ,SAAS,EAAU,cAAe,CAAW,EAEjF,GAAI,OAAO,IAAkB,SAC3B,MAAU,MAAM,sCAAsC,EAGxD,EAAO,KAAK,iCAAiC,EAAc,UAAU,EAAG,EAAE,OAAO,EAErF,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,uCACR,CAAC,EAED,GAAI,EAAE,aAAqB,cAAgB,EAAU,aAAe,EAClE,MAAU,MAAM,yCAAyC,EAG3D,EAAO,KAAK,+BAA+B,EAAU,2BAA2B,EAEpF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,aAAc,CAC5D,OACE,+FACJ,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,SAC/B,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KACL,mCAAmC,KAAK,UAAU,CAAM,EAAE,UAAU,EAAG,GAAG,GAC5E,EAEJ,EACA,CACE,KAAM,uBACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,SAAU,CACxD,MAAO,qCACP,MAAO,CAAC,CAAE,KAAM,oBAAqB,CAAC,EACtC,aAAc,CAChB,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,UAAY,EAAE,SAAU,GACvD,MAAU,MAAM,qDAAqD,EAGvE,GAAI,OAAO,EAAO,OAAS,UAAY,EAAO,KAAK,SAAW,EAC5D,MAAU,MAAM,mDAAmD,EAGrE,EAAO,KACL,kDAAkD,EAAO,KAAK,wBAAwB,EAAO,aAAa,QAAU,GACtH,EAEJ,CACF,CACF,CACF,CACF,EAEe,MenXf,IAAe",
|
|
23
|
-
"debugId": "
|
|
22
|
+
"mappings": "AAgBA,iBAAS,eAAQ,sBCfjB,iBAAS,sBCAT,iBAAS,sBAET,SAAS,EAAW,CAAC,EAAiC,CACpD,GAAI,OAAO,QAAY,KAAe,CAAC,QAAQ,IAC7C,OAEF,IAAM,EAAQ,QAAQ,IAAI,GAC1B,OAAO,IAAU,OAAY,OAAY,OAAO,CAAK,EAGhD,SAAS,CAAU,CACxB,EACA,EACA,EACoB,CACpB,IAAM,EAAQ,EAAQ,WAAW,CAAG,EACpC,GAAI,IAAU,QAAa,IAAU,KACnC,OAAO,OAAO,CAAK,EAErB,OAAO,GAAY,CAAG,GAAK,EActB,SAAS,CAAiB,CAC/B,EACA,EACA,EACQ,CACR,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAS,OAAO,SAAS,EAAO,EAAE,EACxC,GAAI,CAAC,OAAO,SAAS,CAAM,EACzB,MAAU,MAAM,YAAY,oCAAsC,GAAO,EAE3E,OAAO,EAGF,SAAS,CAAiB,CAC/B,EACA,EACA,EACS,CACT,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAa,EAAM,YAAY,EACrC,OAAO,IAAe,QAAU,IAAe,KAAO,IAAe,MAGhE,SAAS,CAAS,EAAY,CACnC,OACE,OAAO,WAAe,KACtB,OAAQ,WAAuC,SAAa,IAIzD,SAAS,CAAW,CAAC,EAAiC,CAC3D,OAAO,EAAU,GAAK,CAAC,CAAC,EAAW,EAAS,yBAAyB,EAGhE,SAAS,CAAS,CAAC,EAA4C,CACpE,OAAO,EAAW,EAAS,gBAAgB,EAGtC,SAAS,EAAkB,CAAC,EAA4C,CAC7E,IAAM,EAAkB,EAAW,EAAS,0BAA0B,EACtE,GAAI,EAEF,OADA,EAAO,MAAM,2CAA2C,EACjD,EAGT,OADA,EAAO,MAAM,yDAAyD,EAC/D,EAAU,CAAO,EAGnB,SAAS,CAAa,CAC3B,EACA,EAAe,GACS,CAIxB,GAAI,EAAU,GAAK,CAAC,EAAkB,EAAS,+BAAgC,EAAK,EAClF,MAAO,CAAC,EAEV,IAAM,EAAM,EAAe,GAAmB,CAAO,EAAI,EAAU,CAAO,EAC1E,OAAO,EAAM,CAAE,cAAe,UAAU,GAAM,EAAI,CAAC,EAG9C,SAAS,CAAU,CAAC,EAAgC,CACzD,IAAM,EAAa,EAAW,EAAS,yBAAyB,EAC1D,EACJ,EAAU,GAAK,EACX,EACC,EAAW,EAAS,iBAAiB,GAAK,4BAEjD,OADA,EAAO,MAAM,sBAAsB,GAAS,EACrC,EAGF,SAAS,CAAmB,CAAC,EAAgC,CAClE,IAAM,EAAe,EAAU,EAC1B,EAAW,EAAS,8BAA8B,GACnD,EAAW,EAAS,yBAAyB,EAC7C,EAAW,EAAS,sBAAsB,EAE9C,GAAI,EAEF,OADA,EAAO,MAAM,sCAAsC,GAAc,EAC1D,EAIT,OADA,EAAO,MAAM,0DAA0D,EAChE,EAAW,CAAO,EAGpB,SAAS,CAAa,CAAC,EAAgC,CAC5D,OACE,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,aAIhF,SAAS,CAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,QAGrF,SAAS,CAAiB,CAAC,EAAgC,CAChE,OAAO,EAAW,EAAS,wBAAwB,GAAK,yBAGnD,SAAS,EAAwB,CAAC,EAAgC,CACvE,OAAO,EAAW,EAAS,gCAAgC,GAAK,aAG3D,SAAS,EAAqB,CAAC,EAAgC,CACpE,OAAO,EAAW,EAAS,4BAA4B,GAAK,wBAGvD,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,QAG7C,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,OAG7C,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAW,EAAS,yBAAyB,GAAK,GAGpD,SAAS,EAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,WAG/C,SAAS,EAAwB,CAAC,EAAiC,CACxE,OAAO,EAAkB,EAAS,gCAAiC,EAAK,EAGnE,SAAS,EAAsB,CAAC,EAAgC,CACrE,OAAO,EAAkB,EAAS,8BAA+B,IAAI,EAGhE,SAAS,EAA4B,CAAC,EAAgC,CAC3E,OAAO,EAAkB,EAAS,sCAAuC,IAAI,EAGxE,SAAS,EAAgB,CAAC,EAAgC,CAC/D,OAAO,EAAW,EAAS,uBAAuB,GAAK,mBAGlD,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAkB,EAAS,0BAA2B,OAAO,EDjLrE,WAAuC,sBAAwB,GAEzD,SAAS,EAAgB,CAC9B,EACA,EACM,CACD,GAA4B,CAAO,EAG1C,eAAe,EAA2B,CAAC,EAAuC,CAChF,GAAI,EAAU,EAAG,CACf,EAAO,MAAM,yDAAyD,EACtE,OAKF,GAAI,CAFW,EAAU,CAAO,EAEnB,CACX,EAAO,KACL,8GAEF,EACA,OAGF,GAAI,CACF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,EAAO,KACL,uCAAuC,EAAS,UAAU,EAAS,2DAErE,EACA,QAEF,MAAO,EAAO,CACd,IAAM,EAAU,aAAiB,MAAQ,EAAM,QAAU,OAAO,CAAK,EACrE,EAAO,KAAK,kCAAkC,yCAA+C,GExCjG,iBAAS,sBCLT,iBAAS,uBAET,IAAM,EAAc,CAClB,IAAK,CACH,OAAQ,CAAC,GAAM,GAAM,GAAM,EAAI,EAC/B,WAAY,CAAC,GAAM,GAAM,GAAM,EAAI,CACrC,EACA,QAAS,CAAC,GAAM,GAAM,EAAI,EAC1B,IAAK,CAAC,GAAM,IAAM,IAAM,EAAI,EAC5B,KAAM,CAAC,IAAM,GAAM,GAAM,EAAI,EAC7B,KAAM,CAAC,IAAM,IAAM,IAAM,GAAI,EAC7B,UAAW,CAAC,GAAM,GAAM,IAAM,GAAI,CACpC,EAEM,GAA4B,GAWlC,SAAS,CAAU,CAAC,EAAgB,EAAgB,EAAsC,CACxF,QAAS,EAAI,EAAG,EAAI,EAAS,OAAQ,IAAK,CACxC,IAAM,EAAe,EAAS,GAC9B,GAAI,IAAiB,QAAa,EAAO,EAAS,KAAO,EACvD,MAAO,GAGX,MAAO,GAGF,SAAS,CAAmB,CAAC,EAA+B,CACjE,GAAI,EAAO,OAAS,GAClB,MAAO,2BAIT,GACE,EAAW,EAAQ,EAAG,EAAY,IAAI,MAAM,GAC5C,EAAW,EAAQ,EAAG,EAAY,IAAI,UAAU,EAEhD,MAAO,YAIT,IAAM,EAAY,EAAO,GACnB,EAAa,EAAO,GAC1B,GACE,EAAW,EAAQ,EAAG,EAAY,OAAO,GACxC,IAAc,KAAQ,IAAe,SAAc,EAAa,OAAU,IAE3E,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,GAAG,EACvC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,SAAS,EAC7C,MAAO,aAIT,OADA,GAAO,KAAK,sEAAsE,EAC3E,2BAGF,SAAS,EAAuB,CAAC,EAAiC,CACvE,OAAQ,OACD,YACH,MAAO,UACJ,aACH,MAAO,UACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,2BACH,MAAO,OAIN,SAAS,EAAsB,CAAC,EAAiC,CAEtE,MAAO,aADK,GAAwB,CAAQ,ID3E9C,SAAS,EAAY,CAAC,EAAsC,CAC1D,OAAO,aAAiB,MAAQ,aAAiB,KAGnD,SAAS,CAAQ,CAAC,EAAiC,CACjD,OAAO,OAAO,SAAS,CAAK,EAG9B,SAAS,EAA0B,CAAC,EAAmD,CACrF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,UAAW,IACV,GAAc,EAAmC,KAAK,GACrD,EAAU,EAAmC,KAAK,GAIxD,SAAS,EAAyB,CAAC,EAAkD,CACnF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,aAAc,GACd,OAAQ,EAAkC,WAAa,SAI3D,eAAe,EAAiB,CAAC,EAA4B,CAC3D,IAAM,EAAW,MAAM,MAAM,CAAG,EAChC,GAAI,CAAC,EAAS,GACZ,MAAU,MAAM,mCAAmC,EAAS,QAAQ,EAEtE,OAAO,EAAS,KAAK,EAEvB,eAAsB,CAAmB,CACvC,EACA,EACiB,CACjB,IAAI,EAAY,GAAsB,CAAO,EACzC,EACA,EAAiD,CAAC,EAEtD,GAAI,OAAO,IAAU,SACnB,EAAO,MAAM,qCAAqC,GAAO,EACzD,EAAO,MAAM,GAAkB,CAAK,EAC/B,QAAI,GAAa,CAAK,EAC3B,EAAO,EACF,QAAI,EAAS,CAAK,EAAG,CAC1B,IAAM,EAAW,EAAoB,CAAK,EAC1C,EAAO,MAAM,2CAA2C,GAAU,EAClE,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,CAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EACtD,QAAI,GAA2B,CAAK,EAAG,CAE5C,GADA,EAAc,EACV,EAAM,MACR,EAAY,EAAM,MAEpB,GAAI,EAAS,EAAM,KAAK,EAAG,CACzB,IAAM,EAAW,EAAM,UAAY,EAAoB,EAAM,KAAK,EAClE,EAAO,MAAM,6BAA6B,GAAU,EACpD,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,EAAM,KAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EAEjE,OAAO,EAAM,MAEV,QAAI,GAA0B,CAAK,EACxC,EAAO,MAAM,qCAAqC,EAAM,UAAU,EAClE,EAAO,MAAM,GAAkB,EAAM,QAAQ,EAC7C,EAAc,CAAE,OAAQ,EAAM,MAAO,EAErC,WAAU,MACR,qFACF,EAGF,EAAO,MAAM,uCAAuC,GAAW,EAE/D,IAAM,EAAY,EAAc,MAAQ,aAClC,EACH,EAAc,MACf,GACE,EAAS,WAAW,QAAQ,EACvB,EACD,YACN,EAEI,EAAW,IAAI,SAIrB,GAHA,EAAS,OAAO,OAAQ,EAAM,CAAQ,EACtC,EAAS,OAAO,QAAS,CAAS,EAE9B,EAAY,SACd,EAAS,OAAO,WAAY,EAAY,QAAQ,EAElD,GAAI,EAAY,eACd,EAAS,OAAO,kBAAmB,EAAY,cAAc,EAE/D,GAAI,EAAY,OACd,EAAS,OAAO,SAAU,EAAY,MAAM,EAE9C,GAAI,EAAY,cAAgB,OAC9B,EAAS,OAAO,cAAe,OAAO,EAAY,WAAW,CAAC,EAEhE,GAAI,EAAY,uBACd,QAAW,KAAe,EAAY,uBACpC,EAAS,OAAO,4BAA6B,CAAW,EAI5D,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,yBAAgC,CAC9D,OAAQ,OACR,QAAS,EAAc,CAAO,EAC9B,KAAM,CACR,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,gCAAgC,EAAS,UAAU,EAAS,gBAAgB,GAC9E,EAIF,OADc,MAAM,EAAS,KAAK,GACtB,KAGd,eAAsB,CAAkB,CACtC,EACA,EACsB,CACtB,IAAI,EACA,EACA,EAA0B,MAC1B,EACA,EAEJ,GAAI,OAAO,IAAU,SACnB,EAAO,EACP,EAAQ,OACH,KAGL,GAFA,EAAO,EAAM,KACb,EAAQ,EAAM,MACV,WAAY,GAAS,EAAM,OAC7B,EAAS,EAAM,OAEjB,GAAI,UAAW,GAAS,EAAM,MAC5B,EAAQ,EAAM,MAEhB,GAAI,iBAAkB,GAAS,EAAM,aACnC,EAAe,EAAM,aAUzB,GANA,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAe,GAAgB,GAAmB,CAAO,EAEzD,EAAO,MAAM,wCAAwC,GAAO,EAExD,CAAC,GAAQ,EAAK,KAAK,EAAE,SAAW,EAClC,MAAU,MAAM,wCAAwC,EAG1D,GAAI,EAAK,OAAS,KAChB,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAA0B,CAAC,QAAS,OAAQ,QAAS,OAAQ,OAAQ,SAAS,EACpF,GAAI,GAAS,CAAC,EAAY,SAAS,CAAiB,EAClD,MAAU,MAAM,kBAAkB,sBAA0B,EAAY,KAAK,IAAI,GAAG,EAGtF,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAsC,CAC1C,QACA,MAAO,EACP,MAAO,EACP,gBAAiB,CACnB,EAEA,GAAI,GAAgB,EAAa,OAAS,EACxC,EAAY,aAAe,EAG7B,IAAM,EAAW,MAAM,MAAM,GAAG,iBAAwB,CACtD,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,sBACZ,IAAW,MAAQ,CAAE,OAAQ,YAAa,EAAI,CAAC,CACrD,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MAAM,sBAAsB,EAAS,UAAU,EAAS,gBAAgB,GAAW,EAG/F,OAAO,EAAS,YAAY,EEhO9B,iBAAS,eAAQ,kBAAW,uBCA5B,oBAAS,uBAGT,IAAM,GAAoB,IAmC1B,SAAS,EAAc,CAAC,EAAwB,CAC9C,GAAI,EAAO,QAAU,GACnB,OAAO,EAET,MAAO,GAAG,EAAO,MAAM,EAAG,EAAiB,KAG7C,SAAS,EAAc,CAAC,EAA+B,CACrD,GAAI,iBAAkB,EAAO,CAC3B,IAAM,EACJ,wBAAyB,EAAQ,EAAM,oBAAsB,OACzD,EAAqB,EAAM,oBAAsB,GAAqB,aAC5E,MAAO,CACL,aAAc,EAAM,cAAgB,EACpC,iBAAkB,EAAM,kBAAoB,EAC5C,YAAa,EAAM,cAAgB,EAAM,cAAgB,IAAM,EAAM,kBAAoB,GACzF,oBACF,EAEF,GAAI,gBAAiB,GAAS,iBAAkB,EAAO,CACrD,IAAM,EAAS,EAAqB,aAAe,EAC7C,EAAU,EAAqB,cAAgB,EAC/C,EAAS,EAAqB,aAAe,EAAQ,EAC3D,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,EACb,mBAAqB,EAAqB,iBAC5C,EAEF,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,CACf,EAGK,SAAS,CAAmB,CACjC,EACA,EACA,EACA,EACM,CACN,IAAM,EAAa,GAAe,CAAK,EAEjC,EAAkC,CACtC,UACA,OAAQ,SACR,SAAU,SACV,OACA,OAAQ,GAAe,CAAM,EAC7B,OAAQ,CACN,OAAQ,EAAW,aACnB,WAAY,EAAW,iBACvB,MAAO,EAAW,eACd,EAAW,qBAAuB,OAClC,CAAE,OAAQ,EAAW,kBAAmB,EACxC,CAAC,CACP,CACF,EAEA,EAAQ,UAAU,GAAU,WAAY,CAAO,EDtFjD,SAAS,EAAiB,CAAC,EAAoC,CAC7D,IAAM,EAAkB,OAAO,OAAO,EAAW,EACjD,GAAI,CAAC,EAAgB,SAAS,CAAS,EACrC,MAAU,MACR,gCAAgC,sBAA8B,EAAgB,KAAK,IAAI,GACzF,EAEF,OAAO,EAGT,SAAS,EAAW,CAAC,EAA4D,CAC/E,GAAI,IAAW,KACb,OAAO,KAET,GAAI,OAAO,IAAW,SACpB,OAAO,EAET,GAAI,OAAO,IAAW,UAAY,OAAO,EAAO,OAAS,SACvD,OAAO,EAAO,KAEhB,MAAU,MAAM,sEAAsE,EAGxF,eAAsB,CAAmB,CACvC,EACA,EACmB,CACnB,IAAM,EAAiB,EAAkB,CAAO,EAC1C,EAAqB,GAAkB,GAAuB,CAAO,CAAC,EAEtE,EAAO,GAAY,CAAM,EAC/B,GAAI,IAAS,KAAM,CACjB,EAAO,MAAM,qDAAqD,EAClE,IAAM,EAAiB,MAAM,CAAkB,EAAE,KAAK,CAAC,EAEvD,OADA,EAAW,GAAK,IACT,EAGT,IAAI,EAAc,EAAK,KAAK,EAC5B,GAAI,EAAY,SAAW,EACzB,MAAU,MAAM,0CAA0C,EAM5D,IAAM,EAAW,MACjB,GAAI,EAAY,OAAS,EACvB,EAAO,KACL,uCAAuC,KAAK,KAAK,EAAY,OAAS,CAAC,uCACzE,EACA,EAAc,EAAY,MAAM,EAAG,CAAQ,EAI7C,IAAM,EAAM,GADI,EAAoB,CAAO,eAG3C,EAAO,MAAM,6CAA6C,GAAgB,EAE1E,IAAM,EAAW,MAAM,MAAM,EAAK,CAChC,OAAQ,OACR,QAAS,IACJ,EAAc,EAAS,EAAI,EAC9B,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CACnB,MAAO,EACP,MAAO,CACT,CAAC,CACH,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,+BAA+B,EAAS,UAAU,EAAS,gBAAgB,GAC7E,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAE5B,EAAc,GAAM,OAAO,GACjC,GAAI,CAAC,GAAa,UAChB,MAAU,MAAM,0DAA0D,EAG5E,IAAM,EAAY,EAAY,UAE9B,GAAI,EAAU,SAAW,EACvB,MAAU,MACR,qCAAqC,EAAU,oBAAoB,+CAErE,EAGF,GAAI,EAAK,MACP,EAAoB,EAAS,GAAU,eAAgB,EAAa,CAClE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAClB,YAAa,EAAK,MAAM,YAC1B,CAAC,EAIH,OADA,EAAO,MAAM,qCAAqC,EAAU,mBAAmB,EACxE,EEpHT,iBAAS,gBAAQ,uBAwBjB,IAAM,GACJ,0EAEF,eAAsB,CAAqB,CACzC,EACA,EACkC,CAClC,IAAM,EAAY,GAAc,CAAO,EACjC,EAAQ,EAAO,OAAS,EACxB,EAAmB,EAAO,MAAsB,YAChD,EAAiB,EAIvB,GAFA,GAAO,MAAM,+BAA+B,GAAW,EAEnD,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,8CAA8C,EAGhE,GAAI,EAAQ,GAAK,EAAQ,GACvB,MAAU,MAAM,sCAAsC,EAGxD,IAAM,EAAU,EAAW,CAAO,EAE5B,EAA+C,CACnD,MAAO,EACP,OAAQ,EAAO,OACf,EAAG,EACH,MACF,EAEA,GAAI,EAAe,QACjB,EAAY,QAAU,EAAe,QAEvC,GAAI,EAAe,MACjB,EAAY,MAAQ,EAAe,MAGrC,IAAM,EAAW,MAAM,MAAM,GAAG,uBAA8B,CAC5D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,mCAAmC,EAAS,UAAU,EAAS,gBAAgB,GACjF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,CAAC,EAAK,MAAQ,EAAK,KAAK,SAAW,EACrC,MAAU,MAAM,+BAA+B,EAGjD,OAAO,EAAK,KAAK,IAAI,CAAC,KAAU,CAC9B,IAAK,EAAK,IACV,cAAe,EAAK,cACtB,EAAE,EAGJ,SAAS,EAAsB,CAAC,EAAyB,CAEvD,OADmB,EAAQ,MAAM,2BAA2B,IACxC,IAAI,KAAK,GAAK,iBAGpC,SAAS,EAA4B,CAAC,EAAyB,CAC7D,OAAO,EAAQ,QAAQ,4BAA6B,EAAE,EAAE,KAAK,EAG/D,eAAsB,CAAsB,CAC1C,EACA,EACiC,CACjC,IAAM,EAAY,GAAyB,CAAO,EAC5C,EAAY,GAA6B,CAAO,EAEtD,GAAO,MAAM,2CAA2C,GAAW,EAEnE,IAAI,EACA,EAEJ,GAAI,OAAO,IAAW,SACpB,EAAW,EACX,EAAa,GAEb,OAAW,EAAO,SAClB,EAAa,EAAO,QAAU,GAGhC,GAAI,CAAC,GAAY,EAAS,KAAK,EAAE,SAAW,EAC1C,MAAU,MAAM,8CAA8C,EAGhE,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAc,CAClB,MAAO,EACP,SAAU,CACR,CACE,KAAM,OACN,QAAS,CACP,CAAE,KAAM,OAAQ,KAAM,CAAW,EACjC,CAAE,KAAM,YAAa,UAAW,CAAE,IAAK,CAAS,CAAE,CACpD,CACF,CACF,EACA,WAAY,CACd,EAEM,EAAW,MAAM,MAAM,GAAG,qBAA4B,CAC1D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,oCAAoC,EAAS,UAAU,EAAS,gBAAgB,GAClF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MACP,EACE,EACA,GAAU,kBACV,OAAO,IAAW,SAAW,EAAU,EAAO,QAAU,GACxD,CACE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAAK,MAAM,kBAC7B,YAAa,EAAK,MAAM,YAC1B,CACF,EAIF,IAAM,EADc,EAAK,UAAU,IACN,SAAS,QAEtC,GAAI,CAAC,EACH,MAAU,MAAM,6CAA6C,EAG/D,MAAO,CACL,MAAO,GAAuB,CAAO,EACrC,YAAa,GAA6B,CAAO,CACnD,EC/KF,iBAAS,gBAAQ,uBACjB,yBAAS,YCPT,uBAAS,wBAIT,IAAM,GAAgB,WAEf,SAAS,CAAkB,CAAC,EAAwC,CACzE,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAS,EAAU,CAAO,EAEhC,GAAI,CAAC,GAAU,EAAY,CAAO,EAChC,OAAO,GAAa,CAClB,OAAQ,GACR,SACF,CAAC,EAGH,GAAI,CAAC,EACH,MAAU,MACR,uFACF,EAGF,OAAO,GAAa,CAClB,SACA,SACF,CAAC,EC1BH,iBAAS,uBACT,yBAAS,YAST,IAAM,GAAwB,CAC5B,cAAe,uBACf,WAAY,YACd,EAEO,SAAS,EAAqB,EAAuB,CAC1D,MAAO,QAAS,OAAM,WAAsD,CAC1E,GAAI,EAAE,aAAiB,IACrB,OAAO,KAET,GAAI,CACF,IAAM,EAAc,EAAK,QAAQ,GAAsB,cAAe,EAAE,EAGxE,OAFA,KAAK,MAAM,CAAW,EACtB,GAAO,MAAM,wEAAwE,EAC9E,EACP,KAAM,CAEN,OADA,GAAO,KAAK,0CAA0C,EAC/C,OFXb,eAAe,EAAyB,CACtC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAIpC,GAFA,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAE1D,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,+CAA+C,EAGjE,GAAI,EAAO,OACT,GAAO,MACL,oGAEF,EAGF,IAAM,EAAQ,EAAO,KAAK,CAAS,GAC3B,SAAQ,SAAU,MAAM,GAAe,CAC7C,QACA,OAAQ,YACR,OAAQ,EAAO,OACf,wBAAyB,GAAsB,CACjD,CAAC,EAED,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,GAAI,OAAO,IAAW,UAAY,IAAW,KAC3C,MAAU,MAAM,8BAA8B,OAAO,oBAAyB,EAGhF,OAAO,EAGT,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EAGzF,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EG7CzF,iBAAS,sBA6ET,SAAS,EAAgB,CAAC,EAAsC,CAC9D,OAAQ,EAAK,UACN,qBACH,MAAO,CAAE,KAAM,oBAAqB,MACjC,cACH,MAAO,CACL,KAAM,cACN,iBAAkB,EAAK,cACzB,MACG,mBACH,MAAO,CACL,KAAM,mBACN,UAAW,EAAK,WAAa,CAAE,KAAM,MAAO,CAC9C,MACG,MACH,MAAO,CACL,KAAM,MACN,aAAc,EAAK,YACnB,WAAY,EAAK,UACjB,iBAAkB,EAAK,iBAAmB,OAC5C,UAEA,MAAU,MAAM,+BAAgC,EAAsB,MAAM,GAOlF,SAAS,EAAiB,CAAC,EAAyD,CAClF,OAAQ,EAAK,UACN,kBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,kBACN,OAAS,EAAK,QAAqC,YACnD,OAAQ,CACN,KAAO,EAAK,QAAQ,MAAoD,SACxE,MAAO,EAAK,QAAQ,MACpB,IAAK,EAAK,QAAQ,GACpB,CACF,MAEG,mBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,mBACN,OAAS,EAAK,QAAqC,YACnD,MAAO,EAAK,OAAS,GACrB,QAAS,EAAK,SAAS,IAAI,CAAC,KAAO,CACjC,OAAQ,EAAE,QACV,SAAU,EAAE,UACZ,MAAO,EAAE,KACX,EAAE,CACJ,MAEG,wBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,wBACN,OAAS,EAAK,QAAqC,YACnD,KAAM,EAAK,MAAQ,GACnB,OAAQ,EAAK,MACf,MAEG,gBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,gBACN,OAAS,EAAK,QAAqC,YACnD,YAAa,EAAK,cAAgB,GAClC,SAAU,EAAK,WAAa,GAC5B,UAAY,EAAK,WAAa,CAAC,EAC/B,OAAQ,EAAK,MACf,MAEG,UACH,MAAO,CACL,KAAM,UACN,QACE,EAAK,SAAS,IAAI,CAAC,KAAO,CACxB,KAAM,cACN,KAAM,EAAE,KACR,YACE,EAAE,aAAa,IAAI,CAAC,KAAO,CACzB,IAAK,EAAE,IACP,MAAO,EAAE,MACT,WAAY,EAAE,YACd,SAAU,EAAE,SACd,EAAE,GAAK,CAAC,CACZ,EAAE,GAAK,CAAC,CACZ,UAIA,OAAO,MAOb,SAAS,EAAyB,CAAC,EAGjC,CAEA,GAAI,EAAS,YAAa,CAExB,IAAM,EAAoC,CAAC,EAC3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QACzB,GAAI,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAOb,MAAO,CAAE,KAAM,EAAS,YAAa,aAAY,EAInD,IAAI,EAAO,GACL,EAAoC,CAAC,EAE3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QAEzB,GADA,GAAQ,EAAQ,KACZ,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAQb,MAAO,CAAE,OAAM,aAAY,EA8B7B,eAAsB,CAAc,CAClC,EACA,EACyB,CACzB,IAAM,EAAS,EAAU,CAAO,EAChC,GAAI,CAAC,EACH,MAAU,MACR,yGACF,EAGF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAY,EAAO,OAAS,GAAiB,CAAO,EACpD,EAAU,GAAmB,CAAO,EAE1C,EAAO,MAAM,+CAA+C,GAAW,EACvE,EAAO,MAAM,4BAA4B,EAAO,MAAM,UAAU,EAAG,GAAG,MAAM,EAG5E,IAAM,EAAkB,EAAO,OAAO,OACpC,CAAC,IAAM,EAAE,OAAS,sBAAwB,EAAE,OAAS,eAAiB,EAAE,OAAS,KACnF,EAEA,GAAI,CAAC,GAAmB,EAAgB,SAAW,EAEjD,EAAO,MAAM,2EAA2E,EACxF,EAAO,MAAQ,CAAC,CAAE,KAAM,oBAAqB,EAAG,GAAI,EAAO,OAAS,CAAC,CAAE,EAIzE,IAAM,EAAuC,CAC3C,MAAO,EACP,MAAO,EAAO,KAChB,EAEA,GAAI,EAAO,aACT,EAAY,aAAe,EAAO,aAGpC,GAAI,EAAO,aAAe,OACxB,EAAY,WAAa,EAAO,WAGlC,GAAI,EAAO,OAAS,EAAO,MAAM,OAAS,EACxC,EAAY,MAAQ,EAAO,MAAM,IAAI,EAAgB,EAGvD,GAAI,EAAO,eAAiB,OAC1B,EAAY,eAAiB,EAAO,aAGtC,GAAI,EAAO,iBACT,EAAY,UAAY,CAAE,QAAS,EAAO,gBAAiB,EAG7D,EAAO,MAAM,mCAAmC,KAAK,UAAU,EAAa,KAAM,CAAC,GAAG,EAGtF,IAAM,EAAW,MAAM,MAAM,GAAG,cAAqB,CACnD,OAAQ,OACR,QAAS,CACP,cAAe,UAAU,IACzB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,EAChC,OAAQ,YAAY,QAAQ,CAAO,CACrC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAEtC,MADA,EAAO,MAAM,qCAAqC,EAAS,UAAU,GAAW,EACtE,MAAM,iCAAiC,EAAS,UAAU,EAAS,YAAY,EAG3F,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MAEP,MADA,EAAO,MAAM,gCAAgC,EAAK,MAAM,SAAS,EACvD,MAAM,wBAAwB,EAAK,MAAM,SAAS,EAG9D,EAAO,MAAM,gDAAgD,EAAK,QAAU,aAAa,EAGzF,IAAQ,OAAM,eAAgB,GAA0B,CAAI,EAGtD,EAAoC,CAAC,EAC3C,GAAI,EAAK,OACP,QAAW,KAAQ,EAAK,OAAQ,CAC9B,IAAM,EAAY,GAAkB,CAAI,EACxC,GAAI,EACF,EAAY,KAAK,CAAS,EAKhC,IAAM,EAAyB,CAC7B,GAAI,EAAK,GACT,OACA,cACA,cACA,OAAQ,EAAK,MACf,EAMA,OAJA,EAAO,KACL,6CAA6C,EAAK,wBAAwB,EAAY,yBAAyB,EAAY,QAC7H,EAEO,ECnYT,iBAAS,gBAAQ,uBACjB,uBAAS,iBAAuC,YAsChD,SAAS,EAAgB,CAAC,EAA6C,CACrE,IAAM,EAQF,CAAC,CAAE,KAAM,OAAQ,KAAM,EAAO,MAAO,CAAC,EAE1C,QAAW,KAAc,EAAO,aAAe,CAAC,EAC9C,EAAQ,KAAK,CACX,KAAM,OACN,KAAM,EAAW,KACjB,UAAW,EAAW,aAClB,EAAW,SAAW,CAAE,SAAU,EAAW,QAAS,EAAI,CAAC,CACjE,CAAC,EAGH,OAAO,EAUT,SAAS,EAAY,CAAC,EAA+D,CACnF,GAAI,CAAC,EACH,OAIF,IAAM,EAAe,EAAM,aAAe,EACpC,EAAmB,EAAM,cAAgB,EACzC,EAAiB,EAEvB,MAAO,CACL,eACA,mBACA,YAAa,EAAe,EAC5B,mBAAoB,EAAe,iBACrC,EAGF,SAAS,EAAyB,CAAC,EAAsD,CACvF,IAAM,EAAoB,EAC1B,MAAO,CACL,eAAgB,EAAkB,iBAAiB,QAAQ,eAC3D,qBAAsB,EAAkB,iBAAiB,QAAQ,oBACnE,EAgBF,eAAe,EAAuB,CACpC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAwB,EACxB,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAEpC,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAC9D,IAAM,EAAqB,GAA0B,CAAM,EAErD,GADkB,EAAsB,aAAa,QAAU,GAAK,EACrC,GAAiB,CAAqB,EAAI,OAGzE,EAAe,EAAQ,UAAU,QAAU,OAO3C,EAAiB,CACrB,MAFY,EAAO,KAAK,CAAS,KAG7B,EACA,CAAE,SAAU,CAAC,CAAE,KAAM,OAAiB,QAAS,CAAY,CAAC,CAAE,EAC9D,CAAE,OAAQ,EAAO,MAAO,EAC5B,OAAQ,EACR,gBAAiB,EAAO,WAAa,KACrC,uBAAwB,CAAE,UAAW,GAAyB,CAAO,CAAE,KACnE,EAAmB,gBAAkB,EAAmB,qBACxD,CACE,gBAAiB,CACf,OAAQ,IACF,EAAmB,eACnB,CAAE,eAAgB,EAAmB,cAAe,EACpD,CAAC,KACD,EAAmB,qBACnB,CAAE,qBAAsB,EAAmB,oBAAqB,EAChE,CAAC,CACP,CACF,CACF,EACA,CAAC,CACP,EAGA,GAAI,EAAO,OAAQ,CACjB,IAAM,EAAS,GAAW,CAAc,EAExC,MAAO,CACL,WAAY,EAAO,WACnB,KAAM,QAAQ,QAAQ,EAAO,IAAI,EACjC,MAAO,QAAQ,QAAQ,EAAO,KAAK,EAAE,KAAK,EAAY,EACtD,aAAc,QAAQ,QAAQ,EAAO,YAAY,EAAE,KAAK,CAAC,KAAM,EAAuB,CACxF,EAIF,IAAQ,OAAM,SAAU,MAAM,GAAa,CAAc,EAEzD,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,OAAO,EAgBT,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,EAYrF,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,ECxNrF,oBAAS,uBCAT,oBAAS,uBACT,2BACE,kBACA,qBASF,SAAS,EAAwB,CAAC,EAA6B,CAE7D,IAAM,EADa,EAAU,YAAY,EACc,SAAS,IAAI,EAChE,aACA,cACJ,GAAI,CACF,OAAO,GAAiB,CAA0B,EAClD,KAAM,CACN,OAAO,GAAY,CAAoC,GAI3D,SAAS,EAAY,CAAC,EAAwB,EAAkC,CAC9E,GAAI,IAAc,GAAU,WAC1B,OAAO,EAAc,CAAO,EAE9B,OAAO,EAAc,CAAO,EAGvB,SAAS,EAAY,CAC1B,EACA,EACA,EACU,CACV,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAI,EAGrB,SAAS,EAAc,CAC5B,EACA,EACA,EACQ,CACR,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAM,ED7C9B,eAAsB,CAAqB,CACzC,EACA,EACmB,CACnB,GAAI,CAAC,EAAO,OACV,MAAU,MAAM,0CAA0C,EAE5D,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAa,EAAS,EAAW,EAAO,MAAM,EAGvD,eAAsB,CAAqB,CACzC,EACA,EACiB,CACjB,GAAI,CAAC,EAAO,QAAU,CAAC,MAAM,QAAQ,EAAO,MAAM,EAChD,MAAU,MAAM,8CAA8C,EAEhE,GAAI,EAAO,OAAO,SAAW,EAC3B,MAAO,GAET,QAAS,EAAI,EAAG,EAAI,EAAO,OAAO,OAAQ,IAAK,CAC7C,IAAM,EAAQ,EAAO,OAAO,GAC5B,GAAI,OAAO,IAAU,UAAY,CAAC,OAAO,SAAS,CAAK,EACrD,MAAU,MAAM,0BAA0B,oBAAoB,EAGlE,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAe,EAAS,EAAW,EAAO,MAAM,EbKzD,SAAS,EAAa,EAAmB,CACvC,GAAI,OAAO,QAAY,IACrB,MAAO,CAAC,EAEV,OAAO,QAAQ,IAGjB,IAAM,EAAM,GAAc,EAEb,GAAuB,CAClC,KAAM,SACN,YAAa,sEAEb,OAAQ,CACN,eAAgB,EAAI,gBAAkB,KACtC,gBAAiB,EAAI,iBAAmB,KACxC,mBAAoB,EAAI,oBAAsB,KAC9C,mBAAoB,EAAI,oBAAsB,KAC9C,YAAa,EAAI,aAAe,KAChC,YAAa,EAAI,aAAe,KAChC,uBAAwB,EAAI,wBAA0B,KACtD,yBAA0B,EAAI,0BAA4B,KAC1D,qBAAsB,EAAI,sBAAwB,KAClD,4BAA6B,EAAI,6BAA+B,KAChE,+BAAgC,EAAI,gCAAkC,KACtE,oCAAqC,EAAI,qCAAuC,KAChF,8BAA+B,EAAI,+BAAiC,KACpE,sBAAuB,EAAI,uBAAyB,KACpD,wBAAyB,EAAI,yBAA2B,IAC1D,OAEM,KAAI,CAAC,EAAgC,EAAuC,CAChF,GAAiB,EAA0C,CAAO,GAGpE,OAAQ,EACL,EAAU,gBAAiB,MAC1B,EACA,IACsB,CACtB,OAAO,EAAoB,EAAS,CAAM,IAG3C,EAAU,uBAAwB,MACjC,EACA,IACsB,CACtB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,uBAAwB,MACjC,EACA,IACoB,CACpB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,OAAQ,MACjB,EACA,IACqC,CACrC,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,mBAAoB,MAC7B,EACA,IACoD,CACpD,OAAO,EAAuB,EAAS,CAAM,IAG9C,EAAU,eAAgB,MACzB,EACA,IACoB,CACpB,OAAO,EAAoB,EAAS,CAAK,IAG1C,EAAU,gBAAiB,MAC1B,EACA,IACyB,CACzB,OAAO,EAAmB,EAAS,CAAK,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,UAAW,MACpB,EACA,IAC4B,CAC5B,OAAO,EAAe,EAAS,CAAM,EAEzC,EAEA,MAAO,CACL,CACE,KAAM,sBACN,MAAO,CACL,CACE,KAAM,+BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GACZ,MAAU,MACR,iCAAiC,EAAS,UAAU,EAAS,YAC/D,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAClC,EAAO,KAAK,gCAAgC,EAAK,MAAM,QAAU,qBAAqB,EAE1F,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,eACR,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAS,GAAK,EAAU,SAAW,EACpD,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,0CAA0C,EAAU,mBAAmB,EAEvF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,+BACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,2CACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,kCACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CACrE,OAHmB,yBAInB,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAM,GAAK,EAAO,SAAW,EAC9C,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAAc,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CAC1E,SACA,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,IAhBiB,yBAiBnB,MAAU,MACR,uEAAgE,IAClE,EAGF,EAAO,KAAK,iDAAiD,EAAO,gBAAgB,EAExF,EACA,CACE,KAAM,wBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAmB,CAAC,EAEpB,EAAS,MAAM,EAAQ,SAAS,EAAU,WAAY,CAC1D,OAAQ,0CACR,OAAQ,GACR,cAAe,CAAC,IAAkB,CAChC,EAAO,KAAK,CAAK,EAErB,CAAC,EAED,GAAI,OAAO,IAAW,UAAY,EAAO,SAAW,EAClD,MAAU,MAAM,0CAA0C,EAG5D,GAAI,EAAO,SAAW,EACpB,MAAU,MAAM,8BAA8B,EAGhD,EAAO,KAAK,iCAAiC,EAAO,wBAAwB,EAEhF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CAInD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,kBAF9C,uIAE6E,EAE/E,GACE,CAAC,GACD,OAAO,IAAW,UAClB,EAAE,UAAW,IACb,EAAE,gBAAiB,GAEnB,MAAU,MAAM,wDAAwD,EAG1E,EAAO,KAAK,mCAAmC,EAAO,QAAQ,EAElE,EACA,CACE,KAAM,4BACN,GAAI,MAAO,IAA0C,CAMnD,IAAM,EAAc,MADH,MAAM,MAFrB,wEAEmC,GACF,YAAY,EACzC,EAAc,OAAO,KAAK,IAAI,WAAW,CAAW,CAAC,EAErD,EAAgB,MAAM,EAAQ,SAAS,EAAU,cAAe,CAAW,EAEjF,GAAI,OAAO,IAAkB,SAC3B,MAAU,MAAM,sCAAsC,EAGxD,EAAO,KAAK,iCAAiC,EAAc,UAAU,EAAG,EAAE,OAAO,EAErF,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,uCACR,CAAC,EAED,GAAI,EAAE,aAAqB,cAAgB,EAAU,aAAe,EAClE,MAAU,MAAM,yCAAyC,EAG3D,EAAO,KAAK,+BAA+B,EAAU,2BAA2B,EAEpF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,aAAc,CAC5D,OACE,+FACJ,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,SAC/B,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KACL,mCAAmC,KAAK,UAAU,CAAM,EAAE,UAAU,EAAG,GAAG,GAC5E,EAEJ,EACA,CACE,KAAM,uBACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,SAAU,CACxD,MAAO,qCACP,MAAO,CAAC,CAAE,KAAM,oBAAqB,CAAC,EACtC,aAAc,CAChB,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,UAAY,EAAE,SAAU,GACvD,MAAU,MAAM,qDAAqD,EAGvE,GAAI,OAAO,EAAO,OAAS,UAAY,EAAO,KAAK,SAAW,EAC5D,MAAU,MAAM,mDAAmD,EAGrE,EAAO,KACL,kDAAkD,EAAO,KAAK,wBAAwB,EAAO,aAAa,QAAU,GACtH,EAEJ,CACF,CACF,CACF,CACF,EAEe,MenXf,IAAe",
|
|
23
|
+
"debugId": "52010ABAEE8E9A2864756E2164756E21",
|
|
24
24
|
"names": []
|
|
25
25
|
}
|
package/dist/cjs/index.node.cjs
CHANGED
|
@@ -540,7 +540,7 @@ async function handleTextEmbedding(runtime, params) {
|
|
|
540
540
|
}
|
|
541
541
|
const data = await response.json();
|
|
542
542
|
const firstResult = data?.data?.[0];
|
|
543
|
-
if (!firstResult
|
|
543
|
+
if (!firstResult?.embedding) {
|
|
544
544
|
throw new Error("OpenAI API returned invalid embedding response structure");
|
|
545
545
|
}
|
|
546
546
|
const embedding = firstResult.embedding;
|
|
@@ -965,6 +965,18 @@ async function handleResearch(runtime, params) {
|
|
|
965
965
|
// models/text.ts
|
|
966
966
|
var import_core11 = require("@elizaos/core");
|
|
967
967
|
var import_ai3 = require("ai");
|
|
968
|
+
function buildUserContent(params) {
|
|
969
|
+
const content = [{ type: "text", text: params.prompt }];
|
|
970
|
+
for (const attachment of params.attachments ?? []) {
|
|
971
|
+
content.push({
|
|
972
|
+
type: "file",
|
|
973
|
+
data: attachment.data,
|
|
974
|
+
mediaType: attachment.mediaType,
|
|
975
|
+
...attachment.filename ? { filename: attachment.filename } : {}
|
|
976
|
+
});
|
|
977
|
+
}
|
|
978
|
+
return content;
|
|
979
|
+
}
|
|
968
980
|
function convertUsage(usage) {
|
|
969
981
|
if (!usage) {
|
|
970
982
|
return;
|
|
@@ -987,15 +999,18 @@ function resolvePromptCacheOptions(params) {
|
|
|
987
999
|
};
|
|
988
1000
|
}
|
|
989
1001
|
async function generateTextByModelType(runtime, params, modelType, getModelFn) {
|
|
1002
|
+
const paramsWithAttachments = params;
|
|
990
1003
|
const openai = createOpenAIClient(runtime);
|
|
991
1004
|
const modelName = getModelFn(runtime);
|
|
992
1005
|
import_core11.logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);
|
|
993
1006
|
const promptCacheOptions = resolvePromptCacheOptions(params);
|
|
1007
|
+
const hasAttachments = (paramsWithAttachments.attachments?.length ?? 0) > 0;
|
|
1008
|
+
const userContent = hasAttachments ? buildUserContent(paramsWithAttachments) : undefined;
|
|
994
1009
|
const systemPrompt = runtime.character.system ?? undefined;
|
|
995
1010
|
const model = openai.chat(modelName);
|
|
996
1011
|
const generateParams = {
|
|
997
1012
|
model,
|
|
998
|
-
prompt: params.prompt,
|
|
1013
|
+
...userContent ? { messages: [{ role: "user", content: userContent }] } : { prompt: params.prompt },
|
|
999
1014
|
system: systemPrompt,
|
|
1000
1015
|
maxOutputTokens: params.maxTokens ?? 8192,
|
|
1001
1016
|
experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },
|
|
@@ -1324,4 +1339,4 @@ var typescript_default = openaiPlugin;
|
|
|
1324
1339
|
// index.node.ts
|
|
1325
1340
|
var index_node_default = typescript_default;
|
|
1326
1341
|
|
|
1327
|
-
//# debugId=
|
|
1342
|
+
//# debugId=E1C63C41A08FD43C64756E2164756E21
|