@elizaos/plugin-openai 2.0.0-alpha.5 → 2.0.0-alpha.7
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/browser/index.browser.js +2 -2
- package/dist/browser/index.browser.js.map +5 -5
- package/dist/cjs/index.node.cjs +30 -6
- package/dist/cjs/index.node.js.map +5 -5
- package/dist/models/embedding.d.ts.map +1 -1
- package/dist/models/text.d.ts.map +1 -1
- package/dist/node/index.node.js +30 -6
- package/dist/node/index.node.js.map +5 -5
- package/dist/types/index.d.ts +16 -0
- package/dist/types/index.d.ts.map +1 -1
- package/dist/utils/events.d.ts +6 -0
- package/dist/utils/events.d.ts.map +1 -1
- package/package.json +4 -4
|
@@ -1,3 +1,3 @@
|
|
|
1
|
-
import{logger as l,ModelType as w}from"@elizaos/core";import{logger as C}from"@elizaos/core";import{logger as L}from"@elizaos/core";function Co(o){if(typeof process>"u"||!process.env)return;let n=process.env[o];return n===void 0?void 0:String(n)}function I(o,n,c){let r=o.getSetting(n);if(r!==void 0&&r!==null)return String(r);return Co(n)??c}function K(o,n,c){let r=I(o,n);if(r===void 0)return c;let f=Number.parseInt(r,10);if(!Number.isFinite(f))throw Error(`Setting '${n}' must be a valid integer, got: ${r}`);return f}function v(o,n,c){let r=I(o,n);if(r===void 0)return c;let f=r.toLowerCase();return f==="true"||f==="1"||f==="yes"}function E(){return typeof globalThis<"u"&&typeof globalThis.document<"u"}function e(o){return E()&&!!I(o,"OPENAI_BROWSER_BASE_URL")}function y(o){return I(o,"OPENAI_API_KEY")}function xo(o){let n=I(o,"OPENAI_EMBEDDING_API_KEY");if(n)return L.debug("[OpenAI] Using specific embedding API key"),n;return L.debug("[OpenAI] Falling back to general API key for embeddings"),y(o)}function R(o,n=!1){if(E()&&!v(o,"OPENAI_ALLOW_BROWSER_API_KEY",!1))return{};let c=n?xo(o):y(o);return c?{Authorization:`Bearer ${c}`}:{}}function P(o){let n=I(o,"OPENAI_BROWSER_BASE_URL"),c=E()&&n?n:I(o,"OPENAI_BASE_URL")??"https://api.openai.com/v1";return L.debug(`[OpenAI] Base URL: ${c}`),c}function u(o){let n=E()?I(o,"OPENAI_BROWSER_EMBEDDING_URL")??I(o,"OPENAI_BROWSER_BASE_URL"):I(o,"OPENAI_EMBEDDING_URL");if(n)return L.debug(`[OpenAI] Using embedding base URL: ${n}`),n;return L.debug("[OpenAI] Falling back to general base URL for embeddings"),P(o)}function F(o){return I(o,"OPENAI_SMALL_MODEL")??I(o,"SMALL_MODEL")??"gpt-5-mini"}function U(o){return I(o,"OPENAI_LARGE_MODEL")??I(o,"LARGE_MODEL")??"gpt-5"}function m(o){return I(o,"OPENAI_EMBEDDING_MODEL")??"text-embedding-3-small"}function oo(o){return I(o,"OPENAI_IMAGE_DESCRIPTION_MODEL")??"gpt-5-mini"}function no(o){return I(o,"OPENAI_TRANSCRIPTION_MODEL")??"gpt-5-mini-transcribe"}function co(o){return I(o,"OPENAI_TTS_MODEL")??"tts-1"}function ro(o){return I(o,"OPENAI_TTS_VOICE")??"nova"}function fo(o){return I(o,"OPENAI_TTS_INSTRUCTIONS")??""}function io(o){return I(o,"OPENAI_IMAGE_MODEL")??"dall-e-3"}function To(o){return v(o,"OPENAI_EXPERIMENTAL_TELEMETRY",!1)}function po(o){return K(o,"OPENAI_EMBEDDING_DIMENSIONS",1536)}function Ao(o){return K(o,"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS",8192)}function ko(o){return I(o,"OPENAI_RESEARCH_MODEL")??"o3-deep-research"}function Oo(o){return K(o,"OPENAI_RESEARCH_TIMEOUT",3600000)}globalThis.AI_SDK_LOG_WARNINGS??=!1;function wo(o,n){Ko(n)}async function Ko(o){if(E()){C.debug("[OpenAI] Skipping API validation in browser environment");return}if(!y(o)){C.warn("[OpenAI] OPENAI_API_KEY is not configured. OpenAI functionality will fail until a valid API key is provided.");return}try{let c=P(o),r=await fetch(`${c}/models`,{headers:R(o)});if(!r.ok){C.warn(`[OpenAI] API key validation failed: ${r.status} ${r.statusText}. Please verify your OPENAI_API_KEY is correct.`);return}}catch(c){let r=c instanceof Error?c.message:String(c);C.warn(`[OpenAI] API validation error: ${r}. OpenAI functionality may be limited.`)}}import{logger as z}from"@elizaos/core";import{logger as Wo}from"@elizaos/core";var S={WAV:{HEADER:[82,73,70,70],IDENTIFIER:[87,65,86,69]},MP3_ID3:[73,68,51],OGG:[79,103,103,83],FLAC:[102,76,97,67],FTYP:[102,116,121,112],WEBM_EBML:[26,69,223,163]},Xo=12;function b(o,n,c){for(let r=0;r<c.length;r++){let f=c[r];if(f===void 0||o[n+r]!==f)return!1}return!0}function W(o){if(o.length<Xo)return"application/octet-stream";if(b(o,0,S.WAV.HEADER)&&b(o,8,S.WAV.IDENTIFIER))return"audio/wav";let n=o[0],c=o[1];if(b(o,0,S.MP3_ID3)||n===255&&c!==void 0&&(c&224)===224)return"audio/mpeg";if(b(o,0,S.OGG))return"audio/ogg";if(b(o,0,S.FLAC))return"audio/flac";if(b(o,4,S.FTYP))return"audio/mp4";if(b(o,0,S.WEBM_EBML))return"audio/webm";return Wo.warn("Could not detect audio format from buffer, using generic binary type"),"application/octet-stream"}function Ho(o){switch(o){case"audio/wav":return"wav";case"audio/mpeg":return"mp3";case"audio/ogg":return"ogg";case"audio/flac":return"flac";case"audio/mp4":return"m4a";case"audio/webm":return"webm";case"application/octet-stream":return"bin"}}function to(o){return`recording.${Ho(o)}`}function Po(o){return o instanceof Blob||o instanceof File}function X(o){return Buffer.isBuffer(o)}function Vo(o){return typeof o==="object"&&o!==null&&"audio"in o&&(Po(o.audio)||X(o.audio))}function Qo(o){return typeof o==="object"&&o!==null&&"audioUrl"in o&&typeof o.audioUrl==="string"}async function Io(o){let n=await fetch(o);if(!n.ok)throw Error(`Failed to fetch audio from URL: ${n.status}`);return n.blob()}async function d(o,n){let c=no(o),r,f={};if(typeof n==="string")z.debug(`[OpenAI] Fetching audio from URL: ${n}`),r=await Io(n);else if(Po(n))r=n;else if(X(n)){let O=W(n);z.debug(`[OpenAI] Auto-detected audio MIME type: ${O}`),r=new Blob([new Uint8Array(n)],{type:O})}else if(Vo(n)){if(f=n,n.model)c=n.model;if(X(n.audio)){let O=n.mimeType??W(n.audio);z.debug(`[OpenAI] Using MIME type: ${O}`),r=new Blob([new Uint8Array(n.audio)],{type:O})}else r=n.audio}else if(Qo(n))z.debug(`[OpenAI] Fetching audio from URL: ${n.audioUrl}`),r=await Io(n.audioUrl),f={prompt:n.prompt};else throw Error("TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object");z.debug(`[OpenAI] Using TRANSCRIPTION model: ${c}`);let i=r.type||"audio/webm",k=r.name||to(i.startsWith("audio/")?i:"audio/webm"),p=new FormData;if(p.append("file",r,k),p.append("model",c),f.language)p.append("language",f.language);if(f.responseFormat)p.append("response_format",f.responseFormat);if(f.prompt)p.append("prompt",f.prompt);if(f.temperature!==void 0)p.append("temperature",String(f.temperature));if(f.timestampGranularities)for(let O of f.timestampGranularities)p.append("timestamp_granularities[]",O);let A=P(o),T=await fetch(`${A}/audio/transcriptions`,{method:"POST",headers:R(o),body:p});if(!T.ok){let O=await T.text().catch(()=>"Unknown error");throw Error(`OpenAI transcription failed: ${T.status} ${T.statusText} - ${O}`)}return(await T.json()).text}async function H(o,n){let c,r,f="mp3",i,k;if(typeof n==="string")c=n,r=void 0;else{if(c=n.text,r=n.voice,"format"in n&&n.format)f=n.format;if("model"in n&&n.model)i=n.model;if("instructions"in n&&n.instructions)k=n.instructions}if(i=i??co(o),r=r??ro(o),k=k??fo(o),z.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${i}`),!c||c.trim().length===0)throw Error("TEXT_TO_SPEECH requires non-empty text");if(c.length>4096)throw Error("TEXT_TO_SPEECH text exceeds 4096 character limit");let p=["alloy","echo","fable","onyx","nova","shimmer"];if(r&&!p.includes(r))throw Error(`Invalid voice: ${r}. Must be one of: ${p.join(", ")}`);let A=P(o),T={model:i,voice:r,input:c,response_format:f};if(k&&k.length>0)T.instructions=k;let t=await fetch(`${A}/audio/speech`,{method:"POST",headers:{...R(o),"Content-Type":"application/json",...f==="mp3"?{Accept:"audio/mpeg"}:{}},body:JSON.stringify(T)});if(!t.ok){let O=await t.text().catch(()=>"Unknown error");throw Error(`OpenAI TTS failed: ${t.status} ${t.statusText} - ${O}`)}return t.arrayBuffer()}import{logger as x,ModelType as Do,VECTOR_DIMS as qo}from"@elizaos/core";import{EventType as Zo}from"@elizaos/core";var so=200;function Yo(o){if(o.length<=so)return o;return`${o.slice(0,so)}…`}function Go(o){if("promptTokens"in o)return{promptTokens:o.promptTokens??0,completionTokens:o.completionTokens??0,totalTokens:o.totalTokens??(o.promptTokens??0)+(o.completionTokens??0)};if("inputTokens"in o||"outputTokens"in o){let n=o.inputTokens??0,c=o.outputTokens??0;return{promptTokens:n,completionTokens:c,totalTokens:n+c}}return{promptTokens:0,completionTokens:0,totalTokens:0}}function J(o,n,c,r){let f=Go(r),i={runtime:o,source:"openai",provider:"openai",type:n,prompt:Yo(c),tokens:{prompt:f.promptTokens,completion:f.completionTokens,total:f.totalTokens}};o.emitEvent(Zo.MODEL_USED,i)}var ho=8000;function Mo(o){let n=Object.values(qo);if(!n.includes(o))throw Error(`Invalid embedding dimension: ${o}. Must be one of: ${n.join(", ")}`);return o}function Bo(o){if(o===null)return null;if(typeof o==="string")return o;if(typeof o==="object"&&typeof o.text==="string")return o.text;throw Error("Invalid embedding params: expected string, { text: string }, or null")}async function V(o,n){let c=m(o),r=Mo(po(o)),f=Bo(n);if(f===null){x.debug("[OpenAI] Creating test embedding for initialization");let j=Array(r).fill(0);return j[0]=0.1,j}let i=f.trim();if(i.length===0)throw Error("Cannot generate embedding for empty text");let k=ho*4;if(i.length>k)x.warn(`[OpenAI] Embedding input too long (~${Math.ceil(i.length/4)} tokens), truncating to ~${ho} tokens`),i=i.slice(0,k);let A=`${u(o)}/embeddings`;x.debug(`[OpenAI] Generating embedding with model: ${c}`);let T=await fetch(A,{method:"POST",headers:{...R(o,!0),"Content-Type":"application/json"},body:JSON.stringify({model:c,input:i})});if(!T.ok){let j=await T.text().catch(()=>"Unknown error");throw Error(`OpenAI embedding API error: ${T.status} ${T.statusText} - ${j}`)}let t=await T.json(),O=t?.data?.[0];if(!O||!O.embedding)throw Error("OpenAI API returned invalid embedding response structure");let h=O.embedding;if(h.length!==r)throw Error(`Embedding dimension mismatch: got ${h.length}, expected ${r}. Check OPENAI_EMBEDDING_DIMENSIONS setting.`);if(t.usage)J(o,Do.TEXT_EMBEDDING,i,{promptTokens:t.usage.prompt_tokens,completionTokens:0,totalTokens:t.usage.total_tokens});return x.debug(`[OpenAI] Generated embedding with ${h.length} dimensions`),h}import{logger as lo,ModelType as ao}from"@elizaos/core";var Ro="Please analyze this image and provide a title and detailed description.";async function Q(o,n){let c=io(o),r=n.count??1,f=n.size??"1024x1024",i=n;if(lo.debug(`[OpenAI] Using IMAGE model: ${c}`),!n.prompt||n.prompt.trim().length===0)throw Error("IMAGE generation requires a non-empty prompt");if(r<1||r>10)throw Error("IMAGE count must be between 1 and 10");let k=P(o),p={model:c,prompt:n.prompt,n:r,size:f};if(i.quality)p.quality=i.quality;if(i.style)p.style=i.style;let A=await fetch(`${k}/images/generations`,{method:"POST",headers:{...R(o),"Content-Type":"application/json"},body:JSON.stringify(p)});if(!A.ok){let t=await A.text().catch(()=>"Unknown error");throw Error(`OpenAI image generation failed: ${A.status} ${A.statusText} - ${t}`)}let T=await A.json();if(!T.data||T.data.length===0)throw Error("OpenAI API returned no images");return T.data.map((t)=>({url:t.url,revisedPrompt:t.revised_prompt}))}function go(o){return o.match(/title[:\s]+(.+?)(?:\n|$)/i)?.[1]?.trim()??"Image Analysis"}function vo(o){return o.replace(/title[:\s]+(.+?)(?:\n|$)/i,"").trim()}async function Z(o,n){let c=oo(o),r=Ao(o);lo.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${c}`);let f,i;if(typeof n==="string")f=n,i=Ro;else f=n.imageUrl,i=n.prompt??Ro;if(!f||f.trim().length===0)throw Error("IMAGE_DESCRIPTION requires a valid image URL");let k=P(o),p={model:c,messages:[{role:"user",content:[{type:"text",text:i},{type:"image_url",image_url:{url:f}}]}],max_tokens:r},A=await fetch(`${k}/chat/completions`,{method:"POST",headers:{...R(o),"Content-Type":"application/json"},body:JSON.stringify(p)});if(!A.ok){let h=await A.text().catch(()=>"Unknown error");throw Error(`OpenAI image description failed: ${A.status} ${A.statusText} - ${h}`)}let T=await A.json();if(T.usage)J(o,ao.IMAGE_DESCRIPTION,typeof n==="string"?n:n.prompt??"",{promptTokens:T.usage.prompt_tokens,completionTokens:T.usage.completion_tokens,totalTokens:T.usage.total_tokens});let O=T.choices?.[0]?.message?.content;if(!O)throw Error("OpenAI API returned empty image description");return{title:go(O),description:vo(O)}}import{logger as No,ModelType as So}from"@elizaos/core";import{generateObject as on}from"ai";import{createOpenAI as yo}from"@ai-sdk/openai";var eo="sk-proxy";function _(o){let n=P(o),c=y(o);if(!c&&e(o))return yo({apiKey:eo,baseURL:n});if(!c)throw Error("OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.");return yo({apiKey:c,baseURL:n})}import{logger as Jo}from"@elizaos/core";import{JSONParseError as uo}from"ai";var mo={MARKDOWN_JSON:/```json\n|\n```|```/g,WHITESPACE:/^\s+|\s+$/g};function $o(){return async({text:o,error:n})=>{if(!(n instanceof uo))return null;try{let c=o.replace(mo.MARKDOWN_JSON,"");return JSON.parse(c),Jo.debug("[JSON Repair] Successfully repaired JSON by removing markdown wrappers"),c}catch{return Jo.warn("[JSON Repair] Unable to repair JSON text"),null}}}async function bo(o,n,c,r){let f=_(o),i=r(o);if(No.debug(`[OpenAI] Using ${c} model: ${i}`),!n.prompt||n.prompt.trim().length===0)throw Error("Object generation requires a non-empty prompt");if(n.schema)No.debug("[OpenAI] Schema provided but using no-schema mode. Structure is determined by prompt instructions.");let k=f.chat(i),{object:p,usage:A}=await on({model:k,output:"no-schema",prompt:n.prompt,experimental_repairText:$o()});if(A)J(o,c,n.prompt,A);if(typeof p!=="object"||p===null)throw Error(`Object generation returned ${typeof p}, expected object`);return p}async function Y(o,n){return bo(o,n,So.OBJECT_SMALL,F)}async function G(o,n){return bo(o,n,So.OBJECT_LARGE,U)}import{logger as $}from"@elizaos/core";function nn(o){switch(o.type){case"web_search_preview":return{type:"web_search_preview"};case"file_search":return{type:"file_search",vector_store_ids:o.vectorStoreIds};case"code_interpreter":return{type:"code_interpreter",container:o.container??{type:"auto"}};case"mcp":return{type:"mcp",server_label:o.serverLabel,server_url:o.serverUrl,require_approval:o.requireApproval??"never"};default:throw Error(`Unknown research tool type: ${o.type}`)}}function cn(o){switch(o.type){case"web_search_call":return{id:o.id??"",type:"web_search_call",status:o.status??"completed",action:{type:o.action?.type??"search",query:o.action?.query,url:o.action?.url}};case"file_search_call":return{id:o.id??"",type:"file_search_call",status:o.status??"completed",query:o.query??"",results:o.results?.map((n)=>({fileId:n.file_id,fileName:n.file_name,score:n.score}))};case"code_interpreter_call":return{id:o.id??"",type:"code_interpreter_call",status:o.status??"completed",code:o.code??"",output:o.output};case"mcp_tool_call":return{id:o.id??"",type:"mcp_tool_call",status:o.status??"completed",serverLabel:o.server_label??"",toolName:o.tool_name??"",arguments:o.arguments??{},result:o.result};case"message":return{type:"message",content:o.content?.map((n)=>({type:"output_text",text:n.text,annotations:n.annotations?.map((c)=>({url:c.url,title:c.title,startIndex:c.start_index,endIndex:c.end_index}))??[]}))??[]};default:return null}}function rn(o){if(o.output_text){let r=[];if(o.output){for(let f of o.output)if(f.type==="message"&&f.content){for(let i of f.content)if(i.annotations)for(let k of i.annotations)r.push({url:k.url,title:k.title,startIndex:k.start_index,endIndex:k.end_index})}}return{text:o.output_text,annotations:r}}let n="",c=[];if(o.output){for(let r of o.output)if(r.type==="message"&&r.content){for(let f of r.content)if(n+=f.text,f.annotations)for(let i of f.annotations)c.push({url:i.url,title:i.title,startIndex:i.start_index,endIndex:i.end_index})}}return{text:n,annotations:c}}async function D(o,n){let c=y(o);if(!c)throw Error("OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.");let r=P(o),f=n.model??ko(o),i=Oo(o);$.debug(`[OpenAI] Starting deep research with model: ${f}`),$.debug(`[OpenAI] Research input: ${n.input.substring(0,100)}...`);let k=n.tools?.filter((N)=>N.type==="web_search_preview"||N.type==="file_search"||N.type==="mcp");if(!k||k.length===0)$.debug("[OpenAI] No data source tools specified, defaulting to web_search_preview"),n.tools=[{type:"web_search_preview"},...n.tools??[]];let p={model:f,input:n.input};if(n.instructions)p.instructions=n.instructions;if(n.background!==void 0)p.background=n.background;if(n.tools&&n.tools.length>0)p.tools=n.tools.map(nn);if(n.maxToolCalls!==void 0)p.max_tool_calls=n.maxToolCalls;if(n.reasoningSummary)p.reasoning={summary:n.reasoningSummary};$.debug(`[OpenAI] Research request body: ${JSON.stringify(p,null,2)}`);let A=await fetch(`${r}/responses`,{method:"POST",headers:{Authorization:`Bearer ${c}`,"Content-Type":"application/json"},body:JSON.stringify(p),signal:AbortSignal.timeout(i)});if(!A.ok){let N=await A.text();throw $.error(`[OpenAI] Research request failed: ${A.status} ${N}`),Error(`Deep research request failed: ${A.status} ${A.statusText}`)}let T=await A.json();if(T.error)throw $.error(`[OpenAI] Research API error: ${T.error.message}`),Error(`Deep research error: ${T.error.message}`);$.debug(`[OpenAI] Research response received. Status: ${T.status??"completed"}`);let{text:t,annotations:O}=rn(T),h=[];if(T.output)for(let N of T.output){let g=cn(N);if(g)h.push(g)}let j={id:T.id,text:t,annotations:O,outputItems:h,status:T.status};return $.info(`[OpenAI] Research completed. Text length: ${t.length}, Annotations: ${O.length}, Output items: ${h.length}`),j}import{logger as fn,ModelType as jo}from"@elizaos/core";import{generateText as Tn,streamText as pn}from"ai";function An(o){if(!o)return;let n=o.inputTokens??0,c=o.outputTokens??0;return{promptTokens:n,completionTokens:c,totalTokens:n+c}}async function Eo(o,n,c,r){let f=_(o),i=r(o);fn.debug(`[OpenAI] Using ${c} model: ${i}`);let k=o.character.system??void 0,A={model:f.chat(i),prompt:n.prompt,system:k,maxOutputTokens:n.maxTokens??8192,experimental_telemetry:{isEnabled:To(o)}};if(n.stream){let O=pn(A);return{textStream:O.textStream,text:Promise.resolve(O.text),usage:Promise.resolve(O.usage).then(An),finishReason:Promise.resolve(O.finishReason).then((h)=>h)}}let{text:T,usage:t}=await Tn(A);if(t)J(o,c,n.prompt,t);return T}async function q(o,n){return Eo(o,n,jo.TEXT_SMALL,F)}async function M(o,n){return Eo(o,n,jo.TEXT_LARGE,U)}import{ModelType as _o}from"@elizaos/core";import{ModelType as kn}from"@elizaos/core";import{encodingForModel as On,getEncoding as wn}from"js-tiktoken";function Fo(o){let c=o.toLowerCase().includes("4o")?"o200k_base":"cl100k_base";try{return On(o)}catch{return wn(c)}}function Uo(o,n){if(n===kn.TEXT_SMALL)return F(o);return U(o)}function zo(o,n,c){let r=Uo(o,n);return Fo(r).encode(c)}function Lo(o,n,c){let r=Uo(o,n);return Fo(r).decode(c)}async function B(o,n){if(!n.prompt)throw Error("Tokenization requires a non-empty prompt");let c=n.modelType??_o.TEXT_LARGE;return zo(o,c,n.prompt)}async function a(o,n){if(!n.tokens||!Array.isArray(n.tokens))throw Error("Detokenization requires a valid tokens array");if(n.tokens.length===0)return"";for(let r=0;r<n.tokens.length;r++){let f=n.tokens[r];if(typeof f!=="number"||!Number.isFinite(f))throw Error(`Invalid token at index ${r}: expected number`)}let c=n.modelType??_o.TEXT_LARGE;return Lo(o,c,n.tokens)}function tn(){if(typeof process>"u")return{};return process.env}var s=tn(),In={name:"openai",description:"OpenAI API integration for text, image, audio, and embedding models",config:{OPENAI_API_KEY:s.OPENAI_API_KEY??null,OPENAI_BASE_URL:s.OPENAI_BASE_URL??null,OPENAI_SMALL_MODEL:s.OPENAI_SMALL_MODEL??null,OPENAI_LARGE_MODEL:s.OPENAI_LARGE_MODEL??null,SMALL_MODEL:s.SMALL_MODEL??null,LARGE_MODEL:s.LARGE_MODEL??null,OPENAI_EMBEDDING_MODEL:s.OPENAI_EMBEDDING_MODEL??null,OPENAI_EMBEDDING_API_KEY:s.OPENAI_EMBEDDING_API_KEY??null,OPENAI_EMBEDDING_URL:s.OPENAI_EMBEDDING_URL??null,OPENAI_EMBEDDING_DIMENSIONS:s.OPENAI_EMBEDDING_DIMENSIONS??null,OPENAI_IMAGE_DESCRIPTION_MODEL:s.OPENAI_IMAGE_DESCRIPTION_MODEL??null,OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS:s.OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS??null,OPENAI_EXPERIMENTAL_TELEMETRY:s.OPENAI_EXPERIMENTAL_TELEMETRY??null,OPENAI_RESEARCH_MODEL:s.OPENAI_RESEARCH_MODEL??null,OPENAI_RESEARCH_TIMEOUT:s.OPENAI_RESEARCH_TIMEOUT??null},async init(o,n){wo(o,n)},models:{[w.TEXT_EMBEDDING]:async(o,n)=>{return V(o,n)},[w.TEXT_TOKENIZER_ENCODE]:async(o,n)=>{return B(o,n)},[w.TEXT_TOKENIZER_DECODE]:async(o,n)=>{return a(o,n)},[w.TEXT_SMALL]:async(o,n)=>{return q(o,n)},[w.TEXT_LARGE]:async(o,n)=>{return M(o,n)},[w.IMAGE]:async(o,n)=>{return Q(o,n)},[w.IMAGE_DESCRIPTION]:async(o,n)=>{return Z(o,n)},[w.TRANSCRIPTION]:async(o,n)=>{return d(o,n)},[w.TEXT_TO_SPEECH]:async(o,n)=>{return H(o,n)},[w.OBJECT_SMALL]:async(o,n)=>{return Y(o,n)},[w.OBJECT_LARGE]:async(o,n)=>{return G(o,n)},[w.RESEARCH]:async(o,n)=>{return D(o,n)}},tests:[{name:"openai_plugin_tests",tests:[{name:"openai_test_api_connectivity",fn:async(o)=>{let n=P(o),c=await fetch(`${n}/models`,{headers:R(o)});if(!c.ok)throw Error(`API connectivity test failed: ${c.status} ${c.statusText}`);let r=await c.json();l.info(`[OpenAI Test] API connected. ${r.data?.length??0} models available.`)}},{name:"openai_test_text_embedding",fn:async(o)=>{let n=await o.useModel(w.TEXT_EMBEDDING,{text:"Hello, world!"});if(!Array.isArray(n)||n.length===0)throw Error("Embedding should return a non-empty array");l.info(`[OpenAI Test] Generated embedding with ${n.length} dimensions`)}},{name:"openai_test_text_small",fn:async(o)=>{let n=await o.useModel(w.TEXT_SMALL,{prompt:"Say hello in exactly 5 words."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_SMALL should return non-empty string");l.info(`[OpenAI Test] TEXT_SMALL generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_text_large",fn:async(o)=>{let n=await o.useModel(w.TEXT_LARGE,{prompt:"Explain quantum computing in 2 sentences."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_LARGE should return non-empty string");l.info(`[OpenAI Test] TEXT_LARGE generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_tokenizer_roundtrip",fn:async(o)=>{let c=await o.useModel(w.TEXT_TOKENIZER_ENCODE,{prompt:"Hello, tokenizer test!",modelType:w.TEXT_SMALL});if(!Array.isArray(c)||c.length===0)throw Error("Tokenization should return non-empty token array");let r=await o.useModel(w.TEXT_TOKENIZER_DECODE,{tokens:c,modelType:w.TEXT_SMALL});if(r!=="Hello, tokenizer test!")throw Error(`Tokenizer roundtrip failed: expected "Hello, tokenizer test!", got "${r}"`);l.info(`[OpenAI Test] Tokenizer roundtrip successful (${c.length} tokens)`)}},{name:"openai_test_streaming",fn:async(o)=>{let n=[],c=await o.useModel(w.TEXT_LARGE,{prompt:"Count from 1 to 5, one number per line.",stream:!0,onStreamChunk:(r)=>{n.push(r)}});if(typeof c!=="string"||c.length===0)throw Error("Streaming should return non-empty result");if(n.length===0)throw Error("No streaming chunks received");l.info(`[OpenAI Test] Streaming test: ${n.length} chunks received`)}},{name:"openai_test_image_description",fn:async(o)=>{let c=await o.useModel(w.IMAGE_DESCRIPTION,"https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Camponotus_flavomarginatus_ant.jpg/440px-Camponotus_flavomarginatus_ant.jpg");if(!c||typeof c!=="object"||!("title"in c)||!("description"in c))throw Error("Image description should return { title, description }");l.info(`[OpenAI Test] Image described: "${c.title}"`)}},{name:"openai_test_transcription",fn:async(o)=>{let r=await(await fetch("https://upload.wikimedia.org/wikipedia/commons/2/25/En-Open_Source.ogg")).arrayBuffer(),f=Buffer.from(new Uint8Array(r)),i=await o.useModel(w.TRANSCRIPTION,f);if(typeof i!=="string")throw Error("Transcription should return a string");l.info(`[OpenAI Test] Transcription: "${i.substring(0,50)}..."`)}},{name:"openai_test_text_to_speech",fn:async(o)=>{let n=await o.useModel(w.TEXT_TO_SPEECH,{text:"Hello, this is a text-to-speech test."});if(!(n instanceof ArrayBuffer)||n.byteLength===0)throw Error("TTS should return non-empty ArrayBuffer");l.info(`[OpenAI Test] TTS generated ${n.byteLength} bytes of audio`)}},{name:"openai_test_object_generation",fn:async(o)=>{let n=await o.useModel(w.OBJECT_SMALL,{prompt:"Return a JSON object with exactly these fields: name (string), age (number), active (boolean)"});if(!n||typeof n!=="object")throw Error("Object generation should return an object");l.info(`[OpenAI Test] Object generated: ${JSON.stringify(n).substring(0,100)}`)}},{name:"openai_test_research",fn:async(o)=>{let n=await o.useModel(w.RESEARCH,{input:"What is the current date and time?",tools:[{type:"web_search_preview"}],maxToolCalls:3});if(!n||typeof n!=="object"||!("text"in n))throw Error("Research should return an object with text property");if(typeof n.text!=="string"||n.text.length===0)throw Error("Research result text should be a non-empty string");l.info(`[OpenAI Test] Research completed. Text length: ${n.text.length}, Annotations: ${n.annotations?.length??0}`)}}]}]},Pn=In;export{In as openaiPlugin,Pn as default};
|
|
1
|
+
import{logger as l,ModelType as I}from"@elizaos/core";import{logger as _}from"@elizaos/core";import{logger as z}from"@elizaos/core";function xo(o){if(typeof process>"u"||!process.env)return;let n=process.env[o];return n===void 0?void 0:String(n)}function s(o,n,c){let r=o.getSetting(n);if(r!==void 0&&r!==null)return String(r);return xo(n)??c}function K(o,n,c){let r=s(o,n);if(r===void 0)return c;let f=Number.parseInt(r,10);if(!Number.isFinite(f))throw Error(`Setting '${n}' must be a valid integer, got: ${r}`);return f}function v(o,n,c){let r=s(o,n);if(r===void 0)return c;let f=r.toLowerCase();return f==="true"||f==="1"||f==="yes"}function E(){return typeof globalThis<"u"&&typeof globalThis.document<"u"}function g(o){return E()&&!!s(o,"OPENAI_BROWSER_BASE_URL")}function J(o){return s(o,"OPENAI_API_KEY")}function Ko(o){let n=s(o,"OPENAI_EMBEDDING_API_KEY");if(n)return z.debug("[OpenAI] Using specific embedding API key"),n;return z.debug("[OpenAI] Falling back to general API key for embeddings"),J(o)}function R(o,n=!1){if(E()&&!v(o,"OPENAI_ALLOW_BROWSER_API_KEY",!1))return{};let c=n?Ko(o):J(o);return c?{Authorization:`Bearer ${c}`}:{}}function O(o){let n=s(o,"OPENAI_BROWSER_BASE_URL"),c=E()&&n?n:s(o,"OPENAI_BASE_URL")??"https://api.openai.com/v1";return z.debug(`[OpenAI] Base URL: ${c}`),c}function u(o){let n=E()?s(o,"OPENAI_BROWSER_EMBEDDING_URL")??s(o,"OPENAI_BROWSER_BASE_URL"):s(o,"OPENAI_EMBEDDING_URL");if(n)return z.debug(`[OpenAI] Using embedding base URL: ${n}`),n;return z.debug("[OpenAI] Falling back to general base URL for embeddings"),O(o)}function F(o){return s(o,"OPENAI_SMALL_MODEL")??s(o,"SMALL_MODEL")??"gpt-5-mini"}function d(o){return s(o,"OPENAI_LARGE_MODEL")??s(o,"LARGE_MODEL")??"gpt-5"}function m(o){return s(o,"OPENAI_EMBEDDING_MODEL")??"text-embedding-3-small"}function oo(o){return s(o,"OPENAI_IMAGE_DESCRIPTION_MODEL")??"gpt-5-mini"}function no(o){return s(o,"OPENAI_TRANSCRIPTION_MODEL")??"gpt-5-mini-transcribe"}function co(o){return s(o,"OPENAI_TTS_MODEL")??"tts-1"}function ro(o){return s(o,"OPENAI_TTS_VOICE")??"nova"}function fo(o){return s(o,"OPENAI_TTS_INSTRUCTIONS")??""}function To(o){return s(o,"OPENAI_IMAGE_MODEL")??"dall-e-3"}function io(o){return v(o,"OPENAI_EXPERIMENTAL_TELEMETRY",!1)}function ko(o){return K(o,"OPENAI_EMBEDDING_DIMENSIONS",1536)}function Ao(o){return K(o,"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS",8192)}function to(o){return s(o,"OPENAI_RESEARCH_MODEL")??"o3-deep-research"}function Po(o){return K(o,"OPENAI_RESEARCH_TIMEOUT",3600000)}globalThis.AI_SDK_LOG_WARNINGS??=!1;function Io(o,n){Wo(n)}async function Wo(o){if(E()){_.debug("[OpenAI] Skipping API validation in browser environment");return}if(!J(o)){_.warn("[OpenAI] OPENAI_API_KEY is not configured. OpenAI functionality will fail until a valid API key is provided.");return}try{let c=O(o),r=await fetch(`${c}/models`,{headers:R(o)});if(!r.ok){_.warn(`[OpenAI] API key validation failed: ${r.status} ${r.statusText}. Please verify your OPENAI_API_KEY is correct.`);return}}catch(c){let r=c instanceof Error?c.message:String(c);_.warn(`[OpenAI] API validation error: ${r}. OpenAI functionality may be limited.`)}}import{logger as U}from"@elizaos/core";import{logger as Xo}from"@elizaos/core";var S={WAV:{HEADER:[82,73,70,70],IDENTIFIER:[87,65,86,69]},MP3_ID3:[73,68,51],OGG:[79,103,103,83],FLAC:[102,76,97,67],FTYP:[102,116,121,112],WEBM_EBML:[26,69,223,163]},Co=12;function j(o,n,c){for(let r=0;r<c.length;r++){let f=c[r];if(f===void 0||o[n+r]!==f)return!1}return!0}function W(o){if(o.length<Co)return"application/octet-stream";if(j(o,0,S.WAV.HEADER)&&j(o,8,S.WAV.IDENTIFIER))return"audio/wav";let n=o[0],c=o[1];if(j(o,0,S.MP3_ID3)||n===255&&c!==void 0&&(c&224)===224)return"audio/mpeg";if(j(o,0,S.OGG))return"audio/ogg";if(j(o,0,S.FLAC))return"audio/flac";if(j(o,4,S.FTYP))return"audio/mp4";if(j(o,0,S.WEBM_EBML))return"audio/webm";return Xo.warn("Could not detect audio format from buffer, using generic binary type"),"application/octet-stream"}function Ho(o){switch(o){case"audio/wav":return"wav";case"audio/mpeg":return"mp3";case"audio/ogg":return"ogg";case"audio/flac":return"flac";case"audio/mp4":return"m4a";case"audio/webm":return"webm";case"application/octet-stream":return"bin"}}function po(o){return`recording.${Ho(o)}`}function wo(o){return o instanceof Blob||o instanceof File}function X(o){return Buffer.isBuffer(o)}function Vo(o){return typeof o==="object"&&o!==null&&"audio"in o&&(wo(o.audio)||X(o.audio))}function Qo(o){return typeof o==="object"&&o!==null&&"audioUrl"in o&&typeof o.audioUrl==="string"}async function so(o){let n=await fetch(o);if(!n.ok)throw Error(`Failed to fetch audio from URL: ${n.status}`);return n.blob()}async function C(o,n){let c=no(o),r,f={};if(typeof n==="string")U.debug(`[OpenAI] Fetching audio from URL: ${n}`),r=await so(n);else if(wo(n))r=n;else if(X(n)){let P=W(n);U.debug(`[OpenAI] Auto-detected audio MIME type: ${P}`),r=new Blob([new Uint8Array(n)],{type:P})}else if(Vo(n)){if(f=n,n.model)c=n.model;if(X(n.audio)){let P=n.mimeType??W(n.audio);U.debug(`[OpenAI] Using MIME type: ${P}`),r=new Blob([new Uint8Array(n.audio)],{type:P})}else r=n.audio}else if(Qo(n))U.debug(`[OpenAI] Fetching audio from URL: ${n.audioUrl}`),r=await so(n.audioUrl),f={prompt:n.prompt};else throw Error("TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object");U.debug(`[OpenAI] Using TRANSCRIPTION model: ${c}`);let T=r.type||"audio/webm",A=r.name||po(T.startsWith("audio/")?T:"audio/webm"),k=new FormData;if(k.append("file",r,A),k.append("model",c),f.language)k.append("language",f.language);if(f.responseFormat)k.append("response_format",f.responseFormat);if(f.prompt)k.append("prompt",f.prompt);if(f.temperature!==void 0)k.append("temperature",String(f.temperature));if(f.timestampGranularities)for(let P of f.timestampGranularities)k.append("timestamp_granularities[]",P);let t=O(o),i=await fetch(`${t}/audio/transcriptions`,{method:"POST",headers:R(o),body:k});if(!i.ok){let P=await i.text().catch(()=>"Unknown error");throw Error(`OpenAI transcription failed: ${i.status} ${i.statusText} - ${P}`)}return(await i.json()).text}async function H(o,n){let c,r,f="mp3",T,A;if(typeof n==="string")c=n,r=void 0;else{if(c=n.text,r=n.voice,"format"in n&&n.format)f=n.format;if("model"in n&&n.model)T=n.model;if("instructions"in n&&n.instructions)A=n.instructions}if(T=T??co(o),r=r??ro(o),A=A??fo(o),U.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${T}`),!c||c.trim().length===0)throw Error("TEXT_TO_SPEECH requires non-empty text");if(c.length>4096)throw Error("TEXT_TO_SPEECH text exceeds 4096 character limit");let k=["alloy","echo","fable","onyx","nova","shimmer"];if(r&&!k.includes(r))throw Error(`Invalid voice: ${r}. Must be one of: ${k.join(", ")}`);let t=O(o),i={model:T,voice:r,input:c,response_format:f};if(A&&A.length>0)i.instructions=A;let p=await fetch(`${t}/audio/speech`,{method:"POST",headers:{...R(o),"Content-Type":"application/json",...f==="mp3"?{Accept:"audio/mpeg"}:{}},body:JSON.stringify(i)});if(!p.ok){let P=await p.text().catch(()=>"Unknown error");throw Error(`OpenAI TTS failed: ${p.status} ${p.statusText} - ${P}`)}return p.arrayBuffer()}import{logger as x,ModelType as Go,VECTOR_DIMS as qo}from"@elizaos/core";import{EventType as Do}from"@elizaos/core";var Oo=200;function Zo(o){if(o.length<=Oo)return o;return`${o.slice(0,Oo)}…`}function Yo(o){if("promptTokens"in o){let n="promptTokensDetails"in o?o.promptTokensDetails:void 0,c=o.cachedPromptTokens??n?.cachedTokens;return{promptTokens:o.promptTokens??0,completionTokens:o.completionTokens??0,totalTokens:o.totalTokens??(o.promptTokens??0)+(o.completionTokens??0),cachedPromptTokens:c}}if("inputTokens"in o||"outputTokens"in o){let n=o.inputTokens??0,c=o.outputTokens??0,r=o.totalTokens??n+c;return{promptTokens:n,completionTokens:c,totalTokens:r,cachedPromptTokens:o.cachedInputTokens}}return{promptTokens:0,completionTokens:0,totalTokens:0}}function $(o,n,c,r){let f=Yo(r),T={runtime:o,source:"openai",provider:"openai",type:n,prompt:Zo(c),tokens:{prompt:f.promptTokens,completion:f.completionTokens,total:f.totalTokens,...f.cachedPromptTokens!==void 0?{cached:f.cachedPromptTokens}:{}}};o.emitEvent(Do.MODEL_USED,T)}var ho=8000;function Mo(o){let n=Object.values(qo);if(!n.includes(o))throw Error(`Invalid embedding dimension: ${o}. Must be one of: ${n.join(", ")}`);return o}function eo(o){if(o===null)return null;if(typeof o==="string")return o;if(typeof o==="object"&&typeof o.text==="string")return o.text;throw Error("Invalid embedding params: expected string, { text: string }, or null")}async function V(o,n){let c=m(o),r=Mo(ko(o)),f=eo(n);if(f===null){x.debug("[OpenAI] Creating test embedding for initialization");let y=Array(r).fill(0);return y[0]=0.1,y}let T=f.trim();if(T.length===0)throw Error("Cannot generate embedding for empty text");let A=ho*4;if(T.length>A)x.warn(`[OpenAI] Embedding input too long (~${Math.ceil(T.length/4)} tokens), truncating to ~${ho} tokens`),T=T.slice(0,A);let t=`${u(o)}/embeddings`;x.debug(`[OpenAI] Generating embedding with model: ${c}`);let i=await fetch(t,{method:"POST",headers:{...R(o,!0),"Content-Type":"application/json"},body:JSON.stringify({model:c,input:T})});if(!i.ok){let y=await i.text().catch(()=>"Unknown error");throw Error(`OpenAI embedding API error: ${i.status} ${i.statusText} - ${y}`)}let p=await i.json(),P=p?.data?.[0];if(!P||!P.embedding)throw Error("OpenAI API returned invalid embedding response structure");let w=P.embedding;if(w.length!==r)throw Error(`Embedding dimension mismatch: got ${w.length}, expected ${r}. Check OPENAI_EMBEDDING_DIMENSIONS setting.`);if(p.usage)$(o,Go.TEXT_EMBEDDING,T,{promptTokens:p.usage.prompt_tokens,completionTokens:0,totalTokens:p.usage.total_tokens});return x.debug(`[OpenAI] Generated embedding with ${w.length} dimensions`),w}import{logger as lo,ModelType as Bo}from"@elizaos/core";var Ro="Please analyze this image and provide a title and detailed description.";async function Q(o,n){let c=To(o),r=n.count??1,f=n.size??"1024x1024",T=n;if(lo.debug(`[OpenAI] Using IMAGE model: ${c}`),!n.prompt||n.prompt.trim().length===0)throw Error("IMAGE generation requires a non-empty prompt");if(r<1||r>10)throw Error("IMAGE count must be between 1 and 10");let A=O(o),k={model:c,prompt:n.prompt,n:r,size:f};if(T.quality)k.quality=T.quality;if(T.style)k.style=T.style;let t=await fetch(`${A}/images/generations`,{method:"POST",headers:{...R(o),"Content-Type":"application/json"},body:JSON.stringify(k)});if(!t.ok){let p=await t.text().catch(()=>"Unknown error");throw Error(`OpenAI image generation failed: ${t.status} ${t.statusText} - ${p}`)}let i=await t.json();if(!i.data||i.data.length===0)throw Error("OpenAI API returned no images");return i.data.map((p)=>({url:p.url,revisedPrompt:p.revised_prompt}))}function ao(o){return o.match(/title[:\s]+(.+?)(?:\n|$)/i)?.[1]?.trim()??"Image Analysis"}function vo(o){return o.replace(/title[:\s]+(.+?)(?:\n|$)/i,"").trim()}async function D(o,n){let c=oo(o),r=Ao(o);lo.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${c}`);let f,T;if(typeof n==="string")f=n,T=Ro;else f=n.imageUrl,T=n.prompt??Ro;if(!f||f.trim().length===0)throw Error("IMAGE_DESCRIPTION requires a valid image URL");let A=O(o),k={model:c,messages:[{role:"user",content:[{type:"text",text:T},{type:"image_url",image_url:{url:f}}]}],max_tokens:r},t=await fetch(`${A}/chat/completions`,{method:"POST",headers:{...R(o),"Content-Type":"application/json"},body:JSON.stringify(k)});if(!t.ok){let w=await t.text().catch(()=>"Unknown error");throw Error(`OpenAI image description failed: ${t.status} ${t.statusText} - ${w}`)}let i=await t.json();if(i.usage)$(o,Bo.IMAGE_DESCRIPTION,typeof n==="string"?n:n.prompt??"",{promptTokens:i.usage.prompt_tokens,completionTokens:i.usage.completion_tokens,totalTokens:i.usage.total_tokens});let P=i.choices?.[0]?.message?.content;if(!P)throw Error("OpenAI API returned empty image description");return{title:ao(P),description:vo(P)}}import{logger as bo,ModelType as No}from"@elizaos/core";import{generateObject as on}from"ai";import{createOpenAI as yo}from"@ai-sdk/openai";var go="sk-proxy";function L(o){let n=O(o),c=J(o);if(!c&&g(o))return yo({apiKey:go,baseURL:n});if(!c)throw Error("OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.");return yo({apiKey:c,baseURL:n})}import{logger as Jo}from"@elizaos/core";import{JSONParseError as uo}from"ai";var mo={MARKDOWN_JSON:/```json\n|\n```|```/g,WHITESPACE:/^\s+|\s+$/g};function $o(){return async({text:o,error:n})=>{if(!(n instanceof uo))return null;try{let c=o.replace(mo.MARKDOWN_JSON,"");return JSON.parse(c),Jo.debug("[JSON Repair] Successfully repaired JSON by removing markdown wrappers"),c}catch{return Jo.warn("[JSON Repair] Unable to repair JSON text"),null}}}async function So(o,n,c,r){let f=L(o),T=r(o);if(bo.debug(`[OpenAI] Using ${c} model: ${T}`),!n.prompt||n.prompt.trim().length===0)throw Error("Object generation requires a non-empty prompt");if(n.schema)bo.debug("[OpenAI] Schema provided but using no-schema mode. Structure is determined by prompt instructions.");let A=f.chat(T),{object:k,usage:t}=await on({model:A,output:"no-schema",prompt:n.prompt,experimental_repairText:$o()});if(t)$(o,c,n.prompt,t);if(typeof k!=="object"||k===null)throw Error(`Object generation returned ${typeof k}, expected object`);return k}async function Z(o,n){return So(o,n,No.OBJECT_SMALL,F)}async function Y(o,n){return So(o,n,No.OBJECT_LARGE,d)}import{logger as b}from"@elizaos/core";function nn(o){switch(o.type){case"web_search_preview":return{type:"web_search_preview"};case"file_search":return{type:"file_search",vector_store_ids:o.vectorStoreIds};case"code_interpreter":return{type:"code_interpreter",container:o.container??{type:"auto"}};case"mcp":return{type:"mcp",server_label:o.serverLabel,server_url:o.serverUrl,require_approval:o.requireApproval??"never"};default:throw Error(`Unknown research tool type: ${o.type}`)}}function cn(o){switch(o.type){case"web_search_call":return{id:o.id??"",type:"web_search_call",status:o.status??"completed",action:{type:o.action?.type??"search",query:o.action?.query,url:o.action?.url}};case"file_search_call":return{id:o.id??"",type:"file_search_call",status:o.status??"completed",query:o.query??"",results:o.results?.map((n)=>({fileId:n.file_id,fileName:n.file_name,score:n.score}))};case"code_interpreter_call":return{id:o.id??"",type:"code_interpreter_call",status:o.status??"completed",code:o.code??"",output:o.output};case"mcp_tool_call":return{id:o.id??"",type:"mcp_tool_call",status:o.status??"completed",serverLabel:o.server_label??"",toolName:o.tool_name??"",arguments:o.arguments??{},result:o.result};case"message":return{type:"message",content:o.content?.map((n)=>({type:"output_text",text:n.text,annotations:n.annotations?.map((c)=>({url:c.url,title:c.title,startIndex:c.start_index,endIndex:c.end_index}))??[]}))??[]};default:return null}}function rn(o){if(o.output_text){let r=[];if(o.output){for(let f of o.output)if(f.type==="message"&&f.content){for(let T of f.content)if(T.annotations)for(let A of T.annotations)r.push({url:A.url,title:A.title,startIndex:A.start_index,endIndex:A.end_index})}}return{text:o.output_text,annotations:r}}let n="",c=[];if(o.output){for(let r of o.output)if(r.type==="message"&&r.content){for(let f of r.content)if(n+=f.text,f.annotations)for(let T of f.annotations)c.push({url:T.url,title:T.title,startIndex:T.start_index,endIndex:T.end_index})}}return{text:n,annotations:c}}async function G(o,n){let c=J(o);if(!c)throw Error("OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.");let r=O(o),f=n.model??to(o),T=Po(o);b.debug(`[OpenAI] Starting deep research with model: ${f}`),b.debug(`[OpenAI] Research input: ${n.input.substring(0,100)}...`);let A=n.tools?.filter((N)=>N.type==="web_search_preview"||N.type==="file_search"||N.type==="mcp");if(!A||A.length===0)b.debug("[OpenAI] No data source tools specified, defaulting to web_search_preview"),n.tools=[{type:"web_search_preview"},...n.tools??[]];let k={model:f,input:n.input};if(n.instructions)k.instructions=n.instructions;if(n.background!==void 0)k.background=n.background;if(n.tools&&n.tools.length>0)k.tools=n.tools.map(nn);if(n.maxToolCalls!==void 0)k.max_tool_calls=n.maxToolCalls;if(n.reasoningSummary)k.reasoning={summary:n.reasoningSummary};b.debug(`[OpenAI] Research request body: ${JSON.stringify(k,null,2)}`);let t=await fetch(`${r}/responses`,{method:"POST",headers:{Authorization:`Bearer ${c}`,"Content-Type":"application/json"},body:JSON.stringify(k),signal:AbortSignal.timeout(T)});if(!t.ok){let N=await t.text();throw b.error(`[OpenAI] Research request failed: ${t.status} ${N}`),Error(`Deep research request failed: ${t.status} ${t.statusText}`)}let i=await t.json();if(i.error)throw b.error(`[OpenAI] Research API error: ${i.error.message}`),Error(`Deep research error: ${i.error.message}`);b.debug(`[OpenAI] Research response received. Status: ${i.status??"completed"}`);let{text:p,annotations:P}=rn(i),w=[];if(i.output)for(let N of i.output){let a=cn(N);if(a)w.push(a)}let y={id:i.id,text:p,annotations:P,outputItems:w,status:i.status};return b.info(`[OpenAI] Research completed. Text length: ${p.length}, Annotations: ${P.length}, Output items: ${w.length}`),y}import{logger as fn,ModelType as jo}from"@elizaos/core";import{generateText as Tn,streamText as kn}from"ai";function An(o){if(!o)return;let n=o.inputTokens??0,c=o.outputTokens??0,r=o;return{promptTokens:n,completionTokens:c,totalTokens:n+c,cachedPromptTokens:r.cachedInputTokens}}function tn(o){let n=o;return{promptCacheKey:n.providerOptions?.openai?.promptCacheKey,promptCacheRetention:n.providerOptions?.openai?.promptCacheRetention}}async function Eo(o,n,c,r){let f=L(o),T=r(o);fn.debug(`[OpenAI] Using ${c} model: ${T}`);let A=tn(n),k=o.character.system??void 0,i={model:f.chat(T),prompt:n.prompt,system:k,maxOutputTokens:n.maxTokens??8192,experimental_telemetry:{isEnabled:io(o)},...A.promptCacheKey||A.promptCacheRetention?{providerOptions:{openai:{...A.promptCacheKey?{promptCacheKey:A.promptCacheKey}:{},...A.promptCacheRetention?{promptCacheRetention:A.promptCacheRetention}:{}}}}:{}};if(n.stream){let w=kn(i);return{textStream:w.textStream,text:Promise.resolve(w.text),usage:Promise.resolve(w.usage).then(An),finishReason:Promise.resolve(w.finishReason).then((y)=>y)}}let{text:p,usage:P}=await Tn(i);if(P)$(o,c,n.prompt,P);return p}async function q(o,n){return Eo(o,n,jo.TEXT_SMALL,F)}async function M(o,n){return Eo(o,n,jo.TEXT_LARGE,d)}import{ModelType as _o}from"@elizaos/core";import{ModelType as Pn}from"@elizaos/core";import{encodingForModel as In,getEncoding as pn}from"js-tiktoken";function Fo(o){let c=o.toLowerCase().includes("4o")?"o200k_base":"cl100k_base";try{return In(o)}catch{return pn(c)}}function Uo(o,n){if(n===Pn.TEXT_SMALL)return F(o);return d(o)}function zo(o,n,c){let r=Uo(o,n);return Fo(r).encode(c)}function Lo(o,n,c){let r=Uo(o,n);return Fo(r).decode(c)}async function e(o,n){if(!n.prompt)throw Error("Tokenization requires a non-empty prompt");let c=n.modelType??_o.TEXT_LARGE;return zo(o,c,n.prompt)}async function B(o,n){if(!n.tokens||!Array.isArray(n.tokens))throw Error("Detokenization requires a valid tokens array");if(n.tokens.length===0)return"";for(let r=0;r<n.tokens.length;r++){let f=n.tokens[r];if(typeof f!=="number"||!Number.isFinite(f))throw Error(`Invalid token at index ${r}: expected number`)}let c=n.modelType??_o.TEXT_LARGE;return Lo(o,c,n.tokens)}function sn(){if(typeof process>"u")return{};return process.env}var h=sn(),wn={name:"openai",description:"OpenAI API integration for text, image, audio, and embedding models",config:{OPENAI_API_KEY:h.OPENAI_API_KEY??null,OPENAI_BASE_URL:h.OPENAI_BASE_URL??null,OPENAI_SMALL_MODEL:h.OPENAI_SMALL_MODEL??null,OPENAI_LARGE_MODEL:h.OPENAI_LARGE_MODEL??null,SMALL_MODEL:h.SMALL_MODEL??null,LARGE_MODEL:h.LARGE_MODEL??null,OPENAI_EMBEDDING_MODEL:h.OPENAI_EMBEDDING_MODEL??null,OPENAI_EMBEDDING_API_KEY:h.OPENAI_EMBEDDING_API_KEY??null,OPENAI_EMBEDDING_URL:h.OPENAI_EMBEDDING_URL??null,OPENAI_EMBEDDING_DIMENSIONS:h.OPENAI_EMBEDDING_DIMENSIONS??null,OPENAI_IMAGE_DESCRIPTION_MODEL:h.OPENAI_IMAGE_DESCRIPTION_MODEL??null,OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS:h.OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS??null,OPENAI_EXPERIMENTAL_TELEMETRY:h.OPENAI_EXPERIMENTAL_TELEMETRY??null,OPENAI_RESEARCH_MODEL:h.OPENAI_RESEARCH_MODEL??null,OPENAI_RESEARCH_TIMEOUT:h.OPENAI_RESEARCH_TIMEOUT??null},async init(o,n){Io(o,n)},models:{[I.TEXT_EMBEDDING]:async(o,n)=>{return V(o,n)},[I.TEXT_TOKENIZER_ENCODE]:async(o,n)=>{return e(o,n)},[I.TEXT_TOKENIZER_DECODE]:async(o,n)=>{return B(o,n)},[I.TEXT_SMALL]:async(o,n)=>{return q(o,n)},[I.TEXT_LARGE]:async(o,n)=>{return M(o,n)},[I.IMAGE]:async(o,n)=>{return Q(o,n)},[I.IMAGE_DESCRIPTION]:async(o,n)=>{return D(o,n)},[I.TRANSCRIPTION]:async(o,n)=>{return C(o,n)},[I.TEXT_TO_SPEECH]:async(o,n)=>{return H(o,n)},[I.OBJECT_SMALL]:async(o,n)=>{return Z(o,n)},[I.OBJECT_LARGE]:async(o,n)=>{return Y(o,n)},[I.RESEARCH]:async(o,n)=>{return G(o,n)}},tests:[{name:"openai_plugin_tests",tests:[{name:"openai_test_api_connectivity",fn:async(o)=>{let n=O(o),c=await fetch(`${n}/models`,{headers:R(o)});if(!c.ok)throw Error(`API connectivity test failed: ${c.status} ${c.statusText}`);let r=await c.json();l.info(`[OpenAI Test] API connected. ${r.data?.length??0} models available.`)}},{name:"openai_test_text_embedding",fn:async(o)=>{let n=await o.useModel(I.TEXT_EMBEDDING,{text:"Hello, world!"});if(!Array.isArray(n)||n.length===0)throw Error("Embedding should return a non-empty array");l.info(`[OpenAI Test] Generated embedding with ${n.length} dimensions`)}},{name:"openai_test_text_small",fn:async(o)=>{let n=await o.useModel(I.TEXT_SMALL,{prompt:"Say hello in exactly 5 words."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_SMALL should return non-empty string");l.info(`[OpenAI Test] TEXT_SMALL generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_text_large",fn:async(o)=>{let n=await o.useModel(I.TEXT_LARGE,{prompt:"Explain quantum computing in 2 sentences."});if(typeof n!=="string"||n.length===0)throw Error("TEXT_LARGE should return non-empty string");l.info(`[OpenAI Test] TEXT_LARGE generated: "${n.substring(0,50)}..."`)}},{name:"openai_test_tokenizer_roundtrip",fn:async(o)=>{let c=await o.useModel(I.TEXT_TOKENIZER_ENCODE,{prompt:"Hello, tokenizer test!",modelType:I.TEXT_SMALL});if(!Array.isArray(c)||c.length===0)throw Error("Tokenization should return non-empty token array");let r=await o.useModel(I.TEXT_TOKENIZER_DECODE,{tokens:c,modelType:I.TEXT_SMALL});if(r!=="Hello, tokenizer test!")throw Error(`Tokenizer roundtrip failed: expected "Hello, tokenizer test!", got "${r}"`);l.info(`[OpenAI Test] Tokenizer roundtrip successful (${c.length} tokens)`)}},{name:"openai_test_streaming",fn:async(o)=>{let n=[],c=await o.useModel(I.TEXT_LARGE,{prompt:"Count from 1 to 5, one number per line.",stream:!0,onStreamChunk:(r)=>{n.push(r)}});if(typeof c!=="string"||c.length===0)throw Error("Streaming should return non-empty result");if(n.length===0)throw Error("No streaming chunks received");l.info(`[OpenAI Test] Streaming test: ${n.length} chunks received`)}},{name:"openai_test_image_description",fn:async(o)=>{let c=await o.useModel(I.IMAGE_DESCRIPTION,"https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Camponotus_flavomarginatus_ant.jpg/440px-Camponotus_flavomarginatus_ant.jpg");if(!c||typeof c!=="object"||!("title"in c)||!("description"in c))throw Error("Image description should return { title, description }");l.info(`[OpenAI Test] Image described: "${c.title}"`)}},{name:"openai_test_transcription",fn:async(o)=>{let r=await(await fetch("https://upload.wikimedia.org/wikipedia/commons/2/25/En-Open_Source.ogg")).arrayBuffer(),f=Buffer.from(new Uint8Array(r)),T=await o.useModel(I.TRANSCRIPTION,f);if(typeof T!=="string")throw Error("Transcription should return a string");l.info(`[OpenAI Test] Transcription: "${T.substring(0,50)}..."`)}},{name:"openai_test_text_to_speech",fn:async(o)=>{let n=await o.useModel(I.TEXT_TO_SPEECH,{text:"Hello, this is a text-to-speech test."});if(!(n instanceof ArrayBuffer)||n.byteLength===0)throw Error("TTS should return non-empty ArrayBuffer");l.info(`[OpenAI Test] TTS generated ${n.byteLength} bytes of audio`)}},{name:"openai_test_object_generation",fn:async(o)=>{let n=await o.useModel(I.OBJECT_SMALL,{prompt:"Return a JSON object with exactly these fields: name (string), age (number), active (boolean)"});if(!n||typeof n!=="object")throw Error("Object generation should return an object");l.info(`[OpenAI Test] Object generated: ${JSON.stringify(n).substring(0,100)}`)}},{name:"openai_test_research",fn:async(o)=>{let n=await o.useModel(I.RESEARCH,{input:"What is the current date and time?",tools:[{type:"web_search_preview"}],maxToolCalls:3});if(!n||typeof n!=="object"||!("text"in n))throw Error("Research should return an object with text property");if(typeof n.text!=="string"||n.text.length===0)throw Error("Research result text should be a non-empty string");l.info(`[OpenAI Test] Research completed. Text length: ${n.text.length}, Annotations: ${n.annotations?.length??0}`)}}]}]},On=wn;export{wn as openaiPlugin,On as default};
|
|
2
2
|
|
|
3
|
-
//# debugId=
|
|
3
|
+
//# debugId=01A431D38E4333C464756E2164756E21
|
|
@@ -7,18 +7,18 @@
|
|
|
7
7
|
"import type { IAgentRuntime } from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\n\nfunction getEnvValue(key: string): string | undefined {\n if (typeof process === \"undefined\" || !process.env) {\n return undefined;\n }\n const value = process.env[key];\n return value === undefined ? undefined : String(value);\n}\n\nexport function getSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue?: string\n): string | undefined {\n const value = runtime.getSetting(key);\n if (value !== undefined && value !== null) {\n return String(value);\n }\n return getEnvValue(key) ?? defaultValue;\n}\nexport function getRequiredSetting(\n runtime: IAgentRuntime,\n key: string,\n errorMessage?: string\n): string {\n const value = getSetting(runtime, key);\n if (value === undefined || value.trim() === \"\") {\n throw new Error(errorMessage ?? `Required setting '${key}' is not configured`);\n }\n return value;\n}\n\nexport function getNumericSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue: number\n): number {\n const value = getSetting(runtime, key);\n if (value === undefined) {\n return defaultValue;\n }\n const parsed = Number.parseInt(value, 10);\n if (!Number.isFinite(parsed)) {\n throw new Error(`Setting '${key}' must be a valid integer, got: ${value}`);\n }\n return parsed;\n}\n\nexport function getBooleanSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue: boolean\n): boolean {\n const value = getSetting(runtime, key);\n if (value === undefined) {\n return defaultValue;\n }\n const normalized = value.toLowerCase();\n return normalized === \"true\" || normalized === \"1\" || normalized === \"yes\";\n}\n\nexport function isBrowser(): boolean {\n return (\n typeof globalThis !== \"undefined\" &&\n typeof (globalThis as { document?: Document }).document !== \"undefined\"\n );\n}\n\nexport function isProxyMode(runtime: IAgentRuntime): boolean {\n return isBrowser() && !!getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n}\n\nexport function getApiKey(runtime: IAgentRuntime): string | undefined {\n return getSetting(runtime, \"OPENAI_API_KEY\");\n}\n\nexport function getEmbeddingApiKey(runtime: IAgentRuntime): string | undefined {\n const embeddingApiKey = getSetting(runtime, \"OPENAI_EMBEDDING_API_KEY\");\n if (embeddingApiKey) {\n logger.debug(\"[OpenAI] Using specific embedding API key\");\n return embeddingApiKey;\n }\n logger.debug(\"[OpenAI] Falling back to general API key for embeddings\");\n return getApiKey(runtime);\n}\n\nexport function getAuthHeader(\n runtime: IAgentRuntime,\n forEmbedding = false\n): Record<string, string> {\n // By default this plugin does NOT send auth headers in the browser. This is safer because\n // frontend builds would otherwise expose secrets. For local demos, you can explicitly\n // opt-in to sending the Authorization header by setting OPENAI_ALLOW_BROWSER_API_KEY=true.\n if (isBrowser() && !getBooleanSetting(runtime, \"OPENAI_ALLOW_BROWSER_API_KEY\", false)) {\n return {};\n }\n const key = forEmbedding ? getEmbeddingApiKey(runtime) : getApiKey(runtime);\n return key ? { Authorization: `Bearer ${key}` } : {};\n}\n\nexport function getBaseURL(runtime: IAgentRuntime): string {\n const browserURL = getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n const baseURL =\n isBrowser() && browserURL\n ? browserURL\n : (getSetting(runtime, \"OPENAI_BASE_URL\") ?? \"https://api.openai.com/v1\");\n logger.debug(`[OpenAI] Base URL: ${baseURL}`);\n return baseURL;\n}\n\nexport function getEmbeddingBaseURL(runtime: IAgentRuntime): string {\n const embeddingURL = isBrowser()\n ? (getSetting(runtime, \"OPENAI_BROWSER_EMBEDDING_URL\") ??\n getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\"))\n : getSetting(runtime, \"OPENAI_EMBEDDING_URL\");\n\n if (embeddingURL) {\n logger.debug(`[OpenAI] Using embedding base URL: ${embeddingURL}`);\n return embeddingURL;\n }\n\n logger.debug(\"[OpenAI] Falling back to general base URL for embeddings\");\n return getBaseURL(runtime);\n}\n\nexport function getSmallModel(runtime: IAgentRuntime): string {\n return (\n getSetting(runtime, \"OPENAI_SMALL_MODEL\") ?? getSetting(runtime, \"SMALL_MODEL\") ?? \"gpt-5-mini\"\n );\n}\n\nexport function getLargeModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_LARGE_MODEL\") ?? getSetting(runtime, \"LARGE_MODEL\") ?? \"gpt-5\";\n}\n\nexport function getEmbeddingModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_EMBEDDING_MODEL\") ?? \"text-embedding-3-small\";\n}\n\nexport function getImageDescriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MODEL\") ?? \"gpt-5-mini\";\n}\n\nexport function getTranscriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TRANSCRIPTION_MODEL\") ?? \"gpt-5-mini-transcribe\";\n}\n\nexport function getTTSModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_MODEL\") ?? \"tts-1\";\n}\n\nexport function getTTSVoice(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_VOICE\") ?? \"nova\";\n}\n\nexport function getTTSInstructions(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_INSTRUCTIONS\") ?? \"\";\n}\n\nexport function getImageModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_MODEL\") ?? \"dall-e-3\";\n}\n\nexport function getExperimentalTelemetry(runtime: IAgentRuntime): boolean {\n return getBooleanSetting(runtime, \"OPENAI_EXPERIMENTAL_TELEMETRY\", false);\n}\n\nexport function getEmbeddingDimensions(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_EMBEDDING_DIMENSIONS\", 1536);\n}\n\nexport function getImageDescriptionMaxTokens(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS\", 8192);\n}\n\nexport function getResearchModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_RESEARCH_MODEL\") ?? \"o3-deep-research\";\n}\n\nexport function getResearchTimeout(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_RESEARCH_TIMEOUT\", 3600000);\n}\n",
|
|
8
8
|
"import type {\n TextToSpeechParams as CoreTextToSpeechParams,\n TranscriptionParams as CoreTranscriptionParams,\n IAgentRuntime,\n} from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport type {\n TextToSpeechParams as LocalTextToSpeechParams,\n TranscriptionParams as LocalTranscriptionParams,\n OpenAITranscriptionResponse,\n TTSOutputFormat,\n TTSVoice,\n} from \"../types\";\nimport { detectAudioMimeType, getFilenameForMimeType } from \"../utils/audio\";\nimport {\n getAuthHeader,\n getBaseURL,\n getTranscriptionModel,\n getTTSInstructions,\n getTTSModel,\n getTTSVoice,\n} from \"../utils/config\";\n\ntype AudioInput = Blob | File | Buffer;\ntype TranscriptionInput = AudioInput | LocalTranscriptionParams | CoreTranscriptionParams | string;\ntype TTSInput = string | LocalTextToSpeechParams | CoreTextToSpeechParams;\n\nfunction isBlobOrFile(value: unknown): value is Blob | File {\n return value instanceof Blob || value instanceof File;\n}\n\nfunction isBuffer(value: unknown): value is Buffer {\n return Buffer.isBuffer(value);\n}\n\nfunction isLocalTranscriptionParams(value: unknown): value is LocalTranscriptionParams {\n return (\n typeof value === \"object\" &&\n value !== null &&\n \"audio\" in value &&\n (isBlobOrFile((value as LocalTranscriptionParams).audio) ||\n isBuffer((value as LocalTranscriptionParams).audio))\n );\n}\n\nfunction isCoreTranscriptionParams(value: unknown): value is CoreTranscriptionParams {\n return (\n typeof value === \"object\" &&\n value !== null &&\n \"audioUrl\" in value &&\n typeof (value as CoreTranscriptionParams).audioUrl === \"string\"\n );\n}\n\nasync function fetchAudioFromUrl(url: string): Promise<Blob> {\n const response = await fetch(url);\n if (!response.ok) {\n throw new Error(`Failed to fetch audio from URL: ${response.status}`);\n }\n return response.blob();\n}\nexport async function handleTranscription(\n runtime: IAgentRuntime,\n input: TranscriptionInput\n): Promise<string> {\n let modelName = getTranscriptionModel(runtime);\n let blob: Blob;\n let extraParams: Partial<LocalTranscriptionParams> = {};\n\n if (typeof input === \"string\") {\n logger.debug(`[OpenAI] Fetching audio from URL: ${input}`);\n blob = await fetchAudioFromUrl(input);\n } else if (isBlobOrFile(input)) {\n blob = input;\n } else if (isBuffer(input)) {\n const mimeType = detectAudioMimeType(input);\n logger.debug(`[OpenAI] Auto-detected audio MIME type: ${mimeType}`);\n blob = new Blob([new Uint8Array(input)], { type: mimeType });\n } else if (isLocalTranscriptionParams(input)) {\n extraParams = input;\n if (input.model) {\n modelName = input.model;\n }\n if (isBuffer(input.audio)) {\n const mimeType = input.mimeType ?? detectAudioMimeType(input.audio);\n logger.debug(`[OpenAI] Using MIME type: ${mimeType}`);\n blob = new Blob([new Uint8Array(input.audio)], { type: mimeType });\n } else {\n blob = input.audio;\n }\n } else if (isCoreTranscriptionParams(input)) {\n logger.debug(`[OpenAI] Fetching audio from URL: ${input.audioUrl}`);\n blob = await fetchAudioFromUrl(input.audioUrl);\n extraParams = { prompt: input.prompt };\n } else {\n throw new Error(\n \"TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object\"\n );\n }\n\n logger.debug(`[OpenAI] Using TRANSCRIPTION model: ${modelName}`);\n\n const mimeType = (blob as File).type || \"audio/webm\";\n const filename =\n (blob as File).name ||\n getFilenameForMimeType(\n mimeType.startsWith(\"audio/\")\n ? (mimeType as ReturnType<typeof detectAudioMimeType>)\n : \"audio/webm\"\n );\n\n const formData = new FormData();\n formData.append(\"file\", blob, filename);\n formData.append(\"model\", modelName);\n\n if (extraParams.language) {\n formData.append(\"language\", extraParams.language);\n }\n if (extraParams.responseFormat) {\n formData.append(\"response_format\", extraParams.responseFormat);\n }\n if (extraParams.prompt) {\n formData.append(\"prompt\", extraParams.prompt);\n }\n if (extraParams.temperature !== undefined) {\n formData.append(\"temperature\", String(extraParams.temperature));\n }\n if (extraParams.timestampGranularities) {\n for (const granularity of extraParams.timestampGranularities) {\n formData.append(\"timestamp_granularities[]\", granularity);\n }\n }\n\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/audio/transcriptions`, {\n method: \"POST\",\n headers: getAuthHeader(runtime),\n body: formData,\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI transcription failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAITranscriptionResponse;\n return data.text;\n}\n\nexport async function handleTextToSpeech(\n runtime: IAgentRuntime,\n input: TTSInput\n): Promise<ArrayBuffer> {\n let text: string;\n let voice: string | undefined;\n let format: TTSOutputFormat = \"mp3\";\n let model: string | undefined;\n let instructions: string | undefined;\n\n if (typeof input === \"string\") {\n text = input;\n voice = undefined;\n } else {\n text = input.text;\n voice = input.voice;\n if (\"format\" in input && input.format) {\n format = input.format;\n }\n if (\"model\" in input && input.model) {\n model = input.model;\n }\n if (\"instructions\" in input && input.instructions) {\n instructions = input.instructions;\n }\n }\n\n model = model ?? getTTSModel(runtime);\n voice = voice ?? getTTSVoice(runtime);\n instructions = instructions ?? getTTSInstructions(runtime);\n\n logger.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${model}`);\n\n if (!text || text.trim().length === 0) {\n throw new Error(\"TEXT_TO_SPEECH requires non-empty text\");\n }\n\n if (text.length > 4096) {\n throw new Error(\"TEXT_TO_SPEECH text exceeds 4096 character limit\");\n }\n\n const validVoices: TTSVoice[] = [\"alloy\", \"echo\", \"fable\", \"onyx\", \"nova\", \"shimmer\"];\n if (voice && !validVoices.includes(voice as TTSVoice)) {\n throw new Error(`Invalid voice: ${voice}. Must be one of: ${validVoices.join(\", \")}`);\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody: Record<string, string> = {\n model,\n voice: voice as TTSVoice,\n input: text,\n response_format: format,\n };\n\n if (instructions && instructions.length > 0) {\n requestBody.instructions = instructions;\n }\n\n const response = await fetch(`${baseURL}/audio/speech`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n ...(format === \"mp3\" ? { Accept: \"audio/mpeg\" } : {}),\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(`OpenAI TTS failed: ${response.status} ${response.statusText} - ${errorText}`);\n }\n\n return response.arrayBuffer();\n}\n",
|
|
9
9
|
"import { logger } from \"@elizaos/core\";\n\nconst MAGIC_BYTES = {\n WAV: {\n HEADER: [0x52, 0x49, 0x46, 0x46] as const, // \"RIFF\"\n IDENTIFIER: [0x57, 0x41, 0x56, 0x45] as const, // \"WAVE\"\n },\n MP3_ID3: [0x49, 0x44, 0x33] as const, // \"ID3\"\n OGG: [0x4f, 0x67, 0x67, 0x53] as const, // \"OggS\"\n FLAC: [0x66, 0x4c, 0x61, 0x43] as const, // \"fLaC\"\n FTYP: [0x66, 0x74, 0x79, 0x70] as const, // \"ftyp\" at offset 4 for mp4/m4a\n WEBM_EBML: [0x1a, 0x45, 0xdf, 0xa3] as const, // EBML header\n} as const;\n\nconst MIN_DETECTION_BUFFER_SIZE = 12;\n\nexport type AudioMimeType =\n | \"audio/wav\"\n | \"audio/mpeg\"\n | \"audio/ogg\"\n | \"audio/flac\"\n | \"audio/mp4\"\n | \"audio/webm\"\n | \"application/octet-stream\";\n\nfunction matchBytes(buffer: Buffer, offset: number, expected: readonly number[]): boolean {\n for (let i = 0; i < expected.length; i++) {\n const expectedByte = expected[i];\n if (expectedByte === undefined || buffer[offset + i] !== expectedByte) {\n return false;\n }\n }\n return true;\n}\n\nexport function detectAudioMimeType(buffer: Buffer): AudioMimeType {\n if (buffer.length < MIN_DETECTION_BUFFER_SIZE) {\n return \"application/octet-stream\";\n }\n\n // WAV: \"RIFF\" + size + \"WAVE\"\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.WAV.HEADER) &&\n matchBytes(buffer, 8, MAGIC_BYTES.WAV.IDENTIFIER)\n ) {\n return \"audio/wav\";\n }\n\n // MP3: ID3 tag or MPEG frame sync\n const firstByte = buffer[0];\n const secondByte = buffer[1];\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.MP3_ID3) ||\n (firstByte === 0xff && secondByte !== undefined && (secondByte & 0xe0) === 0xe0)\n ) {\n return \"audio/mpeg\";\n }\n\n // OGG: \"OggS\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.OGG)) {\n return \"audio/ogg\";\n }\n\n // FLAC: \"fLaC\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.FLAC)) {\n return \"audio/flac\";\n }\n\n // M4A/MP4: \"ftyp\" at offset 4\n if (matchBytes(buffer, 4, MAGIC_BYTES.FTYP)) {\n return \"audio/mp4\";\n }\n\n // WebM: EBML header\n if (matchBytes(buffer, 0, MAGIC_BYTES.WEBM_EBML)) {\n return \"audio/webm\";\n }\n\n logger.warn(\"Could not detect audio format from buffer, using generic binary type\");\n return \"application/octet-stream\";\n}\n\nexport function getExtensionForMimeType(mimeType: AudioMimeType): string {\n switch (mimeType) {\n case \"audio/wav\":\n return \"wav\";\n case \"audio/mpeg\":\n return \"mp3\";\n case \"audio/ogg\":\n return \"ogg\";\n case \"audio/flac\":\n return \"flac\";\n case \"audio/mp4\":\n return \"m4a\";\n case \"audio/webm\":\n return \"webm\";\n case \"application/octet-stream\":\n return \"bin\";\n }\n}\n\nexport function getFilenameForMimeType(mimeType: AudioMimeType): string {\n const ext = getExtensionForMimeType(mimeType);\n return `recording.${ext}`;\n}\n\nexport function validateAudioFormat(buffer: Buffer): AudioMimeType {\n const mimeType = detectAudioMimeType(buffer);\n if (mimeType === \"application/octet-stream\") {\n throw new Error(\n \"Unable to detect audio format. Supported formats: WAV, MP3, OGG, FLAC, M4A, WebM\"\n );\n }\n return mimeType;\n}\n",
|
|
10
|
-
"import type { IAgentRuntime, TextEmbeddingParams } from \"@elizaos/core\";\nimport { logger, ModelType, VECTOR_DIMS } from \"@elizaos/core\";\n\n// Safe default that matches @elizaos/core's MAX_EMBEDDING_TOKENS (8 000).\n// Defined locally so the plugin doesn't break against core versions older\n// than 2.0.0-alpha.4 where the symbol wasn't exported yet.\nconst MAX_EMBEDDING_TOKENS = 8_000;\nimport type { OpenAIEmbeddingResponse } from \"../types\";\nimport {\n getAuthHeader,\n getEmbeddingBaseURL,\n getEmbeddingDimensions,\n getEmbeddingModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ntype VectorDimension = (typeof VECTOR_DIMS)[keyof typeof VECTOR_DIMS];\n\nfunction validateDimension(dimension: number): VectorDimension {\n const validDimensions = Object.values(VECTOR_DIMS) as number[];\n if (!validDimensions.includes(dimension)) {\n throw new Error(\n `Invalid embedding dimension: ${dimension}. Must be one of: ${validDimensions.join(\", \")}`\n );\n }\n return dimension as VectorDimension;\n}\n\nfunction extractText(params: TextEmbeddingParams | string | null): string | null {\n if (params === null) {\n return null;\n }\n if (typeof params === \"string\") {\n return params;\n }\n if (typeof params === \"object\" && typeof params.text === \"string\") {\n return params.text;\n }\n throw new Error(\"Invalid embedding params: expected string, { text: string }, or null\");\n}\n\nexport async function handleTextEmbedding(\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null\n): Promise<number[]> {\n const embeddingModel = getEmbeddingModel(runtime);\n const embeddingDimension = validateDimension(getEmbeddingDimensions(runtime));\n\n const text = extractText(params);\n if (text === null) {\n logger.debug(\"[OpenAI] Creating test embedding for initialization\");\n const testVector = new Array(embeddingDimension).fill(0);\n testVector[0] = 0.1;\n return testVector;\n }\n\n let trimmedText = text.trim();\n if (trimmedText.length === 0) {\n throw new Error(\"Cannot generate embedding for empty text\");\n }\n\n // Truncate to stay within embedding model token limits.\n // OpenAI embedding models support up to 8191 tokens per input;\n // MAX_EMBEDDING_TOKENS (8 000) provides a safe buffer.\n const maxChars = MAX_EMBEDDING_TOKENS * 4; // ~4 chars per token\n if (trimmedText.length > maxChars) {\n logger.warn(\n `[OpenAI] Embedding input too long (~${Math.ceil(trimmedText.length / 4)} tokens), truncating to ~${MAX_EMBEDDING_TOKENS} tokens`\n );\n trimmedText = trimmedText.slice(0, maxChars);\n }\n\n const baseURL = getEmbeddingBaseURL(runtime);\n const url = `${baseURL}/embeddings`;\n\n logger.debug(`[OpenAI] Generating embedding with model: ${embeddingModel}`);\n\n const response = await fetch(url, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime, true),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: embeddingModel,\n input: trimmedText,\n }),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI embedding API error: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIEmbeddingResponse;\n\n const firstResult = data?.data?.[0];\n if (!firstResult || !firstResult.embedding) {\n throw new Error(\"OpenAI API returned invalid embedding response structure\");\n }\n\n const embedding = firstResult.embedding;\n\n if (embedding.length !== embeddingDimension) {\n throw new Error(\n `Embedding dimension mismatch: got ${embedding.length}, expected ${embeddingDimension}. ` +\n `Check OPENAI_EMBEDDING_DIMENSIONS setting.`\n );\n }\n\n if (data.usage) {\n emitModelUsageEvent(runtime, ModelType.TEXT_EMBEDDING, trimmedText, {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: 0,\n totalTokens: data.usage.total_tokens,\n });\n }\n\n logger.debug(`[OpenAI] Generated embedding with ${embedding.length} dimensions`);\n return embedding;\n}\n",
|
|
11
|
-
"import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { EventType } from \"@elizaos/core\";\nimport type { TokenUsage } from \"../types\";\n\nconst MAX_PROMPT_LENGTH = 200;\n\ninterface ModelUsageEventPayload {\n runtime: IAgentRuntime;\n source: \"openai\";\n provider: \"openai\";\n type: ModelTypeName;\n prompt: string;\n tokens: {\n prompt: number;\n completion: number;\n total: number;\n };\n}\n\ninterface AISDKUsage {\n inputTokens?: number;\n outputTokens?: number;\n}\n\ninterface OpenAIAPIUsage {\n promptTokens?: number;\n completionTokens?: number;\n totalTokens?: number;\n}\n\ntype ModelUsage = TokenUsage | AISDKUsage | OpenAIAPIUsage;\n\nfunction truncatePrompt(prompt: string): string {\n if (prompt.length <= MAX_PROMPT_LENGTH) {\n return prompt;\n }\n return `${prompt.slice(0, MAX_PROMPT_LENGTH)}…`;\n}\n\nfunction normalizeUsage(usage: ModelUsage): TokenUsage {\n if (\"promptTokens\" in usage) {\n return {\n promptTokens: usage.promptTokens ?? 0,\n completionTokens: usage.completionTokens ?? 0,\n totalTokens: usage.totalTokens ?? (usage.promptTokens ?? 0) + (usage.completionTokens ?? 0),\n };\n }\n if (\"inputTokens\" in usage || \"outputTokens\" in usage) {\n const input = (usage as AISDKUsage).inputTokens ?? 0;\n const output = (usage as AISDKUsage).outputTokens ?? 0;\n return {\n promptTokens: input,\n completionTokens: output,\n totalTokens:
|
|
10
|
+
"import type { IAgentRuntime, TextEmbeddingParams } from \"@elizaos/core\";\nimport { logger, ModelType, VECTOR_DIMS } from \"@elizaos/core\";\n\n// Safe default that matches @elizaos/core's MAX_EMBEDDING_TOKENS (8 000).\n// Defined locally so the plugin doesn't break against core versions older\n// than 2.0.0-alpha.4 where the symbol wasn't exported yet.\nconst MAX_EMBEDDING_TOKENS = 8_000;\n\nimport type { OpenAIEmbeddingResponse } from \"../types\";\nimport {\n getAuthHeader,\n getEmbeddingBaseURL,\n getEmbeddingDimensions,\n getEmbeddingModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ntype VectorDimension = (typeof VECTOR_DIMS)[keyof typeof VECTOR_DIMS];\n\nfunction validateDimension(dimension: number): VectorDimension {\n const validDimensions = Object.values(VECTOR_DIMS) as number[];\n if (!validDimensions.includes(dimension)) {\n throw new Error(\n `Invalid embedding dimension: ${dimension}. Must be one of: ${validDimensions.join(\", \")}`\n );\n }\n return dimension as VectorDimension;\n}\n\nfunction extractText(params: TextEmbeddingParams | string | null): string | null {\n if (params === null) {\n return null;\n }\n if (typeof params === \"string\") {\n return params;\n }\n if (typeof params === \"object\" && typeof params.text === \"string\") {\n return params.text;\n }\n throw new Error(\"Invalid embedding params: expected string, { text: string }, or null\");\n}\n\nexport async function handleTextEmbedding(\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null\n): Promise<number[]> {\n const embeddingModel = getEmbeddingModel(runtime);\n const embeddingDimension = validateDimension(getEmbeddingDimensions(runtime));\n\n const text = extractText(params);\n if (text === null) {\n logger.debug(\"[OpenAI] Creating test embedding for initialization\");\n const testVector = new Array(embeddingDimension).fill(0);\n testVector[0] = 0.1;\n return testVector;\n }\n\n let trimmedText = text.trim();\n if (trimmedText.length === 0) {\n throw new Error(\"Cannot generate embedding for empty text\");\n }\n\n // Truncate to stay within embedding model token limits.\n // OpenAI embedding models support up to 8191 tokens per input;\n // MAX_EMBEDDING_TOKENS (8 000) provides a safe buffer.\n const maxChars = MAX_EMBEDDING_TOKENS * 4; // ~4 chars per token\n if (trimmedText.length > maxChars) {\n logger.warn(\n `[OpenAI] Embedding input too long (~${Math.ceil(trimmedText.length / 4)} tokens), truncating to ~${MAX_EMBEDDING_TOKENS} tokens`\n );\n trimmedText = trimmedText.slice(0, maxChars);\n }\n\n const baseURL = getEmbeddingBaseURL(runtime);\n const url = `${baseURL}/embeddings`;\n\n logger.debug(`[OpenAI] Generating embedding with model: ${embeddingModel}`);\n\n const response = await fetch(url, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime, true),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: embeddingModel,\n input: trimmedText,\n }),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI embedding API error: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIEmbeddingResponse;\n\n const firstResult = data?.data?.[0];\n if (!firstResult || !firstResult.embedding) {\n throw new Error(\"OpenAI API returned invalid embedding response structure\");\n }\n\n const embedding = firstResult.embedding;\n\n if (embedding.length !== embeddingDimension) {\n throw new Error(\n `Embedding dimension mismatch: got ${embedding.length}, expected ${embeddingDimension}. ` +\n `Check OPENAI_EMBEDDING_DIMENSIONS setting.`\n );\n }\n\n if (data.usage) {\n emitModelUsageEvent(runtime, ModelType.TEXT_EMBEDDING, trimmedText, {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: 0,\n totalTokens: data.usage.total_tokens,\n });\n }\n\n logger.debug(`[OpenAI] Generated embedding with ${embedding.length} dimensions`);\n return embedding;\n}\n",
|
|
11
|
+
"import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { EventType } from \"@elizaos/core\";\nimport type { TokenUsage } from \"../types\";\n\nconst MAX_PROMPT_LENGTH = 200;\n\ninterface ModelUsageEventPayload {\n runtime: IAgentRuntime;\n source: \"openai\";\n provider: \"openai\";\n type: ModelTypeName;\n prompt: string;\n tokens: {\n prompt: number;\n completion: number;\n total: number;\n cached?: number;\n };\n}\n\ninterface AISDKUsage {\n inputTokens?: number;\n outputTokens?: number;\n totalTokens?: number;\n cachedInputTokens?: number;\n}\n\ninterface OpenAIAPIUsage {\n promptTokens?: number;\n completionTokens?: number;\n totalTokens?: number;\n cachedPromptTokens?: number;\n promptTokensDetails?: {\n cachedTokens?: number;\n };\n}\n\ntype ModelUsage = TokenUsage | AISDKUsage | OpenAIAPIUsage;\n\nfunction truncatePrompt(prompt: string): string {\n if (prompt.length <= MAX_PROMPT_LENGTH) {\n return prompt;\n }\n return `${prompt.slice(0, MAX_PROMPT_LENGTH)}…`;\n}\n\nfunction normalizeUsage(usage: ModelUsage): TokenUsage {\n if (\"promptTokens\" in usage) {\n const promptTokensDetails =\n \"promptTokensDetails\" in usage ? usage.promptTokensDetails : undefined;\n const cachedPromptTokens = usage.cachedPromptTokens ?? promptTokensDetails?.cachedTokens;\n return {\n promptTokens: usage.promptTokens ?? 0,\n completionTokens: usage.completionTokens ?? 0,\n totalTokens: usage.totalTokens ?? (usage.promptTokens ?? 0) + (usage.completionTokens ?? 0),\n cachedPromptTokens,\n };\n }\n if (\"inputTokens\" in usage || \"outputTokens\" in usage) {\n const input = (usage as AISDKUsage).inputTokens ?? 0;\n const output = (usage as AISDKUsage).outputTokens ?? 0;\n const total = (usage as AISDKUsage).totalTokens ?? input + output;\n return {\n promptTokens: input,\n completionTokens: output,\n totalTokens: total,\n cachedPromptTokens: (usage as AISDKUsage).cachedInputTokens,\n };\n }\n return {\n promptTokens: 0,\n completionTokens: 0,\n totalTokens: 0,\n };\n}\n\nexport function emitModelUsageEvent(\n runtime: IAgentRuntime,\n type: ModelTypeName,\n prompt: string,\n usage: ModelUsage\n): void {\n const normalized = normalizeUsage(usage);\n\n const payload: ModelUsageEventPayload = {\n runtime,\n source: \"openai\",\n provider: \"openai\",\n type,\n prompt: truncatePrompt(prompt),\n tokens: {\n prompt: normalized.promptTokens,\n completion: normalized.completionTokens,\n total: normalized.totalTokens,\n ...(normalized.cachedPromptTokens !== undefined\n ? { cached: normalized.cachedPromptTokens }\n : {}),\n },\n };\n\n runtime.emitEvent(EventType.MODEL_USED, payload);\n}\n",
|
|
12
12
|
"import type { IAgentRuntime, ImageDescriptionParams, ImageGenerationParams } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport type {\n ImageDescriptionResult,\n ImageGenerationResult,\n ImageQuality,\n ImageSize,\n ImageStyle,\n OpenAIChatCompletionResponse,\n OpenAIImageGenerationResponse,\n} from \"../types\";\nimport {\n getAuthHeader,\n getBaseURL,\n getImageDescriptionMaxTokens,\n getImageDescriptionModel,\n getImageModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ninterface ExtendedImageGenerationParams extends ImageGenerationParams {\n quality?: ImageQuality;\n style?: ImageStyle;\n}\n\nconst DEFAULT_IMAGE_DESCRIPTION_PROMPT =\n \"Please analyze this image and provide a title and detailed description.\";\n\nexport async function handleImageGeneration(\n runtime: IAgentRuntime,\n params: ImageGenerationParams\n): Promise<ImageGenerationResult[]> {\n const modelName = getImageModel(runtime);\n const count = params.count ?? 1;\n const size: ImageSize = (params.size as ImageSize) ?? \"1024x1024\";\n const extendedParams = params as ExtendedImageGenerationParams;\n\n logger.debug(`[OpenAI] Using IMAGE model: ${modelName}`);\n\n if (!params.prompt || params.prompt.trim().length === 0) {\n throw new Error(\"IMAGE generation requires a non-empty prompt\");\n }\n\n if (count < 1 || count > 10) {\n throw new Error(\"IMAGE count must be between 1 and 10\");\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody: Record<string, string | number> = {\n model: modelName,\n prompt: params.prompt,\n n: count,\n size,\n };\n\n if (extendedParams.quality) {\n requestBody.quality = extendedParams.quality;\n }\n if (extendedParams.style) {\n requestBody.style = extendedParams.style;\n }\n\n const response = await fetch(`${baseURL}/images/generations`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI image generation failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIImageGenerationResponse;\n\n if (!data.data || data.data.length === 0) {\n throw new Error(\"OpenAI API returned no images\");\n }\n\n return data.data.map((item) => ({\n url: item.url,\n revisedPrompt: item.revised_prompt,\n }));\n}\n\nfunction parseTitleFromResponse(content: string): string {\n const titleMatch = content.match(/title[:\\s]+(.+?)(?:\\n|$)/i);\n return titleMatch?.[1]?.trim() ?? \"Image Analysis\";\n}\n\nfunction parseDescriptionFromResponse(content: string): string {\n return content.replace(/title[:\\s]+(.+?)(?:\\n|$)/i, \"\").trim();\n}\n\nexport async function handleImageDescription(\n runtime: IAgentRuntime,\n params: ImageDescriptionParams | string\n): Promise<ImageDescriptionResult> {\n const modelName = getImageDescriptionModel(runtime);\n const maxTokens = getImageDescriptionMaxTokens(runtime);\n\n logger.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${modelName}`);\n\n let imageUrl: string;\n let promptText: string;\n\n if (typeof params === \"string\") {\n imageUrl = params;\n promptText = DEFAULT_IMAGE_DESCRIPTION_PROMPT;\n } else {\n imageUrl = params.imageUrl;\n promptText = params.prompt ?? DEFAULT_IMAGE_DESCRIPTION_PROMPT;\n }\n\n if (!imageUrl || imageUrl.trim().length === 0) {\n throw new Error(\"IMAGE_DESCRIPTION requires a valid image URL\");\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody = {\n model: modelName,\n messages: [\n {\n role: \"user\",\n content: [\n { type: \"text\", text: promptText },\n { type: \"image_url\", image_url: { url: imageUrl } },\n ],\n },\n ],\n max_tokens: maxTokens,\n };\n\n const response = await fetch(`${baseURL}/chat/completions`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI image description failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIChatCompletionResponse;\n\n if (data.usage) {\n emitModelUsageEvent(\n runtime,\n ModelType.IMAGE_DESCRIPTION,\n typeof params === \"string\" ? params : (params.prompt ?? \"\"),\n {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: data.usage.completion_tokens,\n totalTokens: data.usage.total_tokens,\n }\n );\n }\n\n const firstChoice = data.choices?.[0];\n const content = firstChoice?.message?.content;\n\n if (!content) {\n throw new Error(\"OpenAI API returned empty image description\");\n }\n\n return {\n title: parseTitleFromResponse(content),\n description: parseDescriptionFromResponse(content),\n };\n}\n",
|
|
13
13
|
"import type {\n IAgentRuntime,\n JsonValue,\n ModelTypeName,\n ObjectGenerationParams,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateObject, type LanguageModel } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport { getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\nimport { getJsonRepairFunction } from \"../utils/json\";\n\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\ntype ChatModelFactory = { chat: (modelName: string) => LanguageModel };\n\nasync function generateObjectByModelType(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<Record<string, JsonValue>> {\n const openai = createOpenAIClient(runtime) as ChatModelFactory;\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n\n if (!params.prompt || params.prompt.trim().length === 0) {\n throw new Error(\"Object generation requires a non-empty prompt\");\n }\n\n if (params.schema) {\n logger.debug(\n \"[OpenAI] Schema provided but using no-schema mode. \" +\n \"Structure is determined by prompt instructions.\"\n );\n }\n\n const model = openai.chat(modelName);\n const { object, usage } = await generateObject({\n model,\n output: \"no-schema\",\n prompt: params.prompt,\n experimental_repairText: getJsonRepairFunction(),\n });\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n if (typeof object !== \"object\" || object === null) {\n throw new Error(`Object generation returned ${typeof object}, expected object`);\n }\n\n return object as Record<string, JsonValue>;\n}\n\nexport async function handleObjectSmall(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n): Promise<Record<string, JsonValue>> {\n return generateObjectByModelType(runtime, params, ModelType.OBJECT_SMALL, getSmallModel);\n}\n\nexport async function handleObjectLarge(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n): Promise<Record<string, JsonValue>> {\n return generateObjectByModelType(runtime, params, ModelType.OBJECT_LARGE, getLargeModel);\n}\n",
|
|
14
14
|
"import { createOpenAI, type OpenAIProvider } from \"@ai-sdk/openai\";\nimport type { IAgentRuntime } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, isProxyMode } from \"../utils/config\";\n\nconst PROXY_API_KEY = \"sk-proxy\";\n\nexport function createOpenAIClient(runtime: IAgentRuntime): OpenAIProvider {\n const baseURL = getBaseURL(runtime);\n const apiKey = getApiKey(runtime);\n\n if (!apiKey && isProxyMode(runtime)) {\n return createOpenAI({\n apiKey: PROXY_API_KEY,\n baseURL,\n });\n }\n\n if (!apiKey) {\n throw new Error(\n \"OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.\"\n );\n }\n\n return createOpenAI({\n apiKey,\n baseURL,\n });\n}\n",
|
|
15
15
|
"import { logger } from \"@elizaos/core\";\nimport { JSONParseError } from \"ai\";\n\ninterface JsonRepairParams {\n text: string;\n error: Error;\n}\n\ntype JsonRepairFunction = (params: JsonRepairParams) => Promise<string | null>;\n\nconst JSON_CLEANUP_PATTERNS = {\n MARKDOWN_JSON: /```json\\n|\\n```|```/g,\n WHITESPACE: /^\\s+|\\s+$/g,\n} as const;\n\nexport function getJsonRepairFunction(): JsonRepairFunction {\n return async ({ text, error }: JsonRepairParams): Promise<string | null> => {\n if (!(error instanceof JSONParseError)) {\n return null;\n }\n try {\n const cleanedText = text.replace(JSON_CLEANUP_PATTERNS.MARKDOWN_JSON, \"\");\n JSON.parse(cleanedText);\n logger.debug(\"[JSON Repair] Successfully repaired JSON by removing markdown wrappers\");\n return cleanedText;\n } catch {\n logger.warn(\"[JSON Repair] Unable to repair JSON text\");\n return null;\n }\n };\n}\n\nexport function parseJsonWithRepair<T>(text: string): T {\n try {\n return JSON.parse(text) as T;\n } catch (firstError) {\n const cleanedText = text.replace(JSON_CLEANUP_PATTERNS.MARKDOWN_JSON, \"\");\n try {\n return JSON.parse(cleanedText) as T;\n } catch {\n const message = firstError instanceof Error ? firstError.message : String(firstError);\n throw new Error(`Failed to parse JSON: ${message}`);\n }\n }\n}\n\nexport function safeStringify(value: unknown, indent = 0): string {\n const seen = new WeakSet();\n\n const replacer = (_key: string, val: unknown): unknown => {\n if (typeof val === \"object\" && val !== null) {\n if (seen.has(val)) {\n return \"[Circular]\";\n }\n seen.add(val);\n }\n\n // Handle special types\n if (typeof val === \"bigint\") {\n return val.toString();\n }\n\n if (val instanceof Error) {\n return {\n name: val.name,\n message: val.message,\n stack: val.stack,\n };\n }\n\n if (val instanceof Date) {\n return val.toISOString();\n }\n\n if (val instanceof Map) {\n return Object.fromEntries(val);\n }\n\n if (val instanceof Set) {\n return Array.from(val);\n }\n\n return val;\n };\n\n return JSON.stringify(value, replacer, indent);\n}\n",
|
|
16
16
|
"/**\n * Deep Research model handler\n *\n * Provides deep research capabilities using OpenAI's o3-deep-research and o4-mini-deep-research models.\n * These models can find, analyze, and synthesize hundreds of sources to create comprehensive reports.\n *\n * @see https://platform.openai.com/docs/guides/deep-research\n */\n\nimport type {\n IAgentRuntime,\n JsonValue,\n ResearchAnnotation,\n ResearchCodeInterpreterCall,\n ResearchFileSearchCall,\n ResearchMcpToolCall,\n ResearchMessageOutput,\n ResearchOutputItem,\n ResearchParams,\n ResearchResult,\n ResearchTool,\n ResearchWebSearchCall,\n} from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, getResearchModel, getResearchTimeout } from \"../utils/config\";\n\n// ============================================================================\n// Types for OpenAI Responses API\n// ============================================================================\n\n/**\n * Tool configuration for the Responses API\n */\ninterface ResponsesApiTool {\n type: \"web_search_preview\" | \"file_search\" | \"code_interpreter\" | \"mcp\";\n vector_store_ids?: string[];\n container?: { type: \"auto\" };\n server_label?: string;\n server_url?: string;\n require_approval?: \"never\";\n}\n\n/**\n * Raw response from the OpenAI Responses API\n */\ninterface ResponsesApiResponse {\n id: string;\n object: string;\n status?: \"queued\" | \"in_progress\" | \"completed\" | \"failed\";\n output?: ResponsesApiOutputItem[];\n output_text?: string;\n error?: {\n message: string;\n code: string;\n };\n}\n\n/**\n * Raw output item from the Responses API\n */\ninterface ResponsesApiOutputItem {\n id?: string;\n type: string;\n status?: string;\n action?: {\n type: string;\n query?: string;\n url?: string;\n };\n query?: string;\n results?: Array<{\n file_id: string;\n file_name: string;\n score: number;\n }>;\n code?: string;\n output?: string;\n server_label?: string;\n tool_name?: string;\n arguments?: Record<string, unknown>;\n result?: unknown;\n content?: Array<{\n type: string;\n text: string;\n annotations?: Array<{\n url: string;\n title: string;\n start_index: number;\n end_index: number;\n }>;\n }>;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts ResearchTool params to Responses API tool format\n */\nfunction convertToolToApi(tool: ResearchTool): ResponsesApiTool {\n switch (tool.type) {\n case \"web_search_preview\":\n return { type: \"web_search_preview\" };\n case \"file_search\":\n return {\n type: \"file_search\",\n vector_store_ids: tool.vectorStoreIds,\n };\n case \"code_interpreter\":\n return {\n type: \"code_interpreter\",\n container: tool.container ?? { type: \"auto\" },\n };\n case \"mcp\":\n return {\n type: \"mcp\",\n server_label: tool.serverLabel,\n server_url: tool.serverUrl,\n require_approval: tool.requireApproval ?? \"never\",\n };\n default:\n throw new Error(`Unknown research tool type: ${(tool as ResearchTool).type}`);\n }\n}\n\n/**\n * Converts raw API output items to typed ResearchOutputItem\n */\nfunction convertOutputItem(item: ResponsesApiOutputItem): ResearchOutputItem | null {\n switch (item.type) {\n case \"web_search_call\":\n return {\n id: item.id ?? \"\",\n type: \"web_search_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n action: {\n type: (item.action?.type as \"search\" | \"open_page\" | \"find_in_page\") ?? \"search\",\n query: item.action?.query,\n url: item.action?.url,\n },\n } satisfies ResearchWebSearchCall;\n\n case \"file_search_call\":\n return {\n id: item.id ?? \"\",\n type: \"file_search_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n query: item.query ?? \"\",\n results: item.results?.map((r) => ({\n fileId: r.file_id,\n fileName: r.file_name,\n score: r.score,\n })),\n } satisfies ResearchFileSearchCall;\n\n case \"code_interpreter_call\":\n return {\n id: item.id ?? \"\",\n type: \"code_interpreter_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n code: item.code ?? \"\",\n output: item.output,\n } satisfies ResearchCodeInterpreterCall;\n\n case \"mcp_tool_call\":\n return {\n id: item.id ?? \"\",\n type: \"mcp_tool_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n serverLabel: item.server_label ?? \"\",\n toolName: item.tool_name ?? \"\",\n arguments: (item.arguments ?? {}) as Record<string, JsonValue>,\n result: item.result as JsonValue,\n } satisfies ResearchMcpToolCall;\n\n case \"message\":\n return {\n type: \"message\",\n content:\n item.content?.map((c) => ({\n type: \"output_text\" as const,\n text: c.text,\n annotations:\n c.annotations?.map((a) => ({\n url: a.url,\n title: a.title,\n startIndex: a.start_index,\n endIndex: a.end_index,\n })) ?? [],\n })) ?? [],\n } satisfies ResearchMessageOutput;\n\n default:\n // Unknown output type, skip\n return null;\n }\n}\n\n/**\n * Extracts text and annotations from the response\n */\nfunction extractTextAndAnnotations(response: ResponsesApiResponse): {\n text: string;\n annotations: ResearchAnnotation[];\n} {\n // Try output_text first (convenience field)\n if (response.output_text) {\n // Find annotations from message output items\n const annotations: ResearchAnnotation[] = [];\n if (response.output) {\n for (const item of response.output) {\n if (item.type === \"message\" && item.content) {\n for (const content of item.content) {\n if (content.annotations) {\n for (const ann of content.annotations) {\n annotations.push({\n url: ann.url,\n title: ann.title,\n startIndex: ann.start_index,\n endIndex: ann.end_index,\n });\n }\n }\n }\n }\n }\n }\n return { text: response.output_text, annotations };\n }\n\n // Fall back to extracting from message output items\n let text = \"\";\n const annotations: ResearchAnnotation[] = [];\n\n if (response.output) {\n for (const item of response.output) {\n if (item.type === \"message\" && item.content) {\n for (const content of item.content) {\n text += content.text;\n if (content.annotations) {\n for (const ann of content.annotations) {\n annotations.push({\n url: ann.url,\n title: ann.title,\n startIndex: ann.start_index,\n endIndex: ann.end_index,\n });\n }\n }\n }\n }\n }\n }\n\n return { text, annotations };\n}\n\n// ============================================================================\n// Main Handler\n// ============================================================================\n\n/**\n * Handles RESEARCH model requests using OpenAI's deep research models.\n *\n * Deep research models can take tens of minutes to complete tasks.\n * Use background mode for long-running tasks.\n *\n * @param runtime - The agent runtime\n * @param params - Research parameters\n * @returns Research result with text, annotations, and output items\n *\n * @example\n * ```typescript\n * const result = await handleResearch(runtime, {\n * input: \"Research the economic impact of AI on global labor markets\",\n * tools: [\n * { type: \"web_search_preview\" },\n * { type: \"code_interpreter\", container: { type: \"auto\" } }\n * ],\n * background: true,\n * });\n * console.log(result.text);\n * ```\n */\nexport async function handleResearch(\n runtime: IAgentRuntime,\n params: ResearchParams\n): Promise<ResearchResult> {\n const apiKey = getApiKey(runtime);\n if (!apiKey) {\n throw new Error(\n \"OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.\"\n );\n }\n\n const baseURL = getBaseURL(runtime);\n const modelName = params.model ?? getResearchModel(runtime);\n const timeout = getResearchTimeout(runtime);\n\n logger.debug(`[OpenAI] Starting deep research with model: ${modelName}`);\n logger.debug(`[OpenAI] Research input: ${params.input.substring(0, 100)}...`);\n\n // Validate that at least one data source tool is provided\n const dataSourceTools = params.tools?.filter(\n (t) => t.type === \"web_search_preview\" || t.type === \"file_search\" || t.type === \"mcp\"\n );\n\n if (!dataSourceTools || dataSourceTools.length === 0) {\n // Default to web search if no tools specified\n logger.debug(\"[OpenAI] No data source tools specified, defaulting to web_search_preview\");\n params.tools = [{ type: \"web_search_preview\" }, ...(params.tools ?? [])];\n }\n\n // Build the request body for the Responses API\n const requestBody: Record<string, unknown> = {\n model: modelName,\n input: params.input,\n };\n\n if (params.instructions) {\n requestBody.instructions = params.instructions;\n }\n\n if (params.background !== undefined) {\n requestBody.background = params.background;\n }\n\n if (params.tools && params.tools.length > 0) {\n requestBody.tools = params.tools.map(convertToolToApi);\n }\n\n if (params.maxToolCalls !== undefined) {\n requestBody.max_tool_calls = params.maxToolCalls;\n }\n\n if (params.reasoningSummary) {\n requestBody.reasoning = { summary: params.reasoningSummary };\n }\n\n logger.debug(`[OpenAI] Research request body: ${JSON.stringify(requestBody, null, 2)}`);\n\n // Make the API request\n const response = await fetch(`${baseURL}/responses`, {\n method: \"POST\",\n headers: {\n Authorization: `Bearer ${apiKey}`,\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n signal: AbortSignal.timeout(timeout),\n });\n\n if (!response.ok) {\n const errorText = await response.text();\n logger.error(`[OpenAI] Research request failed: ${response.status} ${errorText}`);\n throw new Error(`Deep research request failed: ${response.status} ${response.statusText}`);\n }\n\n const data = (await response.json()) as ResponsesApiResponse;\n\n if (data.error) {\n logger.error(`[OpenAI] Research API error: ${data.error.message}`);\n throw new Error(`Deep research error: ${data.error.message}`);\n }\n\n logger.debug(`[OpenAI] Research response received. Status: ${data.status ?? \"completed\"}`);\n\n // Extract text and annotations\n const { text, annotations } = extractTextAndAnnotations(data);\n\n // Convert output items\n const outputItems: ResearchOutputItem[] = [];\n if (data.output) {\n for (const item of data.output) {\n const converted = convertOutputItem(item);\n if (converted) {\n outputItems.push(converted);\n }\n }\n }\n\n const result: ResearchResult = {\n id: data.id,\n text,\n annotations,\n outputItems,\n status: data.status,\n };\n\n logger.info(\n `[OpenAI] Research completed. Text length: ${text.length}, Annotations: ${annotations.length}, Output items: ${outputItems.length}`\n );\n\n return result;\n}\n",
|
|
17
|
-
"/**\n * Text generation model handlers\n *\n * Provides text generation using OpenAI's language models.\n */\n\nimport type { GenerateTextParams, IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateText, type LanguageModelUsage, streamText } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport type { TextStreamResult, TokenUsage } from \"../types\";\nimport { getExperimentalTelemetry, getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n// ============================================================================\n// Types\n// ============================================================================\n\n/**\n * Function to get model name from runtime\n */\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts AI SDK usage to our token usage format\n */\nfunction convertUsage(usage: LanguageModelUsage | undefined): TokenUsage | undefined {\n if (!usage) {\n return undefined;\n }\n\n // The AI SDK uses inputTokens/outputTokens\n const promptTokens = usage.inputTokens ?? 0;\n const completionTokens = usage.outputTokens ?? 0;\n\n return {\n promptTokens,\n completionTokens,\n totalTokens: promptTokens + completionTokens,\n };\n}\n\n// ============================================================================\n// Core Generation Function\n// ============================================================================\n\n/**\n * Generates text using the specified model type.\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @param modelType - The type of model (TEXT_SMALL or TEXT_LARGE)\n * @param getModelFn - Function to get the model name\n * @returns Generated text or stream result\n */\nasync function generateTextByModelType(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<string | TextStreamResult> {\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n\n // Get system prompt from character if available\n const systemPrompt = runtime.character.system ?? undefined;\n\n // Use chat() instead of languageModel() to use the Chat Completions API\n // which has better compatibility than the Responses API\n // gpt-5 and gpt-5-mini (reasoning models) don't support temperature,\n // frequencyPenalty, presencePenalty, or stop parameters - use defaults only\n const model = openai.chat(modelName);\n const generateParams = {\n model,\n prompt: params.prompt,\n system: systemPrompt,\n maxOutputTokens: params.maxTokens ?? 8192,\n experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },\n };\n\n // Handle streaming mode\n if (params.stream) {\n const result = streamText(generateParams);\n\n return {\n textStream: result.textStream,\n text: Promise.resolve(result.text),\n usage: Promise.resolve(result.usage).then(convertUsage),\n finishReason: Promise.resolve(result.finishReason).then((r) => r as string | undefined),\n };\n }\n\n // Non-streaming mode\n const { text, usage } = await generateText(generateParams);\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n return text;\n}\n\n// ============================================================================\n// Public Handlers\n// ============================================================================\n\n/**\n * Handles TEXT_SMALL model requests.\n *\n * Uses the configured small model (default: gpt-5-mini).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextSmall(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_SMALL, getSmallModel);\n}\n\n/**\n * Handles TEXT_LARGE model requests.\n *\n * Uses the configured large model (default: gpt-5).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextLarge(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_LARGE, getLargeModel);\n}\n",
|
|
17
|
+
"/**\n * Text generation model handlers\n *\n * Provides text generation using OpenAI's language models.\n */\n\nimport type { GenerateTextParams, IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateText, type LanguageModelUsage, streamText } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport type { TextStreamResult, TokenUsage } from \"../types\";\nimport { getExperimentalTelemetry, getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n// ============================================================================\n// Types\n// ============================================================================\n\n/**\n * Function to get model name from runtime\n */\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\n\ntype PromptCacheRetention = \"in_memory\" | \"24h\";\n\ninterface OpenAIPromptCacheOptions {\n promptCacheKey?: string;\n promptCacheRetention?: PromptCacheRetention;\n}\n\ninterface GenerateTextParamsWithOpenAIOptions extends GenerateTextParams {\n providerOptions?: {\n openai?: OpenAIPromptCacheOptions;\n };\n}\n\ninterface LanguageModelUsageWithCache extends LanguageModelUsage {\n cachedInputTokens?: number;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts AI SDK usage to our token usage format\n */\nfunction convertUsage(usage: LanguageModelUsage | undefined): TokenUsage | undefined {\n if (!usage) {\n return undefined;\n }\n\n // The AI SDK uses inputTokens/outputTokens\n const promptTokens = usage.inputTokens ?? 0;\n const completionTokens = usage.outputTokens ?? 0;\n const usageWithCache = usage as LanguageModelUsageWithCache;\n\n return {\n promptTokens,\n completionTokens,\n totalTokens: promptTokens + completionTokens,\n cachedPromptTokens: usageWithCache.cachedInputTokens,\n };\n}\n\nfunction resolvePromptCacheOptions(params: GenerateTextParams): OpenAIPromptCacheOptions {\n const withOpenAIOptions = params as GenerateTextParamsWithOpenAIOptions;\n return {\n promptCacheKey: withOpenAIOptions.providerOptions?.openai?.promptCacheKey,\n promptCacheRetention: withOpenAIOptions.providerOptions?.openai?.promptCacheRetention,\n };\n}\n\n// ============================================================================\n// Core Generation Function\n// ============================================================================\n\n/**\n * Generates text using the specified model type.\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @param modelType - The type of model (TEXT_SMALL or TEXT_LARGE)\n * @param getModelFn - Function to get the model name\n * @returns Generated text or stream result\n */\nasync function generateTextByModelType(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<string | TextStreamResult> {\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n const promptCacheOptions = resolvePromptCacheOptions(params);\n\n // Get system prompt from character if available\n const systemPrompt = runtime.character.system ?? undefined;\n\n // Use chat() instead of languageModel() to use the Chat Completions API\n // which has better compatibility than the Responses API\n // gpt-5 and gpt-5-mini (reasoning models) don't support temperature,\n // frequencyPenalty, presencePenalty, or stop parameters - use defaults only\n const model = openai.chat(modelName);\n const generateParams = {\n model,\n prompt: params.prompt,\n system: systemPrompt,\n maxOutputTokens: params.maxTokens ?? 8192,\n experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },\n ...(promptCacheOptions.promptCacheKey || promptCacheOptions.promptCacheRetention\n ? {\n providerOptions: {\n openai: {\n ...(promptCacheOptions.promptCacheKey\n ? { promptCacheKey: promptCacheOptions.promptCacheKey }\n : {}),\n ...(promptCacheOptions.promptCacheRetention\n ? { promptCacheRetention: promptCacheOptions.promptCacheRetention }\n : {}),\n },\n },\n }\n : {}),\n };\n\n // Handle streaming mode\n if (params.stream) {\n const result = streamText(generateParams);\n\n return {\n textStream: result.textStream,\n text: Promise.resolve(result.text),\n usage: Promise.resolve(result.usage).then(convertUsage),\n finishReason: Promise.resolve(result.finishReason).then((r) => r as string | undefined),\n };\n }\n\n // Non-streaming mode\n const { text, usage } = await generateText(generateParams);\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n return text;\n}\n\n// ============================================================================\n// Public Handlers\n// ============================================================================\n\n/**\n * Handles TEXT_SMALL model requests.\n *\n * Uses the configured small model (default: gpt-5-mini).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextSmall(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_SMALL, getSmallModel);\n}\n\n/**\n * Handles TEXT_LARGE model requests.\n *\n * Uses the configured large model (default: gpt-5).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextLarge(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_LARGE, getLargeModel);\n}\n",
|
|
18
18
|
"import type { DetokenizeTextParams, IAgentRuntime, TokenizeTextParams } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport { detokenizeText, tokenizeText } from \"../utils/tokenization\";\n\nexport async function handleTokenizerEncode(\n runtime: IAgentRuntime,\n params: TokenizeTextParams\n): Promise<number[]> {\n if (!params.prompt) {\n throw new Error(\"Tokenization requires a non-empty prompt\");\n }\n const modelType = params.modelType ?? ModelType.TEXT_LARGE;\n return tokenizeText(runtime, modelType, params.prompt);\n}\n\nexport async function handleTokenizerDecode(\n runtime: IAgentRuntime,\n params: DetokenizeTextParams\n): Promise<string> {\n if (!params.tokens || !Array.isArray(params.tokens)) {\n throw new Error(\"Detokenization requires a valid tokens array\");\n }\n if (params.tokens.length === 0) {\n return \"\";\n }\n for (let i = 0; i < params.tokens.length; i++) {\n const token = params.tokens[i];\n if (typeof token !== \"number\" || !Number.isFinite(token)) {\n throw new Error(`Invalid token at index ${i}: expected number`);\n }\n }\n const modelType = params.modelType ?? ModelType.TEXT_LARGE;\n return detokenizeText(runtime, modelType, params.tokens);\n}\n",
|
|
19
19
|
"import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport {\n encodingForModel,\n getEncoding,\n type Tiktoken,\n type TiktokenEncoding,\n type TiktokenModel,\n} from \"js-tiktoken\";\nimport { getLargeModel, getSmallModel } from \"./config\";\n\ntype SupportedEncoding = \"cl100k_base\" | \"o200k_base\";\n\nfunction resolveTokenizerEncoding(modelName: string): Tiktoken {\n const normalized = modelName.toLowerCase();\n const fallbackEncoding: SupportedEncoding = normalized.includes(\"4o\")\n ? \"o200k_base\"\n : \"cl100k_base\";\n try {\n return encodingForModel(modelName as TiktokenModel);\n } catch {\n return getEncoding(fallbackEncoding as TiktokenEncoding);\n }\n}\n\nfunction getModelName(runtime: IAgentRuntime, modelType: ModelTypeName): string {\n if (modelType === ModelType.TEXT_SMALL) {\n return getSmallModel(runtime);\n }\n return getLargeModel(runtime);\n}\n\nexport function tokenizeText(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string\n): number[] {\n const modelName = getModelName(runtime, modelType);\n const encoder = resolveTokenizerEncoding(modelName);\n return encoder.encode(text);\n}\n\nexport function detokenizeText(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n tokens: number[]\n): string {\n const modelName = getModelName(runtime, modelType);\n const encoder = resolveTokenizerEncoding(modelName);\n return encoder.decode(tokens);\n}\n\nexport function countTokens(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string\n): number {\n const tokens = tokenizeText(runtime, modelType, text);\n return tokens.length;\n}\n\nexport function truncateToTokenLimit(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string,\n maxTokens: number\n): string {\n const tokens = tokenizeText(runtime, modelType, text);\n if (tokens.length <= maxTokens) {\n return text;\n }\n const truncatedTokens = tokens.slice(0, maxTokens);\n return detokenizeText(runtime, modelType, truncatedTokens);\n}\n"
|
|
20
20
|
],
|
|
21
|
-
"mappings": "AAgBA,iBAAS,eAAQ,sBCfjB,iBAAS,sBCAT,iBAAS,sBAET,SAAS,EAAW,CAAC,EAAiC,CACpD,GAAI,OAAO,QAAY,KAAe,CAAC,QAAQ,IAC7C,OAEF,IAAM,EAAQ,QAAQ,IAAI,GAC1B,OAAO,IAAU,OAAY,OAAY,OAAO,CAAK,EAGhD,SAAS,CAAU,CACxB,EACA,EACA,EACoB,CACpB,IAAM,EAAQ,EAAQ,WAAW,CAAG,EACpC,GAAI,IAAU,QAAa,IAAU,KACnC,OAAO,OAAO,CAAK,EAErB,OAAO,GAAY,CAAG,GAAK,EActB,SAAS,CAAiB,CAC/B,EACA,EACA,EACQ,CACR,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAS,OAAO,SAAS,EAAO,EAAE,EACxC,GAAI,CAAC,OAAO,SAAS,CAAM,EACzB,MAAU,MAAM,YAAY,oCAAsC,GAAO,EAE3E,OAAO,EAGF,SAAS,CAAiB,CAC/B,EACA,EACA,EACS,CACT,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAa,EAAM,YAAY,EACrC,OAAO,IAAe,QAAU,IAAe,KAAO,IAAe,MAGhE,SAAS,CAAS,EAAY,CACnC,OACE,OAAO,WAAe,KACtB,OAAQ,WAAuC,SAAa,IAIzD,SAAS,CAAW,CAAC,EAAiC,CAC3D,OAAO,EAAU,GAAK,CAAC,CAAC,EAAW,EAAS,yBAAyB,EAGhE,SAAS,CAAS,CAAC,EAA4C,CACpE,OAAO,EAAW,EAAS,gBAAgB,EAGtC,SAAS,EAAkB,CAAC,EAA4C,CAC7E,IAAM,EAAkB,EAAW,EAAS,0BAA0B,EACtE,GAAI,EAEF,OADA,EAAO,MAAM,2CAA2C,EACjD,EAGT,OADA,EAAO,MAAM,yDAAyD,EAC/D,EAAU,CAAO,EAGnB,SAAS,CAAa,CAC3B,EACA,EAAe,GACS,CAIxB,GAAI,EAAU,GAAK,CAAC,EAAkB,EAAS,+BAAgC,EAAK,EAClF,MAAO,CAAC,EAEV,IAAM,EAAM,EAAe,GAAmB,CAAO,EAAI,EAAU,CAAO,EAC1E,OAAO,EAAM,CAAE,cAAe,UAAU,GAAM,EAAI,CAAC,EAG9C,SAAS,CAAU,CAAC,EAAgC,CACzD,IAAM,EAAa,EAAW,EAAS,yBAAyB,EAC1D,EACJ,EAAU,GAAK,EACX,EACC,EAAW,EAAS,iBAAiB,GAAK,4BAEjD,OADA,EAAO,MAAM,sBAAsB,GAAS,EACrC,EAGF,SAAS,CAAmB,CAAC,EAAgC,CAClE,IAAM,EAAe,EAAU,EAC1B,EAAW,EAAS,8BAA8B,GACnD,EAAW,EAAS,yBAAyB,EAC7C,EAAW,EAAS,sBAAsB,EAE9C,GAAI,EAEF,OADA,EAAO,MAAM,sCAAsC,GAAc,EAC1D,EAIT,OADA,EAAO,MAAM,0DAA0D,EAChE,EAAW,CAAO,EAGpB,SAAS,CAAa,CAAC,EAAgC,CAC5D,OACE,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,aAIhF,SAAS,CAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,QAGrF,SAAS,CAAiB,CAAC,EAAgC,CAChE,OAAO,EAAW,EAAS,wBAAwB,GAAK,yBAGnD,SAAS,EAAwB,CAAC,EAAgC,CACvE,OAAO,EAAW,EAAS,gCAAgC,GAAK,aAG3D,SAAS,EAAqB,CAAC,EAAgC,CACpE,OAAO,EAAW,EAAS,4BAA4B,GAAK,wBAGvD,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,QAG7C,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,OAG7C,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAW,EAAS,yBAAyB,GAAK,GAGpD,SAAS,EAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,WAG/C,SAAS,EAAwB,CAAC,EAAiC,CACxE,OAAO,EAAkB,EAAS,gCAAiC,EAAK,EAGnE,SAAS,EAAsB,CAAC,EAAgC,CACrE,OAAO,EAAkB,EAAS,8BAA+B,IAAI,EAGhE,SAAS,EAA4B,CAAC,EAAgC,CAC3E,OAAO,EAAkB,EAAS,sCAAuC,IAAI,EAGxE,SAAS,EAAgB,CAAC,EAAgC,CAC/D,OAAO,EAAW,EAAS,uBAAuB,GAAK,mBAGlD,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAkB,EAAS,0BAA2B,OAAO,EDjLrE,WAAuC,sBAAwB,GAEzD,SAAS,EAAgB,CAC9B,EACA,EACM,CACD,GAA4B,CAAO,EAG1C,eAAe,EAA2B,CAAC,EAAuC,CAChF,GAAI,EAAU,EAAG,CACf,EAAO,MAAM,yDAAyD,EACtE,OAKF,GAAI,CAFW,EAAU,CAAO,EAEnB,CACX,EAAO,KACL,8GAEF,EACA,OAGF,GAAI,CACF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,EAAO,KACL,uCAAuC,EAAS,UAAU,EAAS,2DAErE,EACA,QAEF,MAAO,EAAO,CACd,IAAM,EAAU,aAAiB,MAAQ,EAAM,QAAU,OAAO,CAAK,EACrE,EAAO,KAAK,kCAAkC,yCAA+C,GExCjG,iBAAS,sBCLT,iBAAS,uBAET,IAAM,EAAc,CAClB,IAAK,CACH,OAAQ,CAAC,GAAM,GAAM,GAAM,EAAI,EAC/B,WAAY,CAAC,GAAM,GAAM,GAAM,EAAI,CACrC,EACA,QAAS,CAAC,GAAM,GAAM,EAAI,EAC1B,IAAK,CAAC,GAAM,IAAM,IAAM,EAAI,EAC5B,KAAM,CAAC,IAAM,GAAM,GAAM,EAAI,EAC7B,KAAM,CAAC,IAAM,IAAM,IAAM,GAAI,EAC7B,UAAW,CAAC,GAAM,GAAM,IAAM,GAAI,CACpC,EAEM,GAA4B,GAWlC,SAAS,CAAU,CAAC,EAAgB,EAAgB,EAAsC,CACxF,QAAS,EAAI,EAAG,EAAI,EAAS,OAAQ,IAAK,CACxC,IAAM,EAAe,EAAS,GAC9B,GAAI,IAAiB,QAAa,EAAO,EAAS,KAAO,EACvD,MAAO,GAGX,MAAO,GAGF,SAAS,CAAmB,CAAC,EAA+B,CACjE,GAAI,EAAO,OAAS,GAClB,MAAO,2BAIT,GACE,EAAW,EAAQ,EAAG,EAAY,IAAI,MAAM,GAC5C,EAAW,EAAQ,EAAG,EAAY,IAAI,UAAU,EAEhD,MAAO,YAIT,IAAM,EAAY,EAAO,GACnB,EAAa,EAAO,GAC1B,GACE,EAAW,EAAQ,EAAG,EAAY,OAAO,GACxC,IAAc,KAAQ,IAAe,SAAc,EAAa,OAAU,IAE3E,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,GAAG,EACvC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,SAAS,EAC7C,MAAO,aAIT,OADA,GAAO,KAAK,sEAAsE,EAC3E,2BAGF,SAAS,EAAuB,CAAC,EAAiC,CACvE,OAAQ,OACD,YACH,MAAO,UACJ,aACH,MAAO,UACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,2BACH,MAAO,OAIN,SAAS,EAAsB,CAAC,EAAiC,CAEtE,MAAO,aADK,GAAwB,CAAQ,ID3E9C,SAAS,EAAY,CAAC,EAAsC,CAC1D,OAAO,aAAiB,MAAQ,aAAiB,KAGnD,SAAS,CAAQ,CAAC,EAAiC,CACjD,OAAO,OAAO,SAAS,CAAK,EAG9B,SAAS,EAA0B,CAAC,EAAmD,CACrF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,UAAW,IACV,GAAc,EAAmC,KAAK,GACrD,EAAU,EAAmC,KAAK,GAIxD,SAAS,EAAyB,CAAC,EAAkD,CACnF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,aAAc,GACd,OAAQ,EAAkC,WAAa,SAI3D,eAAe,EAAiB,CAAC,EAA4B,CAC3D,IAAM,EAAW,MAAM,MAAM,CAAG,EAChC,GAAI,CAAC,EAAS,GACZ,MAAU,MAAM,mCAAmC,EAAS,QAAQ,EAEtE,OAAO,EAAS,KAAK,EAEvB,eAAsB,CAAmB,CACvC,EACA,EACiB,CACjB,IAAI,EAAY,GAAsB,CAAO,EACzC,EACA,EAAiD,CAAC,EAEtD,GAAI,OAAO,IAAU,SACnB,EAAO,MAAM,qCAAqC,GAAO,EACzD,EAAO,MAAM,GAAkB,CAAK,EAC/B,QAAI,GAAa,CAAK,EAC3B,EAAO,EACF,QAAI,EAAS,CAAK,EAAG,CAC1B,IAAM,EAAW,EAAoB,CAAK,EAC1C,EAAO,MAAM,2CAA2C,GAAU,EAClE,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,CAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EACtD,QAAI,GAA2B,CAAK,EAAG,CAE5C,GADA,EAAc,EACV,EAAM,MACR,EAAY,EAAM,MAEpB,GAAI,EAAS,EAAM,KAAK,EAAG,CACzB,IAAM,EAAW,EAAM,UAAY,EAAoB,EAAM,KAAK,EAClE,EAAO,MAAM,6BAA6B,GAAU,EACpD,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,EAAM,KAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EAEjE,OAAO,EAAM,MAEV,QAAI,GAA0B,CAAK,EACxC,EAAO,MAAM,qCAAqC,EAAM,UAAU,EAClE,EAAO,MAAM,GAAkB,EAAM,QAAQ,EAC7C,EAAc,CAAE,OAAQ,EAAM,MAAO,EAErC,WAAU,MACR,qFACF,EAGF,EAAO,MAAM,uCAAuC,GAAW,EAE/D,IAAM,EAAY,EAAc,MAAQ,aAClC,EACH,EAAc,MACf,GACE,EAAS,WAAW,QAAQ,EACvB,EACD,YACN,EAEI,EAAW,IAAI,SAIrB,GAHA,EAAS,OAAO,OAAQ,EAAM,CAAQ,EACtC,EAAS,OAAO,QAAS,CAAS,EAE9B,EAAY,SACd,EAAS,OAAO,WAAY,EAAY,QAAQ,EAElD,GAAI,EAAY,eACd,EAAS,OAAO,kBAAmB,EAAY,cAAc,EAE/D,GAAI,EAAY,OACd,EAAS,OAAO,SAAU,EAAY,MAAM,EAE9C,GAAI,EAAY,cAAgB,OAC9B,EAAS,OAAO,cAAe,OAAO,EAAY,WAAW,CAAC,EAEhE,GAAI,EAAY,uBACd,QAAW,KAAe,EAAY,uBACpC,EAAS,OAAO,4BAA6B,CAAW,EAI5D,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,yBAAgC,CAC9D,OAAQ,OACR,QAAS,EAAc,CAAO,EAC9B,KAAM,CACR,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,gCAAgC,EAAS,UAAU,EAAS,gBAAgB,GAC9E,EAIF,OADc,MAAM,EAAS,KAAK,GACtB,KAGd,eAAsB,CAAkB,CACtC,EACA,EACsB,CACtB,IAAI,EACA,EACA,EAA0B,MAC1B,EACA,EAEJ,GAAI,OAAO,IAAU,SACnB,EAAO,EACP,EAAQ,OACH,KAGL,GAFA,EAAO,EAAM,KACb,EAAQ,EAAM,MACV,WAAY,GAAS,EAAM,OAC7B,EAAS,EAAM,OAEjB,GAAI,UAAW,GAAS,EAAM,MAC5B,EAAQ,EAAM,MAEhB,GAAI,iBAAkB,GAAS,EAAM,aACnC,EAAe,EAAM,aAUzB,GANA,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAe,GAAgB,GAAmB,CAAO,EAEzD,EAAO,MAAM,wCAAwC,GAAO,EAExD,CAAC,GAAQ,EAAK,KAAK,EAAE,SAAW,EAClC,MAAU,MAAM,wCAAwC,EAG1D,GAAI,EAAK,OAAS,KAChB,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAA0B,CAAC,QAAS,OAAQ,QAAS,OAAQ,OAAQ,SAAS,EACpF,GAAI,GAAS,CAAC,EAAY,SAAS,CAAiB,EAClD,MAAU,MAAM,kBAAkB,sBAA0B,EAAY,KAAK,IAAI,GAAG,EAGtF,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAsC,CAC1C,QACA,MAAO,EACP,MAAO,EACP,gBAAiB,CACnB,EAEA,GAAI,GAAgB,EAAa,OAAS,EACxC,EAAY,aAAe,EAG7B,IAAM,EAAW,MAAM,MAAM,GAAG,iBAAwB,CACtD,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,sBACZ,IAAW,MAAQ,CAAE,OAAQ,YAAa,EAAI,CAAC,CACrD,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MAAM,sBAAsB,EAAS,UAAU,EAAS,gBAAgB,GAAW,EAG/F,OAAO,EAAS,YAAY,EEhO9B,iBAAS,eAAQ,kBAAW,uBCA5B,oBAAS,uBAGT,IAAM,GAAoB,IA4B1B,SAAS,EAAc,CAAC,EAAwB,CAC9C,GAAI,EAAO,QAAU,GACnB,OAAO,EAET,MAAO,GAAG,EAAO,MAAM,EAAG,EAAiB,KAG7C,SAAS,EAAc,CAAC,EAA+B,CACrD,GAAI,iBAAkB,EACpB,MAAO,CACL,aAAc,EAAM,cAAgB,EACpC,iBAAkB,EAAM,kBAAoB,EAC5C,YAAa,EAAM,cAAgB,EAAM,cAAgB,IAAM,EAAM,kBAAoB,EAC3F,EAEF,GAAI,gBAAiB,GAAS,iBAAkB,EAAO,CACrD,IAAM,EAAS,EAAqB,aAAe,EAC7C,EAAU,EAAqB,cAAgB,EACrD,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,EAAQ,CACvB,EAEF,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,CACf,EAGK,SAAS,CAAmB,CACjC,EACA,EACA,EACA,EACM,CACN,IAAM,EAAa,GAAe,CAAK,EAEjC,EAAkC,CACtC,UACA,OAAQ,SACR,SAAU,SACV,OACA,OAAQ,GAAe,CAAM,EAC7B,OAAQ,CACN,OAAQ,EAAW,aACnB,WAAY,EAAW,iBACvB,MAAO,EAAW,WACpB,CACF,EAEA,EAAQ,UAAU,GAAU,WAAY,CAAO,ED9EjD,IAAM,GAAuB,KAY7B,SAAS,EAAiB,CAAC,EAAoC,CAC7D,IAAM,EAAkB,OAAO,OAAO,EAAW,EACjD,GAAI,CAAC,EAAgB,SAAS,CAAS,EACrC,MAAU,MACR,gCAAgC,sBAA8B,EAAgB,KAAK,IAAI,GACzF,EAEF,OAAO,EAGT,SAAS,EAAW,CAAC,EAA4D,CAC/E,GAAI,IAAW,KACb,OAAO,KAET,GAAI,OAAO,IAAW,SACpB,OAAO,EAET,GAAI,OAAO,IAAW,UAAY,OAAO,EAAO,OAAS,SACvD,OAAO,EAAO,KAEhB,MAAU,MAAM,sEAAsE,EAGxF,eAAsB,CAAmB,CACvC,EACA,EACmB,CACnB,IAAM,EAAiB,EAAkB,CAAO,EAC1C,EAAqB,GAAkB,GAAuB,CAAO,CAAC,EAEtE,EAAO,GAAY,CAAM,EAC/B,GAAI,IAAS,KAAM,CACjB,EAAO,MAAM,qDAAqD,EAClE,IAAM,EAAiB,MAAM,CAAkB,EAAE,KAAK,CAAC,EAEvD,OADA,EAAW,GAAK,IACT,EAGT,IAAI,EAAc,EAAK,KAAK,EAC5B,GAAI,EAAY,SAAW,EACzB,MAAU,MAAM,0CAA0C,EAM5D,IAAM,EAAW,GAAuB,EACxC,GAAI,EAAY,OAAS,EACvB,EAAO,KACL,uCAAuC,KAAK,KAAK,EAAY,OAAS,CAAC,6BAA6B,WACtG,EACA,EAAc,EAAY,MAAM,EAAG,CAAQ,EAI7C,IAAM,EAAM,GADI,EAAoB,CAAO,eAG3C,EAAO,MAAM,6CAA6C,GAAgB,EAE1E,IAAM,EAAW,MAAM,MAAM,EAAK,CAChC,OAAQ,OACR,QAAS,IACJ,EAAc,EAAS,EAAI,EAC9B,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CACnB,MAAO,EACP,MAAO,CACT,CAAC,CACH,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,+BAA+B,EAAS,UAAU,EAAS,gBAAgB,GAC7E,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAE5B,EAAc,GAAM,OAAO,GACjC,GAAI,CAAC,GAAe,CAAC,EAAY,UAC/B,MAAU,MAAM,0DAA0D,EAG5E,IAAM,EAAY,EAAY,UAE9B,GAAI,EAAU,SAAW,EACvB,MAAU,MACR,qCAAqC,EAAU,oBAAoB,+CAErE,EAGF,GAAI,EAAK,MACP,EAAoB,EAAS,GAAU,eAAgB,EAAa,CAClE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAClB,YAAa,EAAK,MAAM,YAC1B,CAAC,EAIH,OADA,EAAO,MAAM,qCAAqC,EAAU,mBAAmB,EACxE,EExHT,iBAAS,gBAAQ,uBAwBjB,IAAM,GACJ,0EAEF,eAAsB,CAAqB,CACzC,EACA,EACkC,CAClC,IAAM,EAAY,GAAc,CAAO,EACjC,EAAQ,EAAO,OAAS,EACxB,EAAmB,EAAO,MAAsB,YAChD,EAAiB,EAIvB,GAFA,GAAO,MAAM,+BAA+B,GAAW,EAEnD,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,8CAA8C,EAGhE,GAAI,EAAQ,GAAK,EAAQ,GACvB,MAAU,MAAM,sCAAsC,EAGxD,IAAM,EAAU,EAAW,CAAO,EAE5B,EAA+C,CACnD,MAAO,EACP,OAAQ,EAAO,OACf,EAAG,EACH,MACF,EAEA,GAAI,EAAe,QACjB,EAAY,QAAU,EAAe,QAEvC,GAAI,EAAe,MACjB,EAAY,MAAQ,EAAe,MAGrC,IAAM,EAAW,MAAM,MAAM,GAAG,uBAA8B,CAC5D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,mCAAmC,EAAS,UAAU,EAAS,gBAAgB,GACjF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,CAAC,EAAK,MAAQ,EAAK,KAAK,SAAW,EACrC,MAAU,MAAM,+BAA+B,EAGjD,OAAO,EAAK,KAAK,IAAI,CAAC,KAAU,CAC9B,IAAK,EAAK,IACV,cAAe,EAAK,cACtB,EAAE,EAGJ,SAAS,EAAsB,CAAC,EAAyB,CAEvD,OADmB,EAAQ,MAAM,2BAA2B,IACxC,IAAI,KAAK,GAAK,iBAGpC,SAAS,EAA4B,CAAC,EAAyB,CAC7D,OAAO,EAAQ,QAAQ,4BAA6B,EAAE,EAAE,KAAK,EAG/D,eAAsB,CAAsB,CAC1C,EACA,EACiC,CACjC,IAAM,EAAY,GAAyB,CAAO,EAC5C,EAAY,GAA6B,CAAO,EAEtD,GAAO,MAAM,2CAA2C,GAAW,EAEnE,IAAI,EACA,EAEJ,GAAI,OAAO,IAAW,SACpB,EAAW,EACX,EAAa,GAEb,OAAW,EAAO,SAClB,EAAa,EAAO,QAAU,GAGhC,GAAI,CAAC,GAAY,EAAS,KAAK,EAAE,SAAW,EAC1C,MAAU,MAAM,8CAA8C,EAGhE,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAc,CAClB,MAAO,EACP,SAAU,CACR,CACE,KAAM,OACN,QAAS,CACP,CAAE,KAAM,OAAQ,KAAM,CAAW,EACjC,CAAE,KAAM,YAAa,UAAW,CAAE,IAAK,CAAS,CAAE,CACpD,CACF,CACF,EACA,WAAY,CACd,EAEM,EAAW,MAAM,MAAM,GAAG,qBAA4B,CAC1D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,oCAAoC,EAAS,UAAU,EAAS,gBAAgB,GAClF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MACP,EACE,EACA,GAAU,kBACV,OAAO,IAAW,SAAW,EAAU,EAAO,QAAU,GACxD,CACE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAAK,MAAM,kBAC7B,YAAa,EAAK,MAAM,YAC1B,CACF,EAIF,IAAM,EADc,EAAK,UAAU,IACN,SAAS,QAEtC,GAAI,CAAC,EACH,MAAU,MAAM,6CAA6C,EAG/D,MAAO,CACL,MAAO,GAAuB,CAAO,EACrC,YAAa,GAA6B,CAAO,CACnD,EC/KF,iBAAS,gBAAQ,uBACjB,yBAAS,YCPT,uBAAS,wBAIT,IAAM,GAAgB,WAEf,SAAS,CAAkB,CAAC,EAAwC,CACzE,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAS,EAAU,CAAO,EAEhC,GAAI,CAAC,GAAU,EAAY,CAAO,EAChC,OAAO,GAAa,CAClB,OAAQ,GACR,SACF,CAAC,EAGH,GAAI,CAAC,EACH,MAAU,MACR,uFACF,EAGF,OAAO,GAAa,CAClB,SACA,SACF,CAAC,EC1BH,iBAAS,uBACT,yBAAS,YAST,IAAM,GAAwB,CAC5B,cAAe,uBACf,WAAY,YACd,EAEO,SAAS,EAAqB,EAAuB,CAC1D,MAAO,QAAS,OAAM,WAAsD,CAC1E,GAAI,EAAE,aAAiB,IACrB,OAAO,KAET,GAAI,CACF,IAAM,EAAc,EAAK,QAAQ,GAAsB,cAAe,EAAE,EAGxE,OAFA,KAAK,MAAM,CAAW,EACtB,GAAO,MAAM,wEAAwE,EAC9E,EACP,KAAM,CAEN,OADA,GAAO,KAAK,0CAA0C,EAC/C,OFXb,eAAe,EAAyB,CACtC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAIpC,GAFA,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAE1D,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,+CAA+C,EAGjE,GAAI,EAAO,OACT,GAAO,MACL,oGAEF,EAGF,IAAM,EAAQ,EAAO,KAAK,CAAS,GAC3B,SAAQ,SAAU,MAAM,GAAe,CAC7C,QACA,OAAQ,YACR,OAAQ,EAAO,OACf,wBAAyB,GAAsB,CACjD,CAAC,EAED,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,GAAI,OAAO,IAAW,UAAY,IAAW,KAC3C,MAAU,MAAM,8BAA8B,OAAO,oBAAyB,EAGhF,OAAO,EAGT,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EAGzF,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EG7CzF,iBAAS,sBA6ET,SAAS,EAAgB,CAAC,EAAsC,CAC9D,OAAQ,EAAK,UACN,qBACH,MAAO,CAAE,KAAM,oBAAqB,MACjC,cACH,MAAO,CACL,KAAM,cACN,iBAAkB,EAAK,cACzB,MACG,mBACH,MAAO,CACL,KAAM,mBACN,UAAW,EAAK,WAAa,CAAE,KAAM,MAAO,CAC9C,MACG,MACH,MAAO,CACL,KAAM,MACN,aAAc,EAAK,YACnB,WAAY,EAAK,UACjB,iBAAkB,EAAK,iBAAmB,OAC5C,UAEA,MAAU,MAAM,+BAAgC,EAAsB,MAAM,GAOlF,SAAS,EAAiB,CAAC,EAAyD,CAClF,OAAQ,EAAK,UACN,kBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,kBACN,OAAS,EAAK,QAAqC,YACnD,OAAQ,CACN,KAAO,EAAK,QAAQ,MAAoD,SACxE,MAAO,EAAK,QAAQ,MACpB,IAAK,EAAK,QAAQ,GACpB,CACF,MAEG,mBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,mBACN,OAAS,EAAK,QAAqC,YACnD,MAAO,EAAK,OAAS,GACrB,QAAS,EAAK,SAAS,IAAI,CAAC,KAAO,CACjC,OAAQ,EAAE,QACV,SAAU,EAAE,UACZ,MAAO,EAAE,KACX,EAAE,CACJ,MAEG,wBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,wBACN,OAAS,EAAK,QAAqC,YACnD,KAAM,EAAK,MAAQ,GACnB,OAAQ,EAAK,MACf,MAEG,gBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,gBACN,OAAS,EAAK,QAAqC,YACnD,YAAa,EAAK,cAAgB,GAClC,SAAU,EAAK,WAAa,GAC5B,UAAY,EAAK,WAAa,CAAC,EAC/B,OAAQ,EAAK,MACf,MAEG,UACH,MAAO,CACL,KAAM,UACN,QACE,EAAK,SAAS,IAAI,CAAC,KAAO,CACxB,KAAM,cACN,KAAM,EAAE,KACR,YACE,EAAE,aAAa,IAAI,CAAC,KAAO,CACzB,IAAK,EAAE,IACP,MAAO,EAAE,MACT,WAAY,EAAE,YACd,SAAU,EAAE,SACd,EAAE,GAAK,CAAC,CACZ,EAAE,GAAK,CAAC,CACZ,UAIA,OAAO,MAOb,SAAS,EAAyB,CAAC,EAGjC,CAEA,GAAI,EAAS,YAAa,CAExB,IAAM,EAAoC,CAAC,EAC3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QACzB,GAAI,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAOb,MAAO,CAAE,KAAM,EAAS,YAAa,aAAY,EAInD,IAAI,EAAO,GACL,EAAoC,CAAC,EAE3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QAEzB,GADA,GAAQ,EAAQ,KACZ,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAQb,MAAO,CAAE,OAAM,aAAY,EA8B7B,eAAsB,CAAc,CAClC,EACA,EACyB,CACzB,IAAM,EAAS,EAAU,CAAO,EAChC,GAAI,CAAC,EACH,MAAU,MACR,yGACF,EAGF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAY,EAAO,OAAS,GAAiB,CAAO,EACpD,EAAU,GAAmB,CAAO,EAE1C,EAAO,MAAM,+CAA+C,GAAW,EACvE,EAAO,MAAM,4BAA4B,EAAO,MAAM,UAAU,EAAG,GAAG,MAAM,EAG5E,IAAM,EAAkB,EAAO,OAAO,OACpC,CAAC,IAAM,EAAE,OAAS,sBAAwB,EAAE,OAAS,eAAiB,EAAE,OAAS,KACnF,EAEA,GAAI,CAAC,GAAmB,EAAgB,SAAW,EAEjD,EAAO,MAAM,2EAA2E,EACxF,EAAO,MAAQ,CAAC,CAAE,KAAM,oBAAqB,EAAG,GAAI,EAAO,OAAS,CAAC,CAAE,EAIzE,IAAM,EAAuC,CAC3C,MAAO,EACP,MAAO,EAAO,KAChB,EAEA,GAAI,EAAO,aACT,EAAY,aAAe,EAAO,aAGpC,GAAI,EAAO,aAAe,OACxB,EAAY,WAAa,EAAO,WAGlC,GAAI,EAAO,OAAS,EAAO,MAAM,OAAS,EACxC,EAAY,MAAQ,EAAO,MAAM,IAAI,EAAgB,EAGvD,GAAI,EAAO,eAAiB,OAC1B,EAAY,eAAiB,EAAO,aAGtC,GAAI,EAAO,iBACT,EAAY,UAAY,CAAE,QAAS,EAAO,gBAAiB,EAG7D,EAAO,MAAM,mCAAmC,KAAK,UAAU,EAAa,KAAM,CAAC,GAAG,EAGtF,IAAM,EAAW,MAAM,MAAM,GAAG,cAAqB,CACnD,OAAQ,OACR,QAAS,CACP,cAAe,UAAU,IACzB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,EAChC,OAAQ,YAAY,QAAQ,CAAO,CACrC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAEtC,MADA,EAAO,MAAM,qCAAqC,EAAS,UAAU,GAAW,EACtE,MAAM,iCAAiC,EAAS,UAAU,EAAS,YAAY,EAG3F,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MAEP,MADA,EAAO,MAAM,gCAAgC,EAAK,MAAM,SAAS,EACvD,MAAM,wBAAwB,EAAK,MAAM,SAAS,EAG9D,EAAO,MAAM,gDAAgD,EAAK,QAAU,aAAa,EAGzF,IAAQ,OAAM,eAAgB,GAA0B,CAAI,EAGtD,EAAoC,CAAC,EAC3C,GAAI,EAAK,OACP,QAAW,KAAQ,EAAK,OAAQ,CAC9B,IAAM,EAAY,GAAkB,CAAI,EACxC,GAAI,EACF,EAAY,KAAK,CAAS,EAKhC,IAAM,EAAyB,CAC7B,GAAI,EAAK,GACT,OACA,cACA,cACA,OAAQ,EAAK,MACf,EAMA,OAJA,EAAO,KACL,6CAA6C,EAAK,wBAAwB,EAAY,yBAAyB,EAAY,QAC7H,EAEO,ECnYT,iBAAS,gBAAQ,uBACjB,uBAAS,iBAAuC,YAsBhD,SAAS,EAAY,CAAC,EAA+D,CACnF,GAAI,CAAC,EACH,OAIF,IAAM,EAAe,EAAM,aAAe,EACpC,EAAmB,EAAM,cAAgB,EAE/C,MAAO,CACL,eACA,mBACA,YAAa,EAAe,CAC9B,EAgBF,eAAe,EAAuB,CACpC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAEpC,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAG9D,IAAM,EAAe,EAAQ,UAAU,QAAU,OAO3C,EAAiB,CACrB,MAFY,EAAO,KAAK,CAAS,EAGjC,OAAQ,EAAO,OACf,OAAQ,EACR,gBAAiB,EAAO,WAAa,KACrC,uBAAwB,CAAE,UAAW,GAAyB,CAAO,CAAE,CACzE,EAGA,GAAI,EAAO,OAAQ,CACjB,IAAM,EAAS,GAAW,CAAc,EAExC,MAAO,CACL,WAAY,EAAO,WACnB,KAAM,QAAQ,QAAQ,EAAO,IAAI,EACjC,MAAO,QAAQ,QAAQ,EAAO,KAAK,EAAE,KAAK,EAAY,EACtD,aAAc,QAAQ,QAAQ,EAAO,YAAY,EAAE,KAAK,CAAC,IAAM,CAAuB,CACxF,EAIF,IAAQ,OAAM,SAAU,MAAM,GAAa,CAAc,EAEzD,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,OAAO,EAgBT,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,EAYrF,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,EC5IrF,oBAAS,uBCAT,oBAAS,uBACT,2BACE,kBACA,qBASF,SAAS,EAAwB,CAAC,EAA6B,CAE7D,IAAM,EADa,EAAU,YAAY,EACc,SAAS,IAAI,EAChE,aACA,cACJ,GAAI,CACF,OAAO,GAAiB,CAA0B,EAClD,KAAM,CACN,OAAO,GAAY,CAAoC,GAI3D,SAAS,EAAY,CAAC,EAAwB,EAAkC,CAC9E,GAAI,IAAc,GAAU,WAC1B,OAAO,EAAc,CAAO,EAE9B,OAAO,EAAc,CAAO,EAGvB,SAAS,EAAY,CAC1B,EACA,EACA,EACU,CACV,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAI,EAGrB,SAAS,EAAc,CAC5B,EACA,EACA,EACQ,CACR,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAM,ED7C9B,eAAsB,CAAqB,CACzC,EACA,EACmB,CACnB,GAAI,CAAC,EAAO,OACV,MAAU,MAAM,0CAA0C,EAE5D,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAa,EAAS,EAAW,EAAO,MAAM,EAGvD,eAAsB,CAAqB,CACzC,EACA,EACiB,CACjB,GAAI,CAAC,EAAO,QAAU,CAAC,MAAM,QAAQ,EAAO,MAAM,EAChD,MAAU,MAAM,8CAA8C,EAEhE,GAAI,EAAO,OAAO,SAAW,EAC3B,MAAO,GAET,QAAS,EAAI,EAAG,EAAI,EAAO,OAAO,OAAQ,IAAK,CAC7C,IAAM,EAAQ,EAAO,OAAO,GAC5B,GAAI,OAAO,IAAU,UAAY,CAAC,OAAO,SAAS,CAAK,EACrD,MAAU,MAAM,0BAA0B,oBAAoB,EAGlE,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAe,EAAS,EAAW,EAAO,MAAM,EbKzD,SAAS,EAAa,EAAmB,CACvC,GAAI,OAAO,QAAY,IACrB,MAAO,CAAC,EAEV,OAAO,QAAQ,IAGjB,IAAM,EAAM,GAAc,EAEb,GAAuB,CAClC,KAAM,SACN,YAAa,sEAEb,OAAQ,CACN,eAAgB,EAAI,gBAAkB,KACtC,gBAAiB,EAAI,iBAAmB,KACxC,mBAAoB,EAAI,oBAAsB,KAC9C,mBAAoB,EAAI,oBAAsB,KAC9C,YAAa,EAAI,aAAe,KAChC,YAAa,EAAI,aAAe,KAChC,uBAAwB,EAAI,wBAA0B,KACtD,yBAA0B,EAAI,0BAA4B,KAC1D,qBAAsB,EAAI,sBAAwB,KAClD,4BAA6B,EAAI,6BAA+B,KAChE,+BAAgC,EAAI,gCAAkC,KACtE,oCAAqC,EAAI,qCAAuC,KAChF,8BAA+B,EAAI,+BAAiC,KACpE,sBAAuB,EAAI,uBAAyB,KACpD,wBAAyB,EAAI,yBAA2B,IAC1D,OAEM,KAAI,CAAC,EAAgC,EAAuC,CAChF,GAAiB,EAA0C,CAAO,GAGpE,OAAQ,EACL,EAAU,gBAAiB,MAC1B,EACA,IACsB,CACtB,OAAO,EAAoB,EAAS,CAAM,IAG3C,EAAU,uBAAwB,MACjC,EACA,IACsB,CACtB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,uBAAwB,MACjC,EACA,IACoB,CACpB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,OAAQ,MACjB,EACA,IACqC,CACrC,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,mBAAoB,MAC7B,EACA,IACoD,CACpD,OAAO,EAAuB,EAAS,CAAM,IAG9C,EAAU,eAAgB,MACzB,EACA,IACoB,CACpB,OAAO,EAAoB,EAAS,CAAK,IAG1C,EAAU,gBAAiB,MAC1B,EACA,IACyB,CACzB,OAAO,EAAmB,EAAS,CAAK,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,UAAW,MACpB,EACA,IAC4B,CAC5B,OAAO,EAAe,EAAS,CAAM,EAEzC,EAEA,MAAO,CACL,CACE,KAAM,sBACN,MAAO,CACL,CACE,KAAM,+BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GACZ,MAAU,MACR,iCAAiC,EAAS,UAAU,EAAS,YAC/D,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAClC,EAAO,KAAK,gCAAgC,EAAK,MAAM,QAAU,qBAAqB,EAE1F,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,eACR,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAS,GAAK,EAAU,SAAW,EACpD,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,0CAA0C,EAAU,mBAAmB,EAEvF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,+BACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,2CACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,kCACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CACrE,OAHmB,yBAInB,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAM,GAAK,EAAO,SAAW,EAC9C,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAAc,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CAC1E,SACA,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,IAhBiB,yBAiBnB,MAAU,MACR,uEAAgE,IAClE,EAGF,EAAO,KAAK,iDAAiD,EAAO,gBAAgB,EAExF,EACA,CACE,KAAM,wBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAmB,CAAC,EAEpB,EAAS,MAAM,EAAQ,SAAS,EAAU,WAAY,CAC1D,OAAQ,0CACR,OAAQ,GACR,cAAe,CAAC,IAAkB,CAChC,EAAO,KAAK,CAAK,EAErB,CAAC,EAED,GAAI,OAAO,IAAW,UAAY,EAAO,SAAW,EAClD,MAAU,MAAM,0CAA0C,EAG5D,GAAI,EAAO,SAAW,EACpB,MAAU,MAAM,8BAA8B,EAGhD,EAAO,KAAK,iCAAiC,EAAO,wBAAwB,EAEhF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CAInD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,kBAF9C,uIAE6E,EAE/E,GACE,CAAC,GACD,OAAO,IAAW,UAClB,EAAE,UAAW,IACb,EAAE,gBAAiB,GAEnB,MAAU,MAAM,wDAAwD,EAG1E,EAAO,KAAK,mCAAmC,EAAO,QAAQ,EAElE,EACA,CACE,KAAM,4BACN,GAAI,MAAO,IAA0C,CAMnD,IAAM,EAAc,MADH,MAAM,MAFrB,wEAEmC,GACF,YAAY,EACzC,EAAc,OAAO,KAAK,IAAI,WAAW,CAAW,CAAC,EAErD,EAAgB,MAAM,EAAQ,SAAS,EAAU,cAAe,CAAW,EAEjF,GAAI,OAAO,IAAkB,SAC3B,MAAU,MAAM,sCAAsC,EAGxD,EAAO,KAAK,iCAAiC,EAAc,UAAU,EAAG,EAAE,OAAO,EAErF,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,uCACR,CAAC,EAED,GAAI,EAAE,aAAqB,cAAgB,EAAU,aAAe,EAClE,MAAU,MAAM,yCAAyC,EAG3D,EAAO,KAAK,+BAA+B,EAAU,2BAA2B,EAEpF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,aAAc,CAC5D,OACE,+FACJ,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,SAC/B,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KACL,mCAAmC,KAAK,UAAU,CAAM,EAAE,UAAU,EAAG,GAAG,GAC5E,EAEJ,EACA,CACE,KAAM,uBACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,SAAU,CACxD,MAAO,qCACP,MAAO,CAAC,CAAE,KAAM,oBAAqB,CAAC,EACtC,aAAc,CAChB,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,UAAY,EAAE,SAAU,GACvD,MAAU,MAAM,qDAAqD,EAGvE,GAAI,OAAO,EAAO,OAAS,UAAY,EAAO,KAAK,SAAW,EAC5D,MAAU,MAAM,mDAAmD,EAGrE,EAAO,KACL,kDAAkD,EAAO,KAAK,wBAAwB,EAAO,aAAa,QAAU,GACtH,EAEJ,CACF,CACF,CACF,CACF,EAEe",
|
|
22
|
-
"debugId": "
|
|
21
|
+
"mappings": "AAgBA,iBAAS,eAAQ,sBCfjB,iBAAS,sBCAT,iBAAS,sBAET,SAAS,EAAW,CAAC,EAAiC,CACpD,GAAI,OAAO,QAAY,KAAe,CAAC,QAAQ,IAC7C,OAEF,IAAM,EAAQ,QAAQ,IAAI,GAC1B,OAAO,IAAU,OAAY,OAAY,OAAO,CAAK,EAGhD,SAAS,CAAU,CACxB,EACA,EACA,EACoB,CACpB,IAAM,EAAQ,EAAQ,WAAW,CAAG,EACpC,GAAI,IAAU,QAAa,IAAU,KACnC,OAAO,OAAO,CAAK,EAErB,OAAO,GAAY,CAAG,GAAK,EActB,SAAS,CAAiB,CAC/B,EACA,EACA,EACQ,CACR,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAS,OAAO,SAAS,EAAO,EAAE,EACxC,GAAI,CAAC,OAAO,SAAS,CAAM,EACzB,MAAU,MAAM,YAAY,oCAAsC,GAAO,EAE3E,OAAO,EAGF,SAAS,CAAiB,CAC/B,EACA,EACA,EACS,CACT,IAAM,EAAQ,EAAW,EAAS,CAAG,EACrC,GAAI,IAAU,OACZ,OAAO,EAET,IAAM,EAAa,EAAM,YAAY,EACrC,OAAO,IAAe,QAAU,IAAe,KAAO,IAAe,MAGhE,SAAS,CAAS,EAAY,CACnC,OACE,OAAO,WAAe,KACtB,OAAQ,WAAuC,SAAa,IAIzD,SAAS,CAAW,CAAC,EAAiC,CAC3D,OAAO,EAAU,GAAK,CAAC,CAAC,EAAW,EAAS,yBAAyB,EAGhE,SAAS,CAAS,CAAC,EAA4C,CACpE,OAAO,EAAW,EAAS,gBAAgB,EAGtC,SAAS,EAAkB,CAAC,EAA4C,CAC7E,IAAM,EAAkB,EAAW,EAAS,0BAA0B,EACtE,GAAI,EAEF,OADA,EAAO,MAAM,2CAA2C,EACjD,EAGT,OADA,EAAO,MAAM,yDAAyD,EAC/D,EAAU,CAAO,EAGnB,SAAS,CAAa,CAC3B,EACA,EAAe,GACS,CAIxB,GAAI,EAAU,GAAK,CAAC,EAAkB,EAAS,+BAAgC,EAAK,EAClF,MAAO,CAAC,EAEV,IAAM,EAAM,EAAe,GAAmB,CAAO,EAAI,EAAU,CAAO,EAC1E,OAAO,EAAM,CAAE,cAAe,UAAU,GAAM,EAAI,CAAC,EAG9C,SAAS,CAAU,CAAC,EAAgC,CACzD,IAAM,EAAa,EAAW,EAAS,yBAAyB,EAC1D,EACJ,EAAU,GAAK,EACX,EACC,EAAW,EAAS,iBAAiB,GAAK,4BAEjD,OADA,EAAO,MAAM,sBAAsB,GAAS,EACrC,EAGF,SAAS,CAAmB,CAAC,EAAgC,CAClE,IAAM,EAAe,EAAU,EAC1B,EAAW,EAAS,8BAA8B,GACnD,EAAW,EAAS,yBAAyB,EAC7C,EAAW,EAAS,sBAAsB,EAE9C,GAAI,EAEF,OADA,EAAO,MAAM,sCAAsC,GAAc,EAC1D,EAIT,OADA,EAAO,MAAM,0DAA0D,EAChE,EAAW,CAAO,EAGpB,SAAS,CAAa,CAAC,EAAgC,CAC5D,OACE,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,aAIhF,SAAS,CAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,EAAW,EAAS,aAAa,GAAK,QAGrF,SAAS,CAAiB,CAAC,EAAgC,CAChE,OAAO,EAAW,EAAS,wBAAwB,GAAK,yBAGnD,SAAS,EAAwB,CAAC,EAAgC,CACvE,OAAO,EAAW,EAAS,gCAAgC,GAAK,aAG3D,SAAS,EAAqB,CAAC,EAAgC,CACpE,OAAO,EAAW,EAAS,4BAA4B,GAAK,wBAGvD,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,QAG7C,SAAS,EAAW,CAAC,EAAgC,CAC1D,OAAO,EAAW,EAAS,kBAAkB,GAAK,OAG7C,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAW,EAAS,yBAAyB,GAAK,GAGpD,SAAS,EAAa,CAAC,EAAgC,CAC5D,OAAO,EAAW,EAAS,oBAAoB,GAAK,WAG/C,SAAS,EAAwB,CAAC,EAAiC,CACxE,OAAO,EAAkB,EAAS,gCAAiC,EAAK,EAGnE,SAAS,EAAsB,CAAC,EAAgC,CACrE,OAAO,EAAkB,EAAS,8BAA+B,IAAI,EAGhE,SAAS,EAA4B,CAAC,EAAgC,CAC3E,OAAO,EAAkB,EAAS,sCAAuC,IAAI,EAGxE,SAAS,EAAgB,CAAC,EAAgC,CAC/D,OAAO,EAAW,EAAS,uBAAuB,GAAK,mBAGlD,SAAS,EAAkB,CAAC,EAAgC,CACjE,OAAO,EAAkB,EAAS,0BAA2B,OAAO,EDjLrE,WAAuC,sBAAwB,GAEzD,SAAS,EAAgB,CAC9B,EACA,EACM,CACD,GAA4B,CAAO,EAG1C,eAAe,EAA2B,CAAC,EAAuC,CAChF,GAAI,EAAU,EAAG,CACf,EAAO,MAAM,yDAAyD,EACtE,OAKF,GAAI,CAFW,EAAU,CAAO,EAEnB,CACX,EAAO,KACL,8GAEF,EACA,OAGF,GAAI,CACF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,EAAO,KACL,uCAAuC,EAAS,UAAU,EAAS,2DAErE,EACA,QAEF,MAAO,EAAO,CACd,IAAM,EAAU,aAAiB,MAAQ,EAAM,QAAU,OAAO,CAAK,EACrE,EAAO,KAAK,kCAAkC,yCAA+C,GExCjG,iBAAS,sBCLT,iBAAS,uBAET,IAAM,EAAc,CAClB,IAAK,CACH,OAAQ,CAAC,GAAM,GAAM,GAAM,EAAI,EAC/B,WAAY,CAAC,GAAM,GAAM,GAAM,EAAI,CACrC,EACA,QAAS,CAAC,GAAM,GAAM,EAAI,EAC1B,IAAK,CAAC,GAAM,IAAM,IAAM,EAAI,EAC5B,KAAM,CAAC,IAAM,GAAM,GAAM,EAAI,EAC7B,KAAM,CAAC,IAAM,IAAM,IAAM,GAAI,EAC7B,UAAW,CAAC,GAAM,GAAM,IAAM,GAAI,CACpC,EAEM,GAA4B,GAWlC,SAAS,CAAU,CAAC,EAAgB,EAAgB,EAAsC,CACxF,QAAS,EAAI,EAAG,EAAI,EAAS,OAAQ,IAAK,CACxC,IAAM,EAAe,EAAS,GAC9B,GAAI,IAAiB,QAAa,EAAO,EAAS,KAAO,EACvD,MAAO,GAGX,MAAO,GAGF,SAAS,CAAmB,CAAC,EAA+B,CACjE,GAAI,EAAO,OAAS,GAClB,MAAO,2BAIT,GACE,EAAW,EAAQ,EAAG,EAAY,IAAI,MAAM,GAC5C,EAAW,EAAQ,EAAG,EAAY,IAAI,UAAU,EAEhD,MAAO,YAIT,IAAM,EAAY,EAAO,GACnB,EAAa,EAAO,GAC1B,GACE,EAAW,EAAQ,EAAG,EAAY,OAAO,GACxC,IAAc,KAAQ,IAAe,SAAc,EAAa,OAAU,IAE3E,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,GAAG,EACvC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,aAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,IAAI,EACxC,MAAO,YAIT,GAAI,EAAW,EAAQ,EAAG,EAAY,SAAS,EAC7C,MAAO,aAIT,OADA,GAAO,KAAK,sEAAsE,EAC3E,2BAGF,SAAS,EAAuB,CAAC,EAAiC,CACvE,OAAQ,OACD,YACH,MAAO,UACJ,aACH,MAAO,UACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,YACH,MAAO,UACJ,aACH,MAAO,WACJ,2BACH,MAAO,OAIN,SAAS,EAAsB,CAAC,EAAiC,CAEtE,MAAO,aADK,GAAwB,CAAQ,ID3E9C,SAAS,EAAY,CAAC,EAAsC,CAC1D,OAAO,aAAiB,MAAQ,aAAiB,KAGnD,SAAS,CAAQ,CAAC,EAAiC,CACjD,OAAO,OAAO,SAAS,CAAK,EAG9B,SAAS,EAA0B,CAAC,EAAmD,CACrF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,UAAW,IACV,GAAc,EAAmC,KAAK,GACrD,EAAU,EAAmC,KAAK,GAIxD,SAAS,EAAyB,CAAC,EAAkD,CACnF,OACE,OAAO,IAAU,UACjB,IAAU,MACV,aAAc,GACd,OAAQ,EAAkC,WAAa,SAI3D,eAAe,EAAiB,CAAC,EAA4B,CAC3D,IAAM,EAAW,MAAM,MAAM,CAAG,EAChC,GAAI,CAAC,EAAS,GACZ,MAAU,MAAM,mCAAmC,EAAS,QAAQ,EAEtE,OAAO,EAAS,KAAK,EAEvB,eAAsB,CAAmB,CACvC,EACA,EACiB,CACjB,IAAI,EAAY,GAAsB,CAAO,EACzC,EACA,EAAiD,CAAC,EAEtD,GAAI,OAAO,IAAU,SACnB,EAAO,MAAM,qCAAqC,GAAO,EACzD,EAAO,MAAM,GAAkB,CAAK,EAC/B,QAAI,GAAa,CAAK,EAC3B,EAAO,EACF,QAAI,EAAS,CAAK,EAAG,CAC1B,IAAM,EAAW,EAAoB,CAAK,EAC1C,EAAO,MAAM,2CAA2C,GAAU,EAClE,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,CAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EACtD,QAAI,GAA2B,CAAK,EAAG,CAE5C,GADA,EAAc,EACV,EAAM,MACR,EAAY,EAAM,MAEpB,GAAI,EAAS,EAAM,KAAK,EAAG,CACzB,IAAM,EAAW,EAAM,UAAY,EAAoB,EAAM,KAAK,EAClE,EAAO,MAAM,6BAA6B,GAAU,EACpD,EAAO,IAAI,KAAK,CAAC,IAAI,WAAW,EAAM,KAAK,CAAC,EAAG,CAAE,KAAM,CAAS,CAAC,EAEjE,OAAO,EAAM,MAEV,QAAI,GAA0B,CAAK,EACxC,EAAO,MAAM,qCAAqC,EAAM,UAAU,EAClE,EAAO,MAAM,GAAkB,EAAM,QAAQ,EAC7C,EAAc,CAAE,OAAQ,EAAM,MAAO,EAErC,WAAU,MACR,qFACF,EAGF,EAAO,MAAM,uCAAuC,GAAW,EAE/D,IAAM,EAAY,EAAc,MAAQ,aAClC,EACH,EAAc,MACf,GACE,EAAS,WAAW,QAAQ,EACvB,EACD,YACN,EAEI,EAAW,IAAI,SAIrB,GAHA,EAAS,OAAO,OAAQ,EAAM,CAAQ,EACtC,EAAS,OAAO,QAAS,CAAS,EAE9B,EAAY,SACd,EAAS,OAAO,WAAY,EAAY,QAAQ,EAElD,GAAI,EAAY,eACd,EAAS,OAAO,kBAAmB,EAAY,cAAc,EAE/D,GAAI,EAAY,OACd,EAAS,OAAO,SAAU,EAAY,MAAM,EAE9C,GAAI,EAAY,cAAgB,OAC9B,EAAS,OAAO,cAAe,OAAO,EAAY,WAAW,CAAC,EAEhE,GAAI,EAAY,uBACd,QAAW,KAAe,EAAY,uBACpC,EAAS,OAAO,4BAA6B,CAAW,EAI5D,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,yBAAgC,CAC9D,OAAQ,OACR,QAAS,EAAc,CAAO,EAC9B,KAAM,CACR,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,gCAAgC,EAAS,UAAU,EAAS,gBAAgB,GAC9E,EAIF,OADc,MAAM,EAAS,KAAK,GACtB,KAGd,eAAsB,CAAkB,CACtC,EACA,EACsB,CACtB,IAAI,EACA,EACA,EAA0B,MAC1B,EACA,EAEJ,GAAI,OAAO,IAAU,SACnB,EAAO,EACP,EAAQ,OACH,KAGL,GAFA,EAAO,EAAM,KACb,EAAQ,EAAM,MACV,WAAY,GAAS,EAAM,OAC7B,EAAS,EAAM,OAEjB,GAAI,UAAW,GAAS,EAAM,MAC5B,EAAQ,EAAM,MAEhB,GAAI,iBAAkB,GAAS,EAAM,aACnC,EAAe,EAAM,aAUzB,GANA,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAQ,GAAS,GAAY,CAAO,EACpC,EAAe,GAAgB,GAAmB,CAAO,EAEzD,EAAO,MAAM,wCAAwC,GAAO,EAExD,CAAC,GAAQ,EAAK,KAAK,EAAE,SAAW,EAClC,MAAU,MAAM,wCAAwC,EAG1D,GAAI,EAAK,OAAS,KAChB,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAA0B,CAAC,QAAS,OAAQ,QAAS,OAAQ,OAAQ,SAAS,EACpF,GAAI,GAAS,CAAC,EAAY,SAAS,CAAiB,EAClD,MAAU,MAAM,kBAAkB,sBAA0B,EAAY,KAAK,IAAI,GAAG,EAGtF,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAsC,CAC1C,QACA,MAAO,EACP,MAAO,EACP,gBAAiB,CACnB,EAEA,GAAI,GAAgB,EAAa,OAAS,EACxC,EAAY,aAAe,EAG7B,IAAM,EAAW,MAAM,MAAM,GAAG,iBAAwB,CACtD,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,sBACZ,IAAW,MAAQ,CAAE,OAAQ,YAAa,EAAI,CAAC,CACrD,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MAAM,sBAAsB,EAAS,UAAU,EAAS,gBAAgB,GAAW,EAG/F,OAAO,EAAS,YAAY,EEhO9B,iBAAS,eAAQ,kBAAW,uBCA5B,oBAAS,uBAGT,IAAM,GAAoB,IAmC1B,SAAS,EAAc,CAAC,EAAwB,CAC9C,GAAI,EAAO,QAAU,GACnB,OAAO,EAET,MAAO,GAAG,EAAO,MAAM,EAAG,EAAiB,KAG7C,SAAS,EAAc,CAAC,EAA+B,CACrD,GAAI,iBAAkB,EAAO,CAC3B,IAAM,EACJ,wBAAyB,EAAQ,EAAM,oBAAsB,OACzD,EAAqB,EAAM,oBAAsB,GAAqB,aAC5E,MAAO,CACL,aAAc,EAAM,cAAgB,EACpC,iBAAkB,EAAM,kBAAoB,EAC5C,YAAa,EAAM,cAAgB,EAAM,cAAgB,IAAM,EAAM,kBAAoB,GACzF,oBACF,EAEF,GAAI,gBAAiB,GAAS,iBAAkB,EAAO,CACrD,IAAM,EAAS,EAAqB,aAAe,EAC7C,EAAU,EAAqB,cAAgB,EAC/C,EAAS,EAAqB,aAAe,EAAQ,EAC3D,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,EACb,mBAAqB,EAAqB,iBAC5C,EAEF,MAAO,CACL,aAAc,EACd,iBAAkB,EAClB,YAAa,CACf,EAGK,SAAS,CAAmB,CACjC,EACA,EACA,EACA,EACM,CACN,IAAM,EAAa,GAAe,CAAK,EAEjC,EAAkC,CACtC,UACA,OAAQ,SACR,SAAU,SACV,OACA,OAAQ,GAAe,CAAM,EAC7B,OAAQ,CACN,OAAQ,EAAW,aACnB,WAAY,EAAW,iBACvB,MAAO,EAAW,eACd,EAAW,qBAAuB,OAClC,CAAE,OAAQ,EAAW,kBAAmB,EACxC,CAAC,CACP,CACF,EAEA,EAAQ,UAAU,GAAU,WAAY,CAAO,ED9FjD,IAAM,GAAuB,KAa7B,SAAS,EAAiB,CAAC,EAAoC,CAC7D,IAAM,EAAkB,OAAO,OAAO,EAAW,EACjD,GAAI,CAAC,EAAgB,SAAS,CAAS,EACrC,MAAU,MACR,gCAAgC,sBAA8B,EAAgB,KAAK,IAAI,GACzF,EAEF,OAAO,EAGT,SAAS,EAAW,CAAC,EAA4D,CAC/E,GAAI,IAAW,KACb,OAAO,KAET,GAAI,OAAO,IAAW,SACpB,OAAO,EAET,GAAI,OAAO,IAAW,UAAY,OAAO,EAAO,OAAS,SACvD,OAAO,EAAO,KAEhB,MAAU,MAAM,sEAAsE,EAGxF,eAAsB,CAAmB,CACvC,EACA,EACmB,CACnB,IAAM,EAAiB,EAAkB,CAAO,EAC1C,EAAqB,GAAkB,GAAuB,CAAO,CAAC,EAEtE,EAAO,GAAY,CAAM,EAC/B,GAAI,IAAS,KAAM,CACjB,EAAO,MAAM,qDAAqD,EAClE,IAAM,EAAiB,MAAM,CAAkB,EAAE,KAAK,CAAC,EAEvD,OADA,EAAW,GAAK,IACT,EAGT,IAAI,EAAc,EAAK,KAAK,EAC5B,GAAI,EAAY,SAAW,EACzB,MAAU,MAAM,0CAA0C,EAM5D,IAAM,EAAW,GAAuB,EACxC,GAAI,EAAY,OAAS,EACvB,EAAO,KACL,uCAAuC,KAAK,KAAK,EAAY,OAAS,CAAC,6BAA6B,WACtG,EACA,EAAc,EAAY,MAAM,EAAG,CAAQ,EAI7C,IAAM,EAAM,GADI,EAAoB,CAAO,eAG3C,EAAO,MAAM,6CAA6C,GAAgB,EAE1E,IAAM,EAAW,MAAM,MAAM,EAAK,CAChC,OAAQ,OACR,QAAS,IACJ,EAAc,EAAS,EAAI,EAC9B,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CACnB,MAAO,EACP,MAAO,CACT,CAAC,CACH,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,+BAA+B,EAAS,UAAU,EAAS,gBAAgB,GAC7E,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAE5B,EAAc,GAAM,OAAO,GACjC,GAAI,CAAC,GAAe,CAAC,EAAY,UAC/B,MAAU,MAAM,0DAA0D,EAG5E,IAAM,EAAY,EAAY,UAE9B,GAAI,EAAU,SAAW,EACvB,MAAU,MACR,qCAAqC,EAAU,oBAAoB,+CAErE,EAGF,GAAI,EAAK,MACP,EAAoB,EAAS,GAAU,eAAgB,EAAa,CAClE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAClB,YAAa,EAAK,MAAM,YAC1B,CAAC,EAIH,OADA,EAAO,MAAM,qCAAqC,EAAU,mBAAmB,EACxE,EEzHT,iBAAS,gBAAQ,uBAwBjB,IAAM,GACJ,0EAEF,eAAsB,CAAqB,CACzC,EACA,EACkC,CAClC,IAAM,EAAY,GAAc,CAAO,EACjC,EAAQ,EAAO,OAAS,EACxB,EAAmB,EAAO,MAAsB,YAChD,EAAiB,EAIvB,GAFA,GAAO,MAAM,+BAA+B,GAAW,EAEnD,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,8CAA8C,EAGhE,GAAI,EAAQ,GAAK,EAAQ,GACvB,MAAU,MAAM,sCAAsC,EAGxD,IAAM,EAAU,EAAW,CAAO,EAE5B,EAA+C,CACnD,MAAO,EACP,OAAQ,EAAO,OACf,EAAG,EACH,MACF,EAEA,GAAI,EAAe,QACjB,EAAY,QAAU,EAAe,QAEvC,GAAI,EAAe,MACjB,EAAY,MAAQ,EAAe,MAGrC,IAAM,EAAW,MAAM,MAAM,GAAG,uBAA8B,CAC5D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,mCAAmC,EAAS,UAAU,EAAS,gBAAgB,GACjF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,CAAC,EAAK,MAAQ,EAAK,KAAK,SAAW,EACrC,MAAU,MAAM,+BAA+B,EAGjD,OAAO,EAAK,KAAK,IAAI,CAAC,KAAU,CAC9B,IAAK,EAAK,IACV,cAAe,EAAK,cACtB,EAAE,EAGJ,SAAS,EAAsB,CAAC,EAAyB,CAEvD,OADmB,EAAQ,MAAM,2BAA2B,IACxC,IAAI,KAAK,GAAK,iBAGpC,SAAS,EAA4B,CAAC,EAAyB,CAC7D,OAAO,EAAQ,QAAQ,4BAA6B,EAAE,EAAE,KAAK,EAG/D,eAAsB,CAAsB,CAC1C,EACA,EACiC,CACjC,IAAM,EAAY,GAAyB,CAAO,EAC5C,EAAY,GAA6B,CAAO,EAEtD,GAAO,MAAM,2CAA2C,GAAW,EAEnE,IAAI,EACA,EAEJ,GAAI,OAAO,IAAW,SACpB,EAAW,EACX,EAAa,GAEb,OAAW,EAAO,SAClB,EAAa,EAAO,QAAU,GAGhC,GAAI,CAAC,GAAY,EAAS,KAAK,EAAE,SAAW,EAC1C,MAAU,MAAM,8CAA8C,EAGhE,IAAM,EAAU,EAAW,CAAO,EAE5B,EAAc,CAClB,MAAO,EACP,SAAU,CACR,CACE,KAAM,OACN,QAAS,CACP,CAAE,KAAM,OAAQ,KAAM,CAAW,EACjC,CAAE,KAAM,YAAa,UAAW,CAAE,IAAK,CAAS,CAAE,CACpD,CACF,CACF,EACA,WAAY,CACd,EAEM,EAAW,MAAM,MAAM,GAAG,qBAA4B,CAC1D,OAAQ,OACR,QAAS,IACJ,EAAc,CAAO,EACxB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,CAClC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAAE,MAAM,IAAM,eAAe,EACnE,MAAU,MACR,oCAAoC,EAAS,UAAU,EAAS,gBAAgB,GAClF,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MACP,EACE,EACA,GAAU,kBACV,OAAO,IAAW,SAAW,EAAU,EAAO,QAAU,GACxD,CACE,aAAc,EAAK,MAAM,cACzB,iBAAkB,EAAK,MAAM,kBAC7B,YAAa,EAAK,MAAM,YAC1B,CACF,EAIF,IAAM,EADc,EAAK,UAAU,IACN,SAAS,QAEtC,GAAI,CAAC,EACH,MAAU,MAAM,6CAA6C,EAG/D,MAAO,CACL,MAAO,GAAuB,CAAO,EACrC,YAAa,GAA6B,CAAO,CACnD,EC/KF,iBAAS,gBAAQ,uBACjB,yBAAS,YCPT,uBAAS,wBAIT,IAAM,GAAgB,WAEf,SAAS,CAAkB,CAAC,EAAwC,CACzE,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAS,EAAU,CAAO,EAEhC,GAAI,CAAC,GAAU,EAAY,CAAO,EAChC,OAAO,GAAa,CAClB,OAAQ,GACR,SACF,CAAC,EAGH,GAAI,CAAC,EACH,MAAU,MACR,uFACF,EAGF,OAAO,GAAa,CAClB,SACA,SACF,CAAC,EC1BH,iBAAS,uBACT,yBAAS,YAST,IAAM,GAAwB,CAC5B,cAAe,uBACf,WAAY,YACd,EAEO,SAAS,EAAqB,EAAuB,CAC1D,MAAO,QAAS,OAAM,WAAsD,CAC1E,GAAI,EAAE,aAAiB,IACrB,OAAO,KAET,GAAI,CACF,IAAM,EAAc,EAAK,QAAQ,GAAsB,cAAe,EAAE,EAGxE,OAFA,KAAK,MAAM,CAAW,EACtB,GAAO,MAAM,wEAAwE,EAC9E,EACP,KAAM,CAEN,OADA,GAAO,KAAK,0CAA0C,EAC/C,OFXb,eAAe,EAAyB,CACtC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAIpC,GAFA,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAE1D,CAAC,EAAO,QAAU,EAAO,OAAO,KAAK,EAAE,SAAW,EACpD,MAAU,MAAM,+CAA+C,EAGjE,GAAI,EAAO,OACT,GAAO,MACL,oGAEF,EAGF,IAAM,EAAQ,EAAO,KAAK,CAAS,GAC3B,SAAQ,SAAU,MAAM,GAAe,CAC7C,QACA,OAAQ,YACR,OAAQ,EAAO,OACf,wBAAyB,GAAsB,CACjD,CAAC,EAED,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,GAAI,OAAO,IAAW,UAAY,IAAW,KAC3C,MAAU,MAAM,8BAA8B,OAAO,oBAAyB,EAGhF,OAAO,EAGT,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EAGzF,eAAsB,CAAiB,CACrC,EACA,EACoC,CACpC,OAAO,GAA0B,EAAS,EAAQ,GAAU,aAAc,CAAa,EG7CzF,iBAAS,sBA6ET,SAAS,EAAgB,CAAC,EAAsC,CAC9D,OAAQ,EAAK,UACN,qBACH,MAAO,CAAE,KAAM,oBAAqB,MACjC,cACH,MAAO,CACL,KAAM,cACN,iBAAkB,EAAK,cACzB,MACG,mBACH,MAAO,CACL,KAAM,mBACN,UAAW,EAAK,WAAa,CAAE,KAAM,MAAO,CAC9C,MACG,MACH,MAAO,CACL,KAAM,MACN,aAAc,EAAK,YACnB,WAAY,EAAK,UACjB,iBAAkB,EAAK,iBAAmB,OAC5C,UAEA,MAAU,MAAM,+BAAgC,EAAsB,MAAM,GAOlF,SAAS,EAAiB,CAAC,EAAyD,CAClF,OAAQ,EAAK,UACN,kBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,kBACN,OAAS,EAAK,QAAqC,YACnD,OAAQ,CACN,KAAO,EAAK,QAAQ,MAAoD,SACxE,MAAO,EAAK,QAAQ,MACpB,IAAK,EAAK,QAAQ,GACpB,CACF,MAEG,mBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,mBACN,OAAS,EAAK,QAAqC,YACnD,MAAO,EAAK,OAAS,GACrB,QAAS,EAAK,SAAS,IAAI,CAAC,KAAO,CACjC,OAAQ,EAAE,QACV,SAAU,EAAE,UACZ,MAAO,EAAE,KACX,EAAE,CACJ,MAEG,wBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,wBACN,OAAS,EAAK,QAAqC,YACnD,KAAM,EAAK,MAAQ,GACnB,OAAQ,EAAK,MACf,MAEG,gBACH,MAAO,CACL,GAAI,EAAK,IAAM,GACf,KAAM,gBACN,OAAS,EAAK,QAAqC,YACnD,YAAa,EAAK,cAAgB,GAClC,SAAU,EAAK,WAAa,GAC5B,UAAY,EAAK,WAAa,CAAC,EAC/B,OAAQ,EAAK,MACf,MAEG,UACH,MAAO,CACL,KAAM,UACN,QACE,EAAK,SAAS,IAAI,CAAC,KAAO,CACxB,KAAM,cACN,KAAM,EAAE,KACR,YACE,EAAE,aAAa,IAAI,CAAC,KAAO,CACzB,IAAK,EAAE,IACP,MAAO,EAAE,MACT,WAAY,EAAE,YACd,SAAU,EAAE,SACd,EAAE,GAAK,CAAC,CACZ,EAAE,GAAK,CAAC,CACZ,UAIA,OAAO,MAOb,SAAS,EAAyB,CAAC,EAGjC,CAEA,GAAI,EAAS,YAAa,CAExB,IAAM,EAAoC,CAAC,EAC3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QACzB,GAAI,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAOb,MAAO,CAAE,KAAM,EAAS,YAAa,aAAY,EAInD,IAAI,EAAO,GACL,EAAoC,CAAC,EAE3C,GAAI,EAAS,QACX,QAAW,KAAQ,EAAS,OAC1B,GAAI,EAAK,OAAS,WAAa,EAAK,SAClC,QAAW,KAAW,EAAK,QAEzB,GADA,GAAQ,EAAQ,KACZ,EAAQ,YACV,QAAW,KAAO,EAAQ,YACxB,EAAY,KAAK,CACf,IAAK,EAAI,IACT,MAAO,EAAI,MACX,WAAY,EAAI,YAChB,SAAU,EAAI,SAChB,CAAC,GAQb,MAAO,CAAE,OAAM,aAAY,EA8B7B,eAAsB,CAAc,CAClC,EACA,EACyB,CACzB,IAAM,EAAS,EAAU,CAAO,EAChC,GAAI,CAAC,EACH,MAAU,MACR,yGACF,EAGF,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAY,EAAO,OAAS,GAAiB,CAAO,EACpD,EAAU,GAAmB,CAAO,EAE1C,EAAO,MAAM,+CAA+C,GAAW,EACvE,EAAO,MAAM,4BAA4B,EAAO,MAAM,UAAU,EAAG,GAAG,MAAM,EAG5E,IAAM,EAAkB,EAAO,OAAO,OACpC,CAAC,IAAM,EAAE,OAAS,sBAAwB,EAAE,OAAS,eAAiB,EAAE,OAAS,KACnF,EAEA,GAAI,CAAC,GAAmB,EAAgB,SAAW,EAEjD,EAAO,MAAM,2EAA2E,EACxF,EAAO,MAAQ,CAAC,CAAE,KAAM,oBAAqB,EAAG,GAAI,EAAO,OAAS,CAAC,CAAE,EAIzE,IAAM,EAAuC,CAC3C,MAAO,EACP,MAAO,EAAO,KAChB,EAEA,GAAI,EAAO,aACT,EAAY,aAAe,EAAO,aAGpC,GAAI,EAAO,aAAe,OACxB,EAAY,WAAa,EAAO,WAGlC,GAAI,EAAO,OAAS,EAAO,MAAM,OAAS,EACxC,EAAY,MAAQ,EAAO,MAAM,IAAI,EAAgB,EAGvD,GAAI,EAAO,eAAiB,OAC1B,EAAY,eAAiB,EAAO,aAGtC,GAAI,EAAO,iBACT,EAAY,UAAY,CAAE,QAAS,EAAO,gBAAiB,EAG7D,EAAO,MAAM,mCAAmC,KAAK,UAAU,EAAa,KAAM,CAAC,GAAG,EAGtF,IAAM,EAAW,MAAM,MAAM,GAAG,cAAqB,CACnD,OAAQ,OACR,QAAS,CACP,cAAe,UAAU,IACzB,eAAgB,kBAClB,EACA,KAAM,KAAK,UAAU,CAAW,EAChC,OAAQ,YAAY,QAAQ,CAAO,CACrC,CAAC,EAED,GAAI,CAAC,EAAS,GAAI,CAChB,IAAM,EAAY,MAAM,EAAS,KAAK,EAEtC,MADA,EAAO,MAAM,qCAAqC,EAAS,UAAU,GAAW,EACtE,MAAM,iCAAiC,EAAS,UAAU,EAAS,YAAY,EAG3F,IAAM,EAAQ,MAAM,EAAS,KAAK,EAElC,GAAI,EAAK,MAEP,MADA,EAAO,MAAM,gCAAgC,EAAK,MAAM,SAAS,EACvD,MAAM,wBAAwB,EAAK,MAAM,SAAS,EAG9D,EAAO,MAAM,gDAAgD,EAAK,QAAU,aAAa,EAGzF,IAAQ,OAAM,eAAgB,GAA0B,CAAI,EAGtD,EAAoC,CAAC,EAC3C,GAAI,EAAK,OACP,QAAW,KAAQ,EAAK,OAAQ,CAC9B,IAAM,EAAY,GAAkB,CAAI,EACxC,GAAI,EACF,EAAY,KAAK,CAAS,EAKhC,IAAM,EAAyB,CAC7B,GAAI,EAAK,GACT,OACA,cACA,cACA,OAAQ,EAAK,MACf,EAMA,OAJA,EAAO,KACL,6CAA6C,EAAK,wBAAwB,EAAY,yBAAyB,EAAY,QAC7H,EAEO,ECnYT,iBAAS,gBAAQ,uBACjB,uBAAS,iBAAuC,YAuChD,SAAS,EAAY,CAAC,EAA+D,CACnF,GAAI,CAAC,EACH,OAIF,IAAM,EAAe,EAAM,aAAe,EACpC,EAAmB,EAAM,cAAgB,EACzC,EAAiB,EAEvB,MAAO,CACL,eACA,mBACA,YAAa,EAAe,EAC5B,mBAAoB,EAAe,iBACrC,EAGF,SAAS,EAAyB,CAAC,EAAsD,CACvF,IAAM,EAAoB,EAC1B,MAAO,CACL,eAAgB,EAAkB,iBAAiB,QAAQ,eAC3D,qBAAsB,EAAkB,iBAAiB,QAAQ,oBACnE,EAgBF,eAAe,EAAuB,CACpC,EACA,EACA,EACA,EACoC,CACpC,IAAM,EAAS,EAAmB,CAAO,EACnC,EAAY,EAAW,CAAO,EAEpC,GAAO,MAAM,kBAAkB,YAAoB,GAAW,EAC9D,IAAM,EAAqB,GAA0B,CAAM,EAGrD,EAAe,EAAQ,UAAU,QAAU,OAO3C,EAAiB,CACrB,MAFY,EAAO,KAAK,CAAS,EAGjC,OAAQ,EAAO,OACf,OAAQ,EACR,gBAAiB,EAAO,WAAa,KACrC,uBAAwB,CAAE,UAAW,GAAyB,CAAO,CAAE,KACnE,EAAmB,gBAAkB,EAAmB,qBACxD,CACE,gBAAiB,CACf,OAAQ,IACF,EAAmB,eACnB,CAAE,eAAgB,EAAmB,cAAe,EACpD,CAAC,KACD,EAAmB,qBACnB,CAAE,qBAAsB,EAAmB,oBAAqB,EAChE,CAAC,CACP,CACF,CACF,EACA,CAAC,CACP,EAGA,GAAI,EAAO,OAAQ,CACjB,IAAM,EAAS,GAAW,CAAc,EAExC,MAAO,CACL,WAAY,EAAO,WACnB,KAAM,QAAQ,QAAQ,EAAO,IAAI,EACjC,MAAO,QAAQ,QAAQ,EAAO,KAAK,EAAE,KAAK,EAAY,EACtD,aAAc,QAAQ,QAAQ,EAAO,YAAY,EAAE,KAAK,CAAC,IAAM,CAAuB,CACxF,EAIF,IAAQ,OAAM,SAAU,MAAM,GAAa,CAAc,EAEzD,GAAI,EACF,EAAoB,EAAS,EAAW,EAAO,OAAQ,CAAK,EAG9D,OAAO,EAgBT,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,EAYrF,eAAsB,CAAe,CACnC,EACA,EACoC,CACpC,OAAO,GAAwB,EAAS,EAAQ,GAAU,WAAY,CAAa,ECtLrF,oBAAS,uBCAT,oBAAS,uBACT,2BACE,kBACA,qBASF,SAAS,EAAwB,CAAC,EAA6B,CAE7D,IAAM,EADa,EAAU,YAAY,EACc,SAAS,IAAI,EAChE,aACA,cACJ,GAAI,CACF,OAAO,GAAiB,CAA0B,EAClD,KAAM,CACN,OAAO,GAAY,CAAoC,GAI3D,SAAS,EAAY,CAAC,EAAwB,EAAkC,CAC9E,GAAI,IAAc,GAAU,WAC1B,OAAO,EAAc,CAAO,EAE9B,OAAO,EAAc,CAAO,EAGvB,SAAS,EAAY,CAC1B,EACA,EACA,EACU,CACV,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAI,EAGrB,SAAS,EAAc,CAC5B,EACA,EACA,EACQ,CACR,IAAM,EAAY,GAAa,EAAS,CAAS,EAEjD,OADgB,GAAyB,CAAS,EACnC,OAAO,CAAM,ED7C9B,eAAsB,CAAqB,CACzC,EACA,EACmB,CACnB,GAAI,CAAC,EAAO,OACV,MAAU,MAAM,0CAA0C,EAE5D,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAa,EAAS,EAAW,EAAO,MAAM,EAGvD,eAAsB,CAAqB,CACzC,EACA,EACiB,CACjB,GAAI,CAAC,EAAO,QAAU,CAAC,MAAM,QAAQ,EAAO,MAAM,EAChD,MAAU,MAAM,8CAA8C,EAEhE,GAAI,EAAO,OAAO,SAAW,EAC3B,MAAO,GAET,QAAS,EAAI,EAAG,EAAI,EAAO,OAAO,OAAQ,IAAK,CAC7C,IAAM,EAAQ,EAAO,OAAO,GAC5B,GAAI,OAAO,IAAU,UAAY,CAAC,OAAO,SAAS,CAAK,EACrD,MAAU,MAAM,0BAA0B,oBAAoB,EAGlE,IAAM,EAAY,EAAO,WAAa,GAAU,WAChD,OAAO,GAAe,EAAS,EAAW,EAAO,MAAM,EbKzD,SAAS,EAAa,EAAmB,CACvC,GAAI,OAAO,QAAY,IACrB,MAAO,CAAC,EAEV,OAAO,QAAQ,IAGjB,IAAM,EAAM,GAAc,EAEb,GAAuB,CAClC,KAAM,SACN,YAAa,sEAEb,OAAQ,CACN,eAAgB,EAAI,gBAAkB,KACtC,gBAAiB,EAAI,iBAAmB,KACxC,mBAAoB,EAAI,oBAAsB,KAC9C,mBAAoB,EAAI,oBAAsB,KAC9C,YAAa,EAAI,aAAe,KAChC,YAAa,EAAI,aAAe,KAChC,uBAAwB,EAAI,wBAA0B,KACtD,yBAA0B,EAAI,0BAA4B,KAC1D,qBAAsB,EAAI,sBAAwB,KAClD,4BAA6B,EAAI,6BAA+B,KAChE,+BAAgC,EAAI,gCAAkC,KACtE,oCAAqC,EAAI,qCAAuC,KAChF,8BAA+B,EAAI,+BAAiC,KACpE,sBAAuB,EAAI,uBAAyB,KACpD,wBAAyB,EAAI,yBAA2B,IAC1D,OAEM,KAAI,CAAC,EAAgC,EAAuC,CAChF,GAAiB,EAA0C,CAAO,GAGpE,OAAQ,EACL,EAAU,gBAAiB,MAC1B,EACA,IACsB,CACtB,OAAO,EAAoB,EAAS,CAAM,IAG3C,EAAU,uBAAwB,MACjC,EACA,IACsB,CACtB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,uBAAwB,MACjC,EACA,IACoB,CACpB,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,YAAa,MACtB,EACA,IACuC,CACvC,OAAO,EAAgB,EAAS,CAAM,IAGvC,EAAU,OAAQ,MACjB,EACA,IACqC,CACrC,OAAO,EAAsB,EAAS,CAAM,IAG7C,EAAU,mBAAoB,MAC7B,EACA,IACoD,CACpD,OAAO,EAAuB,EAAS,CAAM,IAG9C,EAAU,eAAgB,MACzB,EACA,IACoB,CACpB,OAAO,EAAoB,EAAS,CAAK,IAG1C,EAAU,gBAAiB,MAC1B,EACA,IACyB,CACzB,OAAO,EAAmB,EAAS,CAAK,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,cAAe,MACxB,EACA,IACuC,CACvC,OAAO,EAAkB,EAAS,CAAM,IAGzC,EAAU,UAAW,MACpB,EACA,IAC4B,CAC5B,OAAO,EAAe,EAAS,CAAM,EAEzC,EAEA,MAAO,CACL,CACE,KAAM,sBACN,MAAO,CACL,CACE,KAAM,+BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAU,EAAW,CAAO,EAC5B,EAAW,MAAM,MAAM,GAAG,WAAkB,CAChD,QAAS,EAAc,CAAO,CAChC,CAAC,EAED,GAAI,CAAC,EAAS,GACZ,MAAU,MACR,iCAAiC,EAAS,UAAU,EAAS,YAC/D,EAGF,IAAM,EAAQ,MAAM,EAAS,KAAK,EAClC,EAAO,KAAK,gCAAgC,EAAK,MAAM,QAAU,qBAAqB,EAE1F,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,eACR,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAS,GAAK,EAAU,SAAW,EACpD,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,0CAA0C,EAAU,mBAAmB,EAEvF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,+BACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,yBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAO,MAAM,EAAQ,SAAS,EAAU,WAAY,CACxD,OAAQ,2CACV,CAAC,EAED,GAAI,OAAO,IAAS,UAAY,EAAK,SAAW,EAC9C,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KAAK,wCAAwC,EAAK,UAAU,EAAG,EAAE,OAAO,EAEnF,EACA,CACE,KAAM,kCACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CACrE,OAHmB,yBAInB,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,CAAC,MAAM,QAAQ,CAAM,GAAK,EAAO,SAAW,EAC9C,MAAU,MAAM,kDAAkD,EAGpE,IAAM,EAAc,MAAM,EAAQ,SAAS,EAAU,sBAAuB,CAC1E,SACA,UAAW,EAAU,UACvB,CAAC,EAED,GAAI,IAhBiB,yBAiBnB,MAAU,MACR,uEAAgE,IAClE,EAGF,EAAO,KAAK,iDAAiD,EAAO,gBAAgB,EAExF,EACA,CACE,KAAM,wBACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAmB,CAAC,EAEpB,EAAS,MAAM,EAAQ,SAAS,EAAU,WAAY,CAC1D,OAAQ,0CACR,OAAQ,GACR,cAAe,CAAC,IAAkB,CAChC,EAAO,KAAK,CAAK,EAErB,CAAC,EAED,GAAI,OAAO,IAAW,UAAY,EAAO,SAAW,EAClD,MAAU,MAAM,0CAA0C,EAG5D,GAAI,EAAO,SAAW,EACpB,MAAU,MAAM,8BAA8B,EAGhD,EAAO,KAAK,iCAAiC,EAAO,wBAAwB,EAEhF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CAInD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,kBAF9C,uIAE6E,EAE/E,GACE,CAAC,GACD,OAAO,IAAW,UAClB,EAAE,UAAW,IACb,EAAE,gBAAiB,GAEnB,MAAU,MAAM,wDAAwD,EAG1E,EAAO,KAAK,mCAAmC,EAAO,QAAQ,EAElE,EACA,CACE,KAAM,4BACN,GAAI,MAAO,IAA0C,CAMnD,IAAM,EAAc,MADH,MAAM,MAFrB,wEAEmC,GACF,YAAY,EACzC,EAAc,OAAO,KAAK,IAAI,WAAW,CAAW,CAAC,EAErD,EAAgB,MAAM,EAAQ,SAAS,EAAU,cAAe,CAAW,EAEjF,GAAI,OAAO,IAAkB,SAC3B,MAAU,MAAM,sCAAsC,EAGxD,EAAO,KAAK,iCAAiC,EAAc,UAAU,EAAG,EAAE,OAAO,EAErF,EACA,CACE,KAAM,6BACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAY,MAAM,EAAQ,SAAS,EAAU,eAAgB,CACjE,KAAM,uCACR,CAAC,EAED,GAAI,EAAE,aAAqB,cAAgB,EAAU,aAAe,EAClE,MAAU,MAAM,yCAAyC,EAG3D,EAAO,KAAK,+BAA+B,EAAU,2BAA2B,EAEpF,EACA,CACE,KAAM,gCACN,GAAI,MAAO,IAA0C,CACnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,aAAc,CAC5D,OACE,+FACJ,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,SAC/B,MAAU,MAAM,2CAA2C,EAG7D,EAAO,KACL,mCAAmC,KAAK,UAAU,CAAM,EAAE,UAAU,EAAG,GAAG,GAC5E,EAEJ,EACA,CACE,KAAM,uBACN,GAAI,MAAO,IAA0C,CAGnD,IAAM,EAAS,MAAM,EAAQ,SAAS,EAAU,SAAU,CACxD,MAAO,qCACP,MAAO,CAAC,CAAE,KAAM,oBAAqB,CAAC,EACtC,aAAc,CAChB,CAAC,EAED,GAAI,CAAC,GAAU,OAAO,IAAW,UAAY,EAAE,SAAU,GACvD,MAAU,MAAM,qDAAqD,EAGvE,GAAI,OAAO,EAAO,OAAS,UAAY,EAAO,KAAK,SAAW,EAC5D,MAAU,MAAM,mDAAmD,EAGrE,EAAO,KACL,kDAAkD,EAAO,KAAK,wBAAwB,EAAO,aAAa,QAAU,GACtH,EAEJ,CACF,CACF,CACF,CACF,EAEe",
|
|
22
|
+
"debugId": "01A431D38E4333C464756E2164756E21",
|
|
23
23
|
"names": []
|
|
24
24
|
}
|
package/dist/cjs/index.node.cjs
CHANGED
|
@@ -427,19 +427,24 @@ function truncatePrompt(prompt) {
|
|
|
427
427
|
}
|
|
428
428
|
function normalizeUsage(usage) {
|
|
429
429
|
if ("promptTokens" in usage) {
|
|
430
|
+
const promptTokensDetails = "promptTokensDetails" in usage ? usage.promptTokensDetails : undefined;
|
|
431
|
+
const cachedPromptTokens = usage.cachedPromptTokens ?? promptTokensDetails?.cachedTokens;
|
|
430
432
|
return {
|
|
431
433
|
promptTokens: usage.promptTokens ?? 0,
|
|
432
434
|
completionTokens: usage.completionTokens ?? 0,
|
|
433
|
-
totalTokens: usage.totalTokens ?? (usage.promptTokens ?? 0) + (usage.completionTokens ?? 0)
|
|
435
|
+
totalTokens: usage.totalTokens ?? (usage.promptTokens ?? 0) + (usage.completionTokens ?? 0),
|
|
436
|
+
cachedPromptTokens
|
|
434
437
|
};
|
|
435
438
|
}
|
|
436
439
|
if ("inputTokens" in usage || "outputTokens" in usage) {
|
|
437
440
|
const input = usage.inputTokens ?? 0;
|
|
438
441
|
const output = usage.outputTokens ?? 0;
|
|
442
|
+
const total = usage.totalTokens ?? input + output;
|
|
439
443
|
return {
|
|
440
444
|
promptTokens: input,
|
|
441
445
|
completionTokens: output,
|
|
442
|
-
totalTokens:
|
|
446
|
+
totalTokens: total,
|
|
447
|
+
cachedPromptTokens: usage.cachedInputTokens
|
|
443
448
|
};
|
|
444
449
|
}
|
|
445
450
|
return {
|
|
@@ -459,7 +464,8 @@ function emitModelUsageEvent(runtime, type, prompt, usage) {
|
|
|
459
464
|
tokens: {
|
|
460
465
|
prompt: normalized.promptTokens,
|
|
461
466
|
completion: normalized.completionTokens,
|
|
462
|
-
total: normalized.totalTokens
|
|
467
|
+
total: normalized.totalTokens,
|
|
468
|
+
...normalized.cachedPromptTokens !== undefined ? { cached: normalized.cachedPromptTokens } : {}
|
|
463
469
|
}
|
|
464
470
|
};
|
|
465
471
|
runtime.emitEvent(import_core5.EventType.MODEL_USED, payload);
|
|
@@ -956,16 +962,26 @@ function convertUsage(usage) {
|
|
|
956
962
|
}
|
|
957
963
|
const promptTokens = usage.inputTokens ?? 0;
|
|
958
964
|
const completionTokens = usage.outputTokens ?? 0;
|
|
965
|
+
const usageWithCache = usage;
|
|
959
966
|
return {
|
|
960
967
|
promptTokens,
|
|
961
968
|
completionTokens,
|
|
962
|
-
totalTokens: promptTokens + completionTokens
|
|
969
|
+
totalTokens: promptTokens + completionTokens,
|
|
970
|
+
cachedPromptTokens: usageWithCache.cachedInputTokens
|
|
971
|
+
};
|
|
972
|
+
}
|
|
973
|
+
function resolvePromptCacheOptions(params) {
|
|
974
|
+
const withOpenAIOptions = params;
|
|
975
|
+
return {
|
|
976
|
+
promptCacheKey: withOpenAIOptions.providerOptions?.openai?.promptCacheKey,
|
|
977
|
+
promptCacheRetention: withOpenAIOptions.providerOptions?.openai?.promptCacheRetention
|
|
963
978
|
};
|
|
964
979
|
}
|
|
965
980
|
async function generateTextByModelType(runtime, params, modelType, getModelFn) {
|
|
966
981
|
const openai = createOpenAIClient(runtime);
|
|
967
982
|
const modelName = getModelFn(runtime);
|
|
968
983
|
import_core11.logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);
|
|
984
|
+
const promptCacheOptions = resolvePromptCacheOptions(params);
|
|
969
985
|
const systemPrompt = runtime.character.system ?? undefined;
|
|
970
986
|
const model = openai.chat(modelName);
|
|
971
987
|
const generateParams = {
|
|
@@ -973,7 +989,15 @@ async function generateTextByModelType(runtime, params, modelType, getModelFn) {
|
|
|
973
989
|
prompt: params.prompt,
|
|
974
990
|
system: systemPrompt,
|
|
975
991
|
maxOutputTokens: params.maxTokens ?? 8192,
|
|
976
|
-
experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) }
|
|
992
|
+
experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },
|
|
993
|
+
...promptCacheOptions.promptCacheKey || promptCacheOptions.promptCacheRetention ? {
|
|
994
|
+
providerOptions: {
|
|
995
|
+
openai: {
|
|
996
|
+
...promptCacheOptions.promptCacheKey ? { promptCacheKey: promptCacheOptions.promptCacheKey } : {},
|
|
997
|
+
...promptCacheOptions.promptCacheRetention ? { promptCacheRetention: promptCacheOptions.promptCacheRetention } : {}
|
|
998
|
+
}
|
|
999
|
+
}
|
|
1000
|
+
} : {}
|
|
977
1001
|
};
|
|
978
1002
|
if (params.stream) {
|
|
979
1003
|
const result = import_ai3.streamText(generateParams);
|
|
@@ -1288,4 +1312,4 @@ var openaiPlugin = {
|
|
|
1288
1312
|
};
|
|
1289
1313
|
var typescript_default = openaiPlugin;
|
|
1290
1314
|
|
|
1291
|
-
//# debugId=
|
|
1315
|
+
//# debugId=23E20BA2231FE80864756E2164756E21
|