@elizaos/plugin-openai 1.6.0 → 2.0.0-alpha.10

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (55) hide show
  1. package/dist/browser/index.browser.js +2 -2
  2. package/dist/browser/index.browser.js.map +18 -17
  3. package/dist/build.d.ts +13 -0
  4. package/dist/build.d.ts.map +1 -0
  5. package/dist/cjs/index.node.cjs +1027 -658
  6. package/dist/cjs/index.node.js.map +18 -17
  7. package/dist/generated/specs/specs.d.ts +55 -0
  8. package/dist/generated/specs/specs.d.ts.map +1 -0
  9. package/dist/index.browser.d.ts +1 -0
  10. package/dist/index.browser.d.ts.map +1 -0
  11. package/dist/index.d.ts +1 -5
  12. package/dist/index.d.ts.map +1 -0
  13. package/dist/index.node.d.ts +1 -0
  14. package/dist/index.node.d.ts.map +1 -0
  15. package/dist/init.d.ts +4 -5
  16. package/dist/init.d.ts.map +1 -0
  17. package/dist/models/audio.d.ts +9 -10
  18. package/dist/models/audio.d.ts.map +1 -0
  19. package/dist/models/embedding.d.ts +1 -3
  20. package/dist/models/embedding.d.ts.map +1 -0
  21. package/dist/models/image.d.ts +4 -13
  22. package/dist/models/image.d.ts.map +1 -0
  23. package/dist/models/index.d.ts +7 -5
  24. package/dist/models/index.d.ts.map +1 -0
  25. package/dist/models/object.d.ts +4 -9
  26. package/dist/models/object.d.ts.map +1 -0
  27. package/dist/models/research.d.ts +34 -0
  28. package/dist/models/research.d.ts.map +1 -0
  29. package/dist/models/text.d.ts +22 -3
  30. package/dist/models/text.d.ts.map +1 -0
  31. package/dist/models/tokenizer.d.ts +4 -9
  32. package/dist/models/tokenizer.d.ts.map +1 -0
  33. package/dist/node/index.node.js +1016 -644
  34. package/dist/node/index.node.js.map +18 -17
  35. package/dist/providers/index.d.ts +2 -1
  36. package/dist/providers/index.d.ts.map +1 -0
  37. package/dist/providers/openai.d.ts +3 -7
  38. package/dist/providers/openai.d.ts.map +1 -0
  39. package/dist/types/index.d.ts +329 -10
  40. package/dist/types/index.d.ts.map +1 -0
  41. package/dist/utils/audio.d.ts +6 -12
  42. package/dist/utils/audio.d.ts.map +1 -0
  43. package/dist/utils/config.d.ts +16 -59
  44. package/dist/utils/config.d.ts.map +1 -0
  45. package/dist/utils/events.d.ts +20 -9
  46. package/dist/utils/events.d.ts.map +1 -0
  47. package/dist/utils/index.d.ts +2 -1
  48. package/dist/utils/index.d.ts.map +1 -0
  49. package/dist/utils/json.d.ts +9 -6
  50. package/dist/utils/json.d.ts.map +1 -0
  51. package/dist/utils/tokenization.d.ts +5 -16
  52. package/dist/utils/tokenization.d.ts.map +1 -0
  53. package/package.json +37 -29
  54. package/LICENSE +0 -21
  55. package/README.md +0 -160
@@ -1,23 +1,24 @@
1
1
  {
2
2
  "version": 3,
3
- "sources": ["../../src/index.ts", "../../src/init.ts", "../../src/utils/config.ts", "../../src/models/text.ts", "../../src/providers/openai.ts", "../../src/utils/events.ts", "../../src/models/embedding.ts", "../../src/models/image.ts", "../../src/models/audio.ts", "../../src/utils/audio.ts", "../../src/models/object.ts", "../../src/utils/json.ts", "../../src/models/tokenizer.ts", "../../src/utils/tokenization.ts"],
3
+ "sources": ["../../index.ts", "../../init.ts", "../../utils/config.ts", "../../models/audio.ts", "../../utils/audio.ts", "../../models/embedding.ts", "../../utils/events.ts", "../../models/image.ts", "../../models/object.ts", "../../providers/openai.ts", "../../utils/json.ts", "../../models/research.ts", "../../models/text.ts", "../../models/tokenizer.ts", "../../utils/tokenization.ts"],
4
4
  "sourcesContent": [
5
- "import type {\n DetokenizeTextParams,\n GenerateTextParams,\n IAgentRuntime,\n ImageDescriptionParams,\n ObjectGenerationParams,\n Plugin,\n TextEmbeddingParams,\n TokenizeTextParams,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { initializeOpenAI } from \"./init\";\nimport {\n handleTextSmall,\n handleTextLarge,\n handleTextEmbedding,\n handleImageGeneration,\n handleImageDescription,\n handleTranscription,\n handleTextToSpeech,\n handleObjectSmall,\n handleObjectLarge,\n handleTokenizerEncode,\n handleTokenizerDecode,\n} from \"./models\";\nimport { getBaseURL, getAuthHeader } from \"./utils/config\";\n\nexport * from \"./types\";\n\n/**\n * Defines the OpenAI plugin with its name, description, and configuration options.\n * @type {Plugin}\n */\nexport const openaiPlugin: Plugin = {\n name: \"openai\",\n description: \"OpenAI plugin\",\n config: {\n OPENAI_API_KEY: process.env.OPENAI_API_KEY,\n OPENAI_BASE_URL: process.env.OPENAI_BASE_URL,\n OPENAI_SMALL_MODEL: process.env.OPENAI_SMALL_MODEL,\n OPENAI_LARGE_MODEL: process.env.OPENAI_LARGE_MODEL,\n SMALL_MODEL: process.env.SMALL_MODEL,\n LARGE_MODEL: process.env.LARGE_MODEL,\n OPENAI_EMBEDDING_MODEL: process.env.OPENAI_EMBEDDING_MODEL,\n OPENAI_EMBEDDING_API_KEY: process.env.OPENAI_EMBEDDING_API_KEY,\n OPENAI_EMBEDDING_URL: process.env.OPENAI_EMBEDDING_URL,\n OPENAI_EMBEDDING_DIMENSIONS: process.env.OPENAI_EMBEDDING_DIMENSIONS,\n OPENAI_IMAGE_DESCRIPTION_MODEL: process.env.OPENAI_IMAGE_DESCRIPTION_MODEL,\n OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS:\n process.env.OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS,\n OPENAI_EXPERIMENTAL_TELEMETRY: process.env.OPENAI_EXPERIMENTAL_TELEMETRY,\n },\n async init(_config, runtime) {\n // Note: We intentionally don't await here because ElizaOS expects\n // the init method to return quickly. The initializeOpenAI function\n // performs background validation and logging.\n initializeOpenAI(_config, runtime);\n },\n\n models: {\n [ModelType.TEXT_EMBEDDING]: async (\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null,\n ) => {\n return handleTextEmbedding(runtime, params);\n },\n [ModelType.TEXT_TOKENIZER_ENCODE]: async (\n runtime: IAgentRuntime,\n params: TokenizeTextParams,\n ) => {\n return handleTokenizerEncode(runtime, params);\n },\n [ModelType.TEXT_TOKENIZER_DECODE]: async (\n runtime: IAgentRuntime,\n params: DetokenizeTextParams,\n ) => {\n return handleTokenizerDecode(runtime, params);\n },\n [ModelType.TEXT_SMALL]: async (\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n ) => {\n return handleTextSmall(runtime, params);\n },\n [ModelType.TEXT_LARGE]: async (\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n ) => {\n return handleTextLarge(runtime, params);\n },\n [ModelType.IMAGE]: async (\n runtime: IAgentRuntime,\n params: {\n prompt: string;\n n?: number;\n size?: string;\n },\n ) => {\n return handleImageGeneration(runtime, params);\n },\n [ModelType.IMAGE_DESCRIPTION]: async (\n runtime: IAgentRuntime,\n params: ImageDescriptionParams | string,\n ) => {\n return handleImageDescription(runtime, params);\n },\n [ModelType.TRANSCRIPTION]: async (\n runtime: IAgentRuntime,\n input: Blob | File | Buffer | any,\n ) => {\n return handleTranscription(runtime, input);\n },\n [ModelType.TEXT_TO_SPEECH]: async (\n runtime: IAgentRuntime,\n input: string | any,\n ) => {\n return handleTextToSpeech(runtime, input);\n },\n [ModelType.OBJECT_SMALL]: async (\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n ) => {\n return handleObjectSmall(runtime, params);\n },\n [ModelType.OBJECT_LARGE]: async (\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n ) => {\n return handleObjectLarge(runtime, params);\n },\n },\n tests: [\n {\n name: \"openai_plugin_tests\",\n tests: [\n {\n name: \"openai_test_url_and_api_key_validation\",\n fn: async (runtime: IAgentRuntime) => {\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/models`, {\n headers: getAuthHeader(runtime),\n });\n const data = await response.json();\n logger.log(\n { data: (data as { data?: unknown[] })?.data?.length ?? \"N/A\" },\n \"Models Available\",\n );\n if (!response.ok) {\n throw new Error(\n `Failed to validate OpenAI API key: ${response.statusText}`,\n );\n }\n },\n },\n {\n name: \"openai_test_text_embedding\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n const embedding = await runtime.useModel(\n ModelType.TEXT_EMBEDDING,\n {\n text: \"Hello, world!\",\n },\n );\n logger.log({ embedding }, \"embedding\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in test_text_embedding: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"openai_test_text_large\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n const text = await runtime.useModel(ModelType.TEXT_LARGE, {\n prompt: \"What is the nature of reality in 10 words?\",\n });\n if (text.length === 0) {\n throw new Error(\"Failed to generate text\");\n }\n logger.log({ text }, \"generated with test_text_large\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in test_text_large: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"openai_test_text_small\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n const text = await runtime.useModel(ModelType.TEXT_SMALL, {\n prompt: \"What is the nature of reality in 10 words?\",\n });\n if (text.length === 0) {\n throw new Error(\"Failed to generate text\");\n }\n logger.log({ text }, \"generated with test_text_small\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in test_text_small: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"openai_test_image_generation\",\n fn: async (runtime: IAgentRuntime) => {\n logger.log(\"openai_test_image_generation\");\n try {\n const image = await runtime.useModel(ModelType.IMAGE, {\n prompt: \"A beautiful sunset over a calm ocean\",\n count: 1,\n size: \"1024x1024\",\n });\n logger.log({ image }, \"generated with test_image_generation\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in test_image_generation: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"image-description\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n logger.log(\"openai_test_image_description\");\n try {\n const result = await runtime.useModel(\n ModelType.IMAGE_DESCRIPTION,\n \"https://upload.wikimedia.org/wikipedia/commons/thumb/1/1c/Vitalik_Buterin_TechCrunch_London_2015_%28cropped%29.jpg/537px-Vitalik_Buterin_TechCrunch_London_2015_%28cropped%29.jpg\",\n );\n\n if (\n result &&\n typeof result === \"object\" &&\n \"title\" in result &&\n \"description\" in result\n ) {\n logger.log({ result }, \"Image description\");\n } else {\n logger.error(\n \"Invalid image description result format:\",\n result,\n );\n }\n } catch (e: unknown) {\n const message = e instanceof Error ? e.message : String(e);\n logger.error(`Error in image description test: ${message}`);\n }\n } catch (e: unknown) {\n const message = e instanceof Error ? e.message : String(e);\n logger.error(\n `Error in openai_test_image_description: ${message}`,\n );\n }\n },\n },\n {\n name: \"openai_test_transcription\",\n fn: async (runtime: IAgentRuntime) => {\n logger.log(\"openai_test_transcription\");\n try {\n const response = await fetch(\n \"https://upload.wikimedia.org/wikipedia/en/4/40/Chris_Benoit_Voice_Message.ogg\",\n );\n const arrayBuffer = await response.arrayBuffer();\n const transcription = await runtime.useModel(\n ModelType.TRANSCRIPTION,\n Buffer.from(new Uint8Array(arrayBuffer)),\n );\n logger.log(\n { transcription },\n \"generated with test_transcription\",\n );\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in test_transcription: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"openai_test_text_tokenizer_encode\",\n fn: async (runtime: IAgentRuntime) => {\n const prompt = \"Hello tokenizer encode!\";\n const tokens = await runtime.useModel(\n ModelType.TEXT_TOKENIZER_ENCODE,\n { prompt, modelType: ModelType.TEXT_SMALL },\n );\n if (!Array.isArray(tokens) || tokens.length === 0) {\n throw new Error(\n \"Failed to tokenize text: expected non-empty array of tokens\",\n );\n }\n logger.log({ tokens }, \"Tokenized output\");\n },\n },\n {\n name: \"openai_test_text_tokenizer_decode\",\n fn: async (runtime: IAgentRuntime) => {\n const prompt = \"Hello tokenizer decode!\";\n const tokens = await runtime.useModel(\n ModelType.TEXT_TOKENIZER_ENCODE,\n { prompt, modelType: ModelType.TEXT_SMALL },\n );\n const decodedText = await runtime.useModel(\n ModelType.TEXT_TOKENIZER_DECODE,\n {\n tokens,\n modelType: ModelType.TEXT_SMALL,\n },\n );\n if (decodedText !== prompt) {\n throw new Error(\n `Decoded text does not match original. Expected \"${prompt}\", got \"${decodedText}\"`,\n );\n }\n logger.log({ decodedText }, \"Decoded text\");\n },\n },\n {\n name: \"openai_test_text_to_speech\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n const response = await runtime.useModel(\n ModelType.TEXT_TO_SPEECH,\n {\n text: \"Hello, this is a test for text-to-speech.\",\n },\n );\n if (!response) {\n throw new Error(\"Failed to generate speech\");\n }\n logger.log(\"Generated speech successfully\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in openai_test_text_to_speech: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"openai_test_text_generation_large\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n const result = await runtime.useModel(ModelType.TEXT_LARGE, {\n prompt: \"Say hello in 5 words.\",\n });\n if (!result || result.length === 0) {\n throw new Error(\"Text generation returned empty result\");\n }\n logger.log({ result }, \"Text generation test completed\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in openai_test_text_generation_large: ${message}`);\n throw error;\n }\n },\n },\n {\n name: \"openai_test_streaming\",\n fn: async (runtime: IAgentRuntime) => {\n try {\n const chunks: string[] = [];\n const result = await runtime.useModel(ModelType.TEXT_LARGE, {\n prompt: \"Count from 1 to 5.\",\n onStreamChunk: (chunk: string) => {\n chunks.push(chunk);\n },\n });\n if (!result || result.length === 0) {\n throw new Error(\"Streaming returned empty result\");\n }\n if (chunks.length === 0) {\n throw new Error(\"No streaming chunks received\");\n }\n logger.log({ chunks: chunks.length, result: result.substring(0, 50) }, \"Streaming test completed\");\n } catch (error: unknown) {\n const message =\n error instanceof Error ? error.message : String(error);\n logger.error(`Error in openai_test_streaming: ${message}`);\n throw error;\n }\n },\n },\n ],\n },\n ],\n};\n\nexport default openaiPlugin;\n",
6
- "import { logger, type IAgentRuntime } from \"@elizaos/core\";\nimport {\n getApiKey,\n getAuthHeader,\n getBaseURL,\n isBrowser,\n} from \"./utils/config\";\n\n/**\n * Initialize and validate OpenAI configuration\n */\nexport function initializeOpenAI(_config: any, runtime: IAgentRuntime) {\n // Run validation in the background without blocking initialization\n void (async () => {\n try {\n if (!getApiKey(runtime) && !isBrowser()) {\n logger.warn(\n \"OPENAI_API_KEY is not set in environment - OpenAI functionality will be limited\",\n );\n return;\n }\n try {\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/models`, {\n headers: getAuthHeader(runtime),\n });\n if (!response.ok) {\n logger.warn(\n `OpenAI API key validation failed: ${response.statusText}`,\n );\n logger.warn(\n \"OpenAI functionality will be limited until a valid API key is provided\",\n );\n } else {\n logger.log(\"OpenAI API key validated successfully\");\n }\n } catch (fetchError: unknown) {\n const message =\n fetchError instanceof Error ? fetchError.message : String(fetchError);\n logger.warn(`Error validating OpenAI API key: ${message}`);\n logger.warn(\n \"OpenAI functionality will be limited until a valid API key is provided\",\n );\n }\n } catch (error: unknown) {\n const message =\n (error as { errors?: Array<{ message: string }> })?.errors\n ?.map((e) => e.message)\n .join(\", \") ||\n (error instanceof Error ? error.message : String(error));\n logger.warn(\n `OpenAI plugin configuration issue: ${message} - You need to configure the OPENAI_API_KEY in your environment variables`,\n );\n }\n })();\n}\n",
7
- "import type { IAgentRuntime } from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\n\n/**\n * Retrieves a configuration setting from the runtime, falling back to environment variables or a default value if not found.\n *\n * @param key - The name of the setting to retrieve.\n * @param defaultValue - The value to return if the setting is not found in the runtime or environment.\n * @returns The resolved setting value, or {@link defaultValue} if not found.\n */\nexport function getSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue?: string,\n): string | undefined {\n const value = runtime.getSetting(key);\n // Convert to string if value is a number or boolean\n if (value !== undefined && value !== null) {\n return String(value);\n }\n return process.env[key] ?? defaultValue;\n}\n\nexport function isBrowser(): boolean {\n return (\n typeof globalThis !== \"undefined\" &&\n \"document\" in (globalThis as any) &&\n typeof (globalThis as any).document !== \"undefined\"\n );\n}\n\n/**\n * Determines whether we're running in a browser with a server-hosted proxy configured.\n * In this mode, we do not require a real API key on the client and rely on the proxy to inject it.\n */\nexport function isProxyMode(runtime: IAgentRuntime): boolean {\n return isBrowser() && !!getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n}\n\nexport function getAuthHeader(\n runtime: IAgentRuntime,\n forEmbedding = false,\n): Record<string, string> {\n if (isBrowser()) return {};\n const key = forEmbedding ? getEmbeddingApiKey(runtime) : getApiKey(runtime);\n return key ? { Authorization: `Bearer ${key}` } : {};\n}\n\n/**\n * Retrieves the OpenAI API base URL from runtime settings, environment variables, or defaults, using provider-aware resolution.\n *\n * @returns The resolved base URL for OpenAI API requests.\n */\nexport function getBaseURL(runtime: IAgentRuntime): string {\n const browserURL = getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n const baseURL = (\n isBrowser() && browserURL\n ? browserURL\n : getSetting(runtime, \"OPENAI_BASE_URL\", \"https://api.openai.com/v1\")\n ) as string;\n logger.debug(`[OpenAI] Default base URL: ${baseURL}`);\n return baseURL;\n}\n\n/**\n * Retrieves the OpenAI API base URL for embeddings, falling back to the general base URL.\n *\n * @returns The resolved base URL for OpenAI embedding requests.\n */\nexport function getEmbeddingBaseURL(runtime: IAgentRuntime): string {\n const embeddingURL = isBrowser()\n ? getSetting(runtime, \"OPENAI_BROWSER_EMBEDDING_URL\") ||\n getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\")\n : getSetting(runtime, \"OPENAI_EMBEDDING_URL\");\n if (embeddingURL) {\n logger.debug(`[OpenAI] Using specific embedding base URL: ${embeddingURL}`);\n return embeddingURL;\n }\n logger.debug(\"[OpenAI] Falling back to general base URL for embeddings.\");\n return getBaseURL(runtime);\n}\n\n/**\n * Helper function to get the API key for OpenAI\n *\n * @param runtime The runtime context\n * @returns The configured API key\n */\nexport function getApiKey(runtime: IAgentRuntime): string | undefined {\n return getSetting(runtime, \"OPENAI_API_KEY\");\n}\n\n/**\n * Helper function to get the embedding API key for OpenAI, falling back to the general API key if not set.\n *\n * @param runtime The runtime context\n * @returns The configured API key\n */\nexport function getEmbeddingApiKey(runtime: IAgentRuntime): string | undefined {\n const embeddingApiKey = getSetting(runtime, \"OPENAI_EMBEDDING_API_KEY\");\n if (embeddingApiKey) {\n logger.debug(\"[OpenAI] Using specific embedding API key (present)\");\n return embeddingApiKey;\n }\n logger.debug(\"[OpenAI] Falling back to general API key for embeddings.\");\n return getApiKey(runtime);\n}\n\n/**\n * Helper function to get the small model name with fallbacks\n *\n * @param runtime The runtime context\n * @returns The configured small model name\n */\nexport function getSmallModel(runtime: IAgentRuntime): string {\n return (\n getSetting(runtime, \"OPENAI_SMALL_MODEL\") ??\n getSetting(runtime, \"SMALL_MODEL\", \"gpt-4o-mini\")!\n );\n}\n\n/**\n * Helper function to get the large model name with fallbacks\n *\n * @param runtime The runtime context\n * @returns The configured large model name\n */\nexport function getLargeModel(runtime: IAgentRuntime): string {\n return (\n getSetting(runtime, \"OPENAI_LARGE_MODEL\") ??\n getSetting(runtime, \"LARGE_MODEL\", \"gpt-4o\")!\n );\n}\n\n/**\n * Helper function to get the image description model name with fallbacks\n *\n * @param runtime The runtime context\n * @returns The configured image description model name\n */\nexport function getImageDescriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MODEL\", \"gpt-5-nano\")!;\n}\n\n/**\n * Helper function to get experimental telemetry setting\n *\n * @param runtime The runtime context\n * @returns Whether experimental telemetry is enabled\n */\nexport function getExperimentalTelemetry(runtime: IAgentRuntime): boolean {\n const setting = getSetting(runtime, \"OPENAI_EXPERIMENTAL_TELEMETRY\", \"false\");\n // Convert to string and check for truthy values\n const normalizedSetting = String(setting).toLowerCase();\n const result = normalizedSetting === \"true\";\n logger.debug(\n `[OpenAI] Experimental telemetry in function: \"${setting}\" (type: ${typeof setting}, normalized: \"${normalizedSetting}\", result: ${result})`,\n );\n return result;\n}\n",
8
- "import type {\n GenerateTextParams,\n IAgentRuntime,\n ModelTypeName,\n TextStreamResult,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateText, streamText, type LanguageModelUsage } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport {\n getSmallModel,\n getLargeModel,\n getExperimentalTelemetry,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\nasync function generateTextByModelType(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n modelType: ModelTypeName,\n getModelFn: (runtime: IAgentRuntime) => string,\n): Promise<string | TextStreamResult> {\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] ${modelType} model: ${modelName}`);\n\n const generateParams = {\n model: openai.languageModel(modelName),\n prompt: params.prompt,\n system: runtime.character.system ?? undefined,\n temperature: params.temperature ?? 0.7,\n maxOutputTokens: params.maxTokens ?? 8192,\n frequencyPenalty: params.frequencyPenalty ?? 0.7,\n presencePenalty: params.presencePenalty ?? 0.7,\n stopSequences: params.stopSequences ?? [],\n experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },\n };\n\n // Streaming mode\n if (params.stream) {\n const result = streamText(generateParams);\n return {\n textStream: result.textStream,\n text: result.text,\n usage: result.usage.then((u: LanguageModelUsage | undefined) => u ? {\n promptTokens: u.inputTokens ?? 0,\n completionTokens: u.outputTokens ?? 0,\n totalTokens: (u.inputTokens ?? 0) + (u.outputTokens ?? 0),\n } : undefined),\n finishReason: result.finishReason,\n };\n }\n\n // Non-streaming mode\n const { text, usage } = await generateText(generateParams);\n if (usage) emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n return text;\n}\n\n/**\n * TEXT_SMALL model handler\n */\nexport async function handleTextSmall(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(\n runtime,\n params,\n ModelType.TEXT_SMALL,\n getSmallModel,\n );\n}\n\n/**\n * TEXT_LARGE model handler\n */\nexport async function handleTextLarge(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(\n runtime,\n params,\n ModelType.TEXT_LARGE,\n getLargeModel,\n );\n}\n",
9
- "import { createOpenAI } from \"@ai-sdk/openai\";\nimport type { IAgentRuntime } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, isProxyMode } from \"../utils/config\";\n\n/**\n * Create an OpenAI client with proper configuration\n *\n * @param runtime The runtime context\n * @returns Configured OpenAI client\n */\nexport function createOpenAIClient(runtime: IAgentRuntime) {\n const baseURL = getBaseURL(runtime);\n // In proxy mode (browser + proxy base URL), pass a harmless placeholder key.\n // The server proxy replaces Authorization; no secrets leave the server.\n const apiKey =\n getApiKey(runtime) ?? (isProxyMode(runtime) ? \"sk-proxy\" : undefined);\n return createOpenAI({ apiKey: (apiKey ?? \"\") as string, baseURL });\n}\n",
10
- "import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { EventType } from \"@elizaos/core\";\nimport type { LanguageModelUsage } from \"ai\";\n\n/**\n * Emits a model usage event\n * @param runtime The runtime context\n * @param type The model type\n * @param prompt The prompt used\n * @param usage The LLM usage data\n */\nexport function emitModelUsageEvent(\n runtime: IAgentRuntime,\n type: ModelTypeName,\n prompt: string,\n usage: LanguageModelUsage,\n) {\n const promptTokens =\n (\"promptTokens\" in usage\n ? (usage as { promptTokens?: number }).promptTokens\n : undefined) ??\n (\"inputTokens\" in usage\n ? (usage as { inputTokens?: number }).inputTokens\n : undefined) ??\n 0;\n const completionTokens =\n (\"completionTokens\" in usage\n ? (usage as { completionTokens?: number }).completionTokens\n : undefined) ??\n (\"outputTokens\" in usage\n ? (usage as { outputTokens?: number }).outputTokens\n : undefined) ??\n 0;\n const totalTokens =\n (\"totalTokens\" in usage\n ? (usage as { totalTokens?: number }).totalTokens\n : undefined) ?? promptTokens + completionTokens;\n\n // Never emit the full prompt; truncate to avoid leaking secrets/PII\n const truncatedPrompt =\n typeof prompt === \"string\"\n ? prompt.length > 200\n ? `${prompt.slice(0, 200)}…`\n : prompt\n : \"\";\n runtime.emitEvent(EventType.MODEL_USED, {\n runtime,\n source: \"openai\",\n provider: \"openai\",\n type,\n prompt: truncatedPrompt,\n tokens: {\n prompt: promptTokens,\n completion: completionTokens,\n total: totalTokens,\n },\n });\n}\n",
11
- "import type { IAgentRuntime, TextEmbeddingParams } from \"@elizaos/core\";\nimport { logger, ModelType, VECTOR_DIMS } from \"@elizaos/core\";\nimport {\n getSetting,\n getEmbeddingBaseURL,\n getAuthHeader,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n/**\n * TEXT_EMBEDDING model handler\n */\nexport async function handleTextEmbedding(\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null,\n): Promise<number[]> {\n const embeddingModelName = getSetting(\n runtime,\n \"OPENAI_EMBEDDING_MODEL\",\n \"text-embedding-3-small\",\n );\n const embeddingDimension = Number.parseInt(\n getSetting(runtime, \"OPENAI_EMBEDDING_DIMENSIONS\", \"1536\") || \"1536\",\n 10,\n ) as (typeof VECTOR_DIMS)[keyof typeof VECTOR_DIMS];\n\n if (!Object.values(VECTOR_DIMS).includes(embeddingDimension)) {\n const errorMsg = `Invalid embedding dimension: ${embeddingDimension}. Must be one of: ${Object.values(VECTOR_DIMS).join(\", \")}`;\n logger.error(errorMsg);\n throw new Error(errorMsg);\n }\n if (params === null) {\n logger.debug(\"Creating test embedding for initialization\");\n const testVector = Array(embeddingDimension).fill(0);\n testVector[0] = 0.1;\n return testVector;\n }\n let text: string;\n if (typeof params === \"string\") {\n text = params;\n } else if (typeof params === \"object\" && params.text) {\n text = params.text;\n } else {\n const errorMsg = \"Invalid input format for embedding\";\n logger.warn(errorMsg);\n const fallbackVector = Array(embeddingDimension).fill(0);\n fallbackVector[0] = 0.2;\n return fallbackVector;\n }\n if (!text.trim()) {\n const errorMsg = \"Empty text for embedding\";\n logger.warn(errorMsg);\n const fallbackVector = Array(embeddingDimension).fill(0);\n fallbackVector[0] = 0.3;\n return fallbackVector;\n }\n\n const embeddingBaseURL = getEmbeddingBaseURL(runtime);\n\n try {\n const response = await fetch(`${embeddingBaseURL}/embeddings`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime, true),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: embeddingModelName,\n input: text,\n }),\n });\n\n if (!response.ok) {\n logger.error(\n `OpenAI API error: ${response.status} - ${response.statusText}`,\n );\n throw new Error(\n `OpenAI API error: ${response.status} - ${response.statusText}`,\n );\n }\n\n const data = (await response.json()) as {\n data: [{ embedding: number[] }];\n usage?: { prompt_tokens: number; total_tokens: number };\n };\n\n if (!data?.data?.[0]?.embedding) {\n logger.error(\"API returned invalid structure\");\n throw new Error(\"API returned invalid structure\");\n }\n\n const embedding = data.data[0].embedding;\n\n if (!Array.isArray(embedding) || embedding.length !== embeddingDimension) {\n const errorMsg = `Embedding length ${embedding?.length ?? 0} does not match configured dimension ${embeddingDimension}`;\n logger.error(errorMsg);\n const fallbackVector = Array(embeddingDimension).fill(0);\n fallbackVector[0] = 0.4;\n return fallbackVector;\n }\n\n if (data.usage) {\n const usage = {\n inputTokens: data.usage.prompt_tokens,\n outputTokens: 0,\n totalTokens: data.usage.total_tokens,\n };\n\n emitModelUsageEvent(runtime, ModelType.TEXT_EMBEDDING, text, usage);\n }\n\n logger.log(`Got valid embedding with length ${embedding.length}`);\n return embedding;\n } catch (error: unknown) {\n const message = error instanceof Error ? error.message : String(error);\n logger.error(`Error generating embedding: ${message}`);\n throw error instanceof Error ? error : new Error(message);\n }\n}\n",
12
- "import type {\n IAgentRuntime,\n ImageDescriptionParams,\n ImageGenerationParams,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport {\n getSetting,\n getBaseURL,\n getAuthHeader,\n getImageDescriptionModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n/**\n * IMAGE generation model handler\n */\nexport async function handleImageGeneration(\n runtime: IAgentRuntime,\n params: ImageGenerationParams,\n): Promise<{ url: string }[]> {\n const n = params.count || 1;\n const size = params.size || \"1024x1024\";\n const prompt = params.prompt;\n const modelName = getSetting(\n runtime,\n \"OPENAI_IMAGE_MODEL\",\n \"gpt-image-1\",\n ) as string;\n logger.log(`[OpenAI] Using IMAGE model: ${modelName}`);\n\n const baseURL = getBaseURL(runtime);\n\n try {\n const response = await fetch(`${baseURL}/images/generations`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: modelName,\n prompt: prompt,\n n: n,\n size: size,\n }),\n });\n\n if (!response.ok) {\n throw new Error(`Failed to generate image: ${response.statusText}`);\n }\n\n const data = await response.json();\n const typedData = data as { data: { url: string }[] };\n\n return typedData.data;\n } catch (error: unknown) {\n const message = error instanceof Error ? error.message : String(error);\n throw error;\n }\n}\n\n/**\n * IMAGE_DESCRIPTION model handler\n */\nexport async function handleImageDescription(\n runtime: IAgentRuntime,\n params: ImageDescriptionParams | string,\n): Promise<{ title: string; description: string }> {\n let imageUrl: string;\n let promptText: string | undefined;\n const modelName = getImageDescriptionModel(runtime);\n logger.log(`[OpenAI] Using IMAGE_DESCRIPTION model: ${modelName}`);\n const maxTokens = Number.parseInt(\n getSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS\", \"8192\") ||\n \"8192\",\n 10,\n );\n\n const DEFAULT_PROMPT =\n \"Please analyze this image and provide a title and detailed description.\";\n\n if (typeof params === \"string\") {\n imageUrl = params;\n promptText = DEFAULT_PROMPT;\n } else {\n imageUrl = params.imageUrl;\n promptText = params.prompt || DEFAULT_PROMPT;\n }\n\n const messages = [\n {\n role: \"user\",\n content: [\n { type: \"text\", text: promptText },\n { type: \"image_url\", image_url: { url: imageUrl } },\n ],\n },\n ];\n\n const baseURL = getBaseURL(runtime);\n\n try {\n const requestBody: Record<string, any> = {\n model: modelName,\n messages: messages,\n max_tokens: maxTokens,\n };\n\n const response = await fetch(`${baseURL}/chat/completions`, {\n method: \"POST\",\n headers: {\n \"Content-Type\": \"application/json\",\n ...getAuthHeader(runtime),\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n throw new Error(`OpenAI API error: ${response.status}`);\n }\n\n const result: unknown = await response.json();\n\n type OpenAIResponseType = {\n choices?: Array<{\n message?: { content?: string };\n finish_reason?: string;\n }>;\n usage?: {\n prompt_tokens: number;\n completion_tokens: number;\n total_tokens: number;\n };\n };\n\n const typedResult = result as OpenAIResponseType;\n const content = typedResult.choices?.[0]?.message?.content;\n\n if (typedResult.usage) {\n emitModelUsageEvent(\n runtime,\n ModelType.IMAGE_DESCRIPTION,\n typeof params === \"string\" ? params : params.prompt || \"\",\n {\n inputTokens: typedResult.usage.prompt_tokens,\n outputTokens: typedResult.usage.completion_tokens,\n totalTokens: typedResult.usage.total_tokens,\n },\n );\n }\n\n if (!content) {\n return {\n title: \"Failed to analyze image\",\n description: \"No response from API\",\n };\n }\n const titleMatch = content.match(/title[:\\s]+(.+?)(?:\\n|$)/i);\n const title = titleMatch?.[1]?.trim();\n if (!title) {\n logger.warn(\"Could not extract title from image description response\");\n }\n const finalTitle = title || \"Image Analysis\";\n const description = content.replace(/title[:\\s]+(.+?)(?:\\n|$)/i, \"\").trim();\n\n const processedResult = { title: finalTitle, description };\n return processedResult;\n } catch (error: unknown) {\n const message = error instanceof Error ? error.message : String(error);\n logger.error(`Error analyzing image: ${message}`);\n return {\n title: \"Failed to analyze image\",\n description: `Error: ${message}`,\n };\n }\n}\n",
13
- "import type { IAgentRuntime } from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport type {\n OpenAITranscriptionParams,\n OpenAITextToSpeechParams,\n} from \"../types\";\nimport { getSetting, getBaseURL, getAuthHeader } from \"../utils/config\";\nimport { detectAudioMimeType } from \"../utils/audio\";\n\n/**\n * Helper function for text-to-speech\n */\nasync function fetchTextToSpeech(\n runtime: IAgentRuntime,\n options: OpenAITextToSpeechParams,\n): Promise<ArrayBuffer> {\n const defaultModel = getSetting(\n runtime,\n \"OPENAI_TTS_MODEL\",\n \"gpt-4o-mini-tts\",\n );\n const defaultVoice = getSetting(runtime, \"OPENAI_TTS_VOICE\", \"nova\");\n const defaultInstructions = getSetting(\n runtime,\n \"OPENAI_TTS_INSTRUCTIONS\",\n \"\",\n );\n const baseURL = getBaseURL(runtime);\n\n const model = options.model || (defaultModel as string);\n const voice = options.voice || (defaultVoice as string);\n const instructions = options.instructions ?? (defaultInstructions as string);\n const format = options.format || \"mp3\";\n\n try {\n const res = await fetch(`${baseURL}/audio/speech`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n // Hint desired audio format in Accept when possible\n ...(format === \"mp3\" ? { Accept: \"audio/mpeg\" } : {}),\n },\n body: JSON.stringify({\n model,\n voice,\n input: options.text,\n format,\n ...(instructions && { instructions }),\n }),\n });\n\n if (!res.ok) {\n const err = await res.text();\n throw new Error(`OpenAI TTS error ${res.status}: ${err}`);\n }\n\n // Return ArrayBuffer to match core type expectations\n return await res.arrayBuffer();\n } catch (err: unknown) {\n const message = err instanceof Error ? err.message : String(err);\n throw new Error(`Failed to fetch speech from OpenAI TTS: ${message}`);\n }\n}\n\n/**\n * TRANSCRIPTION model handler\n */\nexport async function handleTranscription(\n runtime: IAgentRuntime,\n input: Blob | File | Buffer | OpenAITranscriptionParams,\n): Promise<string> {\n let modelName = getSetting(\n runtime,\n \"OPENAI_TRANSCRIPTION_MODEL\",\n \"gpt-4o-mini-transcribe\",\n );\n logger.log(`[OpenAI] Using TRANSCRIPTION model: ${modelName}`);\n\n const baseURL = getBaseURL(runtime);\n\n // Support Blob/File/Buffer directly, or an object with { audio: Blob/File/Buffer, ...options }\n let blob: Blob;\n let extraParams: OpenAITranscriptionParams | null = null;\n\n if (input instanceof Blob || input instanceof File) {\n blob = input as Blob;\n } else if (Buffer.isBuffer(input)) {\n // Convert Buffer to Blob for Node.js environments\n // Auto-detect MIME type from buffer content\n const detectedMimeType = detectAudioMimeType(input);\n logger.debug(`Auto-detected audio MIME type: ${detectedMimeType}`);\n // Create a new Uint8Array from the Buffer to ensure type compatibility\n const uint8Array = new Uint8Array(input);\n blob = new Blob([uint8Array], { type: detectedMimeType });\n } else if (\n typeof input === \"object\" &&\n input !== null &&\n (input as any).audio != null\n ) {\n const params = input as any;\n if (\n !(params.audio instanceof Blob) &&\n !(params.audio instanceof File) &&\n !Buffer.isBuffer(params.audio)\n ) {\n throw new Error(\n \"TRANSCRIPTION param 'audio' must be a Blob/File/Buffer.\",\n );\n }\n // Convert Buffer to Blob if needed\n if (Buffer.isBuffer(params.audio)) {\n // Use provided mimeType or auto-detect from buffer\n let mimeType = params.mimeType;\n if (!mimeType) {\n mimeType = detectAudioMimeType(params.audio);\n logger.debug(`Auto-detected audio MIME type: ${mimeType}`);\n } else {\n logger.debug(`Using provided MIME type: ${mimeType}`);\n }\n // Create a new Uint8Array from the Buffer to ensure type compatibility\n const uint8Array = new Uint8Array(params.audio);\n blob = new Blob([uint8Array], { type: mimeType });\n } else {\n blob = params.audio as Blob;\n }\n extraParams = params as OpenAITranscriptionParams;\n if (typeof params.model === \"string\" && params.model) {\n modelName = params.model;\n }\n } else {\n throw new Error(\n \"TRANSCRIPTION expects a Blob/File/Buffer or an object { audio: Blob/File/Buffer, mimeType?, language?, response_format?, timestampGranularities?, prompt?, temperature?, model? }\",\n );\n }\n\n const mime = (blob as File).type || \"audio/webm\";\n const filename =\n (blob as File).name ||\n (mime.includes(\"mp3\") || mime.includes(\"mpeg\")\n ? \"recording.mp3\"\n : mime.includes(\"ogg\")\n ? \"recording.ogg\"\n : mime.includes(\"wav\")\n ? \"recording.wav\"\n : mime.includes(\"webm\")\n ? \"recording.webm\"\n : \"recording.bin\");\n\n const formData = new FormData();\n formData.append(\"file\", blob, filename);\n formData.append(\"model\", String(modelName));\n if (extraParams) {\n if (typeof extraParams.language === \"string\") {\n formData.append(\"language\", String(extraParams.language));\n }\n if (typeof extraParams.response_format === \"string\") {\n formData.append(\"response_format\", String(extraParams.response_format));\n }\n if (typeof extraParams.prompt === \"string\") {\n formData.append(\"prompt\", String(extraParams.prompt));\n }\n if (typeof extraParams.temperature === \"number\") {\n formData.append(\"temperature\", String(extraParams.temperature));\n }\n if (Array.isArray(extraParams.timestampGranularities)) {\n for (const g of extraParams.timestampGranularities) {\n formData.append(\"timestamp_granularities[]\", String(g));\n }\n }\n }\n\n try {\n const response = await fetch(`${baseURL}/audio/transcriptions`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n },\n body: formData,\n });\n\n if (!response.ok) {\n throw new Error(\n `Failed to transcribe audio: ${response.status} ${response.statusText}`,\n );\n }\n\n const data = (await response.json()) as { text: string };\n return data.text || \"\";\n } catch (error: unknown) {\n const message = error instanceof Error ? error.message : String(error);\n logger.error(`TRANSCRIPTION error: ${message}`);\n throw error;\n }\n}\n\n/**\n * TEXT_TO_SPEECH model handler\n */\nexport async function handleTextToSpeech(\n runtime: IAgentRuntime,\n input: string | OpenAITextToSpeechParams,\n): Promise<ArrayBuffer> {\n // Normalize input into options with per-call overrides\n const options: OpenAITextToSpeechParams =\n typeof input === \"string\"\n ? { text: input }\n : (input as OpenAITextToSpeechParams);\n\n const resolvedModel =\n options.model ||\n (getSetting(runtime, \"OPENAI_TTS_MODEL\", \"gpt-4o-mini-tts\") as string);\n logger.log(`[OpenAI] Using TEXT_TO_SPEECH model: ${resolvedModel}`);\n try {\n return await fetchTextToSpeech(runtime, options);\n } catch (error: unknown) {\n const message = error instanceof Error ? error.message : String(error);\n logger.error(`Error in TEXT_TO_SPEECH: ${message}`);\n throw error;\n }\n}\n",
14
- "import { logger } from \"@elizaos/core\";\n\nconst MAGIC_BYTES = {\n WAV: {\n HEADER: [0x52, 0x49, 0x46, 0x46] as const,\n IDENTIFIER: [0x57, 0x41, 0x56, 0x45] as const,\n },\n MP3_ID3: [0x49, 0x44, 0x33] as const,\n OGG: [0x4f, 0x67, 0x67, 0x53] as const,\n FLAC: [0x66, 0x4c, 0x61, 0x43] as const,\n FTYP: [0x66, 0x74, 0x79, 0x70] as const, // at offset 4 for mp4/m4a\n WEBM_EBML: [0x1a, 0x45, 0xdf, 0xa3] as const,\n} as const;\n\nfunction matchBytes(\n buffer: Buffer,\n offset: number,\n bytes: readonly number[],\n): boolean {\n for (let i = 0; i < bytes.length; i++) {\n if (buffer[offset + i] !== bytes[i]!) return false;\n }\n return true;\n}\n\n/**\n * Detects audio MIME type from buffer by checking magic bytes (file signature)\n * @param buffer The audio buffer to analyze\n * @returns The detected MIME type or 'application/octet-stream' if unknown\n */\nexport function detectAudioMimeType(buffer: Buffer): string {\n if (buffer.length < 12) {\n return \"application/octet-stream\";\n }\n\n // Check magic bytes for common audio formats\n // WAV: \"RIFF\" + size + \"WAVE\"\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.WAV.HEADER) &&\n matchBytes(buffer, 8, MAGIC_BYTES.WAV.IDENTIFIER)\n ) {\n return \"audio/wav\";\n }\n\n // MP3: ID3 tag or MPEG frame sync\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.MP3_ID3) || // ID3\n (buffer[0] === 0xff && (buffer[1] & 0xe0) === 0xe0) // MPEG sync\n ) {\n return \"audio/mpeg\";\n }\n\n // OGG: \"OggS\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.OGG)) {\n return \"audio/ogg\";\n }\n\n // FLAC: \"fLaC\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.FLAC)) {\n return \"audio/flac\";\n }\n\n // M4A/MP4: \"ftyp\" at offset 4\n if (matchBytes(buffer, 4, MAGIC_BYTES.FTYP)) {\n return \"audio/mp4\";\n }\n\n // WebM: EBML header\n if (matchBytes(buffer, 0, MAGIC_BYTES.WEBM_EBML)) {\n return \"audio/webm\";\n }\n\n // Unknown format - let API try to detect\n logger.warn(\n \"Could not detect audio format from buffer, using generic binary type\",\n );\n return \"application/octet-stream\";\n}\n\n/**\n * Converts a Web ReadableStream to a Node.js Readable stream\n * Handles both browser and Node.js environments\n * Uses dynamic import to avoid bundling node:stream in browser builds\n */\nexport async function webStreamToNodeStream(\n webStream: ReadableStream<Uint8Array>,\n) {\n try {\n // Dynamic import to avoid browser bundling issues\n const { Readable } = await import(\"node:stream\");\n const reader = webStream.getReader();\n\n return new Readable({\n async read() {\n try {\n const { done, value } = await reader.read();\n if (done) {\n this.push(null);\n } else {\n // Push the Uint8Array directly; Node.js Readable can handle it\n this.push(value);\n }\n } catch (error) {\n this.destroy(error as Error);\n }\n },\n destroy(error, callback) {\n reader.cancel().finally(() => callback(error));\n },\n });\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n logger.error(`Failed to load node:stream module: ${message}`);\n throw new Error(\n `Cannot convert stream: node:stream module unavailable. This feature requires a Node.js environment.`,\n );\n }\n}\n",
15
- "import type {\n IAgentRuntime,\n ObjectGenerationParams,\n ModelTypeName,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateObject } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport { getSmallModel, getLargeModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\nimport { getJsonRepairFunction } from \"../utils/json\";\n\n/**\n * Helper function to generate objects using specified model type\n */\nasync function generateObjectByModelType(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n modelType: ModelTypeName,\n getModelFn: (runtime: IAgentRuntime) => string,\n): Promise<Record<string, unknown>> {\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n logger.log(`[OpenAI] Using ${modelType} model: ${modelName}`);\n const temperature = params.temperature ?? 0;\n const schemaPresent = !!params.schema;\n\n if (schemaPresent) {\n logger.warn(\n `Schema provided but ignored: OpenAI object generation currently uses output=no-schema. The schema parameter has no effect.`,\n );\n }\n\n try {\n const { object, usage } = await generateObject({\n model: openai.languageModel(modelName),\n output: \"no-schema\",\n prompt: params.prompt,\n temperature: temperature,\n experimental_repairText: getJsonRepairFunction(),\n });\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n return object as Record<string, unknown>;\n } catch (error: unknown) {\n const message = error instanceof Error ? error.message : String(error);\n logger.error(`[generateObject] Error: ${message}`);\n throw error;\n }\n}\n\n/**\n * OBJECT_SMALL model handler\n */\nexport async function handleObjectSmall(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n): Promise<Record<string, unknown>> {\n return generateObjectByModelType(\n runtime,\n params,\n ModelType.OBJECT_SMALL,\n getSmallModel,\n );\n}\n\n/**\n * OBJECT_LARGE model handler\n */\nexport async function handleObjectLarge(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n): Promise<Record<string, unknown>> {\n return generateObjectByModelType(\n runtime,\n params,\n ModelType.OBJECT_LARGE,\n getLargeModel,\n );\n}\n",
16
- "import { logger } from \"@elizaos/core\";\nimport { JSONParseError } from \"ai\";\n\n/**\n * Returns a function to repair JSON text\n */\nexport function getJsonRepairFunction(): (params: {\n text: string;\n error: unknown;\n}) => Promise<string | null> {\n return async ({ text, error }: { text: string; error: unknown }) => {\n try {\n if (error instanceof JSONParseError) {\n const cleanedText = text.replace(/```json\\n|\\n```|```/g, \"\");\n JSON.parse(cleanedText);\n return cleanedText;\n }\n return null;\n } catch (jsonError: unknown) {\n const message =\n jsonError instanceof Error ? jsonError.message : String(jsonError);\n logger.warn(`Failed to repair JSON text: ${message}`);\n return null;\n }\n };\n}\n",
17
- "import type {\n IAgentRuntime,\n TokenizeTextParams,\n DetokenizeTextParams,\n} from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport { tokenizeText, detokenizeText } from \"../utils/tokenization\";\n\n/**\n * TEXT_TOKENIZER_ENCODE model handler\n */\nexport async function handleTokenizerEncode(\n runtime: IAgentRuntime,\n { prompt, modelType = ModelType.TEXT_LARGE }: TokenizeTextParams,\n): Promise<number[]> {\n return await tokenizeText(runtime, modelType, prompt);\n}\n\n/**\n * TEXT_TOKENIZER_DECODE model handler\n */\nexport async function handleTokenizerDecode(\n runtime: IAgentRuntime,\n { tokens, modelType = ModelType.TEXT_LARGE }: DetokenizeTextParams,\n): Promise<string> {\n return await detokenizeText(runtime, modelType, tokens);\n}\n",
18
- "import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport { encodingForModel, getEncoding, type TiktokenModel, type TiktokenEncoding } from \"js-tiktoken\";\nimport { getLargeModel, getSmallModel } from \"./config\";\n\nfunction resolveTokenizerEncoding(\n modelName: string,\n): ReturnType<typeof encodingForModel> {\n const normalized = modelName.toLowerCase();\n const fallbackEncoding: TiktokenEncoding = normalized.includes(\"4o\")\n ? \"o200k_base\"\n : \"cl100k_base\";\n\n try {\n return encodingForModel(modelName as TiktokenModel);\n } catch (error: unknown) {\n // Use getEncoding for the fallback encoding names\n return getEncoding(fallbackEncoding);\n }\n}\n\n/**\n * Asynchronously tokenizes the given text based on the specified model and prompt.\n *\n * @param {ModelTypeName} model - The type of model to use for tokenization.\n * @param {string} prompt - The text prompt to tokenize.\n * @returns {number[]} - An array of tokens representing the encoded prompt.\n */\nexport async function tokenizeText(\n runtime: IAgentRuntime,\n model: ModelTypeName,\n prompt: string,\n) {\n const modelName =\n model === ModelType.TEXT_SMALL\n ? getSmallModel(runtime)\n : getLargeModel(runtime);\n const tokens = resolveTokenizerEncoding(modelName).encode(prompt);\n return tokens;\n}\n\n/**\n * Detokenize a sequence of tokens back into text using the specified model.\n *\n * @param {ModelTypeName} model - The type of model to use for detokenization.\n * @param {number[]} tokens - The sequence of tokens to detokenize.\n * @returns {string} The detokenized text.\n */\nexport async function detokenizeText(\n runtime: IAgentRuntime,\n model: ModelTypeName,\n tokens: number[],\n) {\n const modelName =\n model === ModelType.TEXT_SMALL\n ? getSmallModel(runtime)\n : getLargeModel(runtime);\n return resolveTokenizerEncoding(modelName).decode(tokens);\n}\n"
5
+ "import type {\n TextToSpeechParams as CoreTextToSpeechParams,\n TranscriptionParams as CoreTranscriptionParams,\n DetokenizeTextParams,\n GenerateTextParams,\n IAgentRuntime,\n ImageDescriptionParams,\n ImageGenerationParams,\n JsonValue,\n ObjectGenerationParams,\n Plugin,\n ResearchParams,\n ResearchResult,\n TextEmbeddingParams,\n TokenizeTextParams,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { initializeOpenAI } from \"./init\";\nimport {\n handleImageDescription,\n handleImageGeneration,\n handleObjectLarge,\n handleObjectSmall,\n handleResearch,\n handleTextEmbedding,\n handleTextLarge,\n handleTextSmall,\n handleTextToSpeech,\n handleTokenizerDecode,\n handleTokenizerEncode,\n handleTranscription,\n} from \"./models\";\nimport type { ImageGenerationResult, OpenAIPluginConfig, TextStreamResult } from \"./types\";\nimport { getAuthHeader, getBaseURL } from \"./utils/config\";\n\ntype ProcessEnvLike = Record<string, string | undefined>;\n\nfunction getProcessEnv(): ProcessEnvLike {\n if (typeof process === \"undefined\") {\n return {};\n }\n return process.env as ProcessEnvLike;\n}\n\nconst env = getProcessEnv();\n\nexport const openaiPlugin: Plugin = {\n name: \"openai\",\n description: \"OpenAI API integration for text, image, audio, and embedding models\",\n\n config: {\n OPENAI_API_KEY: env.OPENAI_API_KEY ?? null,\n OPENAI_BASE_URL: env.OPENAI_BASE_URL ?? null,\n OPENAI_SMALL_MODEL: env.OPENAI_SMALL_MODEL ?? null,\n OPENAI_LARGE_MODEL: env.OPENAI_LARGE_MODEL ?? null,\n SMALL_MODEL: env.SMALL_MODEL ?? null,\n LARGE_MODEL: env.LARGE_MODEL ?? null,\n OPENAI_EMBEDDING_MODEL: env.OPENAI_EMBEDDING_MODEL ?? null,\n OPENAI_EMBEDDING_API_KEY: env.OPENAI_EMBEDDING_API_KEY ?? null,\n OPENAI_EMBEDDING_URL: env.OPENAI_EMBEDDING_URL ?? null,\n OPENAI_EMBEDDING_DIMENSIONS: env.OPENAI_EMBEDDING_DIMENSIONS ?? null,\n OPENAI_IMAGE_DESCRIPTION_MODEL: env.OPENAI_IMAGE_DESCRIPTION_MODEL ?? null,\n OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS: env.OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS ?? null,\n OPENAI_EXPERIMENTAL_TELEMETRY: env.OPENAI_EXPERIMENTAL_TELEMETRY ?? null,\n OPENAI_RESEARCH_MODEL: env.OPENAI_RESEARCH_MODEL ?? null,\n OPENAI_RESEARCH_TIMEOUT: env.OPENAI_RESEARCH_TIMEOUT ?? null,\n },\n\n async init(config: Record<string, string>, runtime: IAgentRuntime): Promise<void> {\n initializeOpenAI(config as OpenAIPluginConfig | undefined, runtime);\n },\n\n models: {\n [ModelType.TEXT_EMBEDDING]: async (\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null\n ): Promise<number[]> => {\n return handleTextEmbedding(runtime, params);\n },\n\n [ModelType.TEXT_TOKENIZER_ENCODE]: async (\n runtime: IAgentRuntime,\n params: TokenizeTextParams\n ): Promise<number[]> => {\n return handleTokenizerEncode(runtime, params);\n },\n\n [ModelType.TEXT_TOKENIZER_DECODE]: async (\n runtime: IAgentRuntime,\n params: DetokenizeTextParams\n ): Promise<string> => {\n return handleTokenizerDecode(runtime, params);\n },\n\n [ModelType.TEXT_SMALL]: async (\n runtime: IAgentRuntime,\n params: GenerateTextParams\n ): Promise<string | TextStreamResult> => {\n return handleTextSmall(runtime, params);\n },\n\n [ModelType.TEXT_LARGE]: async (\n runtime: IAgentRuntime,\n params: GenerateTextParams\n ): Promise<string | TextStreamResult> => {\n return handleTextLarge(runtime, params);\n },\n\n [ModelType.IMAGE]: async (\n runtime: IAgentRuntime,\n params: ImageGenerationParams\n ): Promise<ImageGenerationResult[]> => {\n return handleImageGeneration(runtime, params);\n },\n\n [ModelType.IMAGE_DESCRIPTION]: async (\n runtime: IAgentRuntime,\n params: ImageDescriptionParams | string\n ): Promise<{ title: string; description: string }> => {\n return handleImageDescription(runtime, params);\n },\n\n [ModelType.TRANSCRIPTION]: async (\n runtime: IAgentRuntime,\n input: CoreTranscriptionParams | Buffer | string\n ): Promise<string> => {\n return handleTranscription(runtime, input);\n },\n\n [ModelType.TEXT_TO_SPEECH]: async (\n runtime: IAgentRuntime,\n input: CoreTextToSpeechParams | string\n ): Promise<ArrayBuffer> => {\n return handleTextToSpeech(runtime, input);\n },\n\n [ModelType.OBJECT_SMALL]: async (\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n ): Promise<Record<string, JsonValue>> => {\n return handleObjectSmall(runtime, params);\n },\n\n [ModelType.OBJECT_LARGE]: async (\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n ): Promise<Record<string, JsonValue>> => {\n return handleObjectLarge(runtime, params);\n },\n\n [ModelType.RESEARCH]: async (\n runtime: IAgentRuntime,\n params: ResearchParams\n ): Promise<ResearchResult> => {\n return handleResearch(runtime, params);\n },\n },\n\n tests: [\n {\n name: \"openai_plugin_tests\",\n tests: [\n {\n name: \"openai_test_api_connectivity\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/models`, {\n headers: getAuthHeader(runtime),\n });\n\n if (!response.ok) {\n throw new Error(\n `API connectivity test failed: ${response.status} ${response.statusText}`\n );\n }\n\n const data = (await response.json()) as { data?: unknown[] };\n logger.info(`[OpenAI Test] API connected. ${data.data?.length ?? 0} models available.`);\n },\n },\n {\n name: \"openai_test_text_embedding\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const embedding = await runtime.useModel(ModelType.TEXT_EMBEDDING, {\n text: \"Hello, world!\",\n });\n\n if (!Array.isArray(embedding) || embedding.length === 0) {\n throw new Error(\"Embedding should return a non-empty array\");\n }\n\n logger.info(`[OpenAI Test] Generated embedding with ${embedding.length} dimensions`);\n },\n },\n {\n name: \"openai_test_text_small\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const text = await runtime.useModel(ModelType.TEXT_SMALL, {\n prompt: \"Say hello in exactly 5 words.\",\n });\n\n if (typeof text !== \"string\" || text.length === 0) {\n throw new Error(\"TEXT_SMALL should return non-empty string\");\n }\n\n logger.info(`[OpenAI Test] TEXT_SMALL generated: \"${text.substring(0, 50)}...\"`);\n },\n },\n {\n name: \"openai_test_text_large\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const text = await runtime.useModel(ModelType.TEXT_LARGE, {\n prompt: \"Explain quantum computing in 2 sentences.\",\n });\n\n if (typeof text !== \"string\" || text.length === 0) {\n throw new Error(\"TEXT_LARGE should return non-empty string\");\n }\n\n logger.info(`[OpenAI Test] TEXT_LARGE generated: \"${text.substring(0, 50)}...\"`);\n },\n },\n {\n name: \"openai_test_tokenizer_roundtrip\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const originalText = \"Hello, tokenizer test!\";\n\n const tokens = await runtime.useModel(ModelType.TEXT_TOKENIZER_ENCODE, {\n prompt: originalText,\n modelType: ModelType.TEXT_SMALL,\n });\n\n if (!Array.isArray(tokens) || tokens.length === 0) {\n throw new Error(\"Tokenization should return non-empty token array\");\n }\n\n const decodedText = await runtime.useModel(ModelType.TEXT_TOKENIZER_DECODE, {\n tokens,\n modelType: ModelType.TEXT_SMALL,\n });\n\n if (decodedText !== originalText) {\n throw new Error(\n `Tokenizer roundtrip failed: expected \"${originalText}\", got \"${decodedText}\"`\n );\n }\n\n logger.info(`[OpenAI Test] Tokenizer roundtrip successful (${tokens.length} tokens)`);\n },\n },\n {\n name: \"openai_test_streaming\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const chunks: string[] = [];\n\n const result = await runtime.useModel(ModelType.TEXT_LARGE, {\n prompt: \"Count from 1 to 5, one number per line.\",\n stream: true,\n onStreamChunk: (chunk: string) => {\n chunks.push(chunk);\n },\n });\n\n if (typeof result !== \"string\" || result.length === 0) {\n throw new Error(\"Streaming should return non-empty result\");\n }\n\n if (chunks.length === 0) {\n throw new Error(\"No streaming chunks received\");\n }\n\n logger.info(`[OpenAI Test] Streaming test: ${chunks.length} chunks received`);\n },\n },\n {\n name: \"openai_test_image_description\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const testImageUrl =\n \"https://upload.wikimedia.org/wikipedia/commons/thumb/a/a7/Camponotus_flavomarginatus_ant.jpg/440px-Camponotus_flavomarginatus_ant.jpg\";\n\n const result = await runtime.useModel(ModelType.IMAGE_DESCRIPTION, testImageUrl);\n\n if (\n !result ||\n typeof result !== \"object\" ||\n !(\"title\" in result) ||\n !(\"description\" in result)\n ) {\n throw new Error(\"Image description should return { title, description }\");\n }\n\n logger.info(`[OpenAI Test] Image described: \"${result.title}\"`);\n },\n },\n {\n name: \"openai_test_transcription\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n // Fetch a short audio sample\n const audioUrl =\n \"https://upload.wikimedia.org/wikipedia/commons/2/25/En-Open_Source.ogg\";\n\n const response = await fetch(audioUrl);\n const arrayBuffer = await response.arrayBuffer();\n const audioBuffer = Buffer.from(new Uint8Array(arrayBuffer));\n\n const transcription = await runtime.useModel(ModelType.TRANSCRIPTION, audioBuffer);\n\n if (typeof transcription !== \"string\") {\n throw new Error(\"Transcription should return a string\");\n }\n\n logger.info(`[OpenAI Test] Transcription: \"${transcription.substring(0, 50)}...\"`);\n },\n },\n {\n name: \"openai_test_text_to_speech\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const audioData = await runtime.useModel(ModelType.TEXT_TO_SPEECH, {\n text: \"Hello, this is a text-to-speech test.\",\n });\n\n if (!(audioData instanceof ArrayBuffer) || audioData.byteLength === 0) {\n throw new Error(\"TTS should return non-empty ArrayBuffer\");\n }\n\n logger.info(`[OpenAI Test] TTS generated ${audioData.byteLength} bytes of audio`);\n },\n },\n {\n name: \"openai_test_object_generation\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n const result = await runtime.useModel(ModelType.OBJECT_SMALL, {\n prompt:\n \"Return a JSON object with exactly these fields: name (string), age (number), active (boolean)\",\n });\n\n if (!result || typeof result !== \"object\") {\n throw new Error(\"Object generation should return an object\");\n }\n\n logger.info(\n `[OpenAI Test] Object generated: ${JSON.stringify(result).substring(0, 100)}`\n );\n },\n },\n {\n name: \"openai_test_research\",\n fn: async (runtime: IAgentRuntime): Promise<void> => {\n // Note: Deep research can take a long time (minutes to hours)\n // This test uses a simple query with maxToolCalls to limit execution time\n const result = await runtime.useModel(ModelType.RESEARCH, {\n input: \"What is the current date and time?\",\n tools: [{ type: \"web_search_preview\" }],\n maxToolCalls: 3, // Limit tool calls for faster test execution\n });\n\n if (!result || typeof result !== \"object\" || !(\"text\" in result)) {\n throw new Error(\"Research should return an object with text property\");\n }\n\n if (typeof result.text !== \"string\" || result.text.length === 0) {\n throw new Error(\"Research result text should be a non-empty string\");\n }\n\n logger.info(\n `[OpenAI Test] Research completed. Text length: ${result.text.length}, Annotations: ${result.annotations?.length ?? 0}`\n );\n },\n },\n ],\n },\n ],\n};\n\nexport default openaiPlugin;\n",
6
+ "import type { IAgentRuntime } from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport type { OpenAIPluginConfig } from \"./types\";\nimport { getApiKey, getAuthHeader, getBaseURL, isBrowser } from \"./utils/config\";\n\n(globalThis as Record<string, unknown>).AI_SDK_LOG_WARNINGS ??= false;\n\nexport function initializeOpenAI(\n _config: OpenAIPluginConfig | undefined,\n runtime: IAgentRuntime\n): void {\n void validateOpenAIConfiguration(runtime);\n}\n\nasync function validateOpenAIConfiguration(runtime: IAgentRuntime): Promise<void> {\n if (isBrowser()) {\n logger.debug(\"[OpenAI] Skipping API validation in browser environment\");\n return;\n }\n\n const apiKey = getApiKey(runtime);\n\n if (!apiKey) {\n logger.warn(\n \"[OpenAI] OPENAI_API_KEY is not configured. \" +\n \"OpenAI functionality will fail until a valid API key is provided.\"\n );\n return;\n }\n\n try {\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/models`, {\n headers: getAuthHeader(runtime),\n });\n\n if (!response.ok) {\n logger.warn(\n `[OpenAI] API key validation failed: ${response.status} ${response.statusText}. ` +\n \"Please verify your OPENAI_API_KEY is correct.\"\n );\n return;\n }\n } catch (error) {\n const message = error instanceof Error ? error.message : String(error);\n logger.warn(`[OpenAI] API validation error: ${message}. OpenAI functionality may be limited.`);\n }\n}\n",
7
+ "import type { IAgentRuntime } from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\n\nfunction getEnvValue(key: string): string | undefined {\n if (typeof process === \"undefined\" || !process.env) {\n return undefined;\n }\n const value = process.env[key];\n return value === undefined ? undefined : String(value);\n}\n\nexport function getSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue?: string\n): string | undefined {\n const value = runtime.getSetting(key);\n if (value !== undefined && value !== null) {\n return String(value);\n }\n return getEnvValue(key) ?? defaultValue;\n}\nexport function getRequiredSetting(\n runtime: IAgentRuntime,\n key: string,\n errorMessage?: string\n): string {\n const value = getSetting(runtime, key);\n if (value === undefined || value.trim() === \"\") {\n throw new Error(errorMessage ?? `Required setting '${key}' is not configured`);\n }\n return value;\n}\n\nexport function getNumericSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue: number\n): number {\n const value = getSetting(runtime, key);\n if (value === undefined) {\n return defaultValue;\n }\n const parsed = Number.parseInt(value, 10);\n if (!Number.isFinite(parsed)) {\n throw new Error(`Setting '${key}' must be a valid integer, got: ${value}`);\n }\n return parsed;\n}\n\nexport function getBooleanSetting(\n runtime: IAgentRuntime,\n key: string,\n defaultValue: boolean\n): boolean {\n const value = getSetting(runtime, key);\n if (value === undefined) {\n return defaultValue;\n }\n const normalized = value.toLowerCase();\n return normalized === \"true\" || normalized === \"1\" || normalized === \"yes\";\n}\n\nexport function isBrowser(): boolean {\n return (\n typeof globalThis !== \"undefined\" &&\n typeof (globalThis as { document?: Document }).document !== \"undefined\"\n );\n}\n\nexport function isProxyMode(runtime: IAgentRuntime): boolean {\n return isBrowser() && !!getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n}\n\nexport function getApiKey(runtime: IAgentRuntime): string | undefined {\n return getSetting(runtime, \"OPENAI_API_KEY\");\n}\n\nexport function getEmbeddingApiKey(runtime: IAgentRuntime): string | undefined {\n const embeddingApiKey = getSetting(runtime, \"OPENAI_EMBEDDING_API_KEY\");\n if (embeddingApiKey) {\n logger.debug(\"[OpenAI] Using specific embedding API key\");\n return embeddingApiKey;\n }\n logger.debug(\"[OpenAI] Falling back to general API key for embeddings\");\n return getApiKey(runtime);\n}\n\nexport function getAuthHeader(\n runtime: IAgentRuntime,\n forEmbedding = false\n): Record<string, string> {\n // By default this plugin does NOT send auth headers in the browser. This is safer because\n // frontend builds would otherwise expose secrets. For local demos, you can explicitly\n // opt-in to sending the Authorization header by setting OPENAI_ALLOW_BROWSER_API_KEY=true.\n if (isBrowser() && !getBooleanSetting(runtime, \"OPENAI_ALLOW_BROWSER_API_KEY\", false)) {\n return {};\n }\n const key = forEmbedding ? getEmbeddingApiKey(runtime) : getApiKey(runtime);\n return key ? { Authorization: `Bearer ${key}` } : {};\n}\n\nexport function getBaseURL(runtime: IAgentRuntime): string {\n const browserURL = getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\");\n const baseURL =\n isBrowser() && browserURL\n ? browserURL\n : (getSetting(runtime, \"OPENAI_BASE_URL\") ?? \"https://api.openai.com/v1\");\n logger.debug(`[OpenAI] Base URL: ${baseURL}`);\n return baseURL;\n}\n\nexport function getEmbeddingBaseURL(runtime: IAgentRuntime): string {\n const embeddingURL = isBrowser()\n ? (getSetting(runtime, \"OPENAI_BROWSER_EMBEDDING_URL\") ??\n getSetting(runtime, \"OPENAI_BROWSER_BASE_URL\"))\n : getSetting(runtime, \"OPENAI_EMBEDDING_URL\");\n\n if (embeddingURL) {\n logger.debug(`[OpenAI] Using embedding base URL: ${embeddingURL}`);\n return embeddingURL;\n }\n\n logger.debug(\"[OpenAI] Falling back to general base URL for embeddings\");\n return getBaseURL(runtime);\n}\n\nexport function getSmallModel(runtime: IAgentRuntime): string {\n return (\n getSetting(runtime, \"OPENAI_SMALL_MODEL\") ?? getSetting(runtime, \"SMALL_MODEL\") ?? \"gpt-5-mini\"\n );\n}\n\nexport function getLargeModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_LARGE_MODEL\") ?? getSetting(runtime, \"LARGE_MODEL\") ?? \"gpt-5\";\n}\n\nexport function getEmbeddingModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_EMBEDDING_MODEL\") ?? \"text-embedding-3-small\";\n}\n\nexport function getImageDescriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MODEL\") ?? \"gpt-5-mini\";\n}\n\nexport function getTranscriptionModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TRANSCRIPTION_MODEL\") ?? \"gpt-5-mini-transcribe\";\n}\n\nexport function getTTSModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_MODEL\") ?? \"tts-1\";\n}\n\nexport function getTTSVoice(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_VOICE\") ?? \"nova\";\n}\n\nexport function getTTSInstructions(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_TTS_INSTRUCTIONS\") ?? \"\";\n}\n\nexport function getImageModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_IMAGE_MODEL\") ?? \"dall-e-3\";\n}\n\nexport function getExperimentalTelemetry(runtime: IAgentRuntime): boolean {\n return getBooleanSetting(runtime, \"OPENAI_EXPERIMENTAL_TELEMETRY\", false);\n}\n\nexport function getEmbeddingDimensions(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_EMBEDDING_DIMENSIONS\", 1536);\n}\n\nexport function getImageDescriptionMaxTokens(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_IMAGE_DESCRIPTION_MAX_TOKENS\", 8192);\n}\n\nexport function getResearchModel(runtime: IAgentRuntime): string {\n return getSetting(runtime, \"OPENAI_RESEARCH_MODEL\") ?? \"o3-deep-research\";\n}\n\nexport function getResearchTimeout(runtime: IAgentRuntime): number {\n return getNumericSetting(runtime, \"OPENAI_RESEARCH_TIMEOUT\", 3600000);\n}\n",
8
+ "import type {\n TextToSpeechParams as CoreTextToSpeechParams,\n TranscriptionParams as CoreTranscriptionParams,\n IAgentRuntime,\n} from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport type {\n TextToSpeechParams as LocalTextToSpeechParams,\n TranscriptionParams as LocalTranscriptionParams,\n OpenAITranscriptionResponse,\n TTSOutputFormat,\n TTSVoice,\n} from \"../types\";\nimport { detectAudioMimeType, getFilenameForMimeType } from \"../utils/audio\";\nimport {\n getAuthHeader,\n getBaseURL,\n getTranscriptionModel,\n getTTSInstructions,\n getTTSModel,\n getTTSVoice,\n} from \"../utils/config\";\n\ntype AudioInput = Blob | File | Buffer;\ntype TranscriptionInput = AudioInput | LocalTranscriptionParams | CoreTranscriptionParams | string;\ntype TTSInput = string | LocalTextToSpeechParams | CoreTextToSpeechParams;\n\nfunction isBlobOrFile(value: unknown): value is Blob | File {\n return value instanceof Blob || value instanceof File;\n}\n\nfunction isBuffer(value: unknown): value is Buffer {\n return Buffer.isBuffer(value);\n}\n\nfunction isLocalTranscriptionParams(value: unknown): value is LocalTranscriptionParams {\n return (\n typeof value === \"object\" &&\n value !== null &&\n \"audio\" in value &&\n (isBlobOrFile((value as LocalTranscriptionParams).audio) ||\n isBuffer((value as LocalTranscriptionParams).audio))\n );\n}\n\nfunction isCoreTranscriptionParams(value: unknown): value is CoreTranscriptionParams {\n return (\n typeof value === \"object\" &&\n value !== null &&\n \"audioUrl\" in value &&\n typeof (value as CoreTranscriptionParams).audioUrl === \"string\"\n );\n}\n\nasync function fetchAudioFromUrl(url: string): Promise<Blob> {\n const response = await fetch(url);\n if (!response.ok) {\n throw new Error(`Failed to fetch audio from URL: ${response.status}`);\n }\n return response.blob();\n}\nexport async function handleTranscription(\n runtime: IAgentRuntime,\n input: TranscriptionInput\n): Promise<string> {\n let modelName = getTranscriptionModel(runtime);\n let blob: Blob;\n let extraParams: Partial<LocalTranscriptionParams> = {};\n\n if (typeof input === \"string\") {\n logger.debug(`[OpenAI] Fetching audio from URL: ${input}`);\n blob = await fetchAudioFromUrl(input);\n } else if (isBlobOrFile(input)) {\n blob = input;\n } else if (isBuffer(input)) {\n const mimeType = detectAudioMimeType(input);\n logger.debug(`[OpenAI] Auto-detected audio MIME type: ${mimeType}`);\n blob = new Blob([new Uint8Array(input)], { type: mimeType });\n } else if (isLocalTranscriptionParams(input)) {\n extraParams = input;\n if (input.model) {\n modelName = input.model;\n }\n if (isBuffer(input.audio)) {\n const mimeType = input.mimeType ?? detectAudioMimeType(input.audio);\n logger.debug(`[OpenAI] Using MIME type: ${mimeType}`);\n blob = new Blob([new Uint8Array(input.audio)], { type: mimeType });\n } else {\n blob = input.audio;\n }\n } else if (isCoreTranscriptionParams(input)) {\n logger.debug(`[OpenAI] Fetching audio from URL: ${input.audioUrl}`);\n blob = await fetchAudioFromUrl(input.audioUrl);\n extraParams = { prompt: input.prompt };\n } else {\n throw new Error(\n \"TRANSCRIPTION expects Blob, File, Buffer, URL string, or TranscriptionParams object\"\n );\n }\n\n logger.debug(`[OpenAI] Using TRANSCRIPTION model: ${modelName}`);\n\n const mimeType = (blob as File).type || \"audio/webm\";\n const filename =\n (blob as File).name ||\n getFilenameForMimeType(\n mimeType.startsWith(\"audio/\")\n ? (mimeType as ReturnType<typeof detectAudioMimeType>)\n : \"audio/webm\"\n );\n\n const formData = new FormData();\n formData.append(\"file\", blob, filename);\n formData.append(\"model\", modelName);\n\n if (extraParams.language) {\n formData.append(\"language\", extraParams.language);\n }\n if (extraParams.responseFormat) {\n formData.append(\"response_format\", extraParams.responseFormat);\n }\n if (extraParams.prompt) {\n formData.append(\"prompt\", extraParams.prompt);\n }\n if (extraParams.temperature !== undefined) {\n formData.append(\"temperature\", String(extraParams.temperature));\n }\n if (extraParams.timestampGranularities) {\n for (const granularity of extraParams.timestampGranularities) {\n formData.append(\"timestamp_granularities[]\", granularity);\n }\n }\n\n const baseURL = getBaseURL(runtime);\n const response = await fetch(`${baseURL}/audio/transcriptions`, {\n method: \"POST\",\n headers: getAuthHeader(runtime),\n body: formData,\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI transcription failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAITranscriptionResponse;\n return data.text;\n}\n\nexport async function handleTextToSpeech(\n runtime: IAgentRuntime,\n input: TTSInput\n): Promise<ArrayBuffer> {\n let text: string;\n let voice: string | undefined;\n let format: TTSOutputFormat = \"mp3\";\n let model: string | undefined;\n let instructions: string | undefined;\n\n if (typeof input === \"string\") {\n text = input;\n voice = undefined;\n } else {\n text = input.text;\n voice = input.voice;\n if (\"format\" in input && input.format) {\n format = input.format;\n }\n if (\"model\" in input && input.model) {\n model = input.model;\n }\n if (\"instructions\" in input && input.instructions) {\n instructions = input.instructions;\n }\n }\n\n model = model ?? getTTSModel(runtime);\n voice = voice ?? getTTSVoice(runtime);\n instructions = instructions ?? getTTSInstructions(runtime);\n\n logger.debug(`[OpenAI] Using TEXT_TO_SPEECH model: ${model}`);\n\n if (!text || text.trim().length === 0) {\n throw new Error(\"TEXT_TO_SPEECH requires non-empty text\");\n }\n\n if (text.length > 4096) {\n throw new Error(\"TEXT_TO_SPEECH text exceeds 4096 character limit\");\n }\n\n const validVoices: TTSVoice[] = [\"alloy\", \"echo\", \"fable\", \"onyx\", \"nova\", \"shimmer\"];\n if (voice && !validVoices.includes(voice as TTSVoice)) {\n throw new Error(`Invalid voice: ${voice}. Must be one of: ${validVoices.join(\", \")}`);\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody: Record<string, string> = {\n model,\n voice: voice as TTSVoice,\n input: text,\n response_format: format,\n };\n\n if (instructions && instructions.length > 0) {\n requestBody.instructions = instructions;\n }\n\n const response = await fetch(`${baseURL}/audio/speech`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n ...(format === \"mp3\" ? { Accept: \"audio/mpeg\" } : {}),\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(`OpenAI TTS failed: ${response.status} ${response.statusText} - ${errorText}`);\n }\n\n return response.arrayBuffer();\n}\n",
9
+ "import { logger } from \"@elizaos/core\";\n\nconst MAGIC_BYTES = {\n WAV: {\n HEADER: [0x52, 0x49, 0x46, 0x46] as const, // \"RIFF\"\n IDENTIFIER: [0x57, 0x41, 0x56, 0x45] as const, // \"WAVE\"\n },\n MP3_ID3: [0x49, 0x44, 0x33] as const, // \"ID3\"\n OGG: [0x4f, 0x67, 0x67, 0x53] as const, // \"OggS\"\n FLAC: [0x66, 0x4c, 0x61, 0x43] as const, // \"fLaC\"\n FTYP: [0x66, 0x74, 0x79, 0x70] as const, // \"ftyp\" at offset 4 for mp4/m4a\n WEBM_EBML: [0x1a, 0x45, 0xdf, 0xa3] as const, // EBML header\n} as const;\n\nconst MIN_DETECTION_BUFFER_SIZE = 12;\n\nexport type AudioMimeType =\n | \"audio/wav\"\n | \"audio/mpeg\"\n | \"audio/ogg\"\n | \"audio/flac\"\n | \"audio/mp4\"\n | \"audio/webm\"\n | \"application/octet-stream\";\n\nfunction matchBytes(buffer: Buffer, offset: number, expected: readonly number[]): boolean {\n for (let i = 0; i < expected.length; i++) {\n const expectedByte = expected[i];\n if (expectedByte === undefined || buffer[offset + i] !== expectedByte) {\n return false;\n }\n }\n return true;\n}\n\nexport function detectAudioMimeType(buffer: Buffer): AudioMimeType {\n if (buffer.length < MIN_DETECTION_BUFFER_SIZE) {\n return \"application/octet-stream\";\n }\n\n // WAV: \"RIFF\" + size + \"WAVE\"\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.WAV.HEADER) &&\n matchBytes(buffer, 8, MAGIC_BYTES.WAV.IDENTIFIER)\n ) {\n return \"audio/wav\";\n }\n\n // MP3: ID3 tag or MPEG frame sync\n const firstByte = buffer[0];\n const secondByte = buffer[1];\n if (\n matchBytes(buffer, 0, MAGIC_BYTES.MP3_ID3) ||\n (firstByte === 0xff && secondByte !== undefined && (secondByte & 0xe0) === 0xe0)\n ) {\n return \"audio/mpeg\";\n }\n\n // OGG: \"OggS\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.OGG)) {\n return \"audio/ogg\";\n }\n\n // FLAC: \"fLaC\"\n if (matchBytes(buffer, 0, MAGIC_BYTES.FLAC)) {\n return \"audio/flac\";\n }\n\n // M4A/MP4: \"ftyp\" at offset 4\n if (matchBytes(buffer, 4, MAGIC_BYTES.FTYP)) {\n return \"audio/mp4\";\n }\n\n // WebM: EBML header\n if (matchBytes(buffer, 0, MAGIC_BYTES.WEBM_EBML)) {\n return \"audio/webm\";\n }\n\n logger.warn(\"Could not detect audio format from buffer, using generic binary type\");\n return \"application/octet-stream\";\n}\n\nexport function getExtensionForMimeType(mimeType: AudioMimeType): string {\n switch (mimeType) {\n case \"audio/wav\":\n return \"wav\";\n case \"audio/mpeg\":\n return \"mp3\";\n case \"audio/ogg\":\n return \"ogg\";\n case \"audio/flac\":\n return \"flac\";\n case \"audio/mp4\":\n return \"m4a\";\n case \"audio/webm\":\n return \"webm\";\n case \"application/octet-stream\":\n return \"bin\";\n }\n}\n\nexport function getFilenameForMimeType(mimeType: AudioMimeType): string {\n const ext = getExtensionForMimeType(mimeType);\n return `recording.${ext}`;\n}\n\nexport function validateAudioFormat(buffer: Buffer): AudioMimeType {\n const mimeType = detectAudioMimeType(buffer);\n if (mimeType === \"application/octet-stream\") {\n throw new Error(\n \"Unable to detect audio format. Supported formats: WAV, MP3, OGG, FLAC, M4A, WebM\"\n );\n }\n return mimeType;\n}\n",
10
+ "import type { IAgentRuntime, TextEmbeddingParams } from \"@elizaos/core\";\nimport { logger, ModelType, VECTOR_DIMS } from \"@elizaos/core\";\n\nimport type { OpenAIEmbeddingResponse } from \"../types\";\nimport {\n getAuthHeader,\n getEmbeddingBaseURL,\n getEmbeddingDimensions,\n getEmbeddingModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ntype VectorDimension = (typeof VECTOR_DIMS)[keyof typeof VECTOR_DIMS];\n\nfunction validateDimension(dimension: number): VectorDimension {\n const validDimensions = Object.values(VECTOR_DIMS) as number[];\n if (!validDimensions.includes(dimension)) {\n throw new Error(\n `Invalid embedding dimension: ${dimension}. Must be one of: ${validDimensions.join(\", \")}`\n );\n }\n return dimension as VectorDimension;\n}\n\nfunction extractText(params: TextEmbeddingParams | string | null): string | null {\n if (params === null) {\n return null;\n }\n if (typeof params === \"string\") {\n return params;\n }\n if (typeof params === \"object\" && typeof params.text === \"string\") {\n return params.text;\n }\n throw new Error(\"Invalid embedding params: expected string, { text: string }, or null\");\n}\n\nexport async function handleTextEmbedding(\n runtime: IAgentRuntime,\n params: TextEmbeddingParams | string | null\n): Promise<number[]> {\n const embeddingModel = getEmbeddingModel(runtime);\n const embeddingDimension = validateDimension(getEmbeddingDimensions(runtime));\n\n const text = extractText(params);\n if (text === null) {\n logger.debug(\"[OpenAI] Creating test embedding for initialization\");\n const testVector = new Array(embeddingDimension).fill(0);\n testVector[0] = 0.1;\n return testVector;\n }\n\n let trimmedText = text.trim();\n if (trimmedText.length === 0) {\n throw new Error(\"Cannot generate embedding for empty text\");\n }\n\n // Truncate to stay within embedding model token limits.\n // OpenAI embedding models support up to 8191 tokens per input;\n // 8000 tokens provides a safe buffer (~4 chars per token).\n const maxChars = 8_000 * 4;\n if (trimmedText.length > maxChars) {\n logger.warn(\n `[OpenAI] Embedding input too long (~${Math.ceil(trimmedText.length / 4)} tokens), truncating to ~8000 tokens`\n );\n trimmedText = trimmedText.slice(0, maxChars);\n }\n\n const baseURL = getEmbeddingBaseURL(runtime);\n const url = `${baseURL}/embeddings`;\n\n logger.debug(`[OpenAI] Generating embedding with model: ${embeddingModel}`);\n\n const response = await fetch(url, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime, true),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify({\n model: embeddingModel,\n input: trimmedText,\n }),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI embedding API error: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIEmbeddingResponse;\n\n const firstResult = data?.data?.[0];\n if (!firstResult || !firstResult.embedding) {\n throw new Error(\"OpenAI API returned invalid embedding response structure\");\n }\n\n const embedding = firstResult.embedding;\n\n if (embedding.length !== embeddingDimension) {\n throw new Error(\n `Embedding dimension mismatch: got ${embedding.length}, expected ${embeddingDimension}. ` +\n `Check OPENAI_EMBEDDING_DIMENSIONS setting.`\n );\n }\n\n if (data.usage) {\n emitModelUsageEvent(runtime, ModelType.TEXT_EMBEDDING, trimmedText, {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: 0,\n totalTokens: data.usage.total_tokens,\n });\n }\n\n logger.debug(`[OpenAI] Generated embedding with ${embedding.length} dimensions`);\n return embedding;\n}\n",
11
+ "import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { EventType } from \"@elizaos/core\";\nimport type { TokenUsage } from \"../types\";\n\nconst MAX_PROMPT_LENGTH = 200;\n\ninterface ModelUsageEventPayload {\n runtime: IAgentRuntime;\n source: \"openai\";\n provider: \"openai\";\n type: ModelTypeName;\n prompt: string;\n tokens: {\n prompt: number;\n completion: number;\n total: number;\n cached?: number;\n };\n}\n\ninterface AISDKUsage {\n inputTokens?: number;\n outputTokens?: number;\n totalTokens?: number;\n cachedInputTokens?: number;\n}\n\ninterface OpenAIAPIUsage {\n promptTokens?: number;\n completionTokens?: number;\n totalTokens?: number;\n cachedPromptTokens?: number;\n promptTokensDetails?: {\n cachedTokens?: number;\n };\n}\n\ntype ModelUsage = TokenUsage | AISDKUsage | OpenAIAPIUsage;\n\nfunction truncatePrompt(prompt: string): string {\n if (prompt.length <= MAX_PROMPT_LENGTH) {\n return prompt;\n }\n return `${prompt.slice(0, MAX_PROMPT_LENGTH)}…`;\n}\n\nfunction normalizeUsage(usage: ModelUsage): TokenUsage {\n if (\"promptTokens\" in usage) {\n const promptTokensDetails =\n \"promptTokensDetails\" in usage ? usage.promptTokensDetails : undefined;\n const cachedPromptTokens = usage.cachedPromptTokens ?? promptTokensDetails?.cachedTokens;\n return {\n promptTokens: usage.promptTokens ?? 0,\n completionTokens: usage.completionTokens ?? 0,\n totalTokens: usage.totalTokens ?? (usage.promptTokens ?? 0) + (usage.completionTokens ?? 0),\n cachedPromptTokens,\n };\n }\n if (\"inputTokens\" in usage || \"outputTokens\" in usage) {\n const input = (usage as AISDKUsage).inputTokens ?? 0;\n const output = (usage as AISDKUsage).outputTokens ?? 0;\n const total = (usage as AISDKUsage).totalTokens ?? input + output;\n return {\n promptTokens: input,\n completionTokens: output,\n totalTokens: total,\n cachedPromptTokens: (usage as AISDKUsage).cachedInputTokens,\n };\n }\n return {\n promptTokens: 0,\n completionTokens: 0,\n totalTokens: 0,\n };\n}\n\nexport function emitModelUsageEvent(\n runtime: IAgentRuntime,\n type: ModelTypeName,\n prompt: string,\n usage: ModelUsage\n): void {\n const normalized = normalizeUsage(usage);\n\n const payload: ModelUsageEventPayload = {\n runtime,\n source: \"openai\",\n provider: \"openai\",\n type,\n prompt: truncatePrompt(prompt),\n tokens: {\n prompt: normalized.promptTokens,\n completion: normalized.completionTokens,\n total: normalized.totalTokens,\n ...(normalized.cachedPromptTokens !== undefined\n ? { cached: normalized.cachedPromptTokens }\n : {}),\n },\n };\n\n runtime.emitEvent(EventType.MODEL_USED, payload);\n}\n",
12
+ "import type { IAgentRuntime, ImageDescriptionParams, ImageGenerationParams } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport type {\n ImageDescriptionResult,\n ImageGenerationResult,\n ImageQuality,\n ImageSize,\n ImageStyle,\n OpenAIChatCompletionResponse,\n OpenAIImageGenerationResponse,\n} from \"../types\";\nimport {\n getAuthHeader,\n getBaseURL,\n getImageDescriptionMaxTokens,\n getImageDescriptionModel,\n getImageModel,\n} from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\ninterface ExtendedImageGenerationParams extends ImageGenerationParams {\n quality?: ImageQuality;\n style?: ImageStyle;\n}\n\nconst DEFAULT_IMAGE_DESCRIPTION_PROMPT =\n \"Please analyze this image and provide a title and detailed description.\";\n\nexport async function handleImageGeneration(\n runtime: IAgentRuntime,\n params: ImageGenerationParams\n): Promise<ImageGenerationResult[]> {\n const modelName = getImageModel(runtime);\n const count = params.count ?? 1;\n const size: ImageSize = (params.size as ImageSize) ?? \"1024x1024\";\n const extendedParams = params as ExtendedImageGenerationParams;\n\n logger.debug(`[OpenAI] Using IMAGE model: ${modelName}`);\n\n if (!params.prompt || params.prompt.trim().length === 0) {\n throw new Error(\"IMAGE generation requires a non-empty prompt\");\n }\n\n if (count < 1 || count > 10) {\n throw new Error(\"IMAGE count must be between 1 and 10\");\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody: Record<string, string | number> = {\n model: modelName,\n prompt: params.prompt,\n n: count,\n size,\n };\n\n if (extendedParams.quality) {\n requestBody.quality = extendedParams.quality;\n }\n if (extendedParams.style) {\n requestBody.style = extendedParams.style;\n }\n\n const response = await fetch(`${baseURL}/images/generations`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI image generation failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIImageGenerationResponse;\n\n if (!data.data || data.data.length === 0) {\n throw new Error(\"OpenAI API returned no images\");\n }\n\n return data.data.map((item) => ({\n url: item.url,\n revisedPrompt: item.revised_prompt,\n }));\n}\n\nfunction parseTitleFromResponse(content: string): string {\n const titleMatch = content.match(/title[:\\s]+(.+?)(?:\\n|$)/i);\n return titleMatch?.[1]?.trim() ?? \"Image Analysis\";\n}\n\nfunction parseDescriptionFromResponse(content: string): string {\n return content.replace(/title[:\\s]+(.+?)(?:\\n|$)/i, \"\").trim();\n}\n\nexport async function handleImageDescription(\n runtime: IAgentRuntime,\n params: ImageDescriptionParams | string\n): Promise<ImageDescriptionResult> {\n const modelName = getImageDescriptionModel(runtime);\n const maxTokens = getImageDescriptionMaxTokens(runtime);\n\n logger.debug(`[OpenAI] Using IMAGE_DESCRIPTION model: ${modelName}`);\n\n let imageUrl: string;\n let promptText: string;\n\n if (typeof params === \"string\") {\n imageUrl = params;\n promptText = DEFAULT_IMAGE_DESCRIPTION_PROMPT;\n } else {\n imageUrl = params.imageUrl;\n promptText = params.prompt ?? DEFAULT_IMAGE_DESCRIPTION_PROMPT;\n }\n\n if (!imageUrl || imageUrl.trim().length === 0) {\n throw new Error(\"IMAGE_DESCRIPTION requires a valid image URL\");\n }\n\n const baseURL = getBaseURL(runtime);\n\n const requestBody = {\n model: modelName,\n messages: [\n {\n role: \"user\",\n content: [\n { type: \"text\", text: promptText },\n { type: \"image_url\", image_url: { url: imageUrl } },\n ],\n },\n ],\n max_tokens: maxTokens,\n };\n\n const response = await fetch(`${baseURL}/chat/completions`, {\n method: \"POST\",\n headers: {\n ...getAuthHeader(runtime),\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n });\n\n if (!response.ok) {\n const errorText = await response.text().catch(() => \"Unknown error\");\n throw new Error(\n `OpenAI image description failed: ${response.status} ${response.statusText} - ${errorText}`\n );\n }\n\n const data = (await response.json()) as OpenAIChatCompletionResponse;\n\n if (data.usage) {\n emitModelUsageEvent(\n runtime,\n ModelType.IMAGE_DESCRIPTION,\n typeof params === \"string\" ? params : (params.prompt ?? \"\"),\n {\n promptTokens: data.usage.prompt_tokens,\n completionTokens: data.usage.completion_tokens,\n totalTokens: data.usage.total_tokens,\n }\n );\n }\n\n const firstChoice = data.choices?.[0];\n const content = firstChoice?.message?.content;\n\n if (!content) {\n throw new Error(\"OpenAI API returned empty image description\");\n }\n\n return {\n title: parseTitleFromResponse(content),\n description: parseDescriptionFromResponse(content),\n };\n}\n",
13
+ "import type {\n IAgentRuntime,\n JsonValue,\n ModelTypeName,\n ObjectGenerationParams,\n} from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateObject, type LanguageModel } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport { getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\nimport { getJsonRepairFunction } from \"../utils/json\";\n\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\ntype ChatModelFactory = { chat: (modelName: string) => LanguageModel };\n\nasync function generateObjectByModelType(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<Record<string, JsonValue>> {\n const openai = createOpenAIClient(runtime) as ChatModelFactory;\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n\n if (!params.prompt || params.prompt.trim().length === 0) {\n throw new Error(\"Object generation requires a non-empty prompt\");\n }\n\n if (params.schema) {\n logger.debug(\n \"[OpenAI] Schema provided but using no-schema mode. \" +\n \"Structure is determined by prompt instructions.\"\n );\n }\n\n const model = openai.chat(modelName);\n const { object, usage } = await generateObject({\n model,\n output: \"no-schema\",\n prompt: params.prompt,\n experimental_repairText: getJsonRepairFunction(),\n });\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n if (typeof object !== \"object\" || object === null) {\n throw new Error(`Object generation returned ${typeof object}, expected object`);\n }\n\n return object as Record<string, JsonValue>;\n}\n\nexport async function handleObjectSmall(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n): Promise<Record<string, JsonValue>> {\n return generateObjectByModelType(runtime, params, ModelType.OBJECT_SMALL, getSmallModel);\n}\n\nexport async function handleObjectLarge(\n runtime: IAgentRuntime,\n params: ObjectGenerationParams\n): Promise<Record<string, JsonValue>> {\n return generateObjectByModelType(runtime, params, ModelType.OBJECT_LARGE, getLargeModel);\n}\n",
14
+ "import { createOpenAI, type OpenAIProvider } from \"@ai-sdk/openai\";\nimport type { IAgentRuntime } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, isProxyMode } from \"../utils/config\";\n\nconst PROXY_API_KEY = \"sk-proxy\";\n\nexport function createOpenAIClient(runtime: IAgentRuntime): OpenAIProvider {\n const baseURL = getBaseURL(runtime);\n const apiKey = getApiKey(runtime);\n\n if (!apiKey && isProxyMode(runtime)) {\n return createOpenAI({\n apiKey: PROXY_API_KEY,\n baseURL,\n });\n }\n\n if (!apiKey) {\n throw new Error(\n \"OPENAI_API_KEY is required. Set it in your environment variables or runtime settings.\"\n );\n }\n\n return createOpenAI({\n apiKey,\n baseURL,\n });\n}\n",
15
+ "import { logger } from \"@elizaos/core\";\nimport { JSONParseError } from \"ai\";\n\ninterface JsonRepairParams {\n text: string;\n error: Error;\n}\n\ntype JsonRepairFunction = (params: JsonRepairParams) => Promise<string | null>;\n\nconst JSON_CLEANUP_PATTERNS = {\n MARKDOWN_JSON: /```json\\n|\\n```|```/g,\n WHITESPACE: /^\\s+|\\s+$/g,\n} as const;\n\nexport function getJsonRepairFunction(): JsonRepairFunction {\n return async ({ text, error }: JsonRepairParams): Promise<string | null> => {\n if (!(error instanceof JSONParseError)) {\n return null;\n }\n try {\n const cleanedText = text.replace(JSON_CLEANUP_PATTERNS.MARKDOWN_JSON, \"\");\n JSON.parse(cleanedText);\n logger.debug(\"[JSON Repair] Successfully repaired JSON by removing markdown wrappers\");\n return cleanedText;\n } catch {\n logger.warn(\"[JSON Repair] Unable to repair JSON text\");\n return null;\n }\n };\n}\n\nexport function parseJsonWithRepair<T>(text: string): T {\n try {\n return JSON.parse(text) as T;\n } catch (firstError) {\n const cleanedText = text.replace(JSON_CLEANUP_PATTERNS.MARKDOWN_JSON, \"\");\n try {\n return JSON.parse(cleanedText) as T;\n } catch {\n const message = firstError instanceof Error ? firstError.message : String(firstError);\n throw new Error(`Failed to parse JSON: ${message}`);\n }\n }\n}\n\nexport function safeStringify(value: unknown, indent = 0): string {\n const seen = new WeakSet();\n\n const replacer = (_key: string, val: unknown): unknown => {\n if (typeof val === \"object\" && val !== null) {\n if (seen.has(val)) {\n return \"[Circular]\";\n }\n seen.add(val);\n }\n\n // Handle special types\n if (typeof val === \"bigint\") {\n return val.toString();\n }\n\n if (val instanceof Error) {\n return {\n name: val.name,\n message: val.message,\n stack: val.stack,\n };\n }\n\n if (val instanceof Date) {\n return val.toISOString();\n }\n\n if (val instanceof Map) {\n return Object.fromEntries(val);\n }\n\n if (val instanceof Set) {\n return Array.from(val);\n }\n\n return val;\n };\n\n return JSON.stringify(value, replacer, indent);\n}\n",
16
+ "/**\n * Deep Research model handler\n *\n * Provides deep research capabilities using OpenAI's o3-deep-research and o4-mini-deep-research models.\n * These models can find, analyze, and synthesize hundreds of sources to create comprehensive reports.\n *\n * @see https://platform.openai.com/docs/guides/deep-research\n */\n\nimport type {\n IAgentRuntime,\n JsonValue,\n ResearchAnnotation,\n ResearchCodeInterpreterCall,\n ResearchFileSearchCall,\n ResearchMcpToolCall,\n ResearchMessageOutput,\n ResearchOutputItem,\n ResearchParams,\n ResearchResult,\n ResearchTool,\n ResearchWebSearchCall,\n} from \"@elizaos/core\";\nimport { logger } from \"@elizaos/core\";\nimport { getApiKey, getBaseURL, getResearchModel, getResearchTimeout } from \"../utils/config\";\n\n// ============================================================================\n// Types for OpenAI Responses API\n// ============================================================================\n\n/**\n * Tool configuration for the Responses API\n */\ninterface ResponsesApiTool {\n type: \"web_search_preview\" | \"file_search\" | \"code_interpreter\" | \"mcp\";\n vector_store_ids?: string[];\n container?: { type: \"auto\" };\n server_label?: string;\n server_url?: string;\n require_approval?: \"never\";\n}\n\n/**\n * Raw response from the OpenAI Responses API\n */\ninterface ResponsesApiResponse {\n id: string;\n object: string;\n status?: \"queued\" | \"in_progress\" | \"completed\" | \"failed\";\n output?: ResponsesApiOutputItem[];\n output_text?: string;\n error?: {\n message: string;\n code: string;\n };\n}\n\n/**\n * Raw output item from the Responses API\n */\ninterface ResponsesApiOutputItem {\n id?: string;\n type: string;\n status?: string;\n action?: {\n type: string;\n query?: string;\n url?: string;\n };\n query?: string;\n results?: Array<{\n file_id: string;\n file_name: string;\n score: number;\n }>;\n code?: string;\n output?: string;\n server_label?: string;\n tool_name?: string;\n arguments?: Record<string, unknown>;\n result?: unknown;\n content?: Array<{\n type: string;\n text: string;\n annotations?: Array<{\n url: string;\n title: string;\n start_index: number;\n end_index: number;\n }>;\n }>;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts ResearchTool params to Responses API tool format\n */\nfunction convertToolToApi(tool: ResearchTool): ResponsesApiTool {\n switch (tool.type) {\n case \"web_search_preview\":\n return { type: \"web_search_preview\" };\n case \"file_search\":\n return {\n type: \"file_search\",\n vector_store_ids: tool.vectorStoreIds,\n };\n case \"code_interpreter\":\n return {\n type: \"code_interpreter\",\n container: tool.container ?? { type: \"auto\" },\n };\n case \"mcp\":\n return {\n type: \"mcp\",\n server_label: tool.serverLabel,\n server_url: tool.serverUrl,\n require_approval: tool.requireApproval ?? \"never\",\n };\n default:\n throw new Error(`Unknown research tool type: ${(tool as ResearchTool).type}`);\n }\n}\n\n/**\n * Converts raw API output items to typed ResearchOutputItem\n */\nfunction convertOutputItem(item: ResponsesApiOutputItem): ResearchOutputItem | null {\n switch (item.type) {\n case \"web_search_call\":\n return {\n id: item.id ?? \"\",\n type: \"web_search_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n action: {\n type: (item.action?.type as \"search\" | \"open_page\" | \"find_in_page\") ?? \"search\",\n query: item.action?.query,\n url: item.action?.url,\n },\n } satisfies ResearchWebSearchCall;\n\n case \"file_search_call\":\n return {\n id: item.id ?? \"\",\n type: \"file_search_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n query: item.query ?? \"\",\n results: item.results?.map((r) => ({\n fileId: r.file_id,\n fileName: r.file_name,\n score: r.score,\n })),\n } satisfies ResearchFileSearchCall;\n\n case \"code_interpreter_call\":\n return {\n id: item.id ?? \"\",\n type: \"code_interpreter_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n code: item.code ?? \"\",\n output: item.output,\n } satisfies ResearchCodeInterpreterCall;\n\n case \"mcp_tool_call\":\n return {\n id: item.id ?? \"\",\n type: \"mcp_tool_call\",\n status: (item.status as \"completed\" | \"failed\") ?? \"completed\",\n serverLabel: item.server_label ?? \"\",\n toolName: item.tool_name ?? \"\",\n arguments: (item.arguments ?? {}) as Record<string, JsonValue>,\n result: item.result as JsonValue,\n } satisfies ResearchMcpToolCall;\n\n case \"message\":\n return {\n type: \"message\",\n content:\n item.content?.map((c) => ({\n type: \"output_text\" as const,\n text: c.text,\n annotations:\n c.annotations?.map((a) => ({\n url: a.url,\n title: a.title,\n startIndex: a.start_index,\n endIndex: a.end_index,\n })) ?? [],\n })) ?? [],\n } satisfies ResearchMessageOutput;\n\n default:\n // Unknown output type, skip\n return null;\n }\n}\n\n/**\n * Extracts text and annotations from the response\n */\nfunction extractTextAndAnnotations(response: ResponsesApiResponse): {\n text: string;\n annotations: ResearchAnnotation[];\n} {\n // Try output_text first (convenience field)\n if (response.output_text) {\n // Find annotations from message output items\n const annotations: ResearchAnnotation[] = [];\n if (response.output) {\n for (const item of response.output) {\n if (item.type === \"message\" && item.content) {\n for (const content of item.content) {\n if (content.annotations) {\n for (const ann of content.annotations) {\n annotations.push({\n url: ann.url,\n title: ann.title,\n startIndex: ann.start_index,\n endIndex: ann.end_index,\n });\n }\n }\n }\n }\n }\n }\n return { text: response.output_text, annotations };\n }\n\n // Fall back to extracting from message output items\n let text = \"\";\n const annotations: ResearchAnnotation[] = [];\n\n if (response.output) {\n for (const item of response.output) {\n if (item.type === \"message\" && item.content) {\n for (const content of item.content) {\n text += content.text;\n if (content.annotations) {\n for (const ann of content.annotations) {\n annotations.push({\n url: ann.url,\n title: ann.title,\n startIndex: ann.start_index,\n endIndex: ann.end_index,\n });\n }\n }\n }\n }\n }\n }\n\n return { text, annotations };\n}\n\n// ============================================================================\n// Main Handler\n// ============================================================================\n\n/**\n * Handles RESEARCH model requests using OpenAI's deep research models.\n *\n * Deep research models can take tens of minutes to complete tasks.\n * Use background mode for long-running tasks.\n *\n * @param runtime - The agent runtime\n * @param params - Research parameters\n * @returns Research result with text, annotations, and output items\n *\n * @example\n * ```typescript\n * const result = await handleResearch(runtime, {\n * input: \"Research the economic impact of AI on global labor markets\",\n * tools: [\n * { type: \"web_search_preview\" },\n * { type: \"code_interpreter\", container: { type: \"auto\" } }\n * ],\n * background: true,\n * });\n * console.log(result.text);\n * ```\n */\nexport async function handleResearch(\n runtime: IAgentRuntime,\n params: ResearchParams\n): Promise<ResearchResult> {\n const apiKey = getApiKey(runtime);\n if (!apiKey) {\n throw new Error(\n \"OPENAI_API_KEY is required for deep research. Set it in your environment variables or runtime settings.\"\n );\n }\n\n const baseURL = getBaseURL(runtime);\n const modelName = params.model ?? getResearchModel(runtime);\n const timeout = getResearchTimeout(runtime);\n\n logger.debug(`[OpenAI] Starting deep research with model: ${modelName}`);\n logger.debug(`[OpenAI] Research input: ${params.input.substring(0, 100)}...`);\n\n // Validate that at least one data source tool is provided\n const dataSourceTools = params.tools?.filter(\n (t) => t.type === \"web_search_preview\" || t.type === \"file_search\" || t.type === \"mcp\"\n );\n\n if (!dataSourceTools || dataSourceTools.length === 0) {\n // Default to web search if no tools specified\n logger.debug(\"[OpenAI] No data source tools specified, defaulting to web_search_preview\");\n params.tools = [{ type: \"web_search_preview\" }, ...(params.tools ?? [])];\n }\n\n // Build the request body for the Responses API\n const requestBody: Record<string, unknown> = {\n model: modelName,\n input: params.input,\n };\n\n if (params.instructions) {\n requestBody.instructions = params.instructions;\n }\n\n if (params.background !== undefined) {\n requestBody.background = params.background;\n }\n\n if (params.tools && params.tools.length > 0) {\n requestBody.tools = params.tools.map(convertToolToApi);\n }\n\n if (params.maxToolCalls !== undefined) {\n requestBody.max_tool_calls = params.maxToolCalls;\n }\n\n if (params.reasoningSummary) {\n requestBody.reasoning = { summary: params.reasoningSummary };\n }\n\n logger.debug(`[OpenAI] Research request body: ${JSON.stringify(requestBody, null, 2)}`);\n\n // Make the API request\n const response = await fetch(`${baseURL}/responses`, {\n method: \"POST\",\n headers: {\n Authorization: `Bearer ${apiKey}`,\n \"Content-Type\": \"application/json\",\n },\n body: JSON.stringify(requestBody),\n signal: AbortSignal.timeout(timeout),\n });\n\n if (!response.ok) {\n const errorText = await response.text();\n logger.error(`[OpenAI] Research request failed: ${response.status} ${errorText}`);\n throw new Error(`Deep research request failed: ${response.status} ${response.statusText}`);\n }\n\n const data = (await response.json()) as ResponsesApiResponse;\n\n if (data.error) {\n logger.error(`[OpenAI] Research API error: ${data.error.message}`);\n throw new Error(`Deep research error: ${data.error.message}`);\n }\n\n logger.debug(`[OpenAI] Research response received. Status: ${data.status ?? \"completed\"}`);\n\n // Extract text and annotations\n const { text, annotations } = extractTextAndAnnotations(data);\n\n // Convert output items\n const outputItems: ResearchOutputItem[] = [];\n if (data.output) {\n for (const item of data.output) {\n const converted = convertOutputItem(item);\n if (converted) {\n outputItems.push(converted);\n }\n }\n }\n\n const result: ResearchResult = {\n id: data.id,\n text,\n annotations,\n outputItems,\n status: data.status,\n };\n\n logger.info(\n `[OpenAI] Research completed. Text length: ${text.length}, Annotations: ${annotations.length}, Output items: ${outputItems.length}`\n );\n\n return result;\n}\n",
17
+ "/**\n * Text generation model handlers\n *\n * Provides text generation using OpenAI's language models.\n */\n\nimport type { GenerateTextParams, IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { logger, ModelType } from \"@elizaos/core\";\nimport { generateText, type LanguageModelUsage, streamText } from \"ai\";\nimport { createOpenAIClient } from \"../providers\";\nimport type { TextStreamResult, TokenUsage } from \"../types\";\nimport { getExperimentalTelemetry, getLargeModel, getSmallModel } from \"../utils/config\";\nimport { emitModelUsageEvent } from \"../utils/events\";\n\n// ============================================================================\n// Types\n// ============================================================================\n\n/**\n * Function to get model name from runtime\n */\ntype ModelNameGetter = (runtime: IAgentRuntime) => string;\n\ntype PromptCacheRetention = \"in_memory\" | \"24h\";\n\ninterface OpenAIPromptCacheOptions {\n promptCacheKey?: string;\n promptCacheRetention?: PromptCacheRetention;\n}\n\ninterface GenerateTextParamsWithOpenAIOptions extends GenerateTextParams {\n providerOptions?: {\n openai?: OpenAIPromptCacheOptions;\n };\n}\n\ninterface LanguageModelUsageWithCache extends LanguageModelUsage {\n cachedInputTokens?: number;\n}\n\n// ============================================================================\n// Helper Functions\n// ============================================================================\n\n/**\n * Converts AI SDK usage to our token usage format\n */\nfunction convertUsage(usage: LanguageModelUsage | undefined): TokenUsage | undefined {\n if (!usage) {\n return undefined;\n }\n\n // The AI SDK uses inputTokens/outputTokens\n const promptTokens = usage.inputTokens ?? 0;\n const completionTokens = usage.outputTokens ?? 0;\n const usageWithCache = usage as LanguageModelUsageWithCache;\n\n return {\n promptTokens,\n completionTokens,\n totalTokens: promptTokens + completionTokens,\n cachedPromptTokens: usageWithCache.cachedInputTokens,\n };\n}\n\nfunction resolvePromptCacheOptions(params: GenerateTextParams): OpenAIPromptCacheOptions {\n const withOpenAIOptions = params as GenerateTextParamsWithOpenAIOptions;\n return {\n promptCacheKey: withOpenAIOptions.providerOptions?.openai?.promptCacheKey,\n promptCacheRetention: withOpenAIOptions.providerOptions?.openai?.promptCacheRetention,\n };\n}\n\n// ============================================================================\n// Core Generation Function\n// ============================================================================\n\n/**\n * Generates text using the specified model type.\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @param modelType - The type of model (TEXT_SMALL or TEXT_LARGE)\n * @param getModelFn - Function to get the model name\n * @returns Generated text or stream result\n */\nasync function generateTextByModelType(\n runtime: IAgentRuntime,\n params: GenerateTextParams,\n modelType: ModelTypeName,\n getModelFn: ModelNameGetter\n): Promise<string | TextStreamResult> {\n const openai = createOpenAIClient(runtime);\n const modelName = getModelFn(runtime);\n\n logger.debug(`[OpenAI] Using ${modelType} model: ${modelName}`);\n const promptCacheOptions = resolvePromptCacheOptions(params);\n\n // Get system prompt from character if available\n const systemPrompt = runtime.character.system ?? undefined;\n\n // Use chat() instead of languageModel() to use the Chat Completions API\n // which has better compatibility than the Responses API\n // gpt-5 and gpt-5-mini (reasoning models) don't support temperature,\n // frequencyPenalty, presencePenalty, or stop parameters - use defaults only\n const model = openai.chat(modelName);\n const generateParams = {\n model,\n prompt: params.prompt,\n system: systemPrompt,\n maxOutputTokens: params.maxTokens ?? 8192,\n experimental_telemetry: { isEnabled: getExperimentalTelemetry(runtime) },\n ...(promptCacheOptions.promptCacheKey || promptCacheOptions.promptCacheRetention\n ? {\n providerOptions: {\n openai: {\n ...(promptCacheOptions.promptCacheKey\n ? { promptCacheKey: promptCacheOptions.promptCacheKey }\n : {}),\n ...(promptCacheOptions.promptCacheRetention\n ? { promptCacheRetention: promptCacheOptions.promptCacheRetention }\n : {}),\n },\n },\n }\n : {}),\n };\n\n // Handle streaming mode\n if (params.stream) {\n const result = streamText(generateParams);\n\n return {\n textStream: result.textStream,\n text: Promise.resolve(result.text),\n usage: Promise.resolve(result.usage).then(convertUsage),\n finishReason: Promise.resolve(result.finishReason).then((r) => r as string | undefined),\n };\n }\n\n // Non-streaming mode\n const { text, usage } = await generateText(generateParams);\n\n if (usage) {\n emitModelUsageEvent(runtime, modelType, params.prompt, usage);\n }\n\n return text;\n}\n\n// ============================================================================\n// Public Handlers\n// ============================================================================\n\n/**\n * Handles TEXT_SMALL model requests.\n *\n * Uses the configured small model (default: gpt-5-mini).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextSmall(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_SMALL, getSmallModel);\n}\n\n/**\n * Handles TEXT_LARGE model requests.\n *\n * Uses the configured large model (default: gpt-5).\n *\n * @param runtime - The agent runtime\n * @param params - Generation parameters\n * @returns Generated text or stream result\n */\nexport async function handleTextLarge(\n runtime: IAgentRuntime,\n params: GenerateTextParams\n): Promise<string | TextStreamResult> {\n return generateTextByModelType(runtime, params, ModelType.TEXT_LARGE, getLargeModel);\n}\n",
18
+ "import type { DetokenizeTextParams, IAgentRuntime, TokenizeTextParams } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport { detokenizeText, tokenizeText } from \"../utils/tokenization\";\n\nexport async function handleTokenizerEncode(\n runtime: IAgentRuntime,\n params: TokenizeTextParams\n): Promise<number[]> {\n if (!params.prompt) {\n throw new Error(\"Tokenization requires a non-empty prompt\");\n }\n const modelType = params.modelType ?? ModelType.TEXT_LARGE;\n return tokenizeText(runtime, modelType, params.prompt);\n}\n\nexport async function handleTokenizerDecode(\n runtime: IAgentRuntime,\n params: DetokenizeTextParams\n): Promise<string> {\n if (!params.tokens || !Array.isArray(params.tokens)) {\n throw new Error(\"Detokenization requires a valid tokens array\");\n }\n if (params.tokens.length === 0) {\n return \"\";\n }\n for (let i = 0; i < params.tokens.length; i++) {\n const token = params.tokens[i];\n if (typeof token !== \"number\" || !Number.isFinite(token)) {\n throw new Error(`Invalid token at index ${i}: expected number`);\n }\n }\n const modelType = params.modelType ?? ModelType.TEXT_LARGE;\n return detokenizeText(runtime, modelType, params.tokens);\n}\n",
19
+ "import type { IAgentRuntime, ModelTypeName } from \"@elizaos/core\";\nimport { ModelType } from \"@elizaos/core\";\nimport {\n encodingForModel,\n getEncoding,\n type Tiktoken,\n type TiktokenEncoding,\n type TiktokenModel,\n} from \"js-tiktoken\";\nimport { getLargeModel, getSmallModel } from \"./config\";\n\ntype SupportedEncoding = \"cl100k_base\" | \"o200k_base\";\n\nfunction resolveTokenizerEncoding(modelName: string): Tiktoken {\n const normalized = modelName.toLowerCase();\n const fallbackEncoding: SupportedEncoding = normalized.includes(\"4o\")\n ? \"o200k_base\"\n : \"cl100k_base\";\n try {\n return encodingForModel(modelName as TiktokenModel);\n } catch {\n return getEncoding(fallbackEncoding as TiktokenEncoding);\n }\n}\n\nfunction getModelName(runtime: IAgentRuntime, modelType: ModelTypeName): string {\n if (modelType === ModelType.TEXT_SMALL) {\n return getSmallModel(runtime);\n }\n return getLargeModel(runtime);\n}\n\nexport function tokenizeText(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string\n): number[] {\n const modelName = getModelName(runtime, modelType);\n const encoder = resolveTokenizerEncoding(modelName);\n return encoder.encode(text);\n}\n\nexport function detokenizeText(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n tokens: number[]\n): string {\n const modelName = getModelName(runtime, modelType);\n const encoder = resolveTokenizerEncoding(modelName);\n return encoder.decode(tokens);\n}\n\nexport function countTokens(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string\n): number {\n const tokens = tokenizeText(runtime, modelType, text);\n return tokens.length;\n}\n\nexport function truncateToTokenLimit(\n runtime: IAgentRuntime,\n modelType: ModelTypeName,\n text: string,\n maxTokens: number\n): string {\n const tokens = tokenizeText(runtime, modelType, text);\n if (tokens.length <= maxTokens) {\n return text;\n }\n const truncatedTokens = tokens.slice(0, maxTokens);\n return detokenizeText(runtime, modelType, truncatedTokens);\n}\n"
19
20
  ],
20
- "mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAUkC,IAAlC;;;ACV2C,IAA3C;;;ACCuB,IAAvB;AASO,SAAS,UAAU,CACxB,SACA,KACA,cACoB;AAAA,EACpB,MAAM,QAAQ,QAAQ,WAAW,GAAG;AAAA,EAEpC,IAAI,UAAU,aAAa,UAAU,MAAM;AAAA,IACzC,OAAO,OAAO,KAAK;AAAA,EACrB;AAAA,EACA,OAAO,QAAQ,IAAI,QAAQ;AAAA;AAGtB,SAAS,SAAS,GAAY;AAAA,EACnC,OACE,OAAO,eAAe,eACtB,cAAe,cACf,OAAQ,WAAmB,aAAa;AAAA;AAQrC,SAAS,WAAW,CAAC,SAAiC;AAAA,EAC3D,OAAO,UAAU,KAAK,CAAC,CAAC,WAAW,SAAS,yBAAyB;AAAA;AAGhE,SAAS,aAAa,CAC3B,SACA,eAAe,OACS;AAAA,EACxB,IAAI,UAAU;AAAA,IAAG,OAAO,CAAC;AAAA,EACzB,MAAM,MAAM,eAAe,mBAAmB,OAAO,IAAI,UAAU,OAAO;AAAA,EAC1E,OAAO,MAAM,EAAE,eAAe,UAAU,MAAM,IAAI,CAAC;AAAA;AAQ9C,SAAS,UAAU,CAAC,SAAgC;AAAA,EACzD,MAAM,aAAa,WAAW,SAAS,yBAAyB;AAAA,EAChE,MAAM,UACJ,UAAU,KAAK,aACX,aACA,WAAW,SAAS,mBAAmB,2BAA2B;AAAA,EAExE,mBAAO,MAAM,8BAA8B,SAAS;AAAA,EACpD,OAAO;AAAA;AAQF,SAAS,mBAAmB,CAAC,SAAgC;AAAA,EAClE,MAAM,eAAe,UAAU,IAC3B,WAAW,SAAS,8BAA8B,KAClD,WAAW,SAAS,yBAAyB,IAC7C,WAAW,SAAS,sBAAsB;AAAA,EAC9C,IAAI,cAAc;AAAA,IAChB,mBAAO,MAAM,+CAA+C,cAAc;AAAA,IAC1E,OAAO;AAAA,EACT;AAAA,EACA,mBAAO,MAAM,2DAA2D;AAAA,EACxE,OAAO,WAAW,OAAO;AAAA;AASpB,SAAS,SAAS,CAAC,SAA4C;AAAA,EACpE,OAAO,WAAW,SAAS,gBAAgB;AAAA;AAStC,SAAS,kBAAkB,CAAC,SAA4C;AAAA,EAC7E,MAAM,kBAAkB,WAAW,SAAS,0BAA0B;AAAA,EACtE,IAAI,iBAAiB;AAAA,IACnB,mBAAO,MAAM,qDAAqD;AAAA,IAClE,OAAO;AAAA,EACT;AAAA,EACA,mBAAO,MAAM,0DAA0D;AAAA,EACvE,OAAO,UAAU,OAAO;AAAA;AASnB,SAAS,aAAa,CAAC,SAAgC;AAAA,EAC5D,OACE,WAAW,SAAS,oBAAoB,KACxC,WAAW,SAAS,eAAe,aAAa;AAAA;AAU7C,SAAS,aAAa,CAAC,SAAgC;AAAA,EAC5D,OACE,WAAW,SAAS,oBAAoB,KACxC,WAAW,SAAS,eAAe,QAAQ;AAAA;AAUxC,SAAS,wBAAwB,CAAC,SAAgC;AAAA,EACvE,OAAO,WAAW,SAAS,kCAAkC,YAAY;AAAA;AASpE,SAAS,wBAAwB,CAAC,SAAiC;AAAA,EACxE,MAAM,UAAU,WAAW,SAAS,iCAAiC,OAAO;AAAA,EAE5E,MAAM,oBAAoB,OAAO,OAAO,EAAE,YAAY;AAAA,EACtD,MAAM,SAAS,sBAAsB;AAAA,EACrC,mBAAO,MACL,iDAAiD,mBAAmB,OAAO,yBAAyB,+BAA+B,SACrI;AAAA,EACA,OAAO;AAAA;;;ADnJF,SAAS,gBAAgB,CAAC,SAAc,SAAwB;AAAA,GAE/D,YAAY;AAAA,IAChB,IAAI;AAAA,MACF,IAAI,CAAC,UAAU,OAAO,KAAK,CAAC,UAAU,GAAG;AAAA,QACvC,oBAAO,KACL,iFACF;AAAA,QACA;AAAA,MACF;AAAA,MACA,IAAI;AAAA,QACF,MAAM,UAAU,WAAW,OAAO;AAAA,QAClC,MAAM,WAAW,MAAM,MAAM,GAAG,kBAAkB;AAAA,UAChD,SAAS,cAAc,OAAO;AAAA,QAChC,CAAC;AAAA,QACD,IAAI,CAAC,SAAS,IAAI;AAAA,UAChB,oBAAO,KACL,qCAAqC,SAAS,YAChD;AAAA,UACA,oBAAO,KACL,wEACF;AAAA,QACF,EAAO;AAAA,UACL,oBAAO,IAAI,uCAAuC;AAAA;AAAA,QAEpD,OAAO,YAAqB;AAAA,QAC5B,MAAM,UACJ,sBAAsB,QAAQ,WAAW,UAAU,OAAO,UAAU;AAAA,QACtE,oBAAO,KAAK,oCAAoC,SAAS;AAAA,QACzD,oBAAO,KACL,wEACF;AAAA;AAAA,MAEF,OAAO,OAAgB;AAAA,MACvB,MAAM,UACH,OAAmD,QAChD,IAAI,CAAC,MAAM,EAAE,OAAO,EACrB,KAAK,IAAI,MACX,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MACxD,oBAAO,KACL,sCAAsC,kFACxC;AAAA;AAAA,KAED;AAAA;;;AEhD6B,IAAlC;AACkE,IAAlE;;;ACP6B,IAA7B;AAUO,SAAS,kBAAkB,CAAC,SAAwB;AAAA,EACzD,MAAM,UAAU,WAAW,OAAO;AAAA,EAGlC,MAAM,SACJ,UAAU,OAAO,MAAM,YAAY,OAAO,IAAI,aAAa;AAAA,EAC7D,OAAO,2BAAa,EAAE,QAAS,UAAU,IAAe,QAAQ,CAAC;AAAA;;;ACfzC,IAA1B;AAUO,SAAS,mBAAmB,CACjC,SACA,MACA,QACA,OACA;AAAA,EACA,MAAM,gBACH,kBAAkB,QACd,MAAoC,eACrC,eACH,iBAAiB,QACb,MAAmC,cACpC,cACJ;AAAA,EACF,MAAM,oBACH,sBAAsB,QAClB,MAAwC,mBACzC,eACH,kBAAkB,QACd,MAAoC,eACrC,cACJ;AAAA,EACF,MAAM,eACH,iBAAiB,QACb,MAAmC,cACpC,cAAc,eAAe;AAAA,EAGnC,MAAM,kBACJ,OAAO,WAAW,WACd,OAAO,SAAS,MACd,GAAG,OAAO,MAAM,GAAG,GAAG,OACtB,SACF;AAAA,EACN,QAAQ,UAAU,uBAAU,YAAY;AAAA,IACtC;AAAA,IACA,QAAQ;AAAA,IACR,UAAU;AAAA,IACV;AAAA,IACA,QAAQ;AAAA,IACR,QAAQ;AAAA,MACN,QAAQ;AAAA,MACR,YAAY;AAAA,MACZ,OAAO;AAAA,IACT;AAAA,EACF,CAAC;AAAA;;;AFxCH,eAAe,uBAAuB,CACpC,SACA,QACA,WACA,YACoC;AAAA,EACpC,MAAM,SAAS,mBAAmB,OAAO;AAAA,EACzC,MAAM,YAAY,WAAW,OAAO;AAAA,EAEpC,oBAAO,MAAM,YAAY,oBAAoB,WAAW;AAAA,EAExD,MAAM,iBAAiB;AAAA,IACrB,OAAO,OAAO,cAAc,SAAS;AAAA,IACrC,QAAQ,OAAO;AAAA,IACf,QAAQ,QAAQ,UAAU,UAAU;AAAA,IACpC,aAAa,OAAO,eAAe;AAAA,IACnC,iBAAiB,OAAO,aAAa;AAAA,IACrC,kBAAkB,OAAO,oBAAoB;AAAA,IAC7C,iBAAiB,OAAO,mBAAmB;AAAA,IAC3C,eAAe,OAAO,iBAAiB,CAAC;AAAA,IACxC,wBAAwB,EAAE,WAAW,yBAAyB,OAAO,EAAE;AAAA,EACzE;AAAA,EAGA,IAAI,OAAO,QAAQ;AAAA,IACjB,MAAM,SAAS,qBAAW,cAAc;AAAA,IACxC,OAAO;AAAA,MACL,YAAY,OAAO;AAAA,MACnB,MAAM,OAAO;AAAA,MACb,OAAO,OAAO,MAAM,KAAK,CAAC,MAAsC,IAAI;AAAA,QAClE,cAAc,EAAE,eAAe;AAAA,QAC/B,kBAAkB,EAAE,gBAAgB;AAAA,QACpC,cAAc,EAAE,eAAe,MAAM,EAAE,gBAAgB;AAAA,MACzD,IAAI,SAAS;AAAA,MACb,cAAc,OAAO;AAAA,IACvB;AAAA,EACF;AAAA,EAGA,QAAQ,MAAM,UAAU,MAAM,uBAAa,cAAc;AAAA,EACzD,IAAI;AAAA,IAAO,oBAAoB,SAAS,WAAW,OAAO,QAAQ,KAAK;AAAA,EACvE,OAAO;AAAA;AAMT,eAAsB,eAAe,CACnC,SACA,QACoC;AAAA,EACpC,OAAO,wBACL,SACA,QACA,uBAAU,YACV,aACF;AAAA;AAMF,eAAsB,eAAe,CACnC,SACA,QACoC;AAAA,EACpC,OAAO,wBACL,SACA,QACA,uBAAU,YACV,aACF;AAAA;;AGtF6C,IAA/C;AAWA,eAAsB,mBAAmB,CACvC,SACA,QACmB;AAAA,EACnB,MAAM,qBAAqB,WACzB,SACA,0BACA,wBACF;AAAA,EACA,MAAM,qBAAqB,OAAO,SAChC,WAAW,SAAS,+BAA+B,MAAM,KAAK,QAC9D,EACF;AAAA,EAEA,IAAI,CAAC,OAAO,OAAO,wBAAW,EAAE,SAAS,kBAAkB,GAAG;AAAA,IAC5D,MAAM,WAAW,gCAAgC,uCAAuC,OAAO,OAAO,wBAAW,EAAE,KAAK,IAAI;AAAA,IAC5H,oBAAO,MAAM,QAAQ;AAAA,IACrB,MAAM,IAAI,MAAM,QAAQ;AAAA,EAC1B;AAAA,EACA,IAAI,WAAW,MAAM;AAAA,IACnB,oBAAO,MAAM,4CAA4C;AAAA,IACzD,MAAM,aAAa,MAAM,kBAAkB,EAAE,KAAK,CAAC;AAAA,IACnD,WAAW,KAAK;AAAA,IAChB,OAAO;AAAA,EACT;AAAA,EACA,IAAI;AAAA,EACJ,IAAI,OAAO,WAAW,UAAU;AAAA,IAC9B,OAAO;AAAA,EACT,EAAO,SAAI,OAAO,WAAW,YAAY,OAAO,MAAM;AAAA,IACpD,OAAO,OAAO;AAAA,EAChB,EAAO;AAAA,IACL,MAAM,WAAW;AAAA,IACjB,oBAAO,KAAK,QAAQ;AAAA,IACpB,MAAM,iBAAiB,MAAM,kBAAkB,EAAE,KAAK,CAAC;AAAA,IACvD,eAAe,KAAK;AAAA,IACpB,OAAO;AAAA;AAAA,EAET,IAAI,CAAC,KAAK,KAAK,GAAG;AAAA,IAChB,MAAM,WAAW;AAAA,IACjB,oBAAO,KAAK,QAAQ;AAAA,IACpB,MAAM,iBAAiB,MAAM,kBAAkB,EAAE,KAAK,CAAC;AAAA,IACvD,eAAe,KAAK;AAAA,IACpB,OAAO;AAAA,EACT;AAAA,EAEA,MAAM,mBAAmB,oBAAoB,OAAO;AAAA,EAEpD,IAAI;AAAA,IACF,MAAM,WAAW,MAAM,MAAM,GAAG,+BAA+B;AAAA,MAC7D,QAAQ;AAAA,MACR,SAAS;AAAA,WACJ,cAAc,SAAS,IAAI;AAAA,QAC9B,gBAAgB;AAAA,MAClB;AAAA,MACA,MAAM,KAAK,UAAU;AAAA,QACnB,OAAO;AAAA,QACP,OAAO;AAAA,MACT,CAAC;AAAA,IACH,CAAC;AAAA,IAED,IAAI,CAAC,SAAS,IAAI;AAAA,MAChB,oBAAO,MACL,qBAAqB,SAAS,YAAY,SAAS,YACrD;AAAA,MACA,MAAM,IAAI,MACR,qBAAqB,SAAS,YAAY,SAAS,YACrD;AAAA,IACF;AAAA,IAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,IAKlC,IAAI,CAAC,MAAM,OAAO,IAAI,WAAW;AAAA,MAC/B,oBAAO,MAAM,gCAAgC;AAAA,MAC7C,MAAM,IAAI,MAAM,gCAAgC;AAAA,IAClD;AAAA,IAEA,MAAM,YAAY,KAAK,KAAK,GAAG;AAAA,IAE/B,IAAI,CAAC,MAAM,QAAQ,SAAS,KAAK,UAAU,WAAW,oBAAoB;AAAA,MACxE,MAAM,WAAW,oBAAoB,WAAW,UAAU,yCAAyC;AAAA,MACnG,oBAAO,MAAM,QAAQ;AAAA,MACrB,MAAM,iBAAiB,MAAM,kBAAkB,EAAE,KAAK,CAAC;AAAA,MACvD,eAAe,KAAK;AAAA,MACpB,OAAO;AAAA,IACT;AAAA,IAEA,IAAI,KAAK,OAAO;AAAA,MACd,MAAM,QAAQ;AAAA,QACZ,aAAa,KAAK,MAAM;AAAA,QACxB,cAAc;AAAA,QACd,aAAa,KAAK,MAAM;AAAA,MAC1B;AAAA,MAEA,oBAAoB,SAAS,uBAAU,gBAAgB,MAAM,KAAK;AAAA,IACpE;AAAA,IAEA,oBAAO,IAAI,mCAAmC,UAAU,QAAQ;AAAA,IAChE,OAAO;AAAA,IACP,OAAO,OAAgB;AAAA,IACvB,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,oBAAO,MAAM,+BAA+B,SAAS;AAAA,IACrD,MAAM,iBAAiB,QAAQ,QAAQ,IAAI,MAAM,OAAO;AAAA;AAAA;;AC/G1B,IAAlC;AAYA,eAAsB,qBAAqB,CACzC,SACA,QAC4B;AAAA,EAC5B,MAAM,IAAI,OAAO,SAAS;AAAA,EAC1B,MAAM,OAAO,OAAO,QAAQ;AAAA,EAC5B,MAAM,SAAS,OAAO;AAAA,EACtB,MAAM,YAAY,WAChB,SACA,sBACA,aACF;AAAA,EACA,oBAAO,IAAI,+BAA+B,WAAW;AAAA,EAErD,MAAM,UAAU,WAAW,OAAO;AAAA,EAElC,IAAI;AAAA,IACF,MAAM,WAAW,MAAM,MAAM,GAAG,8BAA8B;AAAA,MAC5D,QAAQ;AAAA,MACR,SAAS;AAAA,WACJ,cAAc,OAAO;AAAA,QACxB,gBAAgB;AAAA,MAClB;AAAA,MACA,MAAM,KAAK,UAAU;AAAA,QACnB,OAAO;AAAA,QACP;AAAA,QACA;AAAA,QACA;AAAA,MACF,CAAC;AAAA,IACH,CAAC;AAAA,IAED,IAAI,CAAC,SAAS,IAAI;AAAA,MAChB,MAAM,IAAI,MAAM,6BAA6B,SAAS,YAAY;AAAA,IACpE;AAAA,IAEA,MAAM,OAAO,MAAM,SAAS,KAAK;AAAA,IACjC,MAAM,YAAY;AAAA,IAElB,OAAO,UAAU;AAAA,IACjB,OAAO,OAAgB;AAAA,IACvB,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,MAAM;AAAA;AAAA;AAOV,eAAsB,sBAAsB,CAC1C,SACA,QACiD;AAAA,EACjD,IAAI;AAAA,EACJ,IAAI;AAAA,EACJ,MAAM,YAAY,yBAAyB,OAAO;AAAA,EAClD,oBAAO,IAAI,2CAA2C,WAAW;AAAA,EACjE,MAAM,YAAY,OAAO,SACvB,WAAW,SAAS,uCAAuC,MAAM,KAC/D,QACF,EACF;AAAA,EAEA,MAAM,iBACJ;AAAA,EAEF,IAAI,OAAO,WAAW,UAAU;AAAA,IAC9B,WAAW;AAAA,IACX,aAAa;AAAA,EACf,EAAO;AAAA,IACL,WAAW,OAAO;AAAA,IAClB,aAAa,OAAO,UAAU;AAAA;AAAA,EAGhC,MAAM,WAAW;AAAA,IACf;AAAA,MACE,MAAM;AAAA,MACN,SAAS;AAAA,QACP,EAAE,MAAM,QAAQ,MAAM,WAAW;AAAA,QACjC,EAAE,MAAM,aAAa,WAAW,EAAE,KAAK,SAAS,EAAE;AAAA,MACpD;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,UAAU,WAAW,OAAO;AAAA,EAElC,IAAI;AAAA,IACF,MAAM,cAAmC;AAAA,MACvC,OAAO;AAAA,MACP;AAAA,MACA,YAAY;AAAA,IACd;AAAA,IAEA,MAAM,WAAW,MAAM,MAAM,GAAG,4BAA4B;AAAA,MAC1D,QAAQ;AAAA,MACR,SAAS;AAAA,QACP,gBAAgB;AAAA,WACb,cAAc,OAAO;AAAA,MAC1B;AAAA,MACA,MAAM,KAAK,UAAU,WAAW;AAAA,IAClC,CAAC;AAAA,IAED,IAAI,CAAC,SAAS,IAAI;AAAA,MAChB,MAAM,IAAI,MAAM,qBAAqB,SAAS,QAAQ;AAAA,IACxD;AAAA,IAEA,MAAM,SAAkB,MAAM,SAAS,KAAK;AAAA,IAc5C,MAAM,cAAc;AAAA,IACpB,MAAM,UAAU,YAAY,UAAU,IAAI,SAAS;AAAA,IAEnD,IAAI,YAAY,OAAO;AAAA,MACrB,oBACE,SACA,uBAAU,mBACV,OAAO,WAAW,WAAW,SAAS,OAAO,UAAU,IACvD;AAAA,QACE,aAAa,YAAY,MAAM;AAAA,QAC/B,cAAc,YAAY,MAAM;AAAA,QAChC,aAAa,YAAY,MAAM;AAAA,MACjC,CACF;AAAA,IACF;AAAA,IAEA,IAAI,CAAC,SAAS;AAAA,MACZ,OAAO;AAAA,QACL,OAAO;AAAA,QACP,aAAa;AAAA,MACf;AAAA,IACF;AAAA,IACA,MAAM,aAAa,QAAQ,MAAM,2BAA2B;AAAA,IAC5D,MAAM,QAAQ,aAAa,IAAI,KAAK;AAAA,IACpC,IAAI,CAAC,OAAO;AAAA,MACV,oBAAO,KAAK,yDAAyD;AAAA,IACvE;AAAA,IACA,MAAM,aAAa,SAAS;AAAA,IAC5B,MAAM,cAAc,QAAQ,QAAQ,6BAA6B,EAAE,EAAE,KAAK;AAAA,IAE1E,MAAM,kBAAkB,EAAE,OAAO,YAAY,YAAY;AAAA,IACzD,OAAO;AAAA,IACP,OAAO,OAAgB;AAAA,IACvB,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,oBAAO,MAAM,0BAA0B,SAAS;AAAA,IAChD,OAAO;AAAA,MACL,OAAO;AAAA,MACP,aAAa,UAAU;AAAA,IACzB;AAAA;AAAA;;AC7KmB,IAAvB;;;ACDuB,IAAvB;AAEA,IAAM,cAAc;AAAA,EAClB,KAAK;AAAA,IACH,QAAQ,CAAC,IAAM,IAAM,IAAM,EAAI;AAAA,IAC/B,YAAY,CAAC,IAAM,IAAM,IAAM,EAAI;AAAA,EACrC;AAAA,EACA,SAAS,CAAC,IAAM,IAAM,EAAI;AAAA,EAC1B,KAAK,CAAC,IAAM,KAAM,KAAM,EAAI;AAAA,EAC5B,MAAM,CAAC,KAAM,IAAM,IAAM,EAAI;AAAA,EAC7B,MAAM,CAAC,KAAM,KAAM,KAAM,GAAI;AAAA,EAC7B,WAAW,CAAC,IAAM,IAAM,KAAM,GAAI;AACpC;AAEA,SAAS,UAAU,CACjB,QACA,QACA,OACS;AAAA,EACT,SAAS,IAAI,EAAG,IAAI,MAAM,QAAQ,KAAK;AAAA,IACrC,IAAI,OAAO,SAAS,OAAO,MAAM;AAAA,MAAK,OAAO;AAAA,EAC/C;AAAA,EACA,OAAO;AAAA;AAQF,SAAS,mBAAmB,CAAC,QAAwB;AAAA,EAC1D,IAAI,OAAO,SAAS,IAAI;AAAA,IACtB,OAAO;AAAA,EACT;AAAA,EAIA,IACE,WAAW,QAAQ,GAAG,YAAY,IAAI,MAAM,KAC5C,WAAW,QAAQ,GAAG,YAAY,IAAI,UAAU,GAChD;AAAA,IACA,OAAO;AAAA,EACT;AAAA,EAGA,IACE,WAAW,QAAQ,GAAG,YAAY,OAAO,KACxC,OAAO,OAAO,QAAS,OAAO,KAAK,SAAU,KAC9C;AAAA,IACA,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,GAAG,GAAG;AAAA,IAC1C,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,IAAI,GAAG;AAAA,IAC3C,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,IAAI,GAAG;AAAA,IAC3C,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,SAAS,GAAG;AAAA,IAChD,OAAO;AAAA,EACT;AAAA,EAGA,oBAAO,KACL,sEACF;AAAA,EACA,OAAO;AAAA;;;ADhET,eAAe,iBAAiB,CAC9B,SACA,SACsB;AAAA,EACtB,MAAM,eAAe,WACnB,SACA,oBACA,iBACF;AAAA,EACA,MAAM,eAAe,WAAW,SAAS,oBAAoB,MAAM;AAAA,EACnE,MAAM,sBAAsB,WAC1B,SACA,2BACA,EACF;AAAA,EACA,MAAM,UAAU,WAAW,OAAO;AAAA,EAElC,MAAM,QAAQ,QAAQ,SAAU;AAAA,EAChC,MAAM,QAAQ,QAAQ,SAAU;AAAA,EAChC,MAAM,eAAe,QAAQ,gBAAiB;AAAA,EAC9C,MAAM,SAAS,QAAQ,UAAU;AAAA,EAEjC,IAAI;AAAA,IACF,MAAM,MAAM,MAAM,MAAM,GAAG,wBAAwB;AAAA,MACjD,QAAQ;AAAA,MACR,SAAS;AAAA,WACJ,cAAc,OAAO;AAAA,QACxB,gBAAgB;AAAA,WAEZ,WAAW,QAAQ,EAAE,QAAQ,aAAa,IAAI,CAAC;AAAA,MACrD;AAAA,MACA,MAAM,KAAK,UAAU;AAAA,QACnB;AAAA,QACA;AAAA,QACA,OAAO,QAAQ;AAAA,QACf;AAAA,WACI,gBAAgB,EAAE,aAAa;AAAA,MACrC,CAAC;AAAA,IACH,CAAC;AAAA,IAED,IAAI,CAAC,IAAI,IAAI;AAAA,MACX,MAAM,MAAM,MAAM,IAAI,KAAK;AAAA,MAC3B,MAAM,IAAI,MAAM,oBAAoB,IAAI,WAAW,KAAK;AAAA,IAC1D;AAAA,IAGA,OAAO,MAAM,IAAI,YAAY;AAAA,IAC7B,OAAO,KAAc;AAAA,IACrB,MAAM,UAAU,eAAe,QAAQ,IAAI,UAAU,OAAO,GAAG;AAAA,IAC/D,MAAM,IAAI,MAAM,2CAA2C,SAAS;AAAA;AAAA;AAOxE,eAAsB,mBAAmB,CACvC,SACA,OACiB;AAAA,EACjB,IAAI,YAAY,WACd,SACA,8BACA,wBACF;AAAA,EACA,oBAAO,IAAI,uCAAuC,WAAW;AAAA,EAE7D,MAAM,UAAU,WAAW,OAAO;AAAA,EAGlC,IAAI;AAAA,EACJ,IAAI,cAAgD;AAAA,EAEpD,IAAI,iBAAiB,QAAQ,iBAAiB,MAAM;AAAA,IAClD,OAAO;AAAA,EACT,EAAO,SAAI,OAAO,SAAS,KAAK,GAAG;AAAA,IAGjC,MAAM,mBAAmB,oBAAoB,KAAK;AAAA,IAClD,oBAAO,MAAM,kCAAkC,kBAAkB;AAAA,IAEjE,MAAM,aAAa,IAAI,WAAW,KAAK;AAAA,IACvC,OAAO,IAAI,KAAK,CAAC,UAAU,GAAG,EAAE,MAAM,iBAAiB,CAAC;AAAA,EAC1D,EAAO,SACL,OAAO,UAAU,YACjB,UAAU,QACT,MAAc,SAAS,MACxB;AAAA,IACA,MAAM,SAAS;AAAA,IACf,IACE,EAAE,OAAO,iBAAiB,SAC1B,EAAE,OAAO,iBAAiB,SAC1B,CAAC,OAAO,SAAS,OAAO,KAAK,GAC7B;AAAA,MACA,MAAM,IAAI,MACR,yDACF;AAAA,IACF;AAAA,IAEA,IAAI,OAAO,SAAS,OAAO,KAAK,GAAG;AAAA,MAEjC,IAAI,WAAW,OAAO;AAAA,MACtB,IAAI,CAAC,UAAU;AAAA,QACb,WAAW,oBAAoB,OAAO,KAAK;AAAA,QAC3C,oBAAO,MAAM,kCAAkC,UAAU;AAAA,MAC3D,EAAO;AAAA,QACL,oBAAO,MAAM,6BAA6B,UAAU;AAAA;AAAA,MAGtD,MAAM,aAAa,IAAI,WAAW,OAAO,KAAK;AAAA,MAC9C,OAAO,IAAI,KAAK,CAAC,UAAU,GAAG,EAAE,MAAM,SAAS,CAAC;AAAA,IAClD,EAAO;AAAA,MACL,OAAO,OAAO;AAAA;AAAA,IAEhB,cAAc;AAAA,IACd,IAAI,OAAO,OAAO,UAAU,YAAY,OAAO,OAAO;AAAA,MACpD,YAAY,OAAO;AAAA,IACrB;AAAA,EACF,EAAO;AAAA,IACL,MAAM,IAAI,MACR,mLACF;AAAA;AAAA,EAGF,MAAM,OAAQ,KAAc,QAAQ;AAAA,EACpC,MAAM,WACH,KAAc,SACd,KAAK,SAAS,KAAK,KAAK,KAAK,SAAS,MAAM,IACzC,kBACA,KAAK,SAAS,KAAK,IACjB,kBACA,KAAK,SAAS,KAAK,IACjB,kBACA,KAAK,SAAS,MAAM,IAClB,mBACA;AAAA,EAEZ,MAAM,WAAW,IAAI;AAAA,EACrB,SAAS,OAAO,QAAQ,MAAM,QAAQ;AAAA,EACtC,SAAS,OAAO,SAAS,OAAO,SAAS,CAAC;AAAA,EAC1C,IAAI,aAAa;AAAA,IACf,IAAI,OAAO,YAAY,aAAa,UAAU;AAAA,MAC5C,SAAS,OAAO,YAAY,OAAO,YAAY,QAAQ,CAAC;AAAA,IAC1D;AAAA,IACA,IAAI,OAAO,YAAY,oBAAoB,UAAU;AAAA,MACnD,SAAS,OAAO,mBAAmB,OAAO,YAAY,eAAe,CAAC;AAAA,IACxE;AAAA,IACA,IAAI,OAAO,YAAY,WAAW,UAAU;AAAA,MAC1C,SAAS,OAAO,UAAU,OAAO,YAAY,MAAM,CAAC;AAAA,IACtD;AAAA,IACA,IAAI,OAAO,YAAY,gBAAgB,UAAU;AAAA,MAC/C,SAAS,OAAO,eAAe,OAAO,YAAY,WAAW,CAAC;AAAA,IAChE;AAAA,IACA,IAAI,MAAM,QAAQ,YAAY,sBAAsB,GAAG;AAAA,MACrD,WAAW,KAAK,YAAY,wBAAwB;AAAA,QAClD,SAAS,OAAO,6BAA6B,OAAO,CAAC,CAAC;AAAA,MACxD;AAAA,IACF;AAAA,EACF;AAAA,EAEA,IAAI;AAAA,IACF,MAAM,WAAW,MAAM,MAAM,GAAG,gCAAgC;AAAA,MAC9D,QAAQ;AAAA,MACR,SAAS;AAAA,WACJ,cAAc,OAAO;AAAA,MAC1B;AAAA,MACA,MAAM;AAAA,IACR,CAAC;AAAA,IAED,IAAI,CAAC,SAAS,IAAI;AAAA,MAChB,MAAM,IAAI,MACR,+BAA+B,SAAS,UAAU,SAAS,YAC7D;AAAA,IACF;AAAA,IAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,IAClC,OAAO,KAAK,QAAQ;AAAA,IACpB,OAAO,OAAgB;AAAA,IACvB,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,oBAAO,MAAM,wBAAwB,SAAS;AAAA,IAC9C,MAAM;AAAA;AAAA;AAOV,eAAsB,kBAAkB,CACtC,SACA,OACsB;AAAA,EAEtB,MAAM,UACJ,OAAO,UAAU,WACb,EAAE,MAAM,MAAM,IACb;AAAA,EAEP,MAAM,gBACJ,QAAQ,SACP,WAAW,SAAS,oBAAoB,iBAAiB;AAAA,EAC5D,oBAAO,IAAI,wCAAwC,eAAe;AAAA,EAClE,IAAI;AAAA,IACF,OAAO,MAAM,kBAAkB,SAAS,OAAO;AAAA,IAC/C,OAAO,OAAgB;AAAA,IACvB,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,oBAAO,MAAM,4BAA4B,SAAS;AAAA,IAClD,MAAM;AAAA;AAAA;;AErNwB,IAAlC;AAC+B,IAA/B;;;ACNuB,IAAvB;AAC+B,IAA/B;AAKO,SAAS,qBAAqB,GAGR;AAAA,EAC3B,OAAO,SAAS,MAAM,YAA8C;AAAA,IAClE,IAAI;AAAA,MACF,IAAI,iBAAiB,2BAAgB;AAAA,QACnC,MAAM,cAAc,KAAK,QAAQ,wBAAwB,EAAE;AAAA,QAC3D,KAAK,MAAM,WAAW;AAAA,QACtB,OAAO;AAAA,MACT;AAAA,MACA,OAAO;AAAA,MACP,OAAO,WAAoB;AAAA,MAC3B,MAAM,UACJ,qBAAqB,QAAQ,UAAU,UAAU,OAAO,SAAS;AAAA,MACnE,oBAAO,KAAK,+BAA+B,SAAS;AAAA,MACpD,OAAO;AAAA;AAAA;AAAA;;;ADPb,eAAe,yBAAyB,CACtC,SACA,QACA,WACA,YACkC;AAAA,EAClC,MAAM,SAAS,mBAAmB,OAAO;AAAA,EACzC,MAAM,YAAY,WAAW,OAAO;AAAA,EACpC,qBAAO,IAAI,kBAAkB,oBAAoB,WAAW;AAAA,EAC5D,MAAM,cAAc,OAAO,eAAe;AAAA,EAC1C,MAAM,gBAAgB,CAAC,CAAC,OAAO;AAAA,EAE/B,IAAI,eAAe;AAAA,IACjB,qBAAO,KACL,4HACF;AAAA,EACF;AAAA,EAEA,IAAI;AAAA,IACF,QAAQ,QAAQ,UAAU,MAAM,0BAAe;AAAA,MAC7C,OAAO,OAAO,cAAc,SAAS;AAAA,MACrC,QAAQ;AAAA,MACR,QAAQ,OAAO;AAAA,MACf;AAAA,MACA,yBAAyB,sBAAsB;AAAA,IACjD,CAAC;AAAA,IAED,IAAI,OAAO;AAAA,MACT,oBAAoB,SAAS,WAAW,OAAO,QAAQ,KAAK;AAAA,IAC9D;AAAA,IACA,OAAO;AAAA,IACP,OAAO,OAAgB;AAAA,IACvB,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,qBAAO,MAAM,2BAA2B,SAAS;AAAA,IACjD,MAAM;AAAA;AAAA;AAOV,eAAsB,iBAAiB,CACrC,SACA,QACkC;AAAA,EAClC,OAAO,0BACL,SACA,QACA,wBAAU,cACV,aACF;AAAA;AAMF,eAAsB,iBAAiB,CACrC,SACA,QACkC;AAAA,EAClC,OAAO,0BACL,SACA,QACA,wBAAU,cACV,aACF;AAAA;;AE3EwB,IAA1B;;;ACJ0B,IAA1B;AACyF,IAAzF;AAGA,SAAS,wBAAwB,CAC/B,WACqC;AAAA,EACrC,MAAM,aAAa,UAAU,YAAY;AAAA,EACzC,MAAM,mBAAqC,WAAW,SAAS,IAAI,IAC/D,eACA;AAAA,EAEJ,IAAI;AAAA,IACF,OAAO,oCAAiB,SAA0B;AAAA,IAClD,OAAO,OAAgB;AAAA,IAEvB,OAAO,+BAAY,gBAAgB;AAAA;AAAA;AAWvC,eAAsB,YAAY,CAChC,SACA,OACA,QACA;AAAA,EACA,MAAM,YACJ,UAAU,wBAAU,aAChB,cAAc,OAAO,IACrB,cAAc,OAAO;AAAA,EAC3B,MAAM,SAAS,yBAAyB,SAAS,EAAE,OAAO,MAAM;AAAA,EAChE,OAAO;AAAA;AAUT,eAAsB,cAAc,CAClC,SACA,OACA,QACA;AAAA,EACA,MAAM,YACJ,UAAU,wBAAU,aAChB,cAAc,OAAO,IACrB,cAAc,OAAO;AAAA,EAC3B,OAAO,yBAAyB,SAAS,EAAE,OAAO,MAAM;AAAA;;;AD9C1D,eAAsB,qBAAqB,CACzC,WACE,QAAQ,YAAY,wBAAU,cACb;AAAA,EACnB,OAAO,MAAM,aAAa,SAAS,WAAW,MAAM;AAAA;AAMtD,eAAsB,qBAAqB,CACzC,WACE,QAAQ,YAAY,wBAAU,cACf;AAAA,EACjB,OAAO,MAAM,eAAe,SAAS,WAAW,MAAM;AAAA;;AZQjD,IAAM,eAAuB;AAAA,EAClC,MAAM;AAAA,EACN,aAAa;AAAA,EACb,QAAQ;AAAA,IACN,gBAAgB,QAAQ,IAAI;AAAA,IAC5B,iBAAiB,QAAQ,IAAI;AAAA,IAC7B,oBAAoB,QAAQ,IAAI;AAAA,IAChC,oBAAoB,QAAQ,IAAI;AAAA,IAChC,aAAa,QAAQ,IAAI;AAAA,IACzB,aAAa,QAAQ,IAAI;AAAA,IACzB,wBAAwB,QAAQ,IAAI;AAAA,IACpC,0BAA0B,QAAQ,IAAI;AAAA,IACtC,sBAAsB,QAAQ,IAAI;AAAA,IAClC,6BAA6B,QAAQ,IAAI;AAAA,IACzC,gCAAgC,QAAQ,IAAI;AAAA,IAC5C,qCACE,QAAQ,IAAI;AAAA,IACd,+BAA+B,QAAQ,IAAI;AAAA,EAC7C;AAAA,OACM,KAAI,CAAC,SAAS,SAAS;AAAA,IAI3B,iBAAiB,SAAS,OAAO;AAAA;AAAA,EAGnC,QAAQ;AAAA,KACL,wBAAU,iBAAiB,OAC1B,SACA,WACG;AAAA,MACH,OAAO,oBAAoB,SAAS,MAAM;AAAA;AAAA,KAE3C,wBAAU,wBAAwB,OACjC,SACA,WACG;AAAA,MACH,OAAO,sBAAsB,SAAS,MAAM;AAAA;AAAA,KAE7C,wBAAU,wBAAwB,OACjC,SACA,WACG;AAAA,MACH,OAAO,sBAAsB,SAAS,MAAM;AAAA;AAAA,KAE7C,wBAAU,aAAa,OACtB,SACA,WACG;AAAA,MACH,OAAO,gBAAgB,SAAS,MAAM;AAAA;AAAA,KAEvC,wBAAU,aAAa,OACtB,SACA,WACG;AAAA,MACH,OAAO,gBAAgB,SAAS,MAAM;AAAA;AAAA,KAEvC,wBAAU,QAAQ,OACjB,SACA,WAKG;AAAA,MACH,OAAO,sBAAsB,SAAS,MAAM;AAAA;AAAA,KAE7C,wBAAU,oBAAoB,OAC7B,SACA,WACG;AAAA,MACH,OAAO,uBAAuB,SAAS,MAAM;AAAA;AAAA,KAE9C,wBAAU,gBAAgB,OACzB,SACA,UACG;AAAA,MACH,OAAO,oBAAoB,SAAS,KAAK;AAAA;AAAA,KAE1C,wBAAU,iBAAiB,OAC1B,SACA,UACG;AAAA,MACH,OAAO,mBAAmB,SAAS,KAAK;AAAA;AAAA,KAEzC,wBAAU,eAAe,OACxB,SACA,WACG;AAAA,MACH,OAAO,kBAAkB,SAAS,MAAM;AAAA;AAAA,KAEzC,wBAAU,eAAe,OACxB,SACA,WACG;AAAA,MACH,OAAO,kBAAkB,SAAS,MAAM;AAAA;AAAA,EAE5C;AAAA,EACA,OAAO;AAAA,IACL;AAAA,MACE,MAAM;AAAA,MACN,OAAO;AAAA,QACL;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,MAAM,UAAU,WAAW,OAAO;AAAA,YAClC,MAAM,WAAW,MAAM,MAAM,GAAG,kBAAkB;AAAA,cAChD,SAAS,cAAc,OAAO;AAAA,YAChC,CAAC;AAAA,YACD,MAAM,OAAO,MAAM,SAAS,KAAK;AAAA,YACjC,qBAAO,IACL,EAAE,MAAO,MAA+B,MAAM,UAAU,MAAM,GAC9D,kBACF;AAAA,YACA,IAAI,CAAC,SAAS,IAAI;AAAA,cAChB,MAAM,IAAI,MACR,sCAAsC,SAAS,YACjD;AAAA,YACF;AAAA;AAAA,QAEJ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,MAAM,YAAY,MAAM,QAAQ,SAC9B,wBAAU,gBACV;AAAA,gBACE,MAAM;AAAA,cACR,CACF;AAAA,cACA,qBAAO,IAAI,EAAE,UAAU,GAAG,WAAW;AAAA,cACrC,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,iCAAiC,SAAS;AAAA,cACvD,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,MAAM,OAAO,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,gBACxD,QAAQ;AAAA,cACV,CAAC;AAAA,cACD,IAAI,KAAK,WAAW,GAAG;AAAA,gBACrB,MAAM,IAAI,MAAM,yBAAyB;AAAA,cAC3C;AAAA,cACA,qBAAO,IAAI,EAAE,KAAK,GAAG,gCAAgC;AAAA,cACrD,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,6BAA6B,SAAS;AAAA,cACnD,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,MAAM,OAAO,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,gBACxD,QAAQ;AAAA,cACV,CAAC;AAAA,cACD,IAAI,KAAK,WAAW,GAAG;AAAA,gBACrB,MAAM,IAAI,MAAM,yBAAyB;AAAA,cAC3C;AAAA,cACA,qBAAO,IAAI,EAAE,KAAK,GAAG,gCAAgC;AAAA,cACrD,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,6BAA6B,SAAS;AAAA,cACnD,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,qBAAO,IAAI,8BAA8B;AAAA,YACzC,IAAI;AAAA,cACF,MAAM,QAAQ,MAAM,QAAQ,SAAS,wBAAU,OAAO;AAAA,gBACpD,QAAQ;AAAA,gBACR,OAAO;AAAA,gBACP,MAAM;AAAA,cACR,CAAC;AAAA,cACD,qBAAO,IAAI,EAAE,MAAM,GAAG,sCAAsC;AAAA,cAC5D,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,mCAAmC,SAAS;AAAA,cACzD,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,qBAAO,IAAI,+BAA+B;AAAA,cAC1C,IAAI;AAAA,gBACF,MAAM,SAAS,MAAM,QAAQ,SAC3B,wBAAU,mBACV,mLACF;AAAA,gBAEA,IACE,UACA,OAAO,WAAW,YAClB,WAAW,UACX,iBAAiB,QACjB;AAAA,kBACA,qBAAO,IAAI,EAAE,OAAO,GAAG,mBAAmB;AAAA,gBAC5C,EAAO;AAAA,kBACL,qBAAO,MACL,4CACA,MACF;AAAA;AAAA,gBAEF,OAAO,GAAY;AAAA,gBACnB,MAAM,UAAU,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AAAA,gBACzD,qBAAO,MAAM,oCAAoC,SAAS;AAAA;AAAA,cAE5D,OAAO,GAAY;AAAA,cACnB,MAAM,UAAU,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AAAA,cACzD,qBAAO,MACL,2CAA2C,SAC7C;AAAA;AAAA;AAAA,QAGN;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,qBAAO,IAAI,2BAA2B;AAAA,YACtC,IAAI;AAAA,cACF,MAAM,WAAW,MAAM,MACrB,+EACF;AAAA,cACA,MAAM,cAAc,MAAM,SAAS,YAAY;AAAA,cAC/C,MAAM,gBAAgB,MAAM,QAAQ,SAClC,wBAAU,eACV,OAAO,KAAK,IAAI,WAAW,WAAW,CAAC,CACzC;AAAA,cACA,qBAAO,IACL,EAAE,cAAc,GAChB,mCACF;AAAA,cACA,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,gCAAgC,SAAS;AAAA,cACtD,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,MAAM,SAAS;AAAA,YACf,MAAM,SAAS,MAAM,QAAQ,SAC3B,wBAAU,uBACV,EAAE,QAAQ,WAAW,wBAAU,WAAW,CAC5C;AAAA,YACA,IAAI,CAAC,MAAM,QAAQ,MAAM,KAAK,OAAO,WAAW,GAAG;AAAA,cACjD,MAAM,IAAI,MACR,6DACF;AAAA,YACF;AAAA,YACA,qBAAO,IAAI,EAAE,OAAO,GAAG,kBAAkB;AAAA;AAAA,QAE7C;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,MAAM,SAAS;AAAA,YACf,MAAM,SAAS,MAAM,QAAQ,SAC3B,wBAAU,uBACV,EAAE,QAAQ,WAAW,wBAAU,WAAW,CAC5C;AAAA,YACA,MAAM,cAAc,MAAM,QAAQ,SAChC,wBAAU,uBACV;AAAA,cACE;AAAA,cACA,WAAW,wBAAU;AAAA,YACvB,CACF;AAAA,YACA,IAAI,gBAAgB,QAAQ;AAAA,cAC1B,MAAM,IAAI,MACR,mDAAmD,iBAAiB,cACtE;AAAA,YACF;AAAA,YACA,qBAAO,IAAI,EAAE,YAAY,GAAG,cAAc;AAAA;AAAA,QAE9C;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,MAAM,WAAW,MAAM,QAAQ,SAC7B,wBAAU,gBACV;AAAA,gBACE,MAAM;AAAA,cACR,CACF;AAAA,cACA,IAAI,CAAC,UAAU;AAAA,gBACb,MAAM,IAAI,MAAM,2BAA2B;AAAA,cAC7C;AAAA,cACA,qBAAO,IAAI,+BAA+B;AAAA,cAC1C,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,wCAAwC,SAAS;AAAA,cAC9D,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,gBAC1D,QAAQ;AAAA,cACV,CAAC;AAAA,cACD,IAAI,CAAC,UAAU,OAAO,WAAW,GAAG;AAAA,gBAClC,MAAM,IAAI,MAAM,uCAAuC;AAAA,cACzD;AAAA,cACA,qBAAO,IAAI,EAAE,OAAO,GAAG,gCAAgC;AAAA,cACvD,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,+CAA+C,SAAS;AAAA,cACrE,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA2B;AAAA,YACpC,IAAI;AAAA,cACF,MAAM,SAAmB,CAAC;AAAA,cAC1B,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,gBAC1D,QAAQ;AAAA,gBACR,eAAe,CAAC,UAAkB;AAAA,kBAChC,OAAO,KAAK,KAAK;AAAA;AAAA,cAErB,CAAC;AAAA,cACD,IAAI,CAAC,UAAU,OAAO,WAAW,GAAG;AAAA,gBAClC,MAAM,IAAI,MAAM,iCAAiC;AAAA,cACnD;AAAA,cACA,IAAI,OAAO,WAAW,GAAG;AAAA,gBACvB,MAAM,IAAI,MAAM,8BAA8B;AAAA,cAChD;AAAA,cACA,qBAAO,IAAI,EAAE,QAAQ,OAAO,QAAQ,QAAQ,OAAO,UAAU,GAAG,EAAE,EAAE,GAAG,0BAA0B;AAAA,cACjG,OAAO,OAAgB;AAAA,cACvB,MAAM,UACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,cACvD,qBAAO,MAAM,mCAAmC,SAAS;AAAA,cACzD,MAAM;AAAA;AAAA;AAAA,QAGZ;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAEA,IAAe;",
21
- "debugId": "F03D6434951BC08A64756E2164756E21",
21
+ "mappings": ";;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;AAgBkC,IAAlC;;;ACfuB,IAAvB;;;ACAuB,IAAvB;AAEA,SAAS,WAAW,CAAC,KAAiC;AAAA,EACpD,IAAI,OAAO,YAAY,eAAe,CAAC,QAAQ,KAAK;AAAA,IAClD;AAAA,EACF;AAAA,EACA,MAAM,QAAQ,QAAQ,IAAI;AAAA,EAC1B,OAAO,UAAU,YAAY,YAAY,OAAO,KAAK;AAAA;AAGhD,SAAS,UAAU,CACxB,SACA,KACA,cACoB;AAAA,EACpB,MAAM,QAAQ,QAAQ,WAAW,GAAG;AAAA,EACpC,IAAI,UAAU,aAAa,UAAU,MAAM;AAAA,IACzC,OAAO,OAAO,KAAK;AAAA,EACrB;AAAA,EACA,OAAO,YAAY,GAAG,KAAK;AAAA;AActB,SAAS,iBAAiB,CAC/B,SACA,KACA,cACQ;AAAA,EACR,MAAM,QAAQ,WAAW,SAAS,GAAG;AAAA,EACrC,IAAI,UAAU,WAAW;AAAA,IACvB,OAAO;AAAA,EACT;AAAA,EACA,MAAM,SAAS,OAAO,SAAS,OAAO,EAAE;AAAA,EACxC,IAAI,CAAC,OAAO,SAAS,MAAM,GAAG;AAAA,IAC5B,MAAM,IAAI,MAAM,YAAY,sCAAsC,OAAO;AAAA,EAC3E;AAAA,EACA,OAAO;AAAA;AAGF,SAAS,iBAAiB,CAC/B,SACA,KACA,cACS;AAAA,EACT,MAAM,QAAQ,WAAW,SAAS,GAAG;AAAA,EACrC,IAAI,UAAU,WAAW;AAAA,IACvB,OAAO;AAAA,EACT;AAAA,EACA,MAAM,aAAa,MAAM,YAAY;AAAA,EACrC,OAAO,eAAe,UAAU,eAAe,OAAO,eAAe;AAAA;AAGhE,SAAS,SAAS,GAAY;AAAA,EACnC,OACE,OAAO,eAAe,eACtB,OAAQ,WAAuC,aAAa;AAAA;AAIzD,SAAS,WAAW,CAAC,SAAiC;AAAA,EAC3D,OAAO,UAAU,KAAK,CAAC,CAAC,WAAW,SAAS,yBAAyB;AAAA;AAGhE,SAAS,SAAS,CAAC,SAA4C;AAAA,EACpE,OAAO,WAAW,SAAS,gBAAgB;AAAA;AAGtC,SAAS,kBAAkB,CAAC,SAA4C;AAAA,EAC7E,MAAM,kBAAkB,WAAW,SAAS,0BAA0B;AAAA,EACtE,IAAI,iBAAiB;AAAA,IACnB,mBAAO,MAAM,2CAA2C;AAAA,IACxD,OAAO;AAAA,EACT;AAAA,EACA,mBAAO,MAAM,yDAAyD;AAAA,EACtE,OAAO,UAAU,OAAO;AAAA;AAGnB,SAAS,aAAa,CAC3B,SACA,eAAe,OACS;AAAA,EAIxB,IAAI,UAAU,KAAK,CAAC,kBAAkB,SAAS,gCAAgC,KAAK,GAAG;AAAA,IACrF,OAAO,CAAC;AAAA,EACV;AAAA,EACA,MAAM,MAAM,eAAe,mBAAmB,OAAO,IAAI,UAAU,OAAO;AAAA,EAC1E,OAAO,MAAM,EAAE,eAAe,UAAU,MAAM,IAAI,CAAC;AAAA;AAG9C,SAAS,UAAU,CAAC,SAAgC;AAAA,EACzD,MAAM,aAAa,WAAW,SAAS,yBAAyB;AAAA,EAChE,MAAM,UACJ,UAAU,KAAK,aACX,aACC,WAAW,SAAS,iBAAiB,KAAK;AAAA,EACjD,mBAAO,MAAM,sBAAsB,SAAS;AAAA,EAC5C,OAAO;AAAA;AAGF,SAAS,mBAAmB,CAAC,SAAgC;AAAA,EAClE,MAAM,eAAe,UAAU,IAC1B,WAAW,SAAS,8BAA8B,KACnD,WAAW,SAAS,yBAAyB,IAC7C,WAAW,SAAS,sBAAsB;AAAA,EAE9C,IAAI,cAAc;AAAA,IAChB,mBAAO,MAAM,sCAAsC,cAAc;AAAA,IACjE,OAAO;AAAA,EACT;AAAA,EAEA,mBAAO,MAAM,0DAA0D;AAAA,EACvE,OAAO,WAAW,OAAO;AAAA;AAGpB,SAAS,aAAa,CAAC,SAAgC;AAAA,EAC5D,OACE,WAAW,SAAS,oBAAoB,KAAK,WAAW,SAAS,aAAa,KAAK;AAAA;AAIhF,SAAS,aAAa,CAAC,SAAgC;AAAA,EAC5D,OAAO,WAAW,SAAS,oBAAoB,KAAK,WAAW,SAAS,aAAa,KAAK;AAAA;AAGrF,SAAS,iBAAiB,CAAC,SAAgC;AAAA,EAChE,OAAO,WAAW,SAAS,wBAAwB,KAAK;AAAA;AAGnD,SAAS,wBAAwB,CAAC,SAAgC;AAAA,EACvE,OAAO,WAAW,SAAS,gCAAgC,KAAK;AAAA;AAG3D,SAAS,qBAAqB,CAAC,SAAgC;AAAA,EACpE,OAAO,WAAW,SAAS,4BAA4B,KAAK;AAAA;AAGvD,SAAS,WAAW,CAAC,SAAgC;AAAA,EAC1D,OAAO,WAAW,SAAS,kBAAkB,KAAK;AAAA;AAG7C,SAAS,WAAW,CAAC,SAAgC;AAAA,EAC1D,OAAO,WAAW,SAAS,kBAAkB,KAAK;AAAA;AAG7C,SAAS,kBAAkB,CAAC,SAAgC;AAAA,EACjE,OAAO,WAAW,SAAS,yBAAyB,KAAK;AAAA;AAGpD,SAAS,aAAa,CAAC,SAAgC;AAAA,EAC5D,OAAO,WAAW,SAAS,oBAAoB,KAAK;AAAA;AAG/C,SAAS,wBAAwB,CAAC,SAAiC;AAAA,EACxE,OAAO,kBAAkB,SAAS,iCAAiC,KAAK;AAAA;AAGnE,SAAS,sBAAsB,CAAC,SAAgC;AAAA,EACrE,OAAO,kBAAkB,SAAS,+BAA+B,IAAI;AAAA;AAGhE,SAAS,4BAA4B,CAAC,SAAgC;AAAA,EAC3E,OAAO,kBAAkB,SAAS,uCAAuC,IAAI;AAAA;AAGxE,SAAS,gBAAgB,CAAC,SAAgC;AAAA,EAC/D,OAAO,WAAW,SAAS,uBAAuB,KAAK;AAAA;AAGlD,SAAS,kBAAkB,CAAC,SAAgC;AAAA,EACjE,OAAO,kBAAkB,SAAS,2BAA2B,OAAO;AAAA;;;ADjLrE,WAAuC,wBAAwB;AAEzD,SAAS,gBAAgB,CAC9B,SACA,SACM;AAAA,EACD,4BAA4B,OAAO;AAAA;AAG1C,eAAe,2BAA2B,CAAC,SAAuC;AAAA,EAChF,IAAI,UAAU,GAAG;AAAA,IACf,oBAAO,MAAM,yDAAyD;AAAA,IACtE;AAAA,EACF;AAAA,EAEA,MAAM,SAAS,UAAU,OAAO;AAAA,EAEhC,IAAI,CAAC,QAAQ;AAAA,IACX,oBAAO,KACL,gDACE,mEACJ;AAAA,IACA;AAAA,EACF;AAAA,EAEA,IAAI;AAAA,IACF,MAAM,UAAU,WAAW,OAAO;AAAA,IAClC,MAAM,WAAW,MAAM,MAAM,GAAG,kBAAkB;AAAA,MAChD,SAAS,cAAc,OAAO;AAAA,IAChC,CAAC;AAAA,IAED,IAAI,CAAC,SAAS,IAAI;AAAA,MAChB,oBAAO,KACL,uCAAuC,SAAS,UAAU,SAAS,iBACjE,+CACJ;AAAA,MACA;AAAA,IACF;AAAA,IACA,OAAO,OAAO;AAAA,IACd,MAAM,UAAU,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,IACrE,oBAAO,KAAK,kCAAkC,+CAA+C;AAAA;AAAA;;;AExC1E,IAAvB;;;ACLuB,IAAvB;AAEA,IAAM,cAAc;AAAA,EAClB,KAAK;AAAA,IACH,QAAQ,CAAC,IAAM,IAAM,IAAM,EAAI;AAAA,IAC/B,YAAY,CAAC,IAAM,IAAM,IAAM,EAAI;AAAA,EACrC;AAAA,EACA,SAAS,CAAC,IAAM,IAAM,EAAI;AAAA,EAC1B,KAAK,CAAC,IAAM,KAAM,KAAM,EAAI;AAAA,EAC5B,MAAM,CAAC,KAAM,IAAM,IAAM,EAAI;AAAA,EAC7B,MAAM,CAAC,KAAM,KAAM,KAAM,GAAI;AAAA,EAC7B,WAAW,CAAC,IAAM,IAAM,KAAM,GAAI;AACpC;AAEA,IAAM,4BAA4B;AAWlC,SAAS,UAAU,CAAC,QAAgB,QAAgB,UAAsC;AAAA,EACxF,SAAS,IAAI,EAAG,IAAI,SAAS,QAAQ,KAAK;AAAA,IACxC,MAAM,eAAe,SAAS;AAAA,IAC9B,IAAI,iBAAiB,aAAa,OAAO,SAAS,OAAO,cAAc;AAAA,MACrE,OAAO;AAAA,IACT;AAAA,EACF;AAAA,EACA,OAAO;AAAA;AAGF,SAAS,mBAAmB,CAAC,QAA+B;AAAA,EACjE,IAAI,OAAO,SAAS,2BAA2B;AAAA,IAC7C,OAAO;AAAA,EACT;AAAA,EAGA,IACE,WAAW,QAAQ,GAAG,YAAY,IAAI,MAAM,KAC5C,WAAW,QAAQ,GAAG,YAAY,IAAI,UAAU,GAChD;AAAA,IACA,OAAO;AAAA,EACT;AAAA,EAGA,MAAM,YAAY,OAAO;AAAA,EACzB,MAAM,aAAa,OAAO;AAAA,EAC1B,IACE,WAAW,QAAQ,GAAG,YAAY,OAAO,KACxC,cAAc,OAAQ,eAAe,cAAc,aAAa,SAAU,KAC3E;AAAA,IACA,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,GAAG,GAAG;AAAA,IAC1C,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,IAAI,GAAG;AAAA,IAC3C,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,IAAI,GAAG;AAAA,IAC3C,OAAO;AAAA,EACT;AAAA,EAGA,IAAI,WAAW,QAAQ,GAAG,YAAY,SAAS,GAAG;AAAA,IAChD,OAAO;AAAA,EACT;AAAA,EAEA,oBAAO,KAAK,sEAAsE;AAAA,EAClF,OAAO;AAAA;AAGF,SAAS,uBAAuB,CAAC,UAAiC;AAAA,EACvE,QAAQ;AAAA,SACD;AAAA,MACH,OAAO;AAAA,SACJ;AAAA,MACH,OAAO;AAAA,SACJ;AAAA,MACH,OAAO;AAAA,SACJ;AAAA,MACH,OAAO;AAAA,SACJ;AAAA,MACH,OAAO;AAAA,SACJ;AAAA,MACH,OAAO;AAAA,SACJ;AAAA,MACH,OAAO;AAAA;AAAA;AAIN,SAAS,sBAAsB,CAAC,UAAiC;AAAA,EACtE,MAAM,MAAM,wBAAwB,QAAQ;AAAA,EAC5C,OAAO,aAAa;AAAA;;;AD5EtB,SAAS,YAAY,CAAC,OAAsC;AAAA,EAC1D,OAAO,iBAAiB,QAAQ,iBAAiB;AAAA;AAGnD,SAAS,QAAQ,CAAC,OAAiC;AAAA,EACjD,OAAO,OAAO,SAAS,KAAK;AAAA;AAG9B,SAAS,0BAA0B,CAAC,OAAmD;AAAA,EACrF,OACE,OAAO,UAAU,YACjB,UAAU,QACV,WAAW,UACV,aAAc,MAAmC,KAAK,KACrD,SAAU,MAAmC,KAAK;AAAA;AAIxD,SAAS,yBAAyB,CAAC,OAAkD;AAAA,EACnF,OACE,OAAO,UAAU,YACjB,UAAU,QACV,cAAc,SACd,OAAQ,MAAkC,aAAa;AAAA;AAI3D,eAAe,iBAAiB,CAAC,KAA4B;AAAA,EAC3D,MAAM,WAAW,MAAM,MAAM,GAAG;AAAA,EAChC,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,IAAI,MAAM,mCAAmC,SAAS,QAAQ;AAAA,EACtE;AAAA,EACA,OAAO,SAAS,KAAK;AAAA;AAEvB,eAAsB,mBAAmB,CACvC,SACA,OACiB;AAAA,EACjB,IAAI,YAAY,sBAAsB,OAAO;AAAA,EAC7C,IAAI;AAAA,EACJ,IAAI,cAAiD,CAAC;AAAA,EAEtD,IAAI,OAAO,UAAU,UAAU;AAAA,IAC7B,oBAAO,MAAM,qCAAqC,OAAO;AAAA,IACzD,OAAO,MAAM,kBAAkB,KAAK;AAAA,EACtC,EAAO,SAAI,aAAa,KAAK,GAAG;AAAA,IAC9B,OAAO;AAAA,EACT,EAAO,SAAI,SAAS,KAAK,GAAG;AAAA,IAC1B,MAAM,YAAW,oBAAoB,KAAK;AAAA,IAC1C,oBAAO,MAAM,2CAA2C,WAAU;AAAA,IAClE,OAAO,IAAI,KAAK,CAAC,IAAI,WAAW,KAAK,CAAC,GAAG,EAAE,MAAM,UAAS,CAAC;AAAA,EAC7D,EAAO,SAAI,2BAA2B,KAAK,GAAG;AAAA,IAC5C,cAAc;AAAA,IACd,IAAI,MAAM,OAAO;AAAA,MACf,YAAY,MAAM;AAAA,IACpB;AAAA,IACA,IAAI,SAAS,MAAM,KAAK,GAAG;AAAA,MACzB,MAAM,YAAW,MAAM,YAAY,oBAAoB,MAAM,KAAK;AAAA,MAClE,oBAAO,MAAM,6BAA6B,WAAU;AAAA,MACpD,OAAO,IAAI,KAAK,CAAC,IAAI,WAAW,MAAM,KAAK,CAAC,GAAG,EAAE,MAAM,UAAS,CAAC;AAAA,IACnE,EAAO;AAAA,MACL,OAAO,MAAM;AAAA;AAAA,EAEjB,EAAO,SAAI,0BAA0B,KAAK,GAAG;AAAA,IAC3C,oBAAO,MAAM,qCAAqC,MAAM,UAAU;AAAA,IAClE,OAAO,MAAM,kBAAkB,MAAM,QAAQ;AAAA,IAC7C,cAAc,EAAE,QAAQ,MAAM,OAAO;AAAA,EACvC,EAAO;AAAA,IACL,MAAM,IAAI,MACR,qFACF;AAAA;AAAA,EAGF,oBAAO,MAAM,uCAAuC,WAAW;AAAA,EAE/D,MAAM,WAAY,KAAc,QAAQ;AAAA,EACxC,MAAM,WACH,KAAc,QACf,uBACE,SAAS,WAAW,QAAQ,IACvB,WACD,YACN;AAAA,EAEF,MAAM,WAAW,IAAI;AAAA,EACrB,SAAS,OAAO,QAAQ,MAAM,QAAQ;AAAA,EACtC,SAAS,OAAO,SAAS,SAAS;AAAA,EAElC,IAAI,YAAY,UAAU;AAAA,IACxB,SAAS,OAAO,YAAY,YAAY,QAAQ;AAAA,EAClD;AAAA,EACA,IAAI,YAAY,gBAAgB;AAAA,IAC9B,SAAS,OAAO,mBAAmB,YAAY,cAAc;AAAA,EAC/D;AAAA,EACA,IAAI,YAAY,QAAQ;AAAA,IACtB,SAAS,OAAO,UAAU,YAAY,MAAM;AAAA,EAC9C;AAAA,EACA,IAAI,YAAY,gBAAgB,WAAW;AAAA,IACzC,SAAS,OAAO,eAAe,OAAO,YAAY,WAAW,CAAC;AAAA,EAChE;AAAA,EACA,IAAI,YAAY,wBAAwB;AAAA,IACtC,WAAW,eAAe,YAAY,wBAAwB;AAAA,MAC5D,SAAS,OAAO,6BAA6B,WAAW;AAAA,IAC1D;AAAA,EACF;AAAA,EAEA,MAAM,UAAU,WAAW,OAAO;AAAA,EAClC,MAAM,WAAW,MAAM,MAAM,GAAG,gCAAgC;AAAA,IAC9D,QAAQ;AAAA,IACR,SAAS,cAAc,OAAO;AAAA,IAC9B,MAAM;AAAA,EACR,CAAC;AAAA,EAED,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,eAAe;AAAA,IACnE,MAAM,IAAI,MACR,gCAAgC,SAAS,UAAU,SAAS,gBAAgB,WAC9E;AAAA,EACF;AAAA,EAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,EAClC,OAAO,KAAK;AAAA;AAGd,eAAsB,kBAAkB,CACtC,SACA,OACsB;AAAA,EACtB,IAAI;AAAA,EACJ,IAAI;AAAA,EACJ,IAAI,SAA0B;AAAA,EAC9B,IAAI;AAAA,EACJ,IAAI;AAAA,EAEJ,IAAI,OAAO,UAAU,UAAU;AAAA,IAC7B,OAAO;AAAA,IACP,QAAQ;AAAA,EACV,EAAO;AAAA,IACL,OAAO,MAAM;AAAA,IACb,QAAQ,MAAM;AAAA,IACd,IAAI,YAAY,SAAS,MAAM,QAAQ;AAAA,MACrC,SAAS,MAAM;AAAA,IACjB;AAAA,IACA,IAAI,WAAW,SAAS,MAAM,OAAO;AAAA,MACnC,QAAQ,MAAM;AAAA,IAChB;AAAA,IACA,IAAI,kBAAkB,SAAS,MAAM,cAAc;AAAA,MACjD,eAAe,MAAM;AAAA,IACvB;AAAA;AAAA,EAGF,QAAQ,SAAS,YAAY,OAAO;AAAA,EACpC,QAAQ,SAAS,YAAY,OAAO;AAAA,EACpC,eAAe,gBAAgB,mBAAmB,OAAO;AAAA,EAEzD,oBAAO,MAAM,wCAAwC,OAAO;AAAA,EAE5D,IAAI,CAAC,QAAQ,KAAK,KAAK,EAAE,WAAW,GAAG;AAAA,IACrC,MAAM,IAAI,MAAM,wCAAwC;AAAA,EAC1D;AAAA,EAEA,IAAI,KAAK,SAAS,MAAM;AAAA,IACtB,MAAM,IAAI,MAAM,kDAAkD;AAAA,EACpE;AAAA,EAEA,MAAM,cAA0B,CAAC,SAAS,QAAQ,SAAS,QAAQ,QAAQ,SAAS;AAAA,EACpF,IAAI,SAAS,CAAC,YAAY,SAAS,KAAiB,GAAG;AAAA,IACrD,MAAM,IAAI,MAAM,kBAAkB,0BAA0B,YAAY,KAAK,IAAI,GAAG;AAAA,EACtF;AAAA,EAEA,MAAM,UAAU,WAAW,OAAO;AAAA,EAElC,MAAM,cAAsC;AAAA,IAC1C;AAAA,IACA;AAAA,IACA,OAAO;AAAA,IACP,iBAAiB;AAAA,EACnB;AAAA,EAEA,IAAI,gBAAgB,aAAa,SAAS,GAAG;AAAA,IAC3C,YAAY,eAAe;AAAA,EAC7B;AAAA,EAEA,MAAM,WAAW,MAAM,MAAM,GAAG,wBAAwB;AAAA,IACtD,QAAQ;AAAA,IACR,SAAS;AAAA,SACJ,cAAc,OAAO;AAAA,MACxB,gBAAgB;AAAA,SACZ,WAAW,QAAQ,EAAE,QAAQ,aAAa,IAAI,CAAC;AAAA,IACrD;AAAA,IACA,MAAM,KAAK,UAAU,WAAW;AAAA,EAClC,CAAC;AAAA,EAED,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,eAAe;AAAA,IACnE,MAAM,IAAI,MAAM,sBAAsB,SAAS,UAAU,SAAS,gBAAgB,WAAW;AAAA,EAC/F;AAAA,EAEA,OAAO,SAAS,YAAY;AAAA;;AEhOiB,IAA/C;;;ACA0B,IAA1B;AAGA,IAAM,oBAAoB;AAmC1B,SAAS,cAAc,CAAC,QAAwB;AAAA,EAC9C,IAAI,OAAO,UAAU,mBAAmB;AAAA,IACtC,OAAO;AAAA,EACT;AAAA,EACA,OAAO,GAAG,OAAO,MAAM,GAAG,iBAAiB;AAAA;AAG7C,SAAS,cAAc,CAAC,OAA+B;AAAA,EACrD,IAAI,kBAAkB,OAAO;AAAA,IAC3B,MAAM,sBACJ,yBAAyB,QAAQ,MAAM,sBAAsB;AAAA,IAC/D,MAAM,qBAAqB,MAAM,sBAAsB,qBAAqB;AAAA,IAC5E,OAAO;AAAA,MACL,cAAc,MAAM,gBAAgB;AAAA,MACpC,kBAAkB,MAAM,oBAAoB;AAAA,MAC5C,aAAa,MAAM,gBAAgB,MAAM,gBAAgB,MAAM,MAAM,oBAAoB;AAAA,MACzF;AAAA,IACF;AAAA,EACF;AAAA,EACA,IAAI,iBAAiB,SAAS,kBAAkB,OAAO;AAAA,IACrD,MAAM,QAAS,MAAqB,eAAe;AAAA,IACnD,MAAM,SAAU,MAAqB,gBAAgB;AAAA,IACrD,MAAM,QAAS,MAAqB,eAAe,QAAQ;AAAA,IAC3D,OAAO;AAAA,MACL,cAAc;AAAA,MACd,kBAAkB;AAAA,MAClB,aAAa;AAAA,MACb,oBAAqB,MAAqB;AAAA,IAC5C;AAAA,EACF;AAAA,EACA,OAAO;AAAA,IACL,cAAc;AAAA,IACd,kBAAkB;AAAA,IAClB,aAAa;AAAA,EACf;AAAA;AAGK,SAAS,mBAAmB,CACjC,SACA,MACA,QACA,OACM;AAAA,EACN,MAAM,aAAa,eAAe,KAAK;AAAA,EAEvC,MAAM,UAAkC;AAAA,IACtC;AAAA,IACA,QAAQ;AAAA,IACR,UAAU;AAAA,IACV;AAAA,IACA,QAAQ,eAAe,MAAM;AAAA,IAC7B,QAAQ;AAAA,MACN,QAAQ,WAAW;AAAA,MACnB,YAAY,WAAW;AAAA,MACvB,OAAO,WAAW;AAAA,SACd,WAAW,uBAAuB,YAClC,EAAE,QAAQ,WAAW,mBAAmB,IACxC,CAAC;AAAA,IACP;AAAA,EACF;AAAA,EAEA,QAAQ,UAAU,uBAAU,YAAY,OAAO;AAAA;;;ADtFjD,SAAS,iBAAiB,CAAC,WAAoC;AAAA,EAC7D,MAAM,kBAAkB,OAAO,OAAO,wBAAW;AAAA,EACjD,IAAI,CAAC,gBAAgB,SAAS,SAAS,GAAG;AAAA,IACxC,MAAM,IAAI,MACR,gCAAgC,8BAA8B,gBAAgB,KAAK,IAAI,GACzF;AAAA,EACF;AAAA,EACA,OAAO;AAAA;AAGT,SAAS,WAAW,CAAC,QAA4D;AAAA,EAC/E,IAAI,WAAW,MAAM;AAAA,IACnB,OAAO;AAAA,EACT;AAAA,EACA,IAAI,OAAO,WAAW,UAAU;AAAA,IAC9B,OAAO;AAAA,EACT;AAAA,EACA,IAAI,OAAO,WAAW,YAAY,OAAO,OAAO,SAAS,UAAU;AAAA,IACjE,OAAO,OAAO;AAAA,EAChB;AAAA,EACA,MAAM,IAAI,MAAM,sEAAsE;AAAA;AAGxF,eAAsB,mBAAmB,CACvC,SACA,QACmB;AAAA,EACnB,MAAM,iBAAiB,kBAAkB,OAAO;AAAA,EAChD,MAAM,qBAAqB,kBAAkB,uBAAuB,OAAO,CAAC;AAAA,EAE5E,MAAM,OAAO,YAAY,MAAM;AAAA,EAC/B,IAAI,SAAS,MAAM;AAAA,IACjB,oBAAO,MAAM,qDAAqD;AAAA,IAClE,MAAM,aAAa,IAAI,MAAM,kBAAkB,EAAE,KAAK,CAAC;AAAA,IACvD,WAAW,KAAK;AAAA,IAChB,OAAO;AAAA,EACT;AAAA,EAEA,IAAI,cAAc,KAAK,KAAK;AAAA,EAC5B,IAAI,YAAY,WAAW,GAAG;AAAA,IAC5B,MAAM,IAAI,MAAM,0CAA0C;AAAA,EAC5D;AAAA,EAKA,MAAM,WAAW,OAAQ;AAAA,EACzB,IAAI,YAAY,SAAS,UAAU;AAAA,IACjC,oBAAO,KACL,uCAAuC,KAAK,KAAK,YAAY,SAAS,CAAC,uCACzE;AAAA,IACA,cAAc,YAAY,MAAM,GAAG,QAAQ;AAAA,EAC7C;AAAA,EAEA,MAAM,UAAU,oBAAoB,OAAO;AAAA,EAC3C,MAAM,MAAM,GAAG;AAAA,EAEf,oBAAO,MAAM,6CAA6C,gBAAgB;AAAA,EAE1E,MAAM,WAAW,MAAM,MAAM,KAAK;AAAA,IAChC,QAAQ;AAAA,IACR,SAAS;AAAA,SACJ,cAAc,SAAS,IAAI;AAAA,MAC9B,gBAAgB;AAAA,IAClB;AAAA,IACA,MAAM,KAAK,UAAU;AAAA,MACnB,OAAO;AAAA,MACP,OAAO;AAAA,IACT,CAAC;AAAA,EACH,CAAC;AAAA,EAED,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,eAAe;AAAA,IACnE,MAAM,IAAI,MACR,+BAA+B,SAAS,UAAU,SAAS,gBAAgB,WAC7E;AAAA,EACF;AAAA,EAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,EAElC,MAAM,cAAc,MAAM,OAAO;AAAA,EACjC,IAAI,CAAC,eAAe,CAAC,YAAY,WAAW;AAAA,IAC1C,MAAM,IAAI,MAAM,0DAA0D;AAAA,EAC5E;AAAA,EAEA,MAAM,YAAY,YAAY;AAAA,EAE9B,IAAI,UAAU,WAAW,oBAAoB;AAAA,IAC3C,MAAM,IAAI,MACR,qCAAqC,UAAU,oBAAoB,yBACjE,4CACJ;AAAA,EACF;AAAA,EAEA,IAAI,KAAK,OAAO;AAAA,IACd,oBAAoB,SAAS,uBAAU,gBAAgB,aAAa;AAAA,MAClE,cAAc,KAAK,MAAM;AAAA,MACzB,kBAAkB;AAAA,MAClB,aAAa,KAAK,MAAM;AAAA,IAC1B,CAAC;AAAA,EACH;AAAA,EAEA,oBAAO,MAAM,qCAAqC,UAAU,mBAAmB;AAAA,EAC/E,OAAO;AAAA;;AEpHyB,IAAlC;AAwBA,IAAM,mCACJ;AAEF,eAAsB,qBAAqB,CACzC,SACA,QACkC;AAAA,EAClC,MAAM,YAAY,cAAc,OAAO;AAAA,EACvC,MAAM,QAAQ,OAAO,SAAS;AAAA,EAC9B,MAAM,OAAmB,OAAO,QAAsB;AAAA,EACtD,MAAM,iBAAiB;AAAA,EAEvB,oBAAO,MAAM,+BAA+B,WAAW;AAAA,EAEvD,IAAI,CAAC,OAAO,UAAU,OAAO,OAAO,KAAK,EAAE,WAAW,GAAG;AAAA,IACvD,MAAM,IAAI,MAAM,8CAA8C;AAAA,EAChE;AAAA,EAEA,IAAI,QAAQ,KAAK,QAAQ,IAAI;AAAA,IAC3B,MAAM,IAAI,MAAM,sCAAsC;AAAA,EACxD;AAAA,EAEA,MAAM,UAAU,WAAW,OAAO;AAAA,EAElC,MAAM,cAA+C;AAAA,IACnD,OAAO;AAAA,IACP,QAAQ,OAAO;AAAA,IACf,GAAG;AAAA,IACH;AAAA,EACF;AAAA,EAEA,IAAI,eAAe,SAAS;AAAA,IAC1B,YAAY,UAAU,eAAe;AAAA,EACvC;AAAA,EACA,IAAI,eAAe,OAAO;AAAA,IACxB,YAAY,QAAQ,eAAe;AAAA,EACrC;AAAA,EAEA,MAAM,WAAW,MAAM,MAAM,GAAG,8BAA8B;AAAA,IAC5D,QAAQ;AAAA,IACR,SAAS;AAAA,SACJ,cAAc,OAAO;AAAA,MACxB,gBAAgB;AAAA,IAClB;AAAA,IACA,MAAM,KAAK,UAAU,WAAW;AAAA,EAClC,CAAC;AAAA,EAED,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,eAAe;AAAA,IACnE,MAAM,IAAI,MACR,mCAAmC,SAAS,UAAU,SAAS,gBAAgB,WACjF;AAAA,EACF;AAAA,EAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,EAElC,IAAI,CAAC,KAAK,QAAQ,KAAK,KAAK,WAAW,GAAG;AAAA,IACxC,MAAM,IAAI,MAAM,+BAA+B;AAAA,EACjD;AAAA,EAEA,OAAO,KAAK,KAAK,IAAI,CAAC,UAAU;AAAA,IAC9B,KAAK,KAAK;AAAA,IACV,eAAe,KAAK;AAAA,EACtB,EAAE;AAAA;AAGJ,SAAS,sBAAsB,CAAC,SAAyB;AAAA,EACvD,MAAM,aAAa,QAAQ,MAAM,2BAA2B;AAAA,EAC5D,OAAO,aAAa,IAAI,KAAK,KAAK;AAAA;AAGpC,SAAS,4BAA4B,CAAC,SAAyB;AAAA,EAC7D,OAAO,QAAQ,QAAQ,6BAA6B,EAAE,EAAE,KAAK;AAAA;AAG/D,eAAsB,sBAAsB,CAC1C,SACA,QACiC;AAAA,EACjC,MAAM,YAAY,yBAAyB,OAAO;AAAA,EAClD,MAAM,YAAY,6BAA6B,OAAO;AAAA,EAEtD,oBAAO,MAAM,2CAA2C,WAAW;AAAA,EAEnE,IAAI;AAAA,EACJ,IAAI;AAAA,EAEJ,IAAI,OAAO,WAAW,UAAU;AAAA,IAC9B,WAAW;AAAA,IACX,aAAa;AAAA,EACf,EAAO;AAAA,IACL,WAAW,OAAO;AAAA,IAClB,aAAa,OAAO,UAAU;AAAA;AAAA,EAGhC,IAAI,CAAC,YAAY,SAAS,KAAK,EAAE,WAAW,GAAG;AAAA,IAC7C,MAAM,IAAI,MAAM,8CAA8C;AAAA,EAChE;AAAA,EAEA,MAAM,UAAU,WAAW,OAAO;AAAA,EAElC,MAAM,cAAc;AAAA,IAClB,OAAO;AAAA,IACP,UAAU;AAAA,MACR;AAAA,QACE,MAAM;AAAA,QACN,SAAS;AAAA,UACP,EAAE,MAAM,QAAQ,MAAM,WAAW;AAAA,UACjC,EAAE,MAAM,aAAa,WAAW,EAAE,KAAK,SAAS,EAAE;AAAA,QACpD;AAAA,MACF;AAAA,IACF;AAAA,IACA,YAAY;AAAA,EACd;AAAA,EAEA,MAAM,WAAW,MAAM,MAAM,GAAG,4BAA4B;AAAA,IAC1D,QAAQ;AAAA,IACR,SAAS;AAAA,SACJ,cAAc,OAAO;AAAA,MACxB,gBAAgB;AAAA,IAClB;AAAA,IACA,MAAM,KAAK,UAAU,WAAW;AAAA,EAClC,CAAC;AAAA,EAED,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,eAAe;AAAA,IACnE,MAAM,IAAI,MACR,oCAAoC,SAAS,UAAU,SAAS,gBAAgB,WAClF;AAAA,EACF;AAAA,EAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,EAElC,IAAI,KAAK,OAAO;AAAA,IACd,oBACE,SACA,uBAAU,mBACV,OAAO,WAAW,WAAW,SAAU,OAAO,UAAU,IACxD;AAAA,MACE,cAAc,KAAK,MAAM;AAAA,MACzB,kBAAkB,KAAK,MAAM;AAAA,MAC7B,aAAa,KAAK,MAAM;AAAA,IAC1B,CACF;AAAA,EACF;AAAA,EAEA,MAAM,cAAc,KAAK,UAAU;AAAA,EACnC,MAAM,UAAU,aAAa,SAAS;AAAA,EAEtC,IAAI,CAAC,SAAS;AAAA,IACZ,MAAM,IAAI,MAAM,6CAA6C;AAAA,EAC/D;AAAA,EAEA,OAAO;AAAA,IACL,OAAO,uBAAuB,OAAO;AAAA,IACrC,aAAa,6BAA6B,OAAO;AAAA,EACnD;AAAA;;AC/KgC,IAAlC;AACmD,IAAnD;;;ACPkD,IAAlD;AAIA,IAAM,gBAAgB;AAEf,SAAS,kBAAkB,CAAC,SAAwC;AAAA,EACzE,MAAM,UAAU,WAAW,OAAO;AAAA,EAClC,MAAM,SAAS,UAAU,OAAO;AAAA,EAEhC,IAAI,CAAC,UAAU,YAAY,OAAO,GAAG;AAAA,IACnC,OAAO,2BAAa;AAAA,MAClB,QAAQ;AAAA,MACR;AAAA,IACF,CAAC;AAAA,EACH;AAAA,EAEA,IAAI,CAAC,QAAQ;AAAA,IACX,MAAM,IAAI,MACR,uFACF;AAAA,EACF;AAAA,EAEA,OAAO,2BAAa;AAAA,IAClB;AAAA,IACA;AAAA,EACF,CAAC;AAAA;;AC1BoB,IAAvB;AAC+B,IAA/B;AASA,IAAM,wBAAwB;AAAA,EAC5B,eAAe;AAAA,EACf,YAAY;AACd;AAEO,SAAS,qBAAqB,GAAuB;AAAA,EAC1D,OAAO,SAAS,MAAM,YAAsD;AAAA,IAC1E,IAAI,EAAE,iBAAiB,2BAAiB;AAAA,MACtC,OAAO;AAAA,IACT;AAAA,IACA,IAAI;AAAA,MACF,MAAM,cAAc,KAAK,QAAQ,sBAAsB,eAAe,EAAE;AAAA,MACxE,KAAK,MAAM,WAAW;AAAA,MACtB,oBAAO,MAAM,wEAAwE;AAAA,MACrF,OAAO;AAAA,MACP,MAAM;AAAA,MACN,oBAAO,KAAK,0CAA0C;AAAA,MACtD,OAAO;AAAA;AAAA;AAAA;;;AFXb,eAAe,yBAAyB,CACtC,SACA,QACA,WACA,YACoC;AAAA,EACpC,MAAM,SAAS,mBAAmB,OAAO;AAAA,EACzC,MAAM,YAAY,WAAW,OAAO;AAAA,EAEpC,oBAAO,MAAM,kBAAkB,oBAAoB,WAAW;AAAA,EAE9D,IAAI,CAAC,OAAO,UAAU,OAAO,OAAO,KAAK,EAAE,WAAW,GAAG;AAAA,IACvD,MAAM,IAAI,MAAM,+CAA+C;AAAA,EACjE;AAAA,EAEA,IAAI,OAAO,QAAQ;AAAA,IACjB,oBAAO,MACL,wDACE,iDACJ;AAAA,EACF;AAAA,EAEA,MAAM,QAAQ,OAAO,KAAK,SAAS;AAAA,EACnC,QAAQ,QAAQ,UAAU,MAAM,0BAAe;AAAA,IAC7C;AAAA,IACA,QAAQ;AAAA,IACR,QAAQ,OAAO;AAAA,IACf,yBAAyB,sBAAsB;AAAA,EACjD,CAAC;AAAA,EAED,IAAI,OAAO;AAAA,IACT,oBAAoB,SAAS,WAAW,OAAO,QAAQ,KAAK;AAAA,EAC9D;AAAA,EAEA,IAAI,OAAO,WAAW,YAAY,WAAW,MAAM;AAAA,IACjD,MAAM,IAAI,MAAM,8BAA8B,OAAO,yBAAyB;AAAA,EAChF;AAAA,EAEA,OAAO;AAAA;AAGT,eAAsB,iBAAiB,CACrC,SACA,QACoC;AAAA,EACpC,OAAO,0BAA0B,SAAS,QAAQ,uBAAU,cAAc,aAAa;AAAA;AAGzF,eAAsB,iBAAiB,CACrC,SACA,QACoC;AAAA,EACpC,OAAO,0BAA0B,SAAS,QAAQ,uBAAU,cAAc,aAAa;AAAA;;AG7ClE,IAAvB;AA6EA,SAAS,gBAAgB,CAAC,MAAsC;AAAA,EAC9D,QAAQ,KAAK;AAAA,SACN;AAAA,MACH,OAAO,EAAE,MAAM,qBAAqB;AAAA,SACjC;AAAA,MACH,OAAO;AAAA,QACL,MAAM;AAAA,QACN,kBAAkB,KAAK;AAAA,MACzB;AAAA,SACG;AAAA,MACH,OAAO;AAAA,QACL,MAAM;AAAA,QACN,WAAW,KAAK,aAAa,EAAE,MAAM,OAAO;AAAA,MAC9C;AAAA,SACG;AAAA,MACH,OAAO;AAAA,QACL,MAAM;AAAA,QACN,cAAc,KAAK;AAAA,QACnB,YAAY,KAAK;AAAA,QACjB,kBAAkB,KAAK,mBAAmB;AAAA,MAC5C;AAAA;AAAA,MAEA,MAAM,IAAI,MAAM,+BAAgC,KAAsB,MAAM;AAAA;AAAA;AAOlF,SAAS,iBAAiB,CAAC,MAAyD;AAAA,EAClF,QAAQ,KAAK;AAAA,SACN;AAAA,MACH,OAAO;AAAA,QACL,IAAI,KAAK,MAAM;AAAA,QACf,MAAM;AAAA,QACN,QAAS,KAAK,UAAqC;AAAA,QACnD,QAAQ;AAAA,UACN,MAAO,KAAK,QAAQ,QAAoD;AAAA,UACxE,OAAO,KAAK,QAAQ;AAAA,UACpB,KAAK,KAAK,QAAQ;AAAA,QACpB;AAAA,MACF;AAAA,SAEG;AAAA,MACH,OAAO;AAAA,QACL,IAAI,KAAK,MAAM;AAAA,QACf,MAAM;AAAA,QACN,QAAS,KAAK,UAAqC;AAAA,QACnD,OAAO,KAAK,SAAS;AAAA,QACrB,SAAS,KAAK,SAAS,IAAI,CAAC,OAAO;AAAA,UACjC,QAAQ,EAAE;AAAA,UACV,UAAU,EAAE;AAAA,UACZ,OAAO,EAAE;AAAA,QACX,EAAE;AAAA,MACJ;AAAA,SAEG;AAAA,MACH,OAAO;AAAA,QACL,IAAI,KAAK,MAAM;AAAA,QACf,MAAM;AAAA,QACN,QAAS,KAAK,UAAqC;AAAA,QACnD,MAAM,KAAK,QAAQ;AAAA,QACnB,QAAQ,KAAK;AAAA,MACf;AAAA,SAEG;AAAA,MACH,OAAO;AAAA,QACL,IAAI,KAAK,MAAM;AAAA,QACf,MAAM;AAAA,QACN,QAAS,KAAK,UAAqC;AAAA,QACnD,aAAa,KAAK,gBAAgB;AAAA,QAClC,UAAU,KAAK,aAAa;AAAA,QAC5B,WAAY,KAAK,aAAa,CAAC;AAAA,QAC/B,QAAQ,KAAK;AAAA,MACf;AAAA,SAEG;AAAA,MACH,OAAO;AAAA,QACL,MAAM;AAAA,QACN,SACE,KAAK,SAAS,IAAI,CAAC,OAAO;AAAA,UACxB,MAAM;AAAA,UACN,MAAM,EAAE;AAAA,UACR,aACE,EAAE,aAAa,IAAI,CAAC,OAAO;AAAA,YACzB,KAAK,EAAE;AAAA,YACP,OAAO,EAAE;AAAA,YACT,YAAY,EAAE;AAAA,YACd,UAAU,EAAE;AAAA,UACd,EAAE,KAAK,CAAC;AAAA,QACZ,EAAE,KAAK,CAAC;AAAA,MACZ;AAAA;AAAA,MAIA,OAAO;AAAA;AAAA;AAOb,SAAS,yBAAyB,CAAC,UAGjC;AAAA,EAEA,IAAI,SAAS,aAAa;AAAA,IAExB,MAAM,eAAoC,CAAC;AAAA,IAC3C,IAAI,SAAS,QAAQ;AAAA,MACnB,WAAW,QAAQ,SAAS,QAAQ;AAAA,QAClC,IAAI,KAAK,SAAS,aAAa,KAAK,SAAS;AAAA,UAC3C,WAAW,WAAW,KAAK,SAAS;AAAA,YAClC,IAAI,QAAQ,aAAa;AAAA,cACvB,WAAW,OAAO,QAAQ,aAAa;AAAA,gBACrC,aAAY,KAAK;AAAA,kBACf,KAAK,IAAI;AAAA,kBACT,OAAO,IAAI;AAAA,kBACX,YAAY,IAAI;AAAA,kBAChB,UAAU,IAAI;AAAA,gBAChB,CAAC;AAAA,cACH;AAAA,YACF;AAAA,UACF;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,IACA,OAAO,EAAE,MAAM,SAAS,aAAa,0BAAY;AAAA,EACnD;AAAA,EAGA,IAAI,OAAO;AAAA,EACX,MAAM,cAAoC,CAAC;AAAA,EAE3C,IAAI,SAAS,QAAQ;AAAA,IACnB,WAAW,QAAQ,SAAS,QAAQ;AAAA,MAClC,IAAI,KAAK,SAAS,aAAa,KAAK,SAAS;AAAA,QAC3C,WAAW,WAAW,KAAK,SAAS;AAAA,UAClC,QAAQ,QAAQ;AAAA,UAChB,IAAI,QAAQ,aAAa;AAAA,YACvB,WAAW,OAAO,QAAQ,aAAa;AAAA,cACrC,YAAY,KAAK;AAAA,gBACf,KAAK,IAAI;AAAA,gBACT,OAAO,IAAI;AAAA,gBACX,YAAY,IAAI;AAAA,gBAChB,UAAU,IAAI;AAAA,cAChB,CAAC;AAAA,YACH;AAAA,UACF;AAAA,QACF;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAAA,EAEA,OAAO,EAAE,MAAM,YAAY;AAAA;AA8B7B,eAAsB,cAAc,CAClC,SACA,QACyB;AAAA,EACzB,MAAM,SAAS,UAAU,OAAO;AAAA,EAChC,IAAI,CAAC,QAAQ;AAAA,IACX,MAAM,IAAI,MACR,yGACF;AAAA,EACF;AAAA,EAEA,MAAM,UAAU,WAAW,OAAO;AAAA,EAClC,MAAM,YAAY,OAAO,SAAS,iBAAiB,OAAO;AAAA,EAC1D,MAAM,UAAU,mBAAmB,OAAO;AAAA,EAE1C,qBAAO,MAAM,+CAA+C,WAAW;AAAA,EACvE,qBAAO,MAAM,4BAA4B,OAAO,MAAM,UAAU,GAAG,GAAG,MAAM;AAAA,EAG5E,MAAM,kBAAkB,OAAO,OAAO,OACpC,CAAC,MAAM,EAAE,SAAS,wBAAwB,EAAE,SAAS,iBAAiB,EAAE,SAAS,KACnF;AAAA,EAEA,IAAI,CAAC,mBAAmB,gBAAgB,WAAW,GAAG;AAAA,IAEpD,qBAAO,MAAM,2EAA2E;AAAA,IACxF,OAAO,QAAQ,CAAC,EAAE,MAAM,qBAAqB,GAAG,GAAI,OAAO,SAAS,CAAC,CAAE;AAAA,EACzE;AAAA,EAGA,MAAM,cAAuC;AAAA,IAC3C,OAAO;AAAA,IACP,OAAO,OAAO;AAAA,EAChB;AAAA,EAEA,IAAI,OAAO,cAAc;AAAA,IACvB,YAAY,eAAe,OAAO;AAAA,EACpC;AAAA,EAEA,IAAI,OAAO,eAAe,WAAW;AAAA,IACnC,YAAY,aAAa,OAAO;AAAA,EAClC;AAAA,EAEA,IAAI,OAAO,SAAS,OAAO,MAAM,SAAS,GAAG;AAAA,IAC3C,YAAY,QAAQ,OAAO,MAAM,IAAI,gBAAgB;AAAA,EACvD;AAAA,EAEA,IAAI,OAAO,iBAAiB,WAAW;AAAA,IACrC,YAAY,iBAAiB,OAAO;AAAA,EACtC;AAAA,EAEA,IAAI,OAAO,kBAAkB;AAAA,IAC3B,YAAY,YAAY,EAAE,SAAS,OAAO,iBAAiB;AAAA,EAC7D;AAAA,EAEA,qBAAO,MAAM,mCAAmC,KAAK,UAAU,aAAa,MAAM,CAAC,GAAG;AAAA,EAGtF,MAAM,WAAW,MAAM,MAAM,GAAG,qBAAqB;AAAA,IACnD,QAAQ;AAAA,IACR,SAAS;AAAA,MACP,eAAe,UAAU;AAAA,MACzB,gBAAgB;AAAA,IAClB;AAAA,IACA,MAAM,KAAK,UAAU,WAAW;AAAA,IAChC,QAAQ,YAAY,QAAQ,OAAO;AAAA,EACrC,CAAC;AAAA,EAED,IAAI,CAAC,SAAS,IAAI;AAAA,IAChB,MAAM,YAAY,MAAM,SAAS,KAAK;AAAA,IACtC,qBAAO,MAAM,qCAAqC,SAAS,UAAU,WAAW;AAAA,IAChF,MAAM,IAAI,MAAM,iCAAiC,SAAS,UAAU,SAAS,YAAY;AAAA,EAC3F;AAAA,EAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,EAElC,IAAI,KAAK,OAAO;AAAA,IACd,qBAAO,MAAM,gCAAgC,KAAK,MAAM,SAAS;AAAA,IACjE,MAAM,IAAI,MAAM,wBAAwB,KAAK,MAAM,SAAS;AAAA,EAC9D;AAAA,EAEA,qBAAO,MAAM,gDAAgD,KAAK,UAAU,aAAa;AAAA,EAGzF,QAAQ,MAAM,gBAAgB,0BAA0B,IAAI;AAAA,EAG5D,MAAM,cAAoC,CAAC;AAAA,EAC3C,IAAI,KAAK,QAAQ;AAAA,IACf,WAAW,QAAQ,KAAK,QAAQ;AAAA,MAC9B,MAAM,YAAY,kBAAkB,IAAI;AAAA,MACxC,IAAI,WAAW;AAAA,QACb,YAAY,KAAK,SAAS;AAAA,MAC5B;AAAA,IACF;AAAA,EACF;AAAA,EAEA,MAAM,SAAyB;AAAA,IAC7B,IAAI,KAAK;AAAA,IACT;AAAA,IACA;AAAA,IACA;AAAA,IACA,QAAQ,KAAK;AAAA,EACf;AAAA,EAEA,qBAAO,KACL,6CAA6C,KAAK,wBAAwB,YAAY,yBAAyB,YAAY,QAC7H;AAAA,EAEA,OAAO;AAAA;;ACnYyB,IAAlC;AACkE,IAAlE;AAuCA,SAAS,YAAY,CAAC,OAA+D;AAAA,EACnF,IAAI,CAAC,OAAO;AAAA,IACV;AAAA,EACF;AAAA,EAGA,MAAM,eAAe,MAAM,eAAe;AAAA,EAC1C,MAAM,mBAAmB,MAAM,gBAAgB;AAAA,EAC/C,MAAM,iBAAiB;AAAA,EAEvB,OAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA,aAAa,eAAe;AAAA,IAC5B,oBAAoB,eAAe;AAAA,EACrC;AAAA;AAGF,SAAS,yBAAyB,CAAC,QAAsD;AAAA,EACvF,MAAM,oBAAoB;AAAA,EAC1B,OAAO;AAAA,IACL,gBAAgB,kBAAkB,iBAAiB,QAAQ;AAAA,IAC3D,sBAAsB,kBAAkB,iBAAiB,QAAQ;AAAA,EACnE;AAAA;AAgBF,eAAe,uBAAuB,CACpC,SACA,QACA,WACA,YACoC;AAAA,EACpC,MAAM,SAAS,mBAAmB,OAAO;AAAA,EACzC,MAAM,YAAY,WAAW,OAAO;AAAA,EAEpC,qBAAO,MAAM,kBAAkB,oBAAoB,WAAW;AAAA,EAC9D,MAAM,qBAAqB,0BAA0B,MAAM;AAAA,EAG3D,MAAM,eAAe,QAAQ,UAAU,UAAU;AAAA,EAMjD,MAAM,QAAQ,OAAO,KAAK,SAAS;AAAA,EACnC,MAAM,iBAAiB;AAAA,IACrB;AAAA,IACA,QAAQ,OAAO;AAAA,IACf,QAAQ;AAAA,IACR,iBAAiB,OAAO,aAAa;AAAA,IACrC,wBAAwB,EAAE,WAAW,yBAAyB,OAAO,EAAE;AAAA,OACnE,mBAAmB,kBAAkB,mBAAmB,uBACxD;AAAA,MACE,iBAAiB;AAAA,QACf,QAAQ;AAAA,aACF,mBAAmB,iBACnB,EAAE,gBAAgB,mBAAmB,eAAe,IACpD,CAAC;AAAA,aACD,mBAAmB,uBACnB,EAAE,sBAAsB,mBAAmB,qBAAqB,IAChE,CAAC;AAAA,QACP;AAAA,MACF;AAAA,IACF,IACA,CAAC;AAAA,EACP;AAAA,EAGA,IAAI,OAAO,QAAQ;AAAA,IACjB,MAAM,SAAS,sBAAW,cAAc;AAAA,IAExC,OAAO;AAAA,MACL,YAAY,OAAO;AAAA,MACnB,MAAM,QAAQ,QAAQ,OAAO,IAAI;AAAA,MACjC,OAAO,QAAQ,QAAQ,OAAO,KAAK,EAAE,KAAK,YAAY;AAAA,MACtD,cAAc,QAAQ,QAAQ,OAAO,YAAY,EAAE,KAAK,CAAC,MAAM,CAAuB;AAAA,IACxF;AAAA,EACF;AAAA,EAGA,QAAQ,MAAM,UAAU,MAAM,wBAAa,cAAc;AAAA,EAEzD,IAAI,OAAO;AAAA,IACT,oBAAoB,SAAS,WAAW,OAAO,QAAQ,KAAK;AAAA,EAC9D;AAAA,EAEA,OAAO;AAAA;AAgBT,eAAsB,eAAe,CACnC,SACA,QACoC;AAAA,EACpC,OAAO,wBAAwB,SAAS,QAAQ,wBAAU,YAAY,aAAa;AAAA;AAYrF,eAAsB,eAAe,CACnC,SACA,QACoC;AAAA,EACpC,OAAO,wBAAwB,SAAS,QAAQ,wBAAU,YAAY,aAAa;AAAA;;ACtL3D,IAA1B;;;ACA0B,IAA1B;AAOO,IANP;AAWA,SAAS,wBAAwB,CAAC,WAA6B;AAAA,EAC7D,MAAM,aAAa,UAAU,YAAY;AAAA,EACzC,MAAM,mBAAsC,WAAW,SAAS,IAAI,IAChE,eACA;AAAA,EACJ,IAAI;AAAA,IACF,OAAO,oCAAiB,SAA0B;AAAA,IAClD,MAAM;AAAA,IACN,OAAO,+BAAY,gBAAoC;AAAA;AAAA;AAI3D,SAAS,YAAY,CAAC,SAAwB,WAAkC;AAAA,EAC9E,IAAI,cAAc,wBAAU,YAAY;AAAA,IACtC,OAAO,cAAc,OAAO;AAAA,EAC9B;AAAA,EACA,OAAO,cAAc,OAAO;AAAA;AAGvB,SAAS,YAAY,CAC1B,SACA,WACA,MACU;AAAA,EACV,MAAM,YAAY,aAAa,SAAS,SAAS;AAAA,EACjD,MAAM,UAAU,yBAAyB,SAAS;AAAA,EAClD,OAAO,QAAQ,OAAO,IAAI;AAAA;AAGrB,SAAS,cAAc,CAC5B,SACA,WACA,QACQ;AAAA,EACR,MAAM,YAAY,aAAa,SAAS,SAAS;AAAA,EACjD,MAAM,UAAU,yBAAyB,SAAS;AAAA,EAClD,OAAO,QAAQ,OAAO,MAAM;AAAA;;;AD7C9B,eAAsB,qBAAqB,CACzC,SACA,QACmB;AAAA,EACnB,IAAI,CAAC,OAAO,QAAQ;AAAA,IAClB,MAAM,IAAI,MAAM,0CAA0C;AAAA,EAC5D;AAAA,EACA,MAAM,YAAY,OAAO,aAAa,wBAAU;AAAA,EAChD,OAAO,aAAa,SAAS,WAAW,OAAO,MAAM;AAAA;AAGvD,eAAsB,qBAAqB,CACzC,SACA,QACiB;AAAA,EACjB,IAAI,CAAC,OAAO,UAAU,CAAC,MAAM,QAAQ,OAAO,MAAM,GAAG;AAAA,IACnD,MAAM,IAAI,MAAM,8CAA8C;AAAA,EAChE;AAAA,EACA,IAAI,OAAO,OAAO,WAAW,GAAG;AAAA,IAC9B,OAAO;AAAA,EACT;AAAA,EACA,SAAS,IAAI,EAAG,IAAI,OAAO,OAAO,QAAQ,KAAK;AAAA,IAC7C,MAAM,QAAQ,OAAO,OAAO;AAAA,IAC5B,IAAI,OAAO,UAAU,YAAY,CAAC,OAAO,SAAS,KAAK,GAAG;AAAA,MACxD,MAAM,IAAI,MAAM,0BAA0B,oBAAoB;AAAA,IAChE;AAAA,EACF;AAAA,EACA,MAAM,YAAY,OAAO,aAAa,wBAAU;AAAA,EAChD,OAAO,eAAe,SAAS,WAAW,OAAO,MAAM;AAAA;;AbKzD,SAAS,aAAa,GAAmB;AAAA,EACvC,IAAI,OAAO,YAAY,aAAa;AAAA,IAClC,OAAO,CAAC;AAAA,EACV;AAAA,EACA,OAAO,QAAQ;AAAA;AAGjB,IAAM,MAAM,cAAc;AAEnB,IAAM,eAAuB;AAAA,EAClC,MAAM;AAAA,EACN,aAAa;AAAA,EAEb,QAAQ;AAAA,IACN,gBAAgB,IAAI,kBAAkB;AAAA,IACtC,iBAAiB,IAAI,mBAAmB;AAAA,IACxC,oBAAoB,IAAI,sBAAsB;AAAA,IAC9C,oBAAoB,IAAI,sBAAsB;AAAA,IAC9C,aAAa,IAAI,eAAe;AAAA,IAChC,aAAa,IAAI,eAAe;AAAA,IAChC,wBAAwB,IAAI,0BAA0B;AAAA,IACtD,0BAA0B,IAAI,4BAA4B;AAAA,IAC1D,sBAAsB,IAAI,wBAAwB;AAAA,IAClD,6BAA6B,IAAI,+BAA+B;AAAA,IAChE,gCAAgC,IAAI,kCAAkC;AAAA,IACtE,qCAAqC,IAAI,uCAAuC;AAAA,IAChF,+BAA+B,IAAI,iCAAiC;AAAA,IACpE,uBAAuB,IAAI,yBAAyB;AAAA,IACpD,yBAAyB,IAAI,2BAA2B;AAAA,EAC1D;AAAA,OAEM,KAAI,CAAC,QAAgC,SAAuC;AAAA,IAChF,iBAAiB,QAA0C,OAAO;AAAA;AAAA,EAGpE,QAAQ;AAAA,KACL,wBAAU,iBAAiB,OAC1B,SACA,WACsB;AAAA,MACtB,OAAO,oBAAoB,SAAS,MAAM;AAAA;AAAA,KAG3C,wBAAU,wBAAwB,OACjC,SACA,WACsB;AAAA,MACtB,OAAO,sBAAsB,SAAS,MAAM;AAAA;AAAA,KAG7C,wBAAU,wBAAwB,OACjC,SACA,WACoB;AAAA,MACpB,OAAO,sBAAsB,SAAS,MAAM;AAAA;AAAA,KAG7C,wBAAU,aAAa,OACtB,SACA,WACuC;AAAA,MACvC,OAAO,gBAAgB,SAAS,MAAM;AAAA;AAAA,KAGvC,wBAAU,aAAa,OACtB,SACA,WACuC;AAAA,MACvC,OAAO,gBAAgB,SAAS,MAAM;AAAA;AAAA,KAGvC,wBAAU,QAAQ,OACjB,SACA,WACqC;AAAA,MACrC,OAAO,sBAAsB,SAAS,MAAM;AAAA;AAAA,KAG7C,wBAAU,oBAAoB,OAC7B,SACA,WACoD;AAAA,MACpD,OAAO,uBAAuB,SAAS,MAAM;AAAA;AAAA,KAG9C,wBAAU,gBAAgB,OACzB,SACA,UACoB;AAAA,MACpB,OAAO,oBAAoB,SAAS,KAAK;AAAA;AAAA,KAG1C,wBAAU,iBAAiB,OAC1B,SACA,UACyB;AAAA,MACzB,OAAO,mBAAmB,SAAS,KAAK;AAAA;AAAA,KAGzC,wBAAU,eAAe,OACxB,SACA,WACuC;AAAA,MACvC,OAAO,kBAAkB,SAAS,MAAM;AAAA;AAAA,KAGzC,wBAAU,eAAe,OACxB,SACA,WACuC;AAAA,MACvC,OAAO,kBAAkB,SAAS,MAAM;AAAA;AAAA,KAGzC,wBAAU,WAAW,OACpB,SACA,WAC4B;AAAA,MAC5B,OAAO,eAAe,SAAS,MAAM;AAAA;AAAA,EAEzC;AAAA,EAEA,OAAO;AAAA,IACL;AAAA,MACE,MAAM;AAAA,MACN,OAAO;AAAA,QACL;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,UAAU,WAAW,OAAO;AAAA,YAClC,MAAM,WAAW,MAAM,MAAM,GAAG,kBAAkB;AAAA,cAChD,SAAS,cAAc,OAAO;AAAA,YAChC,CAAC;AAAA,YAED,IAAI,CAAC,SAAS,IAAI;AAAA,cAChB,MAAM,IAAI,MACR,iCAAiC,SAAS,UAAU,SAAS,YAC/D;AAAA,YACF;AAAA,YAEA,MAAM,OAAQ,MAAM,SAAS,KAAK;AAAA,YAClC,qBAAO,KAAK,gCAAgC,KAAK,MAAM,UAAU,qBAAqB;AAAA;AAAA,QAE1F;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,YAAY,MAAM,QAAQ,SAAS,wBAAU,gBAAgB;AAAA,cACjE,MAAM;AAAA,YACR,CAAC;AAAA,YAED,IAAI,CAAC,MAAM,QAAQ,SAAS,KAAK,UAAU,WAAW,GAAG;AAAA,cACvD,MAAM,IAAI,MAAM,2CAA2C;AAAA,YAC7D;AAAA,YAEA,qBAAO,KAAK,0CAA0C,UAAU,mBAAmB;AAAA;AAAA,QAEvF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,OAAO,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,cACxD,QAAQ;AAAA,YACV,CAAC;AAAA,YAED,IAAI,OAAO,SAAS,YAAY,KAAK,WAAW,GAAG;AAAA,cACjD,MAAM,IAAI,MAAM,2CAA2C;AAAA,YAC7D;AAAA,YAEA,qBAAO,KAAK,wCAAwC,KAAK,UAAU,GAAG,EAAE,OAAO;AAAA;AAAA,QAEnF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,OAAO,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,cACxD,QAAQ;AAAA,YACV,CAAC;AAAA,YAED,IAAI,OAAO,SAAS,YAAY,KAAK,WAAW,GAAG;AAAA,cACjD,MAAM,IAAI,MAAM,2CAA2C;AAAA,YAC7D;AAAA,YAEA,qBAAO,KAAK,wCAAwC,KAAK,UAAU,GAAG,EAAE,OAAO;AAAA;AAAA,QAEnF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,eAAe;AAAA,YAErB,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,uBAAuB;AAAA,cACrE,QAAQ;AAAA,cACR,WAAW,wBAAU;AAAA,YACvB,CAAC;AAAA,YAED,IAAI,CAAC,MAAM,QAAQ,MAAM,KAAK,OAAO,WAAW,GAAG;AAAA,cACjD,MAAM,IAAI,MAAM,kDAAkD;AAAA,YACpE;AAAA,YAEA,MAAM,cAAc,MAAM,QAAQ,SAAS,wBAAU,uBAAuB;AAAA,cAC1E;AAAA,cACA,WAAW,wBAAU;AAAA,YACvB,CAAC;AAAA,YAED,IAAI,gBAAgB,cAAc;AAAA,cAChC,MAAM,IAAI,MACR,yCAAyC,uBAAuB,cAClE;AAAA,YACF;AAAA,YAEA,qBAAO,KAAK,iDAAiD,OAAO,gBAAgB;AAAA;AAAA,QAExF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,SAAmB,CAAC;AAAA,YAE1B,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,YAAY;AAAA,cAC1D,QAAQ;AAAA,cACR,QAAQ;AAAA,cACR,eAAe,CAAC,UAAkB;AAAA,gBAChC,OAAO,KAAK,KAAK;AAAA;AAAA,YAErB,CAAC;AAAA,YAED,IAAI,OAAO,WAAW,YAAY,OAAO,WAAW,GAAG;AAAA,cACrD,MAAM,IAAI,MAAM,0CAA0C;AAAA,YAC5D;AAAA,YAEA,IAAI,OAAO,WAAW,GAAG;AAAA,cACvB,MAAM,IAAI,MAAM,8BAA8B;AAAA,YAChD;AAAA,YAEA,qBAAO,KAAK,iCAAiC,OAAO,wBAAwB;AAAA;AAAA,QAEhF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,eACJ;AAAA,YAEF,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,mBAAmB,YAAY;AAAA,YAE/E,IACE,CAAC,UACD,OAAO,WAAW,YAClB,EAAE,WAAW,WACb,EAAE,iBAAiB,SACnB;AAAA,cACA,MAAM,IAAI,MAAM,wDAAwD;AAAA,YAC1E;AAAA,YAEA,qBAAO,KAAK,mCAAmC,OAAO,QAAQ;AAAA;AAAA,QAElE;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YAEnD,MAAM,WACJ;AAAA,YAEF,MAAM,WAAW,MAAM,MAAM,QAAQ;AAAA,YACrC,MAAM,cAAc,MAAM,SAAS,YAAY;AAAA,YAC/C,MAAM,cAAc,OAAO,KAAK,IAAI,WAAW,WAAW,CAAC;AAAA,YAE3D,MAAM,gBAAgB,MAAM,QAAQ,SAAS,wBAAU,eAAe,WAAW;AAAA,YAEjF,IAAI,OAAO,kBAAkB,UAAU;AAAA,cACrC,MAAM,IAAI,MAAM,sCAAsC;AAAA,YACxD;AAAA,YAEA,qBAAO,KAAK,iCAAiC,cAAc,UAAU,GAAG,EAAE,OAAO;AAAA;AAAA,QAErF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,YAAY,MAAM,QAAQ,SAAS,wBAAU,gBAAgB;AAAA,cACjE,MAAM;AAAA,YACR,CAAC;AAAA,YAED,IAAI,EAAE,qBAAqB,gBAAgB,UAAU,eAAe,GAAG;AAAA,cACrE,MAAM,IAAI,MAAM,yCAAyC;AAAA,YAC3D;AAAA,YAEA,qBAAO,KAAK,+BAA+B,UAAU,2BAA2B;AAAA;AAAA,QAEpF;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YACnD,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,cAAc;AAAA,cAC5D,QACE;AAAA,YACJ,CAAC;AAAA,YAED,IAAI,CAAC,UAAU,OAAO,WAAW,UAAU;AAAA,cACzC,MAAM,IAAI,MAAM,2CAA2C;AAAA,YAC7D;AAAA,YAEA,qBAAO,KACL,mCAAmC,KAAK,UAAU,MAAM,EAAE,UAAU,GAAG,GAAG,GAC5E;AAAA;AAAA,QAEJ;AAAA,QACA;AAAA,UACE,MAAM;AAAA,UACN,IAAI,OAAO,YAA0C;AAAA,YAGnD,MAAM,SAAS,MAAM,QAAQ,SAAS,wBAAU,UAAU;AAAA,cACxD,OAAO;AAAA,cACP,OAAO,CAAC,EAAE,MAAM,qBAAqB,CAAC;AAAA,cACtC,cAAc;AAAA,YAChB,CAAC;AAAA,YAED,IAAI,CAAC,UAAU,OAAO,WAAW,YAAY,EAAE,UAAU,SAAS;AAAA,cAChE,MAAM,IAAI,MAAM,qDAAqD;AAAA,YACvE;AAAA,YAEA,IAAI,OAAO,OAAO,SAAS,YAAY,OAAO,KAAK,WAAW,GAAG;AAAA,cAC/D,MAAM,IAAI,MAAM,mDAAmD;AAAA,YACrE;AAAA,YAEA,qBAAO,KACL,kDAAkD,OAAO,KAAK,wBAAwB,OAAO,aAAa,UAAU,GACtH;AAAA;AAAA,QAEJ;AAAA,MACF;AAAA,IACF;AAAA,EACF;AACF;AAEA,IAAe;",
22
+ "debugId": "063EB1323839B08A64756E2164756E21",
22
23
  "names": []
23
24
  }