@within-7/minto 0.3.10 → 0.4.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (306) hide show
  1. package/dist/Tool.js.map +2 -2
  2. package/dist/commands/agents/AgentsCommand.js +2 -2
  3. package/dist/commands/agents/AgentsCommand.js.map +2 -2
  4. package/dist/commands/ctx_viz.js +1 -1
  5. package/dist/commands/effort.js +87 -0
  6. package/dist/commands/effort.js.map +7 -0
  7. package/dist/commands/export.js +19 -9
  8. package/dist/commands/export.js.map +2 -2
  9. package/dist/commands/ide.js +18 -0
  10. package/dist/commands/ide.js.map +7 -0
  11. package/dist/commands/mcp-interactive.js +14 -8
  12. package/dist/commands/mcp-interactive.js.map +2 -2
  13. package/dist/commands/memory.js +168 -0
  14. package/dist/commands/memory.js.map +7 -0
  15. package/dist/commands/model.js +45 -2
  16. package/dist/commands/model.js.map +2 -2
  17. package/dist/commands/outputStyle.js +64 -0
  18. package/dist/commands/outputStyle.js.map +7 -0
  19. package/dist/commands/plugin/utils.js +33 -1
  20. package/dist/commands/plugin/utils.js.map +2 -2
  21. package/dist/commands/plugin.js +10 -1
  22. package/dist/commands/plugin.js.map +2 -2
  23. package/dist/commands/refreshCommands.js +2 -0
  24. package/dist/commands/refreshCommands.js.map +2 -2
  25. package/dist/commands/review.js +51 -0
  26. package/dist/commands/review.js.map +7 -0
  27. package/dist/commands/terminalSetup.js +6 -0
  28. package/dist/commands/terminalSetup.js.map +2 -2
  29. package/dist/commands/undo.js +8 -0
  30. package/dist/commands/undo.js.map +2 -2
  31. package/dist/commands/vim.js +22 -0
  32. package/dist/commands/vim.js.map +7 -0
  33. package/dist/commands.js +12 -0
  34. package/dist/commands.js.map +2 -2
  35. package/dist/components/HighlightedCode.js +1 -0
  36. package/dist/components/HighlightedCode.js.map +2 -2
  37. package/dist/components/ModelSelector/ModelSelector.js +250 -143
  38. package/dist/components/ModelSelector/ModelSelector.js.map +2 -2
  39. package/dist/components/PromptInput.js +21 -6
  40. package/dist/components/PromptInput.js.map +2 -2
  41. package/dist/components/PulseLabel.js +44 -0
  42. package/dist/components/PulseLabel.js.map +7 -0
  43. package/dist/components/RequestStatusIndicator.js +1 -1
  44. package/dist/components/RequestStatusIndicator.js.map +1 -1
  45. package/dist/components/Spinner.js +12 -42
  46. package/dist/components/Spinner.js.map +3 -3
  47. package/dist/components/StartupStatus.js +57 -0
  48. package/dist/components/StartupStatus.js.map +7 -0
  49. package/dist/components/SubagentBlock.js +43 -6
  50. package/dist/components/SubagentBlock.js.map +2 -2
  51. package/dist/components/TabbedListView/TabBar.js +13 -8
  52. package/dist/components/TabbedListView/TabBar.js.map +2 -2
  53. package/dist/components/TabbedListView/TabbedListView.js +1 -1
  54. package/dist/components/TabbedListView/TabbedListView.js.map +2 -2
  55. package/dist/components/TodoPanel.js +1 -1
  56. package/dist/components/TodoPanel.js.map +1 -1
  57. package/dist/components/ToolUseLoader.js +5 -0
  58. package/dist/components/ToolUseLoader.js.map +2 -2
  59. package/dist/components/TrustDialog.js +0 -2
  60. package/dist/components/TrustDialog.js.map +2 -2
  61. package/dist/components/messages/TaskInModuleView.js +1 -1
  62. package/dist/components/messages/TaskInModuleView.js.map +2 -2
  63. package/dist/components/messages/TaskToolMessage.js +1 -1
  64. package/dist/components/messages/TaskToolMessage.js.map +2 -2
  65. package/dist/components/messages/UserPromptMessage.js +6 -1
  66. package/dist/components/messages/UserPromptMessage.js.map +2 -2
  67. package/dist/constants/modelCapabilities.js +103 -18
  68. package/dist/constants/modelCapabilities.js.map +2 -2
  69. package/dist/constants/product.js +2 -0
  70. package/dist/constants/product.js.map +2 -2
  71. package/dist/constants/prompts/agentPrompt.js +30 -0
  72. package/dist/constants/prompts/agentPrompt.js.map +7 -0
  73. package/dist/constants/prompts/codeConventions.js +27 -0
  74. package/dist/constants/prompts/codeConventions.js.map +7 -0
  75. package/dist/constants/prompts/doingTasks.js +15 -0
  76. package/dist/constants/prompts/doingTasks.js.map +7 -0
  77. package/dist/constants/prompts/envInfo.js +17 -0
  78. package/dist/constants/prompts/envInfo.js.map +7 -0
  79. package/dist/constants/prompts/executingWithCare.js +17 -0
  80. package/dist/constants/prompts/executingWithCare.js.map +7 -0
  81. package/dist/constants/prompts/identity.js +10 -0
  82. package/dist/constants/prompts/identity.js.map +7 -0
  83. package/dist/constants/prompts/index.js +78 -0
  84. package/dist/constants/prompts/index.js.map +7 -0
  85. package/dist/constants/prompts/taskManagement.js +60 -0
  86. package/dist/constants/prompts/taskManagement.js.map +7 -0
  87. package/dist/constants/prompts/toneAndStyle.js +62 -0
  88. package/dist/constants/prompts/toneAndStyle.js.map +7 -0
  89. package/dist/constants/prompts/toolUsagePolicy.js +38 -0
  90. package/dist/constants/prompts/toolUsagePolicy.js.map +7 -0
  91. package/dist/constants/prompts.js +5 -176
  92. package/dist/constants/prompts.js.map +2 -2
  93. package/dist/constants/providerRegistry.js +235 -0
  94. package/dist/constants/providerRegistry.js.map +7 -0
  95. package/dist/constants/providers.js +35 -0
  96. package/dist/constants/providers.js.map +7 -0
  97. package/dist/context/PermissionContext.js +0 -1
  98. package/dist/context/PermissionContext.js.map +2 -2
  99. package/dist/context.js +87 -31
  100. package/dist/context.js.map +2 -2
  101. package/dist/core/backupHook.js +2 -2
  102. package/dist/core/backupHook.js.map +2 -2
  103. package/dist/core/config/defaults.js +4 -1
  104. package/dist/core/config/defaults.js.map +2 -2
  105. package/dist/core/config/schema.js +7 -1
  106. package/dist/core/config/schema.js.map +2 -2
  107. package/dist/core/costTracker.js +18 -0
  108. package/dist/core/costTracker.js.map +2 -2
  109. package/dist/core/index.js +0 -1
  110. package/dist/core/index.js.map +2 -2
  111. package/dist/core/tokenStatsManager.js +22 -4
  112. package/dist/core/tokenStatsManager.js.map +2 -2
  113. package/dist/entrypoints/cli.js +65 -84
  114. package/dist/entrypoints/cli.js.map +2 -2
  115. package/dist/hooks/useAgentTokenStats.js +1 -1
  116. package/dist/hooks/useAgentTokenStats.js.map +2 -2
  117. package/dist/hooks/useAgentTranscripts.js +2 -1
  118. package/dist/hooks/useAgentTranscripts.js.map +2 -2
  119. package/dist/hooks/useBackgroundShells.js +29 -0
  120. package/dist/hooks/useBackgroundShells.js.map +7 -0
  121. package/dist/hooks/useCanUseTool.js +1 -1
  122. package/dist/hooks/useCanUseTool.js.map +2 -2
  123. package/dist/hooks/useDeferredLoading.js +64 -0
  124. package/dist/hooks/useDeferredLoading.js.map +7 -0
  125. package/dist/hooks/useHookStatus.js +1 -1
  126. package/dist/hooks/useHookStatus.js.map +2 -2
  127. package/dist/hooks/useSessionTracking.js +55 -0
  128. package/dist/hooks/useSessionTracking.js.map +7 -0
  129. package/dist/hooks/useTerminalSize.js +21 -0
  130. package/dist/hooks/useTerminalSize.js.map +2 -2
  131. package/dist/hooks/useTextInput.js +1 -0
  132. package/dist/hooks/useTextInput.js.map +2 -2
  133. package/dist/hooks/useUnifiedCompletion.js +3 -2
  134. package/dist/hooks/useUnifiedCompletion.js.map +2 -2
  135. package/dist/i18n/locales/en.js +8 -9
  136. package/dist/i18n/locales/en.js.map +2 -2
  137. package/dist/i18n/locales/zh-CN.js +8 -9
  138. package/dist/i18n/locales/zh-CN.js.map +2 -2
  139. package/dist/i18n/types.js.map +1 -1
  140. package/dist/messages.js +41 -17
  141. package/dist/messages.js.map +2 -2
  142. package/dist/permissions.js +94 -1
  143. package/dist/permissions.js.map +2 -2
  144. package/dist/query.js +27 -19
  145. package/dist/query.js.map +2 -2
  146. package/dist/screens/REPL.js +83 -74
  147. package/dist/screens/REPL.js.map +2 -2
  148. package/dist/services/adapters/responsesAPI.js +6 -0
  149. package/dist/services/adapters/responsesAPI.js.map +2 -2
  150. package/dist/services/agentTeams/index.js +35 -0
  151. package/dist/services/agentTeams/index.js.map +7 -0
  152. package/dist/services/agentTeams/mailbox.js +114 -0
  153. package/dist/services/agentTeams/mailbox.js.map +7 -0
  154. package/dist/services/agentTeams/teamManager.js +149 -0
  155. package/dist/services/agentTeams/teamManager.js.map +7 -0
  156. package/dist/services/agentTeams/teamTaskStore.js +114 -0
  157. package/dist/services/agentTeams/teamTaskStore.js.map +7 -0
  158. package/dist/services/agentTeams/teammateSpawner.js +80 -0
  159. package/dist/services/agentTeams/teammateSpawner.js.map +7 -0
  160. package/dist/services/checkpointManager.js +16 -3
  161. package/dist/services/checkpointManager.js.map +2 -2
  162. package/dist/services/claude.js +19 -1728
  163. package/dist/services/claude.js.map +3 -3
  164. package/dist/services/gpt5ConnectionTest.js +4 -2
  165. package/dist/services/gpt5ConnectionTest.js.map +2 -2
  166. package/dist/services/hookExecutor.js +411 -127
  167. package/dist/services/hookExecutor.js.map +2 -2
  168. package/dist/services/llm/anthropicProvider.js +807 -0
  169. package/dist/services/llm/anthropicProvider.js.map +7 -0
  170. package/dist/services/llm/dispatch.js +218 -0
  171. package/dist/services/llm/dispatch.js.map +7 -0
  172. package/dist/services/llm/index.js +44 -0
  173. package/dist/services/llm/index.js.map +7 -0
  174. package/dist/services/llm/mintoContext.js +69 -0
  175. package/dist/services/llm/mintoContext.js.map +7 -0
  176. package/dist/services/llm/openaiProvider.js +622 -0
  177. package/dist/services/llm/openaiProvider.js.map +7 -0
  178. package/dist/services/llm/types.js +157 -0
  179. package/dist/services/llm/types.js.map +7 -0
  180. package/dist/services/mcpClient.js +183 -33
  181. package/dist/services/mcpClient.js.map +2 -2
  182. package/dist/services/notifier.js +14 -0
  183. package/dist/services/notifier.js.map +2 -2
  184. package/dist/services/oauth.js +4 -2
  185. package/dist/services/oauth.js.map +2 -2
  186. package/dist/services/openai.js +66 -56
  187. package/dist/services/openai.js.map +3 -3
  188. package/dist/services/outputStyles.js +102 -21
  189. package/dist/services/outputStyles.js.map +2 -2
  190. package/dist/services/plugins/skillMarketplace.js +4 -1
  191. package/dist/services/plugins/skillMarketplace.js.map +2 -2
  192. package/dist/services/sentry.js +1 -1
  193. package/dist/services/sentry.js.map +2 -2
  194. package/dist/services/sessionMemory.js +16 -3
  195. package/dist/services/sessionMemory.js.map +2 -2
  196. package/dist/services/systemReminder.js +350 -3
  197. package/dist/services/systemReminder.js.map +2 -2
  198. package/dist/services/taskStore.js +19 -0
  199. package/dist/services/taskStore.js.map +2 -2
  200. package/dist/tools/ArchitectTool/ArchitectTool.js.map +1 -1
  201. package/dist/tools/AskUserQuestionTool/AskUserQuestionTool.js.map +1 -1
  202. package/dist/tools/BashOutputTool/BashOutputTool.js.map +1 -1
  203. package/dist/tools/BashTool/BashTool.js +28 -0
  204. package/dist/tools/BashTool/BashTool.js.map +2 -2
  205. package/dist/tools/FileEditTool/FileEditTool.js +1 -1
  206. package/dist/tools/FileEditTool/FileEditTool.js.map +2 -2
  207. package/dist/tools/FileReadTool/FileReadTool.js +14 -0
  208. package/dist/tools/FileReadTool/FileReadTool.js.map +2 -2
  209. package/dist/tools/FileWriteTool/FileWriteTool.js +3 -1
  210. package/dist/tools/FileWriteTool/FileWriteTool.js.map +2 -2
  211. package/dist/tools/GlobTool/GlobTool.js.map +1 -1
  212. package/dist/tools/GrepTool/GrepTool.js.map +1 -1
  213. package/dist/tools/KillShellTool/KillShellTool.js.map +1 -1
  214. package/dist/tools/ListMcpResourcesTool/ListMcpResourcesTool.js.map +2 -2
  215. package/dist/tools/LspTool/LspTool.js +11 -2
  216. package/dist/tools/LspTool/LspTool.js.map +2 -2
  217. package/dist/tools/MCPTool/MCPTool.js.map +1 -1
  218. package/dist/tools/MemoryReadTool/MemoryReadTool.js +2 -1
  219. package/dist/tools/MemoryReadTool/MemoryReadTool.js.map +2 -2
  220. package/dist/tools/MemoryWriteTool/MemoryWriteTool.js +2 -1
  221. package/dist/tools/MemoryWriteTool/MemoryWriteTool.js.map +2 -2
  222. package/dist/tools/MultiEditTool/MultiEditTool.js.map +1 -1
  223. package/dist/tools/NotebookEditTool/NotebookEditTool.js.map +1 -1
  224. package/dist/tools/NotebookReadTool/NotebookReadTool.js.map +1 -1
  225. package/dist/tools/PlanModeTool/EnterPlanModeTool.js +8 -2
  226. package/dist/tools/PlanModeTool/EnterPlanModeTool.js.map +2 -2
  227. package/dist/tools/PlanModeTool/ExitPlanModeTool.js +2 -0
  228. package/dist/tools/PlanModeTool/ExitPlanModeTool.js.map +2 -2
  229. package/dist/tools/ReadMcpResourceTool/ReadMcpResourceTool.js.map +1 -1
  230. package/dist/tools/SlashCommandTool/SlashCommandTool.js +174 -18
  231. package/dist/tools/SlashCommandTool/SlashCommandTool.js.map +3 -3
  232. package/dist/tools/TaskCreateTool/TaskCreateTool.js.map +1 -1
  233. package/dist/tools/TaskGetTool/TaskGetTool.js.map +1 -1
  234. package/dist/tools/TaskListTool/TaskListTool.js.map +1 -1
  235. package/dist/tools/TaskOutputTool/TaskOutputTool.js.map +1 -1
  236. package/dist/tools/TaskStopTool/TaskStopTool.js.map +1 -1
  237. package/dist/tools/TaskTool/TaskTool.js +75 -5
  238. package/dist/tools/TaskTool/TaskTool.js.map +2 -2
  239. package/dist/tools/TaskTool/prompt.js +12 -6
  240. package/dist/tools/TaskTool/prompt.js.map +2 -2
  241. package/dist/tools/TaskUpdateTool/TaskUpdateTool.js.map +1 -1
  242. package/dist/tools/ThinkTool/ThinkTool.js.map +1 -1
  243. package/dist/tools/TodoWriteTool/TodoWriteTool.js.map +1 -1
  244. package/dist/tools/URLFetcherTool/URLFetcherTool.js.map +1 -1
  245. package/dist/tools/WebSearchTool/WebSearchTool.js.map +1 -1
  246. package/dist/tools/WebSearchTool/searchProviders.js +2 -1
  247. package/dist/tools/WebSearchTool/searchProviders.js.map +2 -2
  248. package/dist/tools/lsTool/lsTool.js.map +2 -2
  249. package/dist/tools/lsTool/prompt.js.map +1 -1
  250. package/dist/tools.js +14 -3
  251. package/dist/tools.js.map +2 -2
  252. package/dist/types/PermissionMode.js +21 -1
  253. package/dist/types/PermissionMode.js.map +2 -2
  254. package/dist/types/agentTeams.js +1 -0
  255. package/dist/types/agentTeams.js.map +7 -0
  256. package/dist/types/hooks.js +8 -2
  257. package/dist/types/hooks.js.map +2 -2
  258. package/dist/types/plugin.js +1 -1
  259. package/dist/types/plugin.js.map +2 -2
  260. package/dist/utils/agentLoader.js +25 -3
  261. package/dist/utils/agentLoader.js.map +2 -2
  262. package/dist/utils/animationManager.js +1 -1
  263. package/dist/utils/animationManager.js.map +2 -2
  264. package/dist/utils/ask.js +1 -1
  265. package/dist/utils/async.js +5 -1
  266. package/dist/utils/async.js.map +2 -2
  267. package/dist/utils/autoCompactCore.js +60 -0
  268. package/dist/utils/autoCompactCore.js.map +2 -2
  269. package/dist/utils/config.js +26 -128
  270. package/dist/utils/config.js.map +2 -2
  271. package/dist/utils/configSchema.js +227 -0
  272. package/dist/utils/configSchema.js.map +7 -0
  273. package/dist/utils/debugLogger.js.map +2 -2
  274. package/dist/utils/env.js +4 -3
  275. package/dist/utils/env.js.map +2 -2
  276. package/dist/utils/envConfig.js +34 -0
  277. package/dist/utils/envConfig.js.map +3 -3
  278. package/dist/utils/gpt5.js +146 -0
  279. package/dist/utils/gpt5.js.map +7 -0
  280. package/dist/utils/hookManager.js +374 -140
  281. package/dist/utils/hookManager.js.map +2 -2
  282. package/dist/utils/markdown.js +47 -0
  283. package/dist/utils/markdown.js.map +2 -2
  284. package/dist/utils/memoizeWithTTL.js +25 -0
  285. package/dist/utils/memoizeWithTTL.js.map +7 -0
  286. package/dist/utils/model.js +34 -9
  287. package/dist/utils/model.js.map +2 -2
  288. package/dist/utils/pluginInstaller.js +34 -5
  289. package/dist/utils/pluginInstaller.js.map +2 -2
  290. package/dist/utils/pluginLoader.js +201 -32
  291. package/dist/utils/pluginLoader.js.map +2 -2
  292. package/dist/utils/safeFetch.js +45 -0
  293. package/dist/utils/safeFetch.js.map +7 -0
  294. package/dist/utils/skillLoader.js +59 -6
  295. package/dist/utils/skillLoader.js.map +2 -2
  296. package/dist/utils/streamingState.js +52 -0
  297. package/dist/utils/streamingState.js.map +7 -0
  298. package/dist/utils/style.js +6 -3
  299. package/dist/utils/style.js.map +2 -2
  300. package/dist/utils/teamConfig.js +9 -3
  301. package/dist/utils/teamConfig.js.map +2 -2
  302. package/dist/utils/toolRiskClassification.js +0 -6
  303. package/dist/utils/toolRiskClassification.js.map +2 -2
  304. package/dist/version.js +2 -2
  305. package/dist/version.js.map +1 -1
  306. package/package.json +2 -1
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "version": 3,
3
3
  "sources": ["../../src/services/claude.ts"],
4
- "sourcesContent": ["import '@anthropic-ai/sdk/shims/node'\nimport Anthropic, { APIConnectionError, APIError } from '@anthropic-ai/sdk'\nimport { AnthropicBedrock } from '@anthropic-ai/bedrock-sdk'\nimport { AnthropicVertex } from '@anthropic-ai/vertex-sdk'\nimport chalk from 'chalk'\nimport { randomUUID, UUID } from 'crypto'\nimport 'dotenv/config'\n\nimport { addToTotalCost } from '@costTracker'\nimport { recordTokenUsage } from '@core/tokenStatsManager'\nimport type { TokenTrackingContext } from '@core/tokenStats'\nimport models from '@constants/models'\nimport type { AssistantMessage, UserMessage } from '@query'\nimport { Tool, getToolDescriptionAsync } from '@tool'\nimport {\n getAnthropicApiKey,\n getOrCreateUserID,\n getGlobalConfig,\n ModelProfile,\n} from '@utils/config'\nimport { getProjectDocs } from '@context'\nimport { logError, SESSION_ID } from '@utils/log'\nimport { abortableDelay } from '@utils/async'\nimport { USER_AGENT } from '@utils/http'\nimport {\n createAssistantAPIErrorMessage,\n normalizeContentFromAPI,\n} from '@utils/messages'\nimport { withVCR } from './vcr'\nimport {\n debug as debugLogger,\n markPhase,\n getCurrentRequest,\n logLLMInteraction,\n logSystemPromptConstruction,\n logErrorWithDiagnosis,\n} from '@utils/debugLogger'\nimport { getModelManager } from '@utils/model'\nimport { zodToJsonSchema } from 'zod-to-json-schema'\nimport { ModelAdapterFactory } from './modelAdapterFactory'\nimport { UnifiedRequestParams } from '@minto-types/modelCapabilities'\nimport { responseStateManager, getConversationId } from './responseStateManager'\nimport type { ToolUseContext } from '@tool'\nimport type {\n MessageParam,\n TextBlockParam,\n} from '@anthropic-ai/sdk/resources/index.mjs'\nimport { USE_BEDROCK, USE_VERTEX } from '@utils/model'\nimport { getCLISyspromptPrefix } from '@constants/prompts'\nimport { getVertexRegionForModel } from '@utils/model'\nimport OpenAI from 'openai'\nimport type { ChatCompletionStream } from 'openai/lib/ChatCompletionStream'\nimport { ContentBlock } from '@anthropic-ai/sdk/resources/messages/messages'\nimport { nanoid } from 'nanoid'\nimport {\n getCompletionWithProfile,\n getGPT5CompletionWithProfile,\n} from './openai'\nimport { getReasoningEffort } from '@utils/thinking'\nimport { generateSystemReminders } from './systemReminder'\nimport { addRetryEventToTranscript } from '@utils/agentTranscripts'\nimport { setStreamingState, getStreamingState } from '@components/Spinner'\nimport type { ThinkingMetadata } from '@minto-types/thinking'\n\n// Helper function to check if a model is GPT-5\nfunction isGPT5Model(modelName: string): boolean {\n return modelName.startsWith('gpt-5')\n}\n\n// MintoContext\u7BA1\u7406\u5668 - \u7528\u4E8E\u9879\u76EE\u6587\u6863\u7684\u540C\u6B65\u7F13\u5B58\u548C\u8BBF\u95EE\nclass MintoContextManager {\n private static instance: MintoContextManager\n private projectDocsCache: string = ''\n private cacheInitialized: boolean = false\n private initPromise: Promise<void> | null = null\n\n private constructor() {}\n\n public static getInstance(): MintoContextManager {\n if (!MintoContextManager.instance) {\n MintoContextManager.instance = new MintoContextManager()\n }\n return MintoContextManager.instance\n }\n\n public async initialize(): Promise<void> {\n if (this.cacheInitialized) return\n\n if (this.initPromise) {\n return this.initPromise\n }\n\n this.initPromise = this.loadProjectDocs()\n await this.initPromise\n }\n\n private async loadProjectDocs(): Promise<void> {\n try {\n const projectDocs = await getProjectDocs()\n this.projectDocsCache = projectDocs || ''\n this.cacheInitialized = true\n\n // \u5728\u8C03\u8BD5\u6A21\u5F0F\u4E0B\u8BB0\u5F55\u52A0\u8F7D\u7ED3\u679C\n if (process.env.NODE_ENV === 'development') {\n debugLogger.info('MINTO_CONTEXT_LOADED', {\n characters: this.projectDocsCache.length,\n })\n }\n } catch (error) {\n debugLogger.warn('MINTO_CONTEXT_LOAD_FAILED', { error: String(error) })\n this.projectDocsCache = ''\n this.cacheInitialized = true\n }\n }\n\n public getMintoContext(): string {\n if (!this.cacheInitialized) {\n // \u5982\u679C\u672A\u521D\u59CB\u5316\uFF0C\u5F02\u6B65\u521D\u59CB\u5316\u4F46\u7ACB\u5373\u8FD4\u56DE\u7A7A\u5B57\u7B26\u4E32\n this.initialize().catch(error => {\n debugLogger.warn('MINTO_CONTEXT_INIT_FAILED', { error: String(error) })\n })\n return ''\n }\n return this.projectDocsCache\n }\n\n public async refreshCache(): Promise<void> {\n this.cacheInitialized = false\n this.initPromise = null\n await this.initialize()\n }\n}\n\n// \u5BFC\u51FA\u51FD\u6570\u4FDD\u6301\u5411\u540E\u517C\u5BB9\nconst mintoContextManager = MintoContextManager.getInstance()\n\n// \u5728\u6A21\u5757\u52A0\u8F7D\u65F6\u5F02\u6B65\u521D\u59CB\u5316\nmintoContextManager.initialize().catch(error => {\n debugLogger.warn('MINTO_CONTEXT_MODULE_INIT_FAILED', { error: String(error) })\n})\n\nexport const generateMintoContext = (): string => {\n return mintoContextManager.getMintoContext()\n}\n\nexport const refreshMintoContext = async (): Promise<void> => {\n await mintoContextManager.refreshCache()\n}\n\nexport const API_ERROR_MESSAGE_PREFIX = 'API Error'\nexport const PROMPT_TOO_LONG_ERROR_MESSAGE = 'Prompt is too long'\nexport const CREDIT_BALANCE_TOO_LOW_ERROR_MESSAGE = 'Credit balance is too low'\nexport const INVALID_API_KEY_ERROR_MESSAGE =\n 'Invalid API key \u00B7 Please run /login'\nexport const NO_CONTENT_MESSAGE = '(no content)'\nconst PROMPT_CACHING_ENABLED = !process.env.DISABLE_PROMPT_CACHING\n\n// @see https://docs.anthropic.com/en/docs/about-claude/models#model-comparison-table\nconst SONNET_COST_PER_MILLION_INPUT_TOKENS = 3\nconst SONNET_COST_PER_MILLION_OUTPUT_TOKENS = 15\nconst SONNET_COST_PER_MILLION_PROMPT_CACHE_WRITE_TOKENS = 3.75\nconst SONNET_COST_PER_MILLION_PROMPT_CACHE_READ_TOKENS = 0.3\n\nexport const MAIN_QUERY_TEMPERATURE = 1 // to get more variation for binary feedback\n\nfunction getMetadata() {\n return {\n user_id: `${getOrCreateUserID()}_${SESSION_ID}`,\n }\n}\n\nconst MAX_RETRIES = process.env.USER_TYPE === 'SWE_BENCH' ? 100 : 10\nconst BASE_DELAY_MS = 500\n\ninterface RetryOptions {\n maxRetries?: number\n signal?: AbortSignal\n /** Callback for retry events. If provided, suppresses console.log output. */\n onRetry?: (info: {\n attempt: number\n maxRetries: number\n error: Error\n delayMs: number\n }) => void\n}\n\nfunction getRetryDelay(\n attempt: number,\n retryAfterHeader?: string | null,\n): number {\n if (retryAfterHeader) {\n const seconds = parseInt(retryAfterHeader, 10)\n if (!isNaN(seconds)) {\n return seconds * 1000\n }\n }\n return Math.min(BASE_DELAY_MS * Math.pow(2, attempt - 1), 32000) // Max 32s delay\n}\n\nfunction shouldRetry(error: APIError): boolean {\n // Check for overloaded errors first and only retry for SWE_BENCH\n if (error.message?.includes('\"type\":\"overloaded_error\"')) {\n return process.env.USER_TYPE === 'SWE_BENCH'\n }\n\n // Note this is not a standard header.\n const shouldRetryHeader = error.headers?.['x-should-retry']\n\n // If the server explicitly says whether or not to retry, obey.\n if (shouldRetryHeader === 'true') return true\n if (shouldRetryHeader === 'false') return false\n\n if (error instanceof APIConnectionError) {\n return true\n }\n\n if (!error.status) return false\n\n // Retry on request timeouts.\n if (error.status === 408) return true\n\n // Retry on lock timeouts.\n if (error.status === 409) return true\n\n // Retry on rate limits.\n if (error.status === 429) return true\n\n // Retry internal errors.\n if (error.status && error.status >= 500) return true\n\n return false\n}\n\nasync function withRetry<T>(\n operation: (attempt: number) => Promise<T>,\n options: RetryOptions = {},\n): Promise<T> {\n const maxRetries = options.maxRetries ?? MAX_RETRIES\n let lastError: unknown\n\n for (let attempt = 1; attempt <= maxRetries + 1; attempt++) {\n try {\n return await operation(attempt)\n } catch (error) {\n lastError = error\n // Only retry if the error indicates we should\n if (\n attempt > maxRetries ||\n !(error instanceof APIError) ||\n !shouldRetry(error)\n ) {\n throw error\n }\n\n if (options.signal?.aborted) {\n throw new Error('Request cancelled by user')\n }\n\n // Get retry-after header if available\n const retryAfter = error.headers?.['retry-after'] ?? null\n const delayMs = getRetryDelay(attempt, retryAfter)\n\n // \u2705 P1 FIX: Update streaming state to show retry in UI\n setStreamingState({\n phase: 'retrying',\n retryCount: attempt,\n maxRetries: maxRetries,\n errorName: error.name,\n })\n\n debugLogger.flow('API_RETRY_ATTEMPT', {\n attempt,\n maxRetries,\n errorName: error.name,\n errorMessage: error.message,\n delayMs,\n })\n\n // Use callback if provided, otherwise fallback to console.log\n if (options.onRetry) {\n options.onRetry({\n attempt,\n maxRetries,\n error,\n delayMs,\n })\n } else {\n console.log(\n ` \u23BF ${chalk.red(`API ${error.name} (${error.message}) \u00B7 Retrying in ${Math.round(delayMs / 1000)} seconds\u2026 (attempt ${attempt}/${maxRetries})`)}`,\n )\n }\n\n try {\n await abortableDelay(delayMs, options.signal)\n } catch (delayError) {\n // If aborted during delay, throw the error to stop retrying\n if (delayError.message === 'Request was aborted') {\n throw new Error('Request cancelled by user')\n }\n throw delayError\n }\n }\n }\n\n throw lastError\n}\n\n/**\n * Fetch available models from Anthropic API\n */\nexport async function fetchAnthropicModels(\n baseURL: string,\n apiKey: string,\n): Promise<any[]> {\n try {\n // Use provided baseURL or default to official Anthropic API\n const modelsURL = baseURL\n ? `${baseURL.replace(/\\/+$/, '')}/v1/models`\n : 'https://api.anthropic.com/v1/models'\n\n const response = await fetch(modelsURL, {\n method: 'GET',\n headers: {\n 'x-api-key': apiKey,\n 'anthropic-version': '2023-06-01',\n 'User-Agent': USER_AGENT,\n },\n })\n\n if (!response.ok) {\n // Provide user-friendly error messages based on status code\n if (response.status === 401) {\n throw new Error(\n 'Invalid API key. Please check your Anthropic API key and try again.',\n )\n } else if (response.status === 403) {\n throw new Error(\n 'API key does not have permission to access models. Please check your API key permissions.',\n )\n } else if (response.status === 429) {\n throw new Error(\n 'Too many requests. Please wait a moment and try again.',\n )\n } else if (response.status >= 500) {\n throw new Error(\n 'Anthropic service is temporarily unavailable. Please try again later.',\n )\n } else {\n throw new Error(\n `Unable to connect to Anthropic API (${response.status}). Please check your internet connection and API key.`,\n )\n }\n }\n\n const data = await response.json()\n return data.data || []\n } catch (error) {\n // If it's already our custom error, pass it through\n if (\n (error instanceof Error && error.message.includes('API key')) ||\n (error instanceof Error && error.message.includes('Anthropic'))\n ) {\n throw error\n }\n\n // For network errors or other issues\n logError(error)\n throw new Error(\n 'Unable to connect to Anthropic API. Please check your internet connection and try again.',\n )\n }\n}\n\nexport async function verifyApiKey(\n apiKey: string,\n baseURL?: string,\n provider?: string,\n): Promise<boolean> {\n if (!apiKey) {\n return false\n }\n\n // For non-Anthropic providers, use OpenAI-compatible verification\n if (provider && provider !== 'anthropic') {\n try {\n const headers: Record<string, string> = {\n Authorization: `Bearer ${apiKey}`,\n 'Content-Type': 'application/json',\n }\n\n if (!baseURL) {\n debugLogger.warn('API_VERIFY', {\n error: 'No baseURL provided for non-Anthropic provider',\n })\n return false\n }\n\n const modelsURL = `${baseURL.replace(/\\/+$/, '')}/models`\n\n const response = await fetch(modelsURL, {\n method: 'GET',\n headers,\n })\n\n return response.ok\n } catch (error) {\n debugLogger.warn('API_VERIFY', {\n error: 'non-Anthropic provider verification failed',\n details: String(error),\n })\n return false\n }\n }\n\n // For Anthropic and Anthropic-compatible APIs\n const clientConfig: any = {\n apiKey,\n dangerouslyAllowBrowser: true,\n maxRetries: 3,\n defaultHeaders: {\n 'User-Agent': USER_AGENT,\n },\n }\n\n // Only add baseURL for true Anthropic-compatible APIs\n if (\n baseURL &&\n (provider === 'anthropic' ||\n provider === 'bigdream' ||\n provider === 'opendev')\n ) {\n clientConfig.baseURL = baseURL\n }\n\n const anthropic = new Anthropic(clientConfig)\n\n try {\n await withRetry(\n async () => {\n const model = 'claude-sonnet-4-20250514'\n const messages: MessageParam[] = [{ role: 'user', content: 'test' }]\n await anthropic.messages.create({\n model,\n max_tokens: 1000, // Simple test token limit for API verification\n messages,\n temperature: 0,\n metadata: getMetadata(),\n })\n return true\n },\n { maxRetries: 2 }, // Use fewer retries for API key verification\n )\n return true\n } catch (error) {\n logError(error)\n // Check for authentication error\n if (\n error instanceof Error &&\n error.message.includes(\n '{\"type\":\"error\",\"error\":{\"type\":\"authentication_error\",\"message\":\"invalid x-api-key\"}}',\n )\n ) {\n return false\n }\n throw error\n }\n}\n\nfunction convertAnthropicMessagesToOpenAIMessages(\n messages: (UserMessage | AssistantMessage)[],\n): (\n | OpenAI.ChatCompletionMessageParam\n | OpenAI.ChatCompletionToolMessageParam\n)[] {\n const openaiMessages: (\n | OpenAI.ChatCompletionMessageParam\n | OpenAI.ChatCompletionToolMessageParam\n )[] = []\n\n const toolResults: Record<string, OpenAI.ChatCompletionToolMessageParam> = {}\n\n for (const message of messages) {\n let contentBlocks = []\n if (typeof message.message.content === 'string') {\n contentBlocks = [\n {\n type: 'text',\n text: message.message.content,\n },\n ]\n } else if (!Array.isArray(message.message.content)) {\n contentBlocks = [message.message.content]\n } else {\n contentBlocks = message.message.content\n }\n\n for (const block of contentBlocks) {\n if (block.type === 'text') {\n openaiMessages.push({\n role: message.message.role,\n content: block.text,\n })\n } else if (block.type === 'tool_use') {\n openaiMessages.push({\n role: 'assistant',\n content: undefined,\n tool_calls: [\n {\n type: 'function',\n function: {\n name: block.name,\n arguments: JSON.stringify(block.input),\n },\n id: block.id,\n },\n ],\n })\n } else if (block.type === 'tool_result') {\n // Ensure content is always a string for role:tool messages\n let toolContent = block.content\n if (typeof toolContent !== 'string') {\n // Convert content to string if it's not already\n toolContent = JSON.stringify(toolContent)\n }\n\n toolResults[block.tool_use_id] = {\n role: 'tool',\n content: toolContent,\n tool_call_id: block.tool_use_id,\n }\n }\n }\n }\n\n const finalMessages: (\n | OpenAI.ChatCompletionMessageParam\n | OpenAI.ChatCompletionToolMessageParam\n )[] = []\n\n for (const message of openaiMessages) {\n finalMessages.push(message)\n\n if ('tool_calls' in message && message.tool_calls) {\n for (const toolCall of message.tool_calls) {\n if (toolResults[toolCall.id]) {\n finalMessages.push(toolResults[toolCall.id])\n }\n }\n }\n }\n\n return finalMessages\n}\n\nfunction messageReducer(\n previous: OpenAI.ChatCompletionMessage,\n item: OpenAI.ChatCompletionChunk,\n): OpenAI.ChatCompletionMessage {\n const reduce = (acc: any, delta: OpenAI.ChatCompletionChunk.Choice.Delta) => {\n acc = { ...acc }\n for (const [key, value] of Object.entries(delta)) {\n if (acc[key] === undefined || acc[key] === null) {\n acc[key] = value\n // OpenAI.Chat.Completions.ChatCompletionMessageToolCall does not have a key, .index\n if (Array.isArray(acc[key])) {\n for (const arr of acc[key]) {\n delete arr.index\n }\n }\n } else if (typeof acc[key] === 'string' && typeof value === 'string') {\n acc[key] += value\n } else if (typeof acc[key] === 'number' && typeof value === 'number') {\n acc[key] = value\n } else if (Array.isArray(acc[key]) && Array.isArray(value)) {\n const accArray = acc[key]\n for (let i = 0; i < value.length; i++) {\n const { index, ...chunkTool } = value[i]\n if (index - accArray.length > 1) {\n throw new Error(\n `Error: An array has an empty value when tool_calls are constructed. tool_calls: ${accArray}; tool: ${value}`,\n )\n }\n accArray[index] = reduce(accArray[index], chunkTool)\n }\n } else if (typeof acc[key] === 'object' && typeof value === 'object') {\n acc[key] = reduce(acc[key], value)\n }\n }\n return acc\n }\n\n const choice = item.choices?.[0]\n if (!choice) {\n // chunk contains information about usage and token counts\n return previous\n }\n return reduce(previous, choice.delta) as OpenAI.ChatCompletionMessage\n}\nasync function handleMessageStream(\n stream: ChatCompletionStream,\n signal?: AbortSignal,\n): Promise<OpenAI.ChatCompletion> {\n const streamStartTime = Date.now()\n let ttftMs: number | undefined\n let chunkCount = 0\n let errorCount = 0\n\n debugLogger.api('OPENAI_STREAM_START', {\n streamStartTime: String(streamStartTime),\n })\n\n let message = {} as OpenAI.ChatCompletionMessage\n\n let id, model, created, object, usage\n try {\n for await (const chunk of stream) {\n if (signal?.aborted) {\n debugLogger.flow('OPENAI_STREAM_ABORTED', {\n chunkCount,\n timestamp: Date.now(),\n })\n throw new Error('Request was cancelled')\n }\n\n chunkCount++\n\n try {\n if (!id) {\n id = chunk.id\n debugLogger.api('OPENAI_STREAM_ID_RECEIVED', {\n id,\n chunkNumber: String(chunkCount),\n })\n }\n if (!model) {\n model = chunk.model\n debugLogger.api('OPENAI_STREAM_MODEL_RECEIVED', {\n model,\n chunkNumber: String(chunkCount),\n })\n }\n if (!created) {\n created = chunk.created\n }\n if (!object) {\n object = chunk.object\n }\n if (!usage && chunk.usage) {\n usage = chunk.usage\n // Update streaming state with token usage (OpenAI provides usage in final chunk)\n setStreamingState({\n inputTokens: chunk.usage.prompt_tokens,\n outputTokens: chunk.usage.completion_tokens,\n })\n }\n\n message = messageReducer(message, chunk)\n\n if (chunk?.choices?.[0]?.delta?.content) {\n if (!ttftMs) {\n ttftMs = Date.now() - streamStartTime\n debugLogger.api('OPENAI_STREAM_FIRST_TOKEN', {\n ttftMs: String(ttftMs),\n chunkNumber: String(chunkCount),\n })\n }\n }\n } catch (chunkError) {\n errorCount++\n debugLogger.error('OPENAI_STREAM_CHUNK_ERROR', {\n chunkNumber: String(chunkCount),\n errorMessage:\n chunkError instanceof Error\n ? chunkError.message\n : String(chunkError),\n errorType:\n chunkError instanceof Error\n ? chunkError.constructor.name\n : typeof chunkError,\n })\n // Continue processing other chunks\n }\n }\n\n debugLogger.api('OPENAI_STREAM_COMPLETE', {\n totalChunks: String(chunkCount),\n errorCount: String(errorCount),\n totalDuration: String(Date.now() - streamStartTime),\n ttftMs: String(ttftMs || 0),\n finalMessageId: id || 'undefined',\n })\n } catch (streamError) {\n debugLogger.error('OPENAI_STREAM_FATAL_ERROR', {\n totalChunks: String(chunkCount),\n errorCount: String(errorCount),\n errorMessage:\n streamError instanceof Error\n ? streamError.message\n : String(streamError),\n errorType:\n streamError instanceof Error\n ? streamError.constructor.name\n : typeof streamError,\n })\n throw streamError\n }\n return {\n id,\n created,\n model,\n object,\n choices: [\n {\n index: 0,\n message,\n finish_reason: 'stop',\n logprobs: undefined,\n },\n ],\n usage,\n }\n}\n\nfunction convertOpenAIResponseToAnthropic(\n response: OpenAI.ChatCompletion,\n _tools?: Tool[],\n) {\n let contentBlocks: ContentBlock[] = []\n const message = response.choices?.[0]?.message\n if (!message) {\n return {\n role: 'assistant',\n content: [],\n stop_reason: response.choices?.[0]?.finish_reason,\n type: 'message',\n usage: response.usage,\n }\n }\n\n if (message?.tool_calls) {\n for (const toolCall of message.tool_calls) {\n const tool = toolCall.function\n const toolName = tool?.name\n let toolArgs = {}\n let parseError: string | null = null\n try {\n toolArgs = tool?.arguments ? JSON.parse(tool.arguments) : {}\n } catch (e) {\n // Invalid JSON in tool arguments - log for debugging\n parseError = e instanceof Error ? e.message : String(e)\n debugLogger.warn('TOOL_ARGUMENTS_PARSE_ERROR', {\n toolName,\n rawArguments: tool?.arguments?.substring(0, 200), // Limit log size\n errorMessage: parseError,\n })\n }\n\n // If parsing failed and we have empty args, add context about the parse failure\n if (parseError && Object.keys(toolArgs).length === 0) {\n debugLogger.error('TOOL_CALL_EMPTY_ARGUMENTS', {\n toolName,\n parseError,\n suggestion:\n 'Model may have sent malformed JSON. The tool call will likely fail validation.',\n })\n }\n\n contentBlocks.push({\n type: 'tool_use',\n input: toolArgs,\n name: toolName,\n id: toolCall.id?.length > 0 ? toolCall.id : nanoid(),\n })\n }\n }\n\n if ((message as any).reasoning) {\n contentBlocks.push({\n type: 'thinking',\n thinking: (message as any).reasoning,\n signature: '',\n })\n }\n\n // NOTE: For deepseek api, the key for its returned reasoning process is reasoning_content\n if ((message as any).reasoning_content) {\n contentBlocks.push({\n type: 'thinking',\n thinking: (message as any).reasoning_content,\n signature: '',\n })\n }\n\n if (message.content) {\n contentBlocks.push({\n type: 'text',\n text: message?.content,\n citations: [],\n })\n }\n\n const finalMessage = {\n role: 'assistant',\n content: contentBlocks,\n stop_reason: response.choices?.[0]?.finish_reason,\n type: 'message',\n usage: response.usage,\n }\n\n return finalMessage\n}\n\nlet anthropicClient: Anthropic | AnthropicBedrock | AnthropicVertex | null =\n null\n\n/**\n * Get the Anthropic client, creating it if it doesn't exist\n */\nexport function getAnthropicClient(\n model?: string,\n): Anthropic | AnthropicBedrock | AnthropicVertex {\n const config = getGlobalConfig()\n const provider = config.primaryProvider\n\n // Reset client if provider has changed to ensure correct configuration\n if (anthropicClient && provider) {\n // Always recreate client for provider-specific configurations\n anthropicClient = null\n }\n\n if (anthropicClient) {\n return anthropicClient\n }\n\n const region = getVertexRegionForModel(model)\n\n const defaultHeaders: { [key: string]: string } = {\n 'x-app': 'cli',\n 'User-Agent': USER_AGENT,\n }\n if (process.env.ANTHROPIC_AUTH_TOKEN) {\n defaultHeaders['Authorization'] =\n `Bearer ${process.env.ANTHROPIC_AUTH_TOKEN}`\n }\n\n const ARGS = {\n defaultHeaders,\n maxRetries: 0, // Disabled auto-retry in favor of manual implementation\n timeout: parseInt(process.env.API_TIMEOUT_MS || String(60 * 1000), 10),\n }\n if (USE_BEDROCK) {\n const client = new AnthropicBedrock(ARGS)\n anthropicClient = client\n return client\n }\n if (USE_VERTEX) {\n const vertexArgs = {\n ...ARGS,\n region: region || process.env.CLOUD_ML_REGION || 'us-east5',\n }\n const client = new AnthropicVertex(vertexArgs)\n anthropicClient = client\n return client\n }\n\n // Get appropriate API key and baseURL from ModelProfile\n const modelManager = getModelManager()\n const modelProfile = modelManager.getModel('main')\n\n let apiKey: string\n let baseURL: string | undefined\n\n if (modelProfile) {\n apiKey = modelProfile.apiKey || ''\n baseURL = modelProfile.baseURL\n } else {\n // Fallback to default anthropic if no ModelProfile\n apiKey = getAnthropicApiKey()\n baseURL = undefined\n }\n\n if (process.env.USER_TYPE === 'ant' && !apiKey && provider === 'anthropic') {\n console.error(\n chalk.red(\n '[ANT-ONLY] Please set the ANTHROPIC_API_KEY environment variable to use the CLI. To create a new key, go to https://console.anthropic.com/settings/keys.',\n ),\n )\n }\n\n // Create client with custom baseURL for BigDream/OpenDev\n // Anthropic SDK will append the appropriate paths (like /v1/messages)\n const clientConfig = {\n apiKey,\n dangerouslyAllowBrowser: true,\n ...ARGS,\n ...(baseURL && { baseURL }), // Use baseURL directly, SDK will handle API versioning\n }\n\n anthropicClient = new Anthropic(clientConfig)\n return anthropicClient\n}\n\n/**\n * Reset the Anthropic client to null, forcing a new client to be created on next use\n */\nexport function resetAnthropicClient(): void {\n anthropicClient = null\n}\n\n/**\n * Environment variables for different client types:\n *\n * Direct API:\n * - ANTHROPIC_API_KEY: Required for direct API access\n *\n * AWS Bedrock:\n * - AWS credentials configured via aws-sdk defaults\n *\n * Vertex AI:\n * - Model-specific region variables (highest priority):\n * - VERTEX_REGION_CLAUDE_3_5_HAIKU: Region for Claude 3.5 Haiku model\n * - VERTEX_REGION_CLAUDE_3_5_SONNET: Region for Claude 3.5 Sonnet model\n * - VERTEX_REGION_CLAUDE_3_7_SONNET: Region for Claude 3.7 Sonnet model\n * - CLOUD_ML_REGION: Optional. The default GCP region to use for all models\n * If specific model region not specified above\n * - ANTHROPIC_VERTEX_PROJECT_ID: Required. Your GCP project ID\n * - Standard GCP credentials configured via google-auth-library\n *\n * Priority for determining region:\n * 1. Hardcoded model-specific environment variables\n * 2. Global CLOUD_ML_REGION variable\n * 3. Default region from config\n * 4. Fallback region (us-east5)\n */\n\n/**\n * Manage cache control to ensure it doesn't exceed Claude's 4 cache block limit\n * Priority:\n * 1. System prompts (high priority)\n * 2. Long documents or reference materials (high priority)\n * 3. Reusable context (medium priority)\n * 4. Short messages or one-time content (no caching)\n */\nfunction applyCacheControlWithLimits(\n systemBlocks: TextBlockParam[],\n messageParams: MessageParam[],\n): { systemBlocks: TextBlockParam[]; messageParams: MessageParam[] } {\n if (!PROMPT_CACHING_ENABLED) {\n return { systemBlocks, messageParams }\n }\n\n const maxCacheBlocks = 4\n let usedCacheBlocks = 0\n\n // 1. Prioritize adding cache to system prompts (highest priority)\n const processedSystemBlocks = systemBlocks.map((block, _index) => {\n if (usedCacheBlocks < maxCacheBlocks && block.text.length > 1000) {\n usedCacheBlocks++\n return {\n ...block,\n cache_control: { type: 'ephemeral' as const },\n }\n }\n const { cache_control, ...blockWithoutCache } = block\n return blockWithoutCache\n })\n\n // 2. Add cache to message content based on priority\n const processedMessageParams = messageParams.map((message, messageIndex) => {\n if (Array.isArray(message.content)) {\n const processedContent = message.content.map(\n (contentBlock, blockIndex) => {\n // Determine whether this content block should be cached\n const shouldCache =\n usedCacheBlocks < maxCacheBlocks &&\n contentBlock.type === 'text' &&\n typeof contentBlock.text === 'string' &&\n // Long documents (over 2000 characters)\n (contentBlock.text.length > 2000 ||\n // Last content block of the last message (may be important context)\n (messageIndex === messageParams.length - 1 &&\n blockIndex === message.content.length - 1 &&\n contentBlock.text.length > 500))\n\n if (shouldCache) {\n usedCacheBlocks++\n return {\n ...contentBlock,\n cache_control: { type: 'ephemeral' as const },\n }\n }\n\n // Remove existing cache_control\n const { cache_control, ...blockWithoutCache } = contentBlock as any\n return blockWithoutCache\n },\n )\n\n return {\n ...message,\n content: processedContent,\n }\n }\n\n return message\n })\n\n return {\n systemBlocks: processedSystemBlocks,\n messageParams: processedMessageParams,\n }\n}\n\nexport function userMessageToMessageParam(\n message: UserMessage,\n addCache = false,\n): MessageParam {\n if (addCache) {\n if (typeof message.message.content === 'string') {\n return {\n role: 'user',\n content: [\n {\n type: 'text',\n text: message.message.content,\n },\n ],\n }\n } else {\n return {\n role: 'user',\n content: message.message.content.map(_ => ({ ..._ })),\n }\n }\n }\n return {\n role: 'user',\n content: message.message.content,\n }\n}\n\nexport function assistantMessageToMessageParam(\n message: AssistantMessage,\n addCache = false,\n): MessageParam {\n if (addCache) {\n if (typeof message.message.content === 'string') {\n return {\n role: 'assistant',\n content: [\n {\n type: 'text',\n text: message.message.content,\n },\n ],\n }\n } else {\n return {\n role: 'assistant',\n content: message.message.content.map(_ => ({ ..._ })),\n }\n }\n }\n return {\n role: 'assistant',\n content: message.message.content,\n }\n}\n\nfunction splitSysPromptPrefix(systemPrompt: string[]): string[] {\n // split out the first block of the system prompt as the \"prefix\" for API\n\n const systemPromptFirstBlock = systemPrompt[0] || ''\n const systemPromptRest = systemPrompt.slice(1)\n return [systemPromptFirstBlock, systemPromptRest.join('\\n')].filter(Boolean)\n}\n\nexport async function queryLLM(\n messages: (UserMessage | AssistantMessage)[],\n systemPrompt: string[],\n maxThinkingTokens: number,\n tools: Tool[],\n signal: AbortSignal,\n options: {\n safeMode: boolean\n model: string | import('@utils/config').ModelPointerType\n prependCLISysprompt: boolean\n toolUseContext?: ToolUseContext\n },\n): Promise<AssistantMessage> {\n const modelManager = getModelManager()\n const modelResolution = modelManager.resolveModelWithInfo(options.model)\n\n if (!modelResolution.success || !modelResolution.profile) {\n throw new Error(\n modelResolution.error || `Failed to resolve model: ${options.model}`,\n )\n }\n\n const modelProfile = modelResolution.profile\n const resolvedModel = modelProfile.modelName\n\n // Initialize response state if toolUseContext is provided\n const toolUseContext = options.toolUseContext\n if (toolUseContext && !toolUseContext.responseState) {\n const conversationId = getConversationId(\n toolUseContext.agentId,\n toolUseContext.messageId,\n )\n const previousResponseId =\n responseStateManager.getPreviousResponseId(conversationId)\n\n toolUseContext.responseState = {\n previousResponseId,\n conversationId,\n }\n }\n\n debugLogger.api('MODEL_RESOLVED', {\n inputParam: options.model,\n resolvedModelName: resolvedModel,\n provider: modelProfile.provider,\n isPointer: ['main', 'task', 'reasoning', 'quick'].includes(options.model),\n hasResponseState: !!toolUseContext?.responseState,\n conversationId: toolUseContext?.responseState?.conversationId,\n requestId: getCurrentRequest()?.id,\n })\n\n const currentRequest = getCurrentRequest()\n debugLogger.api('LLM_REQUEST_START', {\n messageCount: messages.length,\n systemPromptLength: systemPrompt.join(' ').length,\n toolCount: tools.length,\n model: resolvedModel,\n originalModelParam: options.model,\n requestId: getCurrentRequest()?.id,\n })\n\n markPhase('LLM_CALL')\n\n try {\n const result = await withVCR(messages, () =>\n queryLLMWithPromptCaching(\n messages,\n systemPrompt,\n maxThinkingTokens,\n tools,\n signal,\n { ...options, model: resolvedModel, modelProfile, toolUseContext }, // Pass resolved ModelProfile and toolUseContext\n ),\n )\n\n debugLogger.api('LLM_REQUEST_SUCCESS', {\n costUSD: result.costUSD,\n durationMs: result.durationMs,\n responseLength: result.message.content?.length || 0,\n requestId: getCurrentRequest()?.id,\n })\n\n // Update response state for GPT-5 Responses API continuation\n if (toolUseContext?.responseState?.conversationId && result.responseId) {\n responseStateManager.setPreviousResponseId(\n toolUseContext.responseState.conversationId,\n result.responseId,\n )\n\n debugLogger.api('RESPONSE_STATE_UPDATED', {\n conversationId: toolUseContext.responseState.conversationId,\n responseId: result.responseId,\n requestId: getCurrentRequest()?.id,\n })\n }\n\n return result\n } catch (error) {\n // \u4F7F\u7528\u9519\u8BEF\u8BCA\u65AD\u7CFB\u7EDF\u8BB0\u5F55 LLM \u76F8\u5173\u9519\u8BEF\n logErrorWithDiagnosis(\n error,\n {\n messageCount: messages.length,\n systemPromptLength: systemPrompt.join(' ').length,\n model: options.model,\n toolCount: tools.length,\n phase: 'LLM_CALL',\n },\n currentRequest?.id,\n )\n\n throw error\n }\n}\n\nexport function formatSystemPromptWithContext(\n systemPrompt: string[],\n context: { [k: string]: string },\n agentId?: string,\n skipContextReminders = false, // Parameter kept for API compatibility but not used anymore\n): { systemPrompt: string[]; reminders: string } {\n // \u6784\u5EFA\u589E\u5F3A\u7684\u7CFB\u7EDF\u63D0\u793A\uFF0C\u4FDD\u6301\u4E0E\u539F\u5148\u76F4\u63A5\u6CE8\u5165\u65B9\u5F0F\u7684\u517C\u5BB9\n const enhancedPrompt = [...systemPrompt]\n let reminders = ''\n\n // Step 0: Add GPT-5 Agent persistence support for coding tasks\n const modelManager = getModelManager()\n const modelProfile = modelManager.getModel('main')\n if (modelProfile && isGPT5Model(modelProfile.modelName)) {\n // Add coding-specific persistence instructions based on GPT-5 documentation\n const persistencePrompts = [\n '\\n# Agent Persistence for Long-Running Coding Tasks',\n 'You are working on a coding project that may involve multiple steps and iterations. Please maintain context and continuity throughout the session:',\n '- Remember architectural decisions and design patterns established earlier',\n '- Keep track of file modifications and their relationships',\n '- Maintain awareness of the overall project structure and goals',\n '- Reference previous implementations when making related changes',\n '- Ensure consistency with existing code style and conventions',\n '- Build incrementally on previous work rather than starting from scratch',\n ]\n enhancedPrompt.push(...persistencePrompts)\n }\n\n // \u53EA\u6709\u5F53\u4E0A\u4E0B\u6587\u5B58\u5728\u65F6\u624D\u5904\u7406\n const hasContext = Object.entries(context).length > 0\n\n if (hasContext) {\n // \u6B65\u9AA41: \u76F4\u63A5\u6CE8\u5165 Minto \u4E0A\u4E0B\u6587\u5230\u7CFB\u7EDF\u63D0\u793A - \u5BF9\u9F50\u5B98\u65B9\u8BBE\u8BA1\n if (!skipContextReminders) {\n const mintoContext = generateMintoContext()\n if (mintoContext) {\n // \u6DFB\u52A0\u5206\u9694\u7B26\u548C\u6807\u8BC6\uFF0C\u4F7F\u9879\u76EE\u6587\u6863\u5728\u7CFB\u7EDF\u63D0\u793A\u4E2D\u66F4\u6E05\u6670\n enhancedPrompt.push('\\n---\\n# \u9879\u76EE\u4E0A\u4E0B\u6587\\n')\n enhancedPrompt.push(mintoContext)\n enhancedPrompt.push('\\n---\\n')\n }\n }\n\n // \u6B65\u9AA42: \u751F\u6210\u5176\u4ED6\u52A8\u6001\u63D0\u9192\u8FD4\u56DE\u7ED9\u8C03\u7528\u65B9 - \u4FDD\u6301\u73B0\u6709\u52A8\u6001\u63D0\u9192\u529F\u80FD\n const reminderMessages = generateSystemReminders(hasContext, agentId)\n if (reminderMessages.length > 0) {\n reminders = reminderMessages.map(r => r.content).join('\\n') + '\\n'\n }\n\n // \u6B65\u9AA43: \u6DFB\u52A0\u5176\u4ED6\u4E0A\u4E0B\u6587\u5230\u7CFB\u7EDF\u63D0\u793A\n enhancedPrompt.push(\n `\\nAs you answer the user's questions, you can use the following context:\\n`,\n )\n\n // \u8FC7\u6EE4\u6389\u5DF2\u7ECF\u7531 Minto \u4E0A\u4E0B\u6587\u5904\u7406\u7684\u9879\u76EE\u6587\u6863\uFF08\u907F\u514D\u91CD\u590D\uFF09\n const filteredContext = Object.fromEntries(\n Object.entries(context).filter(\n ([key]) => key !== 'projectDocs' && key !== 'userDocs',\n ),\n )\n\n enhancedPrompt.push(\n ...Object.entries(filteredContext).map(\n ([key, value]) => `<context name=\"${key}\">${value}</context>`,\n ),\n )\n }\n\n return { systemPrompt: enhancedPrompt, reminders }\n}\n\nasync function queryLLMWithPromptCaching(\n messages: (UserMessage | AssistantMessage)[],\n systemPrompt: string[],\n maxThinkingTokens: number,\n tools: Tool[],\n signal: AbortSignal,\n options: {\n safeMode: boolean\n model: string\n prependCLISysprompt: boolean\n modelProfile?: ModelProfile | null\n toolUseContext?: ToolUseContext\n },\n): Promise<AssistantMessage> {\n const config = getGlobalConfig()\n const modelManager = getModelManager()\n const toolUseContext = options.toolUseContext\n\n const modelProfile = options.modelProfile || modelManager.getModel('main')\n let provider: string\n\n if (modelProfile) {\n provider = modelProfile.provider || config.primaryProvider || 'anthropic'\n } else {\n provider = config.primaryProvider || 'anthropic'\n }\n\n // Use native Anthropic SDK for Anthropic and some Anthropic-compatible providers\n if (\n provider === 'anthropic' ||\n provider === 'bigdream' ||\n provider === 'opendev'\n ) {\n return queryAnthropicNative(\n messages,\n systemPrompt,\n maxThinkingTokens,\n tools,\n signal,\n { ...options, modelProfile, toolUseContext },\n )\n }\n\n // Use OpenAI-compatible interface for all other providers\n return queryOpenAI(messages, systemPrompt, maxThinkingTokens, tools, signal, {\n ...options,\n modelProfile,\n toolUseContext,\n })\n}\n\nasync function queryAnthropicNative(\n messages: (UserMessage | AssistantMessage)[],\n systemPrompt: string[],\n maxThinkingTokens: number,\n tools: Tool[],\n signal: AbortSignal,\n options?: {\n safeMode: boolean\n model: string\n prependCLISysprompt: boolean\n modelProfile?: ModelProfile | null\n toolUseContext?: ToolUseContext\n },\n): Promise<AssistantMessage> {\n const config = getGlobalConfig()\n const modelManager = getModelManager()\n const toolUseContext = options?.toolUseContext\n\n const modelProfile = options?.modelProfile || modelManager.getModel('main')\n let anthropic: Anthropic | AnthropicBedrock | AnthropicVertex\n let model: string\n let provider: string\n\n // \uD83D\uDD0D Debug: \u8BB0\u5F55\u6A21\u578B\u914D\u7F6E\u8BE6\u60C5\n debugLogger.api('MODEL_CONFIG_ANTHROPIC', {\n modelProfileFound: !!modelProfile,\n modelProfileId: modelProfile?.modelName,\n modelProfileName: modelProfile?.name,\n modelProfileModelName: modelProfile?.modelName,\n modelProfileProvider: modelProfile?.provider,\n modelProfileBaseURL: modelProfile?.baseURL,\n modelProfileApiKeyExists: !!modelProfile?.apiKey,\n optionsModel: options?.model,\n requestId: getCurrentRequest()?.id,\n })\n\n if (modelProfile) {\n // \u4F7F\u7528ModelProfile\u7684\u5B8C\u6574\u914D\u7F6E\n model = modelProfile.modelName\n provider = modelProfile.provider || config.primaryProvider || 'anthropic'\n\n // \u57FA\u4E8EModelProfile\u521B\u5EFA\u4E13\u7528\u7684API\u5BA2\u6237\u7AEF\n if (\n modelProfile.provider === 'anthropic' ||\n modelProfile.provider === 'bigdream' ||\n modelProfile.provider === 'opendev'\n ) {\n const clientConfig: any = {\n apiKey: modelProfile.apiKey,\n dangerouslyAllowBrowser: true,\n maxRetries: 0,\n timeout: parseInt(process.env.API_TIMEOUT_MS || String(60 * 1000), 10),\n defaultHeaders: {\n 'x-app': 'cli',\n 'User-Agent': USER_AGENT,\n },\n }\n\n // \u4F7F\u7528ModelProfile\u7684baseURL\u800C\u4E0D\u662F\u5168\u5C40\u914D\u7F6E\n if (modelProfile.baseURL) {\n clientConfig.baseURL = modelProfile.baseURL\n }\n\n anthropic = new Anthropic(clientConfig)\n } else {\n // \u5176\u4ED6\u63D0\u4F9B\u5546\u7684\u5904\u7406\u903B\u8F91\n anthropic = getAnthropicClient(model)\n }\n } else {\n // \uD83D\uDEA8 \u964D\u7EA7\uFF1A\u6CA1\u6709\u6709\u6548\u7684ModelProfile\u65F6\uFF0C\u5E94\u8BE5\u629B\u51FA\u9519\u8BEF\n const errorDetails = {\n modelProfileExists: !!modelProfile,\n modelProfileModelName: modelProfile?.modelName,\n requestedModel: options?.model,\n requestId: getCurrentRequest()?.id,\n }\n debugLogger.error('ANTHROPIC_FALLBACK_ERROR', errorDetails)\n throw new Error(\n `No valid ModelProfile available for Anthropic provider. Please configure model through /model command. Debug: ${JSON.stringify(errorDetails)}`,\n )\n }\n\n // Prepend system prompt block for easy API identification\n if (options?.prependCLISysprompt) {\n // Log stats about first block for analyzing prefix matching config\n const [firstSyspromptBlock] = splitSysPromptPrefix(systemPrompt)\n\n systemPrompt = [getCLISyspromptPrefix(), ...systemPrompt]\n }\n\n const system: TextBlockParam[] = splitSysPromptPrefix(systemPrompt).map(\n _ => ({\n text: _,\n type: 'text',\n }),\n )\n\n const toolSchemas = await Promise.all(\n tools.map(\n async tool =>\n ({\n name: tool.name,\n // Use getToolDescriptionAsync for consistent async access\n description: await getToolDescriptionAsync(tool),\n input_schema:\n 'inputJSONSchema' in tool && tool.inputJSONSchema\n ? tool.inputJSONSchema\n : zodToJsonSchema(tool.inputSchema),\n }) as unknown as Anthropic.Beta.Messages.BetaTool,\n ),\n )\n\n const anthropicMessages = addCacheBreakpoints(messages)\n\n // apply cache control\n const { systemBlocks: processedSystem, messageParams: processedMessages } =\n applyCacheControlWithLimits(system, anthropicMessages)\n const startIncludingRetries = Date.now()\n\n // \u8BB0\u5F55\u7CFB\u7EDF\u63D0\u793A\u6784\u5EFA\u8FC7\u7A0B\n logSystemPromptConstruction({\n basePrompt: systemPrompt.join('\\n'),\n mintoContext: generateMintoContext() || '',\n reminders: [], // \u8FD9\u91CC\u53EF\u4EE5\u4ECE generateSystemReminders \u83B7\u53D6\n finalPrompt: systemPrompt.join('\\n'),\n })\n\n let start = Date.now()\n let attemptNumber = 0\n let response\n\n try {\n response = await withRetry(\n async attempt => {\n attemptNumber = attempt\n start = Date.now()\n\n const params: Anthropic.Beta.Messages.MessageCreateParams = {\n model,\n max_tokens: getMaxTokensFromProfile(modelProfile),\n messages: processedMessages,\n system: processedSystem,\n tools: toolSchemas.length > 0 ? toolSchemas : undefined,\n tool_choice: toolSchemas.length > 0 ? { type: 'auto' } : undefined,\n }\n\n if (maxThinkingTokens > 0) {\n ;(params as any).extra_headers = {\n 'anthropic-beta': 'max-tokens-3-5-sonnet-2024-07-15',\n }\n ;(params as any).thinking = { max_tokens: maxThinkingTokens }\n }\n\n // \uD83D\uDD25 REAL-TIME API CALL DEBUG - \u4F7F\u7528\u5168\u5C40\u65E5\u5FD7\u7CFB\u7EDF (Anthropic Streaming)\n debugLogger.api('ANTHROPIC_API_CALL_START_STREAMING', {\n endpoint: modelProfile?.baseURL || 'DEFAULT_ANTHROPIC',\n model,\n provider,\n apiKeyConfigured: !!modelProfile?.apiKey,\n apiKeyPrefix: modelProfile?.apiKey\n ? modelProfile.apiKey.substring(0, 8)\n : null,\n maxTokens: params.max_tokens,\n temperature: MAIN_QUERY_TEMPERATURE,\n params: params,\n messageCount: params.messages?.length || 0,\n streamMode: true,\n toolsCount: toolSchemas.length,\n thinkingTokens: maxThinkingTokens,\n timestamp: new Date().toISOString(),\n modelProfileId: modelProfile?.modelName,\n modelProfileName: modelProfile?.name,\n })\n\n if (config.stream) {\n const stream = await anthropic.beta.messages.create(\n {\n ...params,\n stream: true,\n },\n {\n signal: signal, // \u2190 CRITICAL: Connect the AbortSignal to API call\n },\n )\n\n let finalResponse: any | null = null\n let messageStartEvent: any = null\n const contentBlocks: any[] = []\n const inputJSONBuffers = new Map<number, string>()\n let usage: any = null\n let stopReason: string | null = null\n let stopSequence: string | null = null\n\n for await (const event of stream) {\n if (signal.aborted) {\n debugLogger.flow('STREAM_ABORTED', {\n eventType: event.type,\n timestamp: Date.now(),\n })\n throw new Error('Request was cancelled')\n }\n\n switch (event.type) {\n case 'message_start':\n messageStartEvent = event\n finalResponse = {\n ...event.message,\n content: [], // Will be populated from content blocks\n }\n // NEW: Set initial input tokens from message_start and reset receivedChars\n setStreamingState({\n inputTokens: event.message.usage?.input_tokens,\n receivedChars: 0, // Reset for new request\n })\n break\n\n case 'content_block_start':\n contentBlocks[event.index] = { ...event.content_block }\n // Initialize JSON buffer for tool_use blocks\n if (event.content_block.type === 'tool_use') {\n inputJSONBuffers.set(event.index, '')\n }\n // \u2705 P0 FIX: Detect thinking block start and update UI\n if (event.content_block.type === 'thinking') {\n const thinkingStartTime = Date.now()\n setStreamingState({\n phase: 'deep_thinking',\n tokenCount: 0,\n thinkingMaxTokens: maxThinkingTokens,\n thinkingStartTime, // NEW: Record start time for duration tracking\n })\n debugLogger.flow('THINKING_BLOCK_START', {\n index: event.index,\n maxTokens: maxThinkingTokens,\n startTime: thinkingStartTime,\n })\n }\n break\n\n case 'content_block_delta':\n const blockIndex = event.index\n\n // Ensure content block exists\n if (!contentBlocks[blockIndex]) {\n contentBlocks[blockIndex] = {\n type:\n event.delta.type === 'text_delta' ? 'text' : 'tool_use',\n text: event.delta.type === 'text_delta' ? '' : undefined,\n }\n if (event.delta.type === 'input_json_delta') {\n inputJSONBuffers.set(blockIndex, '')\n }\n }\n\n if (event.delta.type === 'text_delta') {\n contentBlocks[blockIndex].text += event.delta.text\n // Update received chars for approximate token display\n const currentState = getStreamingState()\n setStreamingState({\n receivedChars:\n (currentState.receivedChars || 0) +\n event.delta.text.length,\n })\n } else if (event.delta.type === 'input_json_delta') {\n const currentBuffer = inputJSONBuffers.get(blockIndex) || ''\n inputJSONBuffers.set(\n blockIndex,\n currentBuffer + event.delta.partial_json,\n )\n }\n // \u2705 P0 FIX: Handle thinking_delta events to update token count\n else if (event.delta.type === 'thinking_delta') {\n const currentThinking =\n contentBlocks[blockIndex].thinking || ''\n const updatedThinking =\n currentThinking + (event.delta.thinking || '')\n contentBlocks[blockIndex] = {\n ...contentBlocks[blockIndex],\n thinking: updatedThinking,\n }\n\n // Update UI with current thinking token count\n setStreamingState({\n phase: 'deep_thinking',\n tokenCount: updatedThinking.length,\n thinkingMaxTokens: maxThinkingTokens,\n })\n\n debugLogger.trace('THINKING_DELTA', {\n index: blockIndex,\n tokenCount: updatedThinking.length,\n })\n }\n break\n\n case 'message_delta':\n if (event.delta.stop_reason)\n stopReason = event.delta.stop_reason\n if (event.delta.stop_sequence)\n stopSequence = event.delta.stop_sequence\n if (event.usage) {\n usage = { ...usage, ...event.usage }\n // NEW: Update UI with real-time output token count\n // Note: message_delta only provides output_tokens, not input_tokens\n setStreamingState({\n outputTokens: event.usage.output_tokens,\n })\n }\n break\n\n case 'content_block_stop':\n const stopIndex = event.index\n const block = contentBlocks[stopIndex]\n\n // \u2705 P0 FIX: Handle thinking block completion\n if (block?.type === 'thinking') {\n const streamState = getStreamingState()\n const thinkingDurationMs = streamState.thinkingStartTime\n ? Date.now() - streamState.thinkingStartTime\n : 0\n setStreamingState({\n phase: 'generating',\n thinkingDurationMs, // NEW: Store duration for display\n })\n debugLogger.flow('THINKING_BLOCK_COMPLETE', {\n index: stopIndex,\n finalTokenCount: block.thinking?.length || 0,\n durationMs: thinkingDurationMs,\n })\n }\n\n if (\n block?.type === 'tool_use' &&\n inputJSONBuffers.has(stopIndex)\n ) {\n const jsonStr = inputJSONBuffers.get(stopIndex)\n if (jsonStr) {\n try {\n block.input = JSON.parse(jsonStr)\n } catch (error) {\n const errorMsg =\n error instanceof Error ? error.message : String(error)\n debugLogger.error('JSON_PARSE_ERROR', {\n blockIndex: stopIndex,\n jsonStr:\n jsonStr.length > 500\n ? jsonStr.slice(0, 500) + '...'\n : jsonStr,\n error: errorMsg,\n })\n // Preserve parse error info for downstream error handling\n // Instead of empty {}, provide error context that query.ts can detect\n block.input = {\n __parse_error__: true,\n __error_message__: `JSON parse failed: ${errorMsg}`,\n __raw_json_preview__:\n jsonStr.length > 200\n ? jsonStr.slice(0, 200) + '...'\n : jsonStr,\n }\n }\n inputJSONBuffers.delete(stopIndex)\n }\n }\n break\n\n case 'message_stop':\n // Clear any remaining buffers\n inputJSONBuffers.clear()\n break\n }\n\n if (event.type === 'message_stop') {\n break\n }\n }\n\n if (!finalResponse || !messageStartEvent) {\n throw new Error('Stream ended without proper message structure')\n }\n\n // Construct the final response\n finalResponse = {\n ...messageStartEvent.message,\n content: contentBlocks.filter(Boolean),\n stop_reason: stopReason,\n stop_sequence: stopSequence,\n usage: {\n ...messageStartEvent.message.usage,\n ...usage,\n },\n }\n\n return finalResponse\n } else {\n // \uD83D\uDD25 REAL-TIME API CALL DEBUG - \u4F7F\u7528\u5168\u5C40\u65E5\u5FD7\u7CFB\u7EDF (Anthropic Non-Streaming)\n debugLogger.api('ANTHROPIC_API_CALL_START_NON_STREAMING', {\n endpoint: modelProfile?.baseURL || 'DEFAULT_ANTHROPIC',\n model,\n provider,\n apiKeyConfigured: !!modelProfile?.apiKey,\n apiKeyPrefix: modelProfile?.apiKey\n ? modelProfile.apiKey.substring(0, 8)\n : null,\n maxTokens: params.max_tokens,\n temperature: MAIN_QUERY_TEMPERATURE,\n messageCount: params.messages?.length || 0,\n streamMode: false,\n toolsCount: toolSchemas.length,\n thinkingTokens: maxThinkingTokens,\n timestamp: new Date().toISOString(),\n modelProfileId: modelProfile?.modelName,\n modelProfileName: modelProfile?.name,\n })\n\n return await anthropic.beta.messages.create(params, {\n signal: signal, // \u2190 CRITICAL: Connect the AbortSignal to API call\n })\n }\n },\n {\n signal,\n // Pass retry events to agent transcript for UI display\n onRetry: toolUseContext?.agentId\n ? ({ attempt, maxRetries, error, delayMs }) => {\n addRetryEventToTranscript(toolUseContext.agentId!, {\n attempt,\n maxRetries,\n errorMessage: error.message,\n delayMs,\n })\n }\n : undefined,\n },\n )\n\n debugLogger.api('ANTHROPIC_API_CALL_SUCCESS', {\n content: response.content,\n })\n\n const ttftMs = start - Date.now()\n const durationMs = Date.now() - startIncludingRetries\n\n const content = response.content.map((block: ContentBlock) => {\n if (block.type === 'text') {\n return {\n type: 'text' as const,\n text: block.text,\n }\n } else if (block.type === 'tool_use') {\n return {\n type: 'tool_use' as const,\n id: block.id,\n name: block.name,\n input: block.input,\n }\n }\n return block\n })\n\n // Build thinking metadata if thinking blocks are present\n const thinkingBlocks = content.filter(\n (block): block is ContentBlock & { type: 'thinking'; thinking: string } =>\n block.type === 'thinking',\n )\n const streamState = getStreamingState()\n const thinkingMetadata: ThinkingMetadata | undefined =\n thinkingBlocks.length > 0\n ? {\n charCount: thinkingBlocks.reduce(\n (sum, b) => sum + (b.thinking?.length || 0),\n 0,\n ),\n durationMs: streamState.thinkingDurationMs || 0,\n startTime: streamState.thinkingStartTime || Date.now(),\n isComplete: true,\n }\n : undefined\n\n const assistantMessage: AssistantMessage = {\n message: {\n id: response.id,\n content,\n model: response.model,\n role: 'assistant',\n stop_reason: response.stop_reason,\n stop_sequence: response.stop_sequence,\n type: 'message',\n usage: response.usage,\n },\n type: 'assistant',\n uuid: nanoid() as UUID,\n durationMs,\n costUSD: 0, // Will be calculated below\n thinkingMetadata,\n }\n\n // \u8BB0\u5F55\u5B8C\u6574\u7684 LLM \u4EA4\u4E92\u8C03\u8BD5\u4FE1\u606F (Anthropic path)\n // \u6CE8\u610F\uFF1AAnthropic API\u5C06system prompt\u548Cmessages\u5206\u5F00\uFF0C\u8FD9\u91CC\u91CD\u6784\u4E3A\u5B8C\u6574\u7684API\u8C03\u7528\u89C6\u56FE\n const systemMessages = system.map(block => ({\n role: 'system',\n content: block.text,\n }))\n\n logLLMInteraction({\n systemPrompt: systemPrompt.join('\\n'),\n messages: [...systemMessages, ...anthropicMessages],\n response: response,\n usage: response.usage\n ? {\n inputTokens: response.usage.input_tokens,\n outputTokens: response.usage.output_tokens,\n }\n : undefined,\n timing: {\n start: start,\n end: Date.now(),\n },\n apiFormat: 'anthropic',\n })\n\n // Calculate cost using native Anthropic usage data\n const inputTokens = response.usage.input_tokens\n const outputTokens = response.usage.output_tokens\n const cacheCreationInputTokens =\n response.usage.cache_creation_input_tokens ?? 0\n const cacheReadInputTokens = response.usage.cache_read_input_tokens ?? 0\n\n const costUSD =\n (inputTokens / 1_000_000) * getModelInputTokenCostUSD(model) +\n (outputTokens / 1_000_000) * getModelOutputTokenCostUSD(model) +\n (cacheCreationInputTokens / 1_000_000) *\n getModelInputTokenCostUSD(model) +\n (cacheReadInputTokens / 1_000_000) *\n (getModelInputTokenCostUSD(model) * 0.1) // Cache reads are 10% of input cost\n\n assistantMessage.costUSD = costUSD\n addToTotalCost(costUSD, durationMs)\n\n // Record token usage to unified stats manager\n recordTokenUsage(\n {\n inputTokens,\n outputTokens,\n cacheCreationTokens: cacheCreationInputTokens,\n cacheReadTokens: cacheReadInputTokens,\n },\n costUSD,\n model,\n toolUseContext\n ? ({\n agentId: toolUseContext.agentId,\n toolUseId: toolUseContext.toolUseId,\n model,\n } as TokenTrackingContext)\n : undefined,\n )\n\n return assistantMessage\n } catch (error) {\n return getAssistantMessageFromError(error)\n }\n}\n\nfunction getAssistantMessageFromError(error: unknown): AssistantMessage {\n if (error instanceof Error && error.message.includes('prompt is too long')) {\n return createAssistantAPIErrorMessage(PROMPT_TOO_LONG_ERROR_MESSAGE)\n }\n if (\n error instanceof Error &&\n error.message.includes('Your credit balance is too low')\n ) {\n return createAssistantAPIErrorMessage(CREDIT_BALANCE_TOO_LOW_ERROR_MESSAGE)\n }\n if (\n error instanceof Error &&\n error.message.toLowerCase().includes('x-api-key')\n ) {\n return createAssistantAPIErrorMessage(INVALID_API_KEY_ERROR_MESSAGE)\n }\n if (error instanceof Error) {\n if (process.env.NODE_ENV === 'development') {\n debugLogger.error('ANTHROPIC_API_ERROR', {\n message: error.message,\n stack: error.stack,\n })\n }\n return createAssistantAPIErrorMessage(\n `${API_ERROR_MESSAGE_PREFIX}: ${error.message}`,\n )\n }\n return createAssistantAPIErrorMessage(API_ERROR_MESSAGE_PREFIX)\n}\n\nfunction addCacheBreakpoints(\n messages: (UserMessage | AssistantMessage)[],\n): MessageParam[] {\n return messages.map((msg, index) => {\n return msg.type === 'user'\n ? userMessageToMessageParam(msg, index > messages.length - 3)\n : assistantMessageToMessageParam(msg, index > messages.length - 3)\n })\n}\n\nasync function queryOpenAI(\n messages: (UserMessage | AssistantMessage)[],\n systemPrompt: string[],\n maxThinkingTokens: number,\n tools: Tool[],\n signal: AbortSignal,\n options?: {\n safeMode: boolean\n model: string\n prependCLISysprompt: boolean\n modelProfile?: ModelProfile | null\n toolUseContext?: ToolUseContext\n },\n): Promise<AssistantMessage> {\n const config = getGlobalConfig()\n const modelManager = getModelManager()\n const toolUseContext = options?.toolUseContext\n\n const modelProfile = options?.modelProfile || modelManager.getModel('main')\n let model: string\n\n // \uD83D\uDD0D Debug: \u8BB0\u5F55\u6A21\u578B\u914D\u7F6E\u8BE6\u60C5\n const currentRequest = getCurrentRequest()\n debugLogger.api('MODEL_CONFIG_OPENAI', {\n modelProfileFound: !!modelProfile,\n modelProfileId: modelProfile?.modelName,\n modelProfileName: modelProfile?.name,\n modelProfileModelName: modelProfile?.modelName,\n modelProfileProvider: modelProfile?.provider,\n modelProfileBaseURL: modelProfile?.baseURL,\n modelProfileApiKeyExists: !!modelProfile?.apiKey,\n optionsModel: options?.model,\n requestId: getCurrentRequest()?.id,\n })\n\n if (modelProfile) {\n model = modelProfile.modelName\n } else {\n model = options?.model || modelProfile?.modelName || ''\n }\n // Prepend system prompt block for easy API identification\n if (options?.prependCLISysprompt) {\n const [firstSyspromptBlock] = splitSysPromptPrefix(systemPrompt)\n\n systemPrompt = [getCLISyspromptPrefix() + systemPrompt] // some openai-like providers need the entire system prompt as a single block\n }\n\n const system: TextBlockParam[] = splitSysPromptPrefix(systemPrompt).map(\n _ => ({\n ...(PROMPT_CACHING_ENABLED\n ? { cache_control: { type: 'ephemeral' } }\n : {}),\n text: _,\n type: 'text',\n }),\n )\n\n const toolSchemas = await Promise.all(\n tools.map(\n async _ =>\n ({\n type: 'function',\n function: {\n name: _.name,\n description: await _.prompt({\n safeMode: options?.safeMode,\n }),\n // Use tool's JSON schema directly if provided, otherwise convert Zod schema\n parameters:\n 'inputJSONSchema' in _ && _.inputJSONSchema\n ? _.inputJSONSchema\n : zodToJsonSchema(_.inputSchema),\n },\n }) as OpenAI.ChatCompletionTool,\n ),\n )\n\n const openaiSystem = system.map(\n s =>\n ({\n role: 'system',\n content: s.text,\n }) as OpenAI.ChatCompletionMessageParam,\n )\n\n const openaiMessages = convertAnthropicMessagesToOpenAIMessages(messages)\n const startIncludingRetries = Date.now()\n\n // \u8BB0\u5F55\u7CFB\u7EDF\u63D0\u793A\u6784\u5EFA\u8FC7\u7A0B (OpenAI path)\n logSystemPromptConstruction({\n basePrompt: systemPrompt.join('\\n'),\n mintoContext: generateMintoContext() || '',\n reminders: [], // \u8FD9\u91CC\u53EF\u4EE5\u4ECE generateSystemReminders \u83B7\u53D6\n finalPrompt: systemPrompt.join('\\n'),\n })\n\n let start = Date.now()\n let attemptNumber = 0\n let response\n\n try {\n response = await withRetry(\n async attempt => {\n attemptNumber = attempt\n start = Date.now()\n // \uD83D\uDD25 GPT-5 Enhanced Parameter Construction\n const maxTokens = getMaxTokensFromProfile(modelProfile)\n const isGPT5 = isGPT5Model(model)\n\n const opts: OpenAI.ChatCompletionCreateParams = {\n model,\n\n ...(isGPT5\n ? { max_completion_tokens: maxTokens }\n : { max_tokens: maxTokens }),\n messages: [...openaiSystem, ...openaiMessages],\n\n temperature: isGPT5 ? 1 : MAIN_QUERY_TEMPERATURE,\n }\n if (config.stream) {\n ;(opts as OpenAI.ChatCompletionCreateParams).stream = true\n opts.stream_options = {\n include_usage: true,\n }\n }\n\n if (toolSchemas.length > 0) {\n opts.tools = toolSchemas\n opts.tool_choice = 'auto'\n }\n const reasoningEffort = await getReasoningEffort(modelProfile, messages)\n if (reasoningEffort) {\n opts.reasoning_effort = reasoningEffort\n }\n\n if (modelProfile && modelProfile.modelName) {\n debugLogger.api('USING_MODEL_PROFILE_PATH', {\n modelProfileName: modelProfile.modelName,\n modelName: modelProfile.modelName,\n provider: modelProfile.provider,\n baseURL: modelProfile.baseURL,\n apiKeyExists: !!modelProfile.apiKey,\n requestId: getCurrentRequest()?.id,\n })\n\n // Enable new adapter system with environment variable\n const USE_NEW_ADAPTER_SYSTEM =\n process.env.USE_NEW_ADAPTERS !== 'false'\n\n if (USE_NEW_ADAPTER_SYSTEM) {\n // New adapter system\n const adapter = ModelAdapterFactory.createAdapter(modelProfile)\n\n // Build unified request parameters\n const unifiedParams: UnifiedRequestParams = {\n messages: openaiMessages,\n systemPrompt: openaiSystem.map(s => s.content as string),\n tools: tools,\n maxTokens: getMaxTokensFromProfile(modelProfile),\n stream: config.stream,\n reasoningEffort: reasoningEffort as any,\n temperature: isGPT5Model(model) ? 1 : MAIN_QUERY_TEMPERATURE,\n previousResponseId:\n toolUseContext?.responseState?.previousResponseId,\n verbosity: 'high', // High verbosity for coding tasks\n }\n\n // Create request using adapter\n const request = adapter.createRequest(unifiedParams)\n\n // Determine which API to use\n if (ModelAdapterFactory.shouldUseResponsesAPI(modelProfile)) {\n // Use Responses API for GPT-5 and similar models\n const { callGPT5ResponsesAPI } = await import('./openai')\n const response = await callGPT5ResponsesAPI(\n modelProfile,\n request,\n signal,\n )\n const unifiedResponse = adapter.parseResponse(response)\n\n // Convert unified response back to Anthropic format\n const apiMessage = {\n role: 'assistant' as const,\n content: unifiedResponse.content,\n tool_calls: unifiedResponse.toolCalls,\n usage: {\n prompt_tokens: unifiedResponse.usage.promptTokens,\n completion_tokens: unifiedResponse.usage.completionTokens,\n },\n }\n const assistantMsg: AssistantMessage = {\n type: 'assistant',\n message: apiMessage as any,\n costUSD: 0, // Will be calculated later\n durationMs: Date.now() - start,\n uuid: `${Date.now()}-${Math.random().toString(36).substr(2, 9)}` as any,\n responseId: unifiedResponse.responseId, // For state management\n }\n return assistantMsg\n } else {\n // Use existing Chat Completions flow\n const s = await getCompletionWithProfile(\n modelProfile,\n request,\n 0,\n 10,\n signal,\n )\n let finalResponse\n if (config.stream) {\n finalResponse = await handleMessageStream(\n s as ChatCompletionStream,\n signal,\n )\n } else {\n finalResponse = s\n }\n const r = convertOpenAIResponseToAnthropic(finalResponse, tools)\n return r\n }\n } else {\n // Legacy system (preserved for fallback)\n const completionFunction = isGPT5Model(modelProfile.modelName)\n ? getGPT5CompletionWithProfile\n : getCompletionWithProfile\n const s = await completionFunction(\n modelProfile,\n opts,\n 0,\n 10,\n signal,\n )\n let finalResponse\n if (opts.stream) {\n finalResponse = await handleMessageStream(\n s as ChatCompletionStream,\n signal,\n )\n } else {\n finalResponse = s\n }\n const r = convertOpenAIResponseToAnthropic(finalResponse, tools)\n return r\n }\n } else {\n // \uD83D\uDEA8 \u8B66\u544A\uFF1AModelProfile\u4E0D\u53EF\u7528\uFF0C\u4F7F\u7528\u65E7\u903B\u8F91\u8DEF\u5F84\n debugLogger.api('USING_LEGACY_PATH', {\n modelProfileExists: !!modelProfile,\n modelProfileId: modelProfile?.modelName,\n modelNameExists: !!modelProfile?.modelName,\n fallbackModel: 'main',\n actualModel: model,\n requestId: getCurrentRequest()?.id,\n })\n\n // \uD83D\uDEA8 FALLBACK: \u6CA1\u6709\u6709\u6548\u7684ModelProfile\u65F6\uFF0C\u5E94\u8BE5\u629B\u51FA\u9519\u8BEF\u800C\u4E0D\u662F\u4F7F\u7528\u9057\u7559\u7CFB\u7EDF\n const errorDetails = {\n modelProfileExists: !!modelProfile,\n modelProfileId: modelProfile?.modelName,\n modelNameExists: !!modelProfile?.modelName,\n requestedModel: model,\n requestId: getCurrentRequest()?.id,\n }\n debugLogger.error('NO_VALID_MODEL_PROFILE', errorDetails)\n throw new Error(\n `No valid ModelProfile available for model: ${model}. Please configure model through /model command. Debug: ${JSON.stringify(errorDetails)}`,\n )\n }\n },\n {\n signal,\n // Pass retry events to agent transcript for UI display\n onRetry: toolUseContext?.agentId\n ? ({ attempt, maxRetries, error, delayMs }) => {\n addRetryEventToTranscript(toolUseContext.agentId!, {\n attempt,\n maxRetries,\n errorMessage: error.message,\n delayMs,\n })\n }\n : undefined,\n },\n )\n } catch (error) {\n logError(error)\n return getAssistantMessageFromError(error)\n }\n const durationMs = Date.now() - start\n const durationMsIncludingRetries = Date.now() - startIncludingRetries\n\n const inputTokens = response.usage?.prompt_tokens ?? 0\n const outputTokens = response.usage?.completion_tokens ?? 0\n const cacheReadInputTokens =\n response.usage?.prompt_token_details?.cached_tokens ?? 0\n const cacheCreationInputTokens =\n response.usage?.prompt_token_details?.cached_tokens ?? 0\n const costUSD =\n (inputTokens / 1_000_000) * SONNET_COST_PER_MILLION_INPUT_TOKENS +\n (outputTokens / 1_000_000) * SONNET_COST_PER_MILLION_OUTPUT_TOKENS +\n (cacheReadInputTokens / 1_000_000) *\n SONNET_COST_PER_MILLION_PROMPT_CACHE_READ_TOKENS +\n (cacheCreationInputTokens / 1_000_000) *\n SONNET_COST_PER_MILLION_PROMPT_CACHE_WRITE_TOKENS\n\n addToTotalCost(costUSD, durationMsIncludingRetries)\n\n // Record token usage to unified stats manager\n recordTokenUsage(\n {\n inputTokens,\n outputTokens,\n cacheCreationTokens: cacheCreationInputTokens,\n cacheReadTokens: cacheReadInputTokens,\n },\n costUSD,\n model,\n toolUseContext\n ? ({\n agentId: toolUseContext.agentId,\n toolUseId: toolUseContext.toolUseId,\n model,\n } as TokenTrackingContext)\n : undefined,\n )\n\n // \u8BB0\u5F55\u5B8C\u6574\u7684 LLM \u4EA4\u4E92\u8C03\u8BD5\u4FE1\u606F (OpenAI path)\n logLLMInteraction({\n systemPrompt: systemPrompt.join('\\n'),\n messages: [...openaiSystem, ...openaiMessages],\n response: response,\n usage: {\n inputTokens: inputTokens,\n outputTokens: outputTokens,\n },\n timing: {\n start: start,\n end: Date.now(),\n },\n apiFormat: 'openai',\n })\n\n return {\n message: {\n ...response,\n content: normalizeContentFromAPI(response.content),\n usage: {\n input_tokens: inputTokens,\n output_tokens: outputTokens,\n cache_read_input_tokens: cacheReadInputTokens,\n cache_creation_input_tokens: 0,\n },\n },\n costUSD,\n durationMs,\n type: 'assistant',\n uuid: randomUUID(),\n }\n}\n\nfunction getMaxTokensFromProfile(modelProfile: any): number {\n // Use ModelProfile maxTokens or reasonable default\n return modelProfile?.maxTokens || 8000\n}\n\nfunction getModelInputTokenCostUSD(model: string): number {\n // Find the model in the models object\n for (const providerModels of Object.values(models)) {\n const modelInfo = providerModels.find((m: any) => m.model === model)\n if (modelInfo) {\n return modelInfo.input_cost_per_token || 0\n }\n }\n // Default fallback cost for unknown models\n return 0.000003 // Default to Claude 3 Haiku cost\n}\n\nfunction getModelOutputTokenCostUSD(model: string): number {\n // Find the model in the models object\n for (const providerModels of Object.values(models)) {\n const modelInfo = providerModels.find((m: any) => m.model === model)\n if (modelInfo) {\n return modelInfo.output_cost_per_token || 0\n }\n }\n // Default fallback cost for unknown models\n return 0.000015 // Default to Claude 3 Haiku cost\n}\n\n// New unified query functions for model pointer system\nexport async function queryModel(\n modelPointer: import('@utils/config').ModelPointerType,\n messages: (UserMessage | AssistantMessage)[],\n systemPrompt: string[] = [],\n signal?: AbortSignal,\n): Promise<AssistantMessage> {\n // Use queryLLM with the pointer directly\n return queryLLM(\n messages,\n systemPrompt,\n 0, // maxThinkingTokens\n [], // tools\n signal || new AbortController().signal,\n {\n safeMode: false,\n model: modelPointer,\n prependCLISysprompt: true,\n },\n )\n}\n\n// Note: Use queryModel(pointer, ...) directly instead of these convenience functions\n\n// Simplified query function using quick model pointer\nexport async function queryQuick({\n systemPrompt = [],\n userPrompt,\n assistantPrompt,\n enablePromptCaching = false,\n signal,\n}: {\n systemPrompt?: string[]\n userPrompt: string\n assistantPrompt?: string\n enablePromptCaching?: boolean\n signal?: AbortSignal\n}): Promise<AssistantMessage> {\n const messages = [\n {\n message: { role: 'user', content: userPrompt },\n type: 'user',\n uuid: randomUUID(),\n },\n ] as (UserMessage | AssistantMessage)[]\n\n return queryModel('quick', messages, systemPrompt, signal)\n}\n"],
5
- "mappings": "AAAA,OAAO;AACP,OAAO,aAAa,oBAAoB,gBAAgB;AACxD,SAAS,wBAAwB;AACjC,SAAS,uBAAuB;AAChC,OAAO,WAAW;AAClB,SAAS,kBAAwB;AACjC,OAAO;AAEP,SAAS,sBAAsB;AAC/B,SAAS,wBAAwB;AAEjC,OAAO,YAAY;AAEnB,SAAe,+BAA+B;AAC9C;AAAA,EACE;AAAA,EACA;AAAA,EACA;AAAA,OAEK;AACP,SAAS,sBAAsB;AAC/B,SAAS,UAAU,kBAAkB;AACrC,SAAS,sBAAsB;AAC/B,SAAS,kBAAkB;AAC3B;AAAA,EACE;AAAA,EACA;AAAA,OACK;AACP,SAAS,eAAe;AACxB;AAAA,EACE,SAAS;AAAA,EACT;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,OACK;AACP,SAAS,uBAAuB;AAChC,SAAS,uBAAuB;AAChC,SAAS,2BAA2B;AAEpC,SAAS,sBAAsB,yBAAyB;AAMxD,SAAS,aAAa,kBAAkB;AACxC,SAAS,6BAA6B;AACtC,SAAS,+BAA+B;AAIxC,SAAS,cAAc;AACvB;AAAA,EACE;AAAA,EACA;AAAA,OACK;AACP,SAAS,0BAA0B;AACnC,SAAS,+BAA+B;AACxC,SAAS,iCAAiC;AAC1C,SAAS,mBAAmB,yBAAyB;AAIrD,SAAS,YAAY,WAA4B;AAC/C,SAAO,UAAU,WAAW,OAAO;AACrC;AAGA,MAAM,oBAAoB;AAAA,EACxB,OAAe;AAAA,EACP,mBAA2B;AAAA,EAC3B,mBAA4B;AAAA,EAC5B,cAAoC;AAAA,EAEpC,cAAc;AAAA,EAAC;AAAA,EAEvB,OAAc,cAAmC;AAC/C,QAAI,CAAC,oBAAoB,UAAU;AACjC,0BAAoB,WAAW,IAAI,oBAAoB;AAAA,IACzD;AACA,WAAO,oBAAoB;AAAA,EAC7B;AAAA,EAEA,MAAa,aAA4B;AACvC,QAAI,KAAK,iBAAkB;AAE3B,QAAI,KAAK,aAAa;AACpB,aAAO,KAAK;AAAA,IACd;AAEA,SAAK,cAAc,KAAK,gBAAgB;AACxC,UAAM,KAAK;AAAA,EACb;AAAA,EAEA,MAAc,kBAAiC;AAC7C,QAAI;AACF,YAAM,cAAc,MAAM,eAAe;AACzC,WAAK,mBAAmB,eAAe;AACvC,WAAK,mBAAmB;AAGxB,UAAI,QAAQ,IAAI,aAAa,eAAe;AAC1C,oBAAY,KAAK,wBAAwB;AAAA,UACvC,YAAY,KAAK,iBAAiB;AAAA,QACpC,CAAC;AAAA,MACH;AAAA,IACF,SAAS,OAAO;AACd,kBAAY,KAAK,6BAA6B,EAAE,OAAO,OAAO,KAAK,EAAE,CAAC;AACtE,WAAK,mBAAmB;AACxB,WAAK,mBAAmB;AAAA,IAC1B;AAAA,EACF;AAAA,EAEO,kBAA0B;AAC/B,QAAI,CAAC,KAAK,kBAAkB;AAE1B,WAAK,WAAW,EAAE,MAAM,WAAS;AAC/B,oBAAY,KAAK,6BAA6B,EAAE,OAAO,OAAO,KAAK,EAAE,CAAC;AAAA,MACxE,CAAC;AACD,aAAO;AAAA,IACT;AACA,WAAO,KAAK;AAAA,EACd;AAAA,EAEA,MAAa,eAA8B;AACzC,SAAK,mBAAmB;AACxB,SAAK,cAAc;AACnB,UAAM,KAAK,WAAW;AAAA,EACxB;AACF;AAGA,MAAM,sBAAsB,oBAAoB,YAAY;AAG5D,oBAAoB,WAAW,EAAE,MAAM,WAAS;AAC9C,cAAY,KAAK,oCAAoC,EAAE,OAAO,OAAO,KAAK,EAAE,CAAC;AAC/E,CAAC;AAEM,MAAM,uBAAuB,MAAc;AAChD,SAAO,oBAAoB,gBAAgB;AAC7C;AAEO,MAAM,sBAAsB,YAA2B;AAC5D,QAAM,oBAAoB,aAAa;AACzC;AAEO,MAAM,2BAA2B;AACjC,MAAM,gCAAgC;AACtC,MAAM,uCAAuC;AAC7C,MAAM,gCACX;AACK,MAAM,qBAAqB;AAClC,MAAM,yBAAyB,CAAC,QAAQ,IAAI;AAG5C,MAAM,uCAAuC;AAC7C,MAAM,wCAAwC;AAC9C,MAAM,oDAAoD;AAC1D,MAAM,mDAAmD;AAElD,MAAM,yBAAyB;AAEtC,SAAS,cAAc;AACrB,SAAO;AAAA,IACL,SAAS,GAAG,kBAAkB,CAAC,IAAI,UAAU;AAAA,EAC/C;AACF;AAEA,MAAM,cAAc,QAAQ,IAAI,cAAc,cAAc,MAAM;AAClE,MAAM,gBAAgB;AActB,SAAS,cACP,SACA,kBACQ;AACR,MAAI,kBAAkB;AACpB,UAAM,UAAU,SAAS,kBAAkB,EAAE;AAC7C,QAAI,CAAC,MAAM,OAAO,GAAG;AACnB,aAAO,UAAU;AAAA,IACnB;AAAA,EACF;AACA,SAAO,KAAK,IAAI,gBAAgB,KAAK,IAAI,GAAG,UAAU,CAAC,GAAG,IAAK;AACjE;AAEA,SAAS,YAAY,OAA0B;AAE7C,MAAI,MAAM,SAAS,SAAS,2BAA2B,GAAG;AACxD,WAAO,QAAQ,IAAI,cAAc;AAAA,EACnC;AAGA,QAAM,oBAAoB,MAAM,UAAU,gBAAgB;AAG1D,MAAI,sBAAsB,OAAQ,QAAO;AACzC,MAAI,sBAAsB,QAAS,QAAO;AAE1C,MAAI,iBAAiB,oBAAoB;AACvC,WAAO;AAAA,EACT;AAEA,MAAI,CAAC,MAAM,OAAQ,QAAO;AAG1B,MAAI,MAAM,WAAW,IAAK,QAAO;AAGjC,MAAI,MAAM,WAAW,IAAK,QAAO;AAGjC,MAAI,MAAM,WAAW,IAAK,QAAO;AAGjC,MAAI,MAAM,UAAU,MAAM,UAAU,IAAK,QAAO;AAEhD,SAAO;AACT;AAEA,eAAe,UACb,WACA,UAAwB,CAAC,GACb;AACZ,QAAM,aAAa,QAAQ,cAAc;AACzC,MAAI;AAEJ,WAAS,UAAU,GAAG,WAAW,aAAa,GAAG,WAAW;AAC1D,QAAI;AACF,aAAO,MAAM,UAAU,OAAO;AAAA,IAChC,SAAS,OAAO;AACd,kBAAY;AAEZ,UACE,UAAU,cACV,EAAE,iBAAiB,aACnB,CAAC,YAAY,KAAK,GAClB;AACA,cAAM;AAAA,MACR;AAEA,UAAI,QAAQ,QAAQ,SAAS;AAC3B,cAAM,IAAI,MAAM,2BAA2B;AAAA,MAC7C;AAGA,YAAM,aAAa,MAAM,UAAU,aAAa,KAAK;AACrD,YAAM,UAAU,cAAc,SAAS,UAAU;AAGjD,wBAAkB;AAAA,QAChB,OAAO;AAAA,QACP,YAAY;AAAA,QACZ;AAAA,QACA,WAAW,MAAM;AAAA,MACnB,CAAC;AAED,kBAAY,KAAK,qBAAqB;AAAA,QACpC;AAAA,QACA;AAAA,QACA,WAAW,MAAM;AAAA,QACjB,cAAc,MAAM;AAAA,QACpB;AAAA,MACF,CAAC;AAGD,UAAI,QAAQ,SAAS;AACnB,gBAAQ,QAAQ;AAAA,UACd;AAAA,UACA;AAAA,UACA;AAAA,UACA;AAAA,QACF,CAAC;AAAA,MACH,OAAO;AACL,gBAAQ;AAAA,UACN,aAAQ,MAAM,IAAI,OAAO,MAAM,IAAI,KAAK,MAAM,OAAO,sBAAmB,KAAK,MAAM,UAAU,GAAI,CAAC,2BAAsB,OAAO,IAAI,UAAU,GAAG,CAAC;AAAA,QACnJ;AAAA,MACF;AAEA,UAAI;AACF,cAAM,eAAe,SAAS,QAAQ,MAAM;AAAA,MAC9C,SAAS,YAAY;AAEnB,YAAI,WAAW,YAAY,uBAAuB;AAChD,gBAAM,IAAI,MAAM,2BAA2B;AAAA,QAC7C;AACA,cAAM;AAAA,MACR;AAAA,IACF;AAAA,EACF;AAEA,QAAM;AACR;AAKA,eAAsB,qBACpB,SACA,QACgB;AAChB,MAAI;AAEF,UAAM,YAAY,UACd,GAAG,QAAQ,QAAQ,QAAQ,EAAE,CAAC,eAC9B;AAEJ,UAAM,WAAW,MAAM,MAAM,WAAW;AAAA,MACtC,QAAQ;AAAA,MACR,SAAS;AAAA,QACP,aAAa;AAAA,QACb,qBAAqB;AAAA,QACrB,cAAc;AAAA,MAChB;AAAA,IACF,CAAC;AAED,QAAI,CAAC,SAAS,IAAI;AAEhB,UAAI,SAAS,WAAW,KAAK;AAC3B,cAAM,IAAI;AAAA,UACR;AAAA,QACF;AAAA,MACF,WAAW,SAAS,WAAW,KAAK;AAClC,cAAM,IAAI;AAAA,UACR;AAAA,QACF;AAAA,MACF,WAAW,SAAS,WAAW,KAAK;AAClC,cAAM,IAAI;AAAA,UACR;AAAA,QACF;AAAA,MACF,WAAW,SAAS,UAAU,KAAK;AACjC,cAAM,IAAI;AAAA,UACR;AAAA,QACF;AAAA,MACF,OAAO;AACL,cAAM,IAAI;AAAA,UACR,uCAAuC,SAAS,MAAM;AAAA,QACxD;AAAA,MACF;AAAA,IACF;AAEA,UAAM,OAAO,MAAM,SAAS,KAAK;AACjC,WAAO,KAAK,QAAQ,CAAC;AAAA,EACvB,SAAS,OAAO;AAEd,QACG,iBAAiB,SAAS,MAAM,QAAQ,SAAS,SAAS,KAC1D,iBAAiB,SAAS,MAAM,QAAQ,SAAS,WAAW,GAC7D;AACA,YAAM;AAAA,IACR;AAGA,aAAS,KAAK;AACd,UAAM,IAAI;AAAA,MACR;AAAA,IACF;AAAA,EACF;AACF;AAEA,eAAsB,aACpB,QACA,SACA,UACkB;AAClB,MAAI,CAAC,QAAQ;AACX,WAAO;AAAA,EACT;AAGA,MAAI,YAAY,aAAa,aAAa;AACxC,QAAI;AACF,YAAM,UAAkC;AAAA,QACtC,eAAe,UAAU,MAAM;AAAA,QAC/B,gBAAgB;AAAA,MAClB;AAEA,UAAI,CAAC,SAAS;AACZ,oBAAY,KAAK,cAAc;AAAA,UAC7B,OAAO;AAAA,QACT,CAAC;AACD,eAAO;AAAA,MACT;AAEA,YAAM,YAAY,GAAG,QAAQ,QAAQ,QAAQ,EAAE,CAAC;AAEhD,YAAM,WAAW,MAAM,MAAM,WAAW;AAAA,QACtC,QAAQ;AAAA,QACR;AAAA,MACF,CAAC;AAED,aAAO,SAAS;AAAA,IAClB,SAAS,OAAO;AACd,kBAAY,KAAK,cAAc;AAAA,QAC7B,OAAO;AAAA,QACP,SAAS,OAAO,KAAK;AAAA,MACvB,CAAC;AACD,aAAO;AAAA,IACT;AAAA,EACF;AAGA,QAAM,eAAoB;AAAA,IACxB;AAAA,IACA,yBAAyB;AAAA,IACzB,YAAY;AAAA,IACZ,gBAAgB;AAAA,MACd,cAAc;AAAA,IAChB;AAAA,EACF;AAGA,MACE,YACC,aAAa,eACZ,aAAa,cACb,aAAa,YACf;AACA,iBAAa,UAAU;AAAA,EACzB;AAEA,QAAM,YAAY,IAAI,UAAU,YAAY;AAE5C,MAAI;AACF,UAAM;AAAA,MACJ,YAAY;AACV,cAAM,QAAQ;AACd,cAAM,WAA2B,CAAC,EAAE,MAAM,QAAQ,SAAS,OAAO,CAAC;AACnE,cAAM,UAAU,SAAS,OAAO;AAAA,UAC9B;AAAA,UACA,YAAY;AAAA;AAAA,UACZ;AAAA,UACA,aAAa;AAAA,UACb,UAAU,YAAY;AAAA,QACxB,CAAC;AACD,eAAO;AAAA,MACT;AAAA,MACA,EAAE,YAAY,EAAE;AAAA;AAAA,IAClB;AACA,WAAO;AAAA,EACT,SAAS,OAAO;AACd,aAAS,KAAK;AAEd,QACE,iBAAiB,SACjB,MAAM,QAAQ;AAAA,MACZ;AAAA,IACF,GACA;AACA,aAAO;AAAA,IACT;AACA,UAAM;AAAA,EACR;AACF;AAEA,SAAS,yCACP,UAIE;AACF,QAAM,iBAGA,CAAC;AAEP,QAAM,cAAqE,CAAC;AAE5E,aAAW,WAAW,UAAU;AAC9B,QAAI,gBAAgB,CAAC;AACrB,QAAI,OAAO,QAAQ,QAAQ,YAAY,UAAU;AAC/C,sBAAgB;AAAA,QACd;AAAA,UACE,MAAM;AAAA,UACN,MAAM,QAAQ,QAAQ;AAAA,QACxB;AAAA,MACF;AAAA,IACF,WAAW,CAAC,MAAM,QAAQ,QAAQ,QAAQ,OAAO,GAAG;AAClD,sBAAgB,CAAC,QAAQ,QAAQ,OAAO;AAAA,IAC1C,OAAO;AACL,sBAAgB,QAAQ,QAAQ;AAAA,IAClC;AAEA,eAAW,SAAS,eAAe;AACjC,UAAI,MAAM,SAAS,QAAQ;AACzB,uBAAe,KAAK;AAAA,UAClB,MAAM,QAAQ,QAAQ;AAAA,UACtB,SAAS,MAAM;AAAA,QACjB,CAAC;AAAA,MACH,WAAW,MAAM,SAAS,YAAY;AACpC,uBAAe,KAAK;AAAA,UAClB,MAAM;AAAA,UACN,SAAS;AAAA,UACT,YAAY;AAAA,YACV;AAAA,cACE,MAAM;AAAA,cACN,UAAU;AAAA,gBACR,MAAM,MAAM;AAAA,gBACZ,WAAW,KAAK,UAAU,MAAM,KAAK;AAAA,cACvC;AAAA,cACA,IAAI,MAAM;AAAA,YACZ;AAAA,UACF;AAAA,QACF,CAAC;AAAA,MACH,WAAW,MAAM,SAAS,eAAe;AAEvC,YAAI,cAAc,MAAM;AACxB,YAAI,OAAO,gBAAgB,UAAU;AAEnC,wBAAc,KAAK,UAAU,WAAW;AAAA,QAC1C;AAEA,oBAAY,MAAM,WAAW,IAAI;AAAA,UAC/B,MAAM;AAAA,UACN,SAAS;AAAA,UACT,cAAc,MAAM;AAAA,QACtB;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAEA,QAAM,gBAGA,CAAC;AAEP,aAAW,WAAW,gBAAgB;AACpC,kBAAc,KAAK,OAAO;AAE1B,QAAI,gBAAgB,WAAW,QAAQ,YAAY;AACjD,iBAAW,YAAY,QAAQ,YAAY;AACzC,YAAI,YAAY,SAAS,EAAE,GAAG;AAC5B,wBAAc,KAAK,YAAY,SAAS,EAAE,CAAC;AAAA,QAC7C;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,eACP,UACA,MAC8B;AAC9B,QAAM,SAAS,CAAC,KAAU,UAAmD;AAC3E,UAAM,EAAE,GAAG,IAAI;AACf,eAAW,CAAC,KAAK,KAAK,KAAK,OAAO,QAAQ,KAAK,GAAG;AAChD,UAAI,IAAI,GAAG,MAAM,UAAa,IAAI,GAAG,MAAM,MAAM;AAC/C,YAAI,GAAG,IAAI;AAEX,YAAI,MAAM,QAAQ,IAAI,GAAG,CAAC,GAAG;AAC3B,qBAAW,OAAO,IAAI,GAAG,GAAG;AAC1B,mBAAO,IAAI;AAAA,UACb;AAAA,QACF;AAAA,MACF,WAAW,OAAO,IAAI,GAAG,MAAM,YAAY,OAAO,UAAU,UAAU;AACpE,YAAI,GAAG,KAAK;AAAA,MACd,WAAW,OAAO,IAAI,GAAG,MAAM,YAAY,OAAO,UAAU,UAAU;AACpE,YAAI,GAAG,IAAI;AAAA,MACb,WAAW,MAAM,QAAQ,IAAI,GAAG,CAAC,KAAK,MAAM,QAAQ,KAAK,GAAG;AAC1D,cAAM,WAAW,IAAI,GAAG;AACxB,iBAAS,IAAI,GAAG,IAAI,MAAM,QAAQ,KAAK;AACrC,gBAAM,EAAE,OAAO,GAAG,UAAU,IAAI,MAAM,CAAC;AACvC,cAAI,QAAQ,SAAS,SAAS,GAAG;AAC/B,kBAAM,IAAI;AAAA,cACR,mFAAmF,QAAQ,WAAW,KAAK;AAAA,YAC7G;AAAA,UACF;AACA,mBAAS,KAAK,IAAI,OAAO,SAAS,KAAK,GAAG,SAAS;AAAA,QACrD;AAAA,MACF,WAAW,OAAO,IAAI,GAAG,MAAM,YAAY,OAAO,UAAU,UAAU;AACpE,YAAI,GAAG,IAAI,OAAO,IAAI,GAAG,GAAG,KAAK;AAAA,MACnC;AAAA,IACF;AACA,WAAO;AAAA,EACT;AAEA,QAAM,SAAS,KAAK,UAAU,CAAC;AAC/B,MAAI,CAAC,QAAQ;AAEX,WAAO;AAAA,EACT;AACA,SAAO,OAAO,UAAU,OAAO,KAAK;AACtC;AACA,eAAe,oBACb,QACA,QACgC;AAChC,QAAM,kBAAkB,KAAK,IAAI;AACjC,MAAI;AACJ,MAAI,aAAa;AACjB,MAAI,aAAa;AAEjB,cAAY,IAAI,uBAAuB;AAAA,IACrC,iBAAiB,OAAO,eAAe;AAAA,EACzC,CAAC;AAED,MAAI,UAAU,CAAC;AAEf,MAAI,IAAI,OAAO,SAAS,QAAQ;AAChC,MAAI;AACF,qBAAiB,SAAS,QAAQ;AAChC,UAAI,QAAQ,SAAS;AACnB,oBAAY,KAAK,yBAAyB;AAAA,UACxC;AAAA,UACA,WAAW,KAAK,IAAI;AAAA,QACtB,CAAC;AACD,cAAM,IAAI,MAAM,uBAAuB;AAAA,MACzC;AAEA;AAEA,UAAI;AACF,YAAI,CAAC,IAAI;AACP,eAAK,MAAM;AACX,sBAAY,IAAI,6BAA6B;AAAA,YAC3C;AAAA,YACA,aAAa,OAAO,UAAU;AAAA,UAChC,CAAC;AAAA,QACH;AACA,YAAI,CAAC,OAAO;AACV,kBAAQ,MAAM;AACd,sBAAY,IAAI,gCAAgC;AAAA,YAC9C;AAAA,YACA,aAAa,OAAO,UAAU;AAAA,UAChC,CAAC;AAAA,QACH;AACA,YAAI,CAAC,SAAS;AACZ,oBAAU,MAAM;AAAA,QAClB;AACA,YAAI,CAAC,QAAQ;AACX,mBAAS,MAAM;AAAA,QACjB;AACA,YAAI,CAAC,SAAS,MAAM,OAAO;AACzB,kBAAQ,MAAM;AAEd,4BAAkB;AAAA,YAChB,aAAa,MAAM,MAAM;AAAA,YACzB,cAAc,MAAM,MAAM;AAAA,UAC5B,CAAC;AAAA,QACH;AAEA,kBAAU,eAAe,SAAS,KAAK;AAEvC,YAAI,OAAO,UAAU,CAAC,GAAG,OAAO,SAAS;AACvC,cAAI,CAAC,QAAQ;AACX,qBAAS,KAAK,IAAI,IAAI;AACtB,wBAAY,IAAI,6BAA6B;AAAA,cAC3C,QAAQ,OAAO,MAAM;AAAA,cACrB,aAAa,OAAO,UAAU;AAAA,YAChC,CAAC;AAAA,UACH;AAAA,QACF;AAAA,MACF,SAAS,YAAY;AACnB;AACA,oBAAY,MAAM,6BAA6B;AAAA,UAC7C,aAAa,OAAO,UAAU;AAAA,UAC9B,cACE,sBAAsB,QAClB,WAAW,UACX,OAAO,UAAU;AAAA,UACvB,WACE,sBAAsB,QAClB,WAAW,YAAY,OACvB,OAAO;AAAA,QACf,CAAC;AAAA,MAEH;AAAA,IACF;AAEA,gBAAY,IAAI,0BAA0B;AAAA,MACxC,aAAa,OAAO,UAAU;AAAA,MAC9B,YAAY,OAAO,UAAU;AAAA,MAC7B,eAAe,OAAO,KAAK,IAAI,IAAI,eAAe;AAAA,MAClD,QAAQ,OAAO,UAAU,CAAC;AAAA,MAC1B,gBAAgB,MAAM;AAAA,IACxB,CAAC;AAAA,EACH,SAAS,aAAa;AACpB,gBAAY,MAAM,6BAA6B;AAAA,MAC7C,aAAa,OAAO,UAAU;AAAA,MAC9B,YAAY,OAAO,UAAU;AAAA,MAC7B,cACE,uBAAuB,QACnB,YAAY,UACZ,OAAO,WAAW;AAAA,MACxB,WACE,uBAAuB,QACnB,YAAY,YAAY,OACxB,OAAO;AAAA,IACf,CAAC;AACD,UAAM;AAAA,EACR;AACA,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA,IACA;AAAA,IACA,SAAS;AAAA,MACP;AAAA,QACE,OAAO;AAAA,QACP;AAAA,QACA,eAAe;AAAA,QACf,UAAU;AAAA,MACZ;AAAA,IACF;AAAA,IACA;AAAA,EACF;AACF;AAEA,SAAS,iCACP,UACA,QACA;AACA,MAAI,gBAAgC,CAAC;AACrC,QAAM,UAAU,SAAS,UAAU,CAAC,GAAG;AACvC,MAAI,CAAC,SAAS;AACZ,WAAO;AAAA,MACL,MAAM;AAAA,MACN,SAAS,CAAC;AAAA,MACV,aAAa,SAAS,UAAU,CAAC,GAAG;AAAA,MACpC,MAAM;AAAA,MACN,OAAO,SAAS;AAAA,IAClB;AAAA,EACF;AAEA,MAAI,SAAS,YAAY;AACvB,eAAW,YAAY,QAAQ,YAAY;AACzC,YAAM,OAAO,SAAS;AACtB,YAAM,WAAW,MAAM;AACvB,UAAI,WAAW,CAAC;AAChB,UAAI,aAA4B;AAChC,UAAI;AACF,mBAAW,MAAM,YAAY,KAAK,MAAM,KAAK,SAAS,IAAI,CAAC;AAAA,MAC7D,SAAS,GAAG;AAEV,qBAAa,aAAa,QAAQ,EAAE,UAAU,OAAO,CAAC;AACtD,oBAAY,KAAK,8BAA8B;AAAA,UAC7C;AAAA,UACA,cAAc,MAAM,WAAW,UAAU,GAAG,GAAG;AAAA;AAAA,UAC/C,cAAc;AAAA,QAChB,CAAC;AAAA,MACH;AAGA,UAAI,cAAc,OAAO,KAAK,QAAQ,EAAE,WAAW,GAAG;AACpD,oBAAY,MAAM,6BAA6B;AAAA,UAC7C;AAAA,UACA;AAAA,UACA,YACE;AAAA,QACJ,CAAC;AAAA,MACH;AAEA,oBAAc,KAAK;AAAA,QACjB,MAAM;AAAA,QACN,OAAO;AAAA,QACP,MAAM;AAAA,QACN,IAAI,SAAS,IAAI,SAAS,IAAI,SAAS,KAAK,OAAO;AAAA,MACrD,CAAC;AAAA,IACH;AAAA,EACF;AAEA,MAAK,QAAgB,WAAW;AAC9B,kBAAc,KAAK;AAAA,MACjB,MAAM;AAAA,MACN,UAAW,QAAgB;AAAA,MAC3B,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAGA,MAAK,QAAgB,mBAAmB;AACtC,kBAAc,KAAK;AAAA,MACjB,MAAM;AAAA,MACN,UAAW,QAAgB;AAAA,MAC3B,WAAW;AAAA,IACb,CAAC;AAAA,EACH;AAEA,MAAI,QAAQ,SAAS;AACnB,kBAAc,KAAK;AAAA,MACjB,MAAM;AAAA,MACN,MAAM,SAAS;AAAA,MACf,WAAW,CAAC;AAAA,IACd,CAAC;AAAA,EACH;AAEA,QAAM,eAAe;AAAA,IACnB,MAAM;AAAA,IACN,SAAS;AAAA,IACT,aAAa,SAAS,UAAU,CAAC,GAAG;AAAA,IACpC,MAAM;AAAA,IACN,OAAO,SAAS;AAAA,EAClB;AAEA,SAAO;AACT;AAEA,IAAI,kBACF;AAKK,SAAS,mBACd,OACgD;AAChD,QAAM,SAAS,gBAAgB;AAC/B,QAAM,WAAW,OAAO;AAGxB,MAAI,mBAAmB,UAAU;AAE/B,sBAAkB;AAAA,EACpB;AAEA,MAAI,iBAAiB;AACnB,WAAO;AAAA,EACT;AAEA,QAAM,SAAS,wBAAwB,KAAK;AAE5C,QAAM,iBAA4C;AAAA,IAChD,SAAS;AAAA,IACT,cAAc;AAAA,EAChB;AACA,MAAI,QAAQ,IAAI,sBAAsB;AACpC,mBAAe,eAAe,IAC5B,UAAU,QAAQ,IAAI,oBAAoB;AAAA,EAC9C;AAEA,QAAM,OAAO;AAAA,IACX;AAAA,IACA,YAAY;AAAA;AAAA,IACZ,SAAS,SAAS,QAAQ,IAAI,kBAAkB,OAAO,KAAK,GAAI,GAAG,EAAE;AAAA,EACvE;AACA,MAAI,aAAa;AACf,UAAM,SAAS,IAAI,iBAAiB,IAAI;AACxC,sBAAkB;AAClB,WAAO;AAAA,EACT;AACA,MAAI,YAAY;AACd,UAAM,aAAa;AAAA,MACjB,GAAG;AAAA,MACH,QAAQ,UAAU,QAAQ,IAAI,mBAAmB;AAAA,IACnD;AACA,UAAM,SAAS,IAAI,gBAAgB,UAAU;AAC7C,sBAAkB;AAClB,WAAO;AAAA,EACT;AAGA,QAAM,eAAe,gBAAgB;AACrC,QAAM,eAAe,aAAa,SAAS,MAAM;AAEjD,MAAI;AACJ,MAAI;AAEJ,MAAI,cAAc;AAChB,aAAS,aAAa,UAAU;AAChC,cAAU,aAAa;AAAA,EACzB,OAAO;AAEL,aAAS,mBAAmB;AAC5B,cAAU;AAAA,EACZ;AAEA,MAAI,QAAQ,IAAI,cAAc,SAAS,CAAC,UAAU,aAAa,aAAa;AAC1E,YAAQ;AAAA,MACN,MAAM;AAAA,QACJ;AAAA,MACF;AAAA,IACF;AAAA,EACF;AAIA,QAAM,eAAe;AAAA,IACnB;AAAA,IACA,yBAAyB;AAAA,IACzB,GAAG;AAAA,IACH,GAAI,WAAW,EAAE,QAAQ;AAAA;AAAA,EAC3B;AAEA,oBAAkB,IAAI,UAAU,YAAY;AAC5C,SAAO;AACT;AAKO,SAAS,uBAA6B;AAC3C,oBAAkB;AACpB;AAoCA,SAAS,4BACP,cACA,eACmE;AACnE,MAAI,CAAC,wBAAwB;AAC3B,WAAO,EAAE,cAAc,cAAc;AAAA,EACvC;AAEA,QAAM,iBAAiB;AACvB,MAAI,kBAAkB;AAGtB,QAAM,wBAAwB,aAAa,IAAI,CAAC,OAAO,WAAW;AAChE,QAAI,kBAAkB,kBAAkB,MAAM,KAAK,SAAS,KAAM;AAChE;AACA,aAAO;AAAA,QACL,GAAG;AAAA,QACH,eAAe,EAAE,MAAM,YAAqB;AAAA,MAC9C;AAAA,IACF;AACA,UAAM,EAAE,eAAe,GAAG,kBAAkB,IAAI;AAChD,WAAO;AAAA,EACT,CAAC;AAGD,QAAM,yBAAyB,cAAc,IAAI,CAAC,SAAS,iBAAiB;AAC1E,QAAI,MAAM,QAAQ,QAAQ,OAAO,GAAG;AAClC,YAAM,mBAAmB,QAAQ,QAAQ;AAAA,QACvC,CAAC,cAAc,eAAe;AAE5B,gBAAM,cACJ,kBAAkB,kBAClB,aAAa,SAAS,UACtB,OAAO,aAAa,SAAS;AAAA,WAE5B,aAAa,KAAK,SAAS;AAAA,UAEzB,iBAAiB,cAAc,SAAS,KACvC,eAAe,QAAQ,QAAQ,SAAS,KACxC,aAAa,KAAK,SAAS;AAEjC,cAAI,aAAa;AACf;AACA,mBAAO;AAAA,cACL,GAAG;AAAA,cACH,eAAe,EAAE,MAAM,YAAqB;AAAA,YAC9C;AAAA,UACF;AAGA,gBAAM,EAAE,eAAe,GAAG,kBAAkB,IAAI;AAChD,iBAAO;AAAA,QACT;AAAA,MACF;AAEA,aAAO;AAAA,QACL,GAAG;AAAA,QACH,SAAS;AAAA,MACX;AAAA,IACF;AAEA,WAAO;AAAA,EACT,CAAC;AAED,SAAO;AAAA,IACL,cAAc;AAAA,IACd,eAAe;AAAA,EACjB;AACF;AAEO,SAAS,0BACd,SACA,WAAW,OACG;AACd,MAAI,UAAU;AACZ,QAAI,OAAO,QAAQ,QAAQ,YAAY,UAAU;AAC/C,aAAO;AAAA,QACL,MAAM;AAAA,QACN,SAAS;AAAA,UACP;AAAA,YACE,MAAM;AAAA,YACN,MAAM,QAAQ,QAAQ;AAAA,UACxB;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,aAAO;AAAA,QACL,MAAM;AAAA,QACN,SAAS,QAAQ,QAAQ,QAAQ,IAAI,QAAM,EAAE,GAAG,EAAE,EAAE;AAAA,MACtD;AAAA,IACF;AAAA,EACF;AACA,SAAO;AAAA,IACL,MAAM;AAAA,IACN,SAAS,QAAQ,QAAQ;AAAA,EAC3B;AACF;AAEO,SAAS,+BACd,SACA,WAAW,OACG;AACd,MAAI,UAAU;AACZ,QAAI,OAAO,QAAQ,QAAQ,YAAY,UAAU;AAC/C,aAAO;AAAA,QACL,MAAM;AAAA,QACN,SAAS;AAAA,UACP;AAAA,YACE,MAAM;AAAA,YACN,MAAM,QAAQ,QAAQ;AAAA,UACxB;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,aAAO;AAAA,QACL,MAAM;AAAA,QACN,SAAS,QAAQ,QAAQ,QAAQ,IAAI,QAAM,EAAE,GAAG,EAAE,EAAE;AAAA,MACtD;AAAA,IACF;AAAA,EACF;AACA,SAAO;AAAA,IACL,MAAM;AAAA,IACN,SAAS,QAAQ,QAAQ;AAAA,EAC3B;AACF;AAEA,SAAS,qBAAqB,cAAkC;AAG9D,QAAM,yBAAyB,aAAa,CAAC,KAAK;AAClD,QAAM,mBAAmB,aAAa,MAAM,CAAC;AAC7C,SAAO,CAAC,wBAAwB,iBAAiB,KAAK,IAAI,CAAC,EAAE,OAAO,OAAO;AAC7E;AAEA,eAAsB,SACpB,UACA,cACA,mBACA,OACA,QACA,SAM2B;AAC3B,QAAM,eAAe,gBAAgB;AACrC,QAAM,kBAAkB,aAAa,qBAAqB,QAAQ,KAAK;AAEvE,MAAI,CAAC,gBAAgB,WAAW,CAAC,gBAAgB,SAAS;AACxD,UAAM,IAAI;AAAA,MACR,gBAAgB,SAAS,4BAA4B,QAAQ,KAAK;AAAA,IACpE;AAAA,EACF;AAEA,QAAM,eAAe,gBAAgB;AACrC,QAAM,gBAAgB,aAAa;AAGnC,QAAM,iBAAiB,QAAQ;AAC/B,MAAI,kBAAkB,CAAC,eAAe,eAAe;AACnD,UAAM,iBAAiB;AAAA,MACrB,eAAe;AAAA,MACf,eAAe;AAAA,IACjB;AACA,UAAM,qBACJ,qBAAqB,sBAAsB,cAAc;AAE3D,mBAAe,gBAAgB;AAAA,MAC7B;AAAA,MACA;AAAA,IACF;AAAA,EACF;AAEA,cAAY,IAAI,kBAAkB;AAAA,IAChC,YAAY,QAAQ;AAAA,IACpB,mBAAmB;AAAA,IACnB,UAAU,aAAa;AAAA,IACvB,WAAW,CAAC,QAAQ,QAAQ,aAAa,OAAO,EAAE,SAAS,QAAQ,KAAK;AAAA,IACxE,kBAAkB,CAAC,CAAC,gBAAgB;AAAA,IACpC,gBAAgB,gBAAgB,eAAe;AAAA,IAC/C,WAAW,kBAAkB,GAAG;AAAA,EAClC,CAAC;AAED,QAAM,iBAAiB,kBAAkB;AACzC,cAAY,IAAI,qBAAqB;AAAA,IACnC,cAAc,SAAS;AAAA,IACvB,oBAAoB,aAAa,KAAK,GAAG,EAAE;AAAA,IAC3C,WAAW,MAAM;AAAA,IACjB,OAAO;AAAA,IACP,oBAAoB,QAAQ;AAAA,IAC5B,WAAW,kBAAkB,GAAG;AAAA,EAClC,CAAC;AAED,YAAU,UAAU;AAEpB,MAAI;AACF,UAAM,SAAS,MAAM;AAAA,MAAQ;AAAA,MAAU,MACrC;AAAA,QACE;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,QACA;AAAA,QACA,EAAE,GAAG,SAAS,OAAO,eAAe,cAAc,eAAe;AAAA;AAAA,MACnE;AAAA,IACF;AAEA,gBAAY,IAAI,uBAAuB;AAAA,MACrC,SAAS,OAAO;AAAA,MAChB,YAAY,OAAO;AAAA,MACnB,gBAAgB,OAAO,QAAQ,SAAS,UAAU;AAAA,MAClD,WAAW,kBAAkB,GAAG;AAAA,IAClC,CAAC;AAGD,QAAI,gBAAgB,eAAe,kBAAkB,OAAO,YAAY;AACtE,2BAAqB;AAAA,QACnB,eAAe,cAAc;AAAA,QAC7B,OAAO;AAAA,MACT;AAEA,kBAAY,IAAI,0BAA0B;AAAA,QACxC,gBAAgB,eAAe,cAAc;AAAA,QAC7C,YAAY,OAAO;AAAA,QACnB,WAAW,kBAAkB,GAAG;AAAA,MAClC,CAAC;AAAA,IACH;AAEA,WAAO;AAAA,EACT,SAAS,OAAO;AAEd;AAAA,MACE;AAAA,MACA;AAAA,QACE,cAAc,SAAS;AAAA,QACvB,oBAAoB,aAAa,KAAK,GAAG,EAAE;AAAA,QAC3C,OAAO,QAAQ;AAAA,QACf,WAAW,MAAM;AAAA,QACjB,OAAO;AAAA,MACT;AAAA,MACA,gBAAgB;AAAA,IAClB;AAEA,UAAM;AAAA,EACR;AACF;AAEO,SAAS,8BACd,cACA,SACA,SACA,uBAAuB,OACwB;AAE/C,QAAM,iBAAiB,CAAC,GAAG,YAAY;AACvC,MAAI,YAAY;AAGhB,QAAM,eAAe,gBAAgB;AACrC,QAAM,eAAe,aAAa,SAAS,MAAM;AACjD,MAAI,gBAAgB,YAAY,aAAa,SAAS,GAAG;AAEvD,UAAM,qBAAqB;AAAA,MACzB;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,IACF;AACA,mBAAe,KAAK,GAAG,kBAAkB;AAAA,EAC3C;AAGA,QAAM,aAAa,OAAO,QAAQ,OAAO,EAAE,SAAS;AAEpD,MAAI,YAAY;AAEd,QAAI,CAAC,sBAAsB;AACzB,YAAM,eAAe,qBAAqB;AAC1C,UAAI,cAAc;AAEhB,uBAAe,KAAK,2CAAkB;AACtC,uBAAe,KAAK,YAAY;AAChC,uBAAe,KAAK,SAAS;AAAA,MAC/B;AAAA,IACF;AAGA,UAAM,mBAAmB,wBAAwB,YAAY,OAAO;AACpE,QAAI,iBAAiB,SAAS,GAAG;AAC/B,kBAAY,iBAAiB,IAAI,OAAK,EAAE,OAAO,EAAE,KAAK,IAAI,IAAI;AAAA,IAChE;AAGA,mBAAe;AAAA,MACb;AAAA;AAAA;AAAA,IACF;AAGA,UAAM,kBAAkB,OAAO;AAAA,MAC7B,OAAO,QAAQ,OAAO,EAAE;AAAA,QACtB,CAAC,CAAC,GAAG,MAAM,QAAQ,iBAAiB,QAAQ;AAAA,MAC9C;AAAA,IACF;AAEA,mBAAe;AAAA,MACb,GAAG,OAAO,QAAQ,eAAe,EAAE;AAAA,QACjC,CAAC,CAAC,KAAK,KAAK,MAAM,kBAAkB,GAAG,KAAK,KAAK;AAAA,MACnD;AAAA,IACF;AAAA,EACF;AAEA,SAAO,EAAE,cAAc,gBAAgB,UAAU;AACnD;AAEA,eAAe,0BACb,UACA,cACA,mBACA,OACA,QACA,SAO2B;AAC3B,QAAM,SAAS,gBAAgB;AAC/B,QAAM,eAAe,gBAAgB;AACrC,QAAM,iBAAiB,QAAQ;AAE/B,QAAM,eAAe,QAAQ,gBAAgB,aAAa,SAAS,MAAM;AACzE,MAAI;AAEJ,MAAI,cAAc;AAChB,eAAW,aAAa,YAAY,OAAO,mBAAmB;AAAA,EAChE,OAAO;AACL,eAAW,OAAO,mBAAmB;AAAA,EACvC;AAGA,MACE,aAAa,eACb,aAAa,cACb,aAAa,WACb;AACA,WAAO;AAAA,MACL;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA;AAAA,MACA,EAAE,GAAG,SAAS,cAAc,eAAe;AAAA,IAC7C;AAAA,EACF;AAGA,SAAO,YAAY,UAAU,cAAc,mBAAmB,OAAO,QAAQ;AAAA,IAC3E,GAAG;AAAA,IACH;AAAA,IACA;AAAA,EACF,CAAC;AACH;AAEA,eAAe,qBACb,UACA,cACA,mBACA,OACA,QACA,SAO2B;AAC3B,QAAM,SAAS,gBAAgB;AAC/B,QAAM,eAAe,gBAAgB;AACrC,QAAM,iBAAiB,SAAS;AAEhC,QAAM,eAAe,SAAS,gBAAgB,aAAa,SAAS,MAAM;AAC1E,MAAI;AACJ,MAAI;AACJ,MAAI;AAGJ,cAAY,IAAI,0BAA0B;AAAA,IACxC,mBAAmB,CAAC,CAAC;AAAA,IACrB,gBAAgB,cAAc;AAAA,IAC9B,kBAAkB,cAAc;AAAA,IAChC,uBAAuB,cAAc;AAAA,IACrC,sBAAsB,cAAc;AAAA,IACpC,qBAAqB,cAAc;AAAA,IACnC,0BAA0B,CAAC,CAAC,cAAc;AAAA,IAC1C,cAAc,SAAS;AAAA,IACvB,WAAW,kBAAkB,GAAG;AAAA,EAClC,CAAC;AAED,MAAI,cAAc;AAEhB,YAAQ,aAAa;AACrB,eAAW,aAAa,YAAY,OAAO,mBAAmB;AAG9D,QACE,aAAa,aAAa,eAC1B,aAAa,aAAa,cAC1B,aAAa,aAAa,WAC1B;AACA,YAAM,eAAoB;AAAA,QACxB,QAAQ,aAAa;AAAA,QACrB,yBAAyB;AAAA,QACzB,YAAY;AAAA,QACZ,SAAS,SAAS,QAAQ,IAAI,kBAAkB,OAAO,KAAK,GAAI,GAAG,EAAE;AAAA,QACrE,gBAAgB;AAAA,UACd,SAAS;AAAA,UACT,cAAc;AAAA,QAChB;AAAA,MACF;AAGA,UAAI,aAAa,SAAS;AACxB,qBAAa,UAAU,aAAa;AAAA,MACtC;AAEA,kBAAY,IAAI,UAAU,YAAY;AAAA,IACxC,OAAO;AAEL,kBAAY,mBAAmB,KAAK;AAAA,IACtC;AAAA,EACF,OAAO;AAEL,UAAM,eAAe;AAAA,MACnB,oBAAoB,CAAC,CAAC;AAAA,MACtB,uBAAuB,cAAc;AAAA,MACrC,gBAAgB,SAAS;AAAA,MACzB,WAAW,kBAAkB,GAAG;AAAA,IAClC;AACA,gBAAY,MAAM,4BAA4B,YAAY;AAC1D,UAAM,IAAI;AAAA,MACR,iHAAiH,KAAK,UAAU,YAAY,CAAC;AAAA,IAC/I;AAAA,EACF;AAGA,MAAI,SAAS,qBAAqB;AAEhC,UAAM,CAAC,mBAAmB,IAAI,qBAAqB,YAAY;AAE/D,mBAAe,CAAC,sBAAsB,GAAG,GAAG,YAAY;AAAA,EAC1D;AAEA,QAAM,SAA2B,qBAAqB,YAAY,EAAE;AAAA,IAClE,QAAM;AAAA,MACJ,MAAM;AAAA,MACN,MAAM;AAAA,IACR;AAAA,EACF;AAEA,QAAM,cAAc,MAAM,QAAQ;AAAA,IAChC,MAAM;AAAA,MACJ,OAAM,UACH;AAAA,QACC,MAAM,KAAK;AAAA;AAAA,QAEX,aAAa,MAAM,wBAAwB,IAAI;AAAA,QAC/C,cACE,qBAAqB,QAAQ,KAAK,kBAC9B,KAAK,kBACL,gBAAgB,KAAK,WAAW;AAAA,MACxC;AAAA,IACJ;AAAA,EACF;AAEA,QAAM,oBAAoB,oBAAoB,QAAQ;AAGtD,QAAM,EAAE,cAAc,iBAAiB,eAAe,kBAAkB,IACtE,4BAA4B,QAAQ,iBAAiB;AACvD,QAAM,wBAAwB,KAAK,IAAI;AAGvC,8BAA4B;AAAA,IAC1B,YAAY,aAAa,KAAK,IAAI;AAAA,IAClC,cAAc,qBAAqB,KAAK;AAAA,IACxC,WAAW,CAAC;AAAA;AAAA,IACZ,aAAa,aAAa,KAAK,IAAI;AAAA,EACrC,CAAC;AAED,MAAI,QAAQ,KAAK,IAAI;AACrB,MAAI,gBAAgB;AACpB,MAAI;AAEJ,MAAI;AACF,eAAW,MAAM;AAAA,MACf,OAAM,YAAW;AACf,wBAAgB;AAChB,gBAAQ,KAAK,IAAI;AAEjB,cAAM,SAAsD;AAAA,UAC1D;AAAA,UACA,YAAY,wBAAwB,YAAY;AAAA,UAChD,UAAU;AAAA,UACV,QAAQ;AAAA,UACR,OAAO,YAAY,SAAS,IAAI,cAAc;AAAA,UAC9C,aAAa,YAAY,SAAS,IAAI,EAAE,MAAM,OAAO,IAAI;AAAA,QAC3D;AAEA,YAAI,oBAAoB,GAAG;AACzB;AAAC,UAAC,OAAe,gBAAgB;AAAA,YAC/B,kBAAkB;AAAA,UACpB;AACC,UAAC,OAAe,WAAW,EAAE,YAAY,kBAAkB;AAAA,QAC9D;AAGA,oBAAY,IAAI,sCAAsC;AAAA,UACpD,UAAU,cAAc,WAAW;AAAA,UACnC;AAAA,UACA;AAAA,UACA,kBAAkB,CAAC,CAAC,cAAc;AAAA,UAClC,cAAc,cAAc,SACxB,aAAa,OAAO,UAAU,GAAG,CAAC,IAClC;AAAA,UACJ,WAAW,OAAO;AAAA,UAClB,aAAa;AAAA,UACb;AAAA,UACA,cAAc,OAAO,UAAU,UAAU;AAAA,UACzC,YAAY;AAAA,UACZ,YAAY,YAAY;AAAA,UACxB,gBAAgB;AAAA,UAChB,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,UAClC,gBAAgB,cAAc;AAAA,UAC9B,kBAAkB,cAAc;AAAA,QAClC,CAAC;AAED,YAAI,OAAO,QAAQ;AACjB,gBAAM,SAAS,MAAM,UAAU,KAAK,SAAS;AAAA,YAC3C;AAAA,cACE,GAAG;AAAA,cACH,QAAQ;AAAA,YACV;AAAA,YACA;AAAA,cACE;AAAA;AAAA,YACF;AAAA,UACF;AAEA,cAAI,gBAA4B;AAChC,cAAI,oBAAyB;AAC7B,gBAAM,gBAAuB,CAAC;AAC9B,gBAAM,mBAAmB,oBAAI,IAAoB;AACjD,cAAI,QAAa;AACjB,cAAI,aAA4B;AAChC,cAAI,eAA8B;AAElC,2BAAiB,SAAS,QAAQ;AAChC,gBAAI,OAAO,SAAS;AAClB,0BAAY,KAAK,kBAAkB;AAAA,gBACjC,WAAW,MAAM;AAAA,gBACjB,WAAW,KAAK,IAAI;AAAA,cACtB,CAAC;AACD,oBAAM,IAAI,MAAM,uBAAuB;AAAA,YACzC;AAEA,oBAAQ,MAAM,MAAM;AAAA,cAClB,KAAK;AACH,oCAAoB;AACpB,gCAAgB;AAAA,kBACd,GAAG,MAAM;AAAA,kBACT,SAAS,CAAC;AAAA;AAAA,gBACZ;AAEA,kCAAkB;AAAA,kBAChB,aAAa,MAAM,QAAQ,OAAO;AAAA,kBAClC,eAAe;AAAA;AAAA,gBACjB,CAAC;AACD;AAAA,cAEF,KAAK;AACH,8BAAc,MAAM,KAAK,IAAI,EAAE,GAAG,MAAM,cAAc;AAEtD,oBAAI,MAAM,cAAc,SAAS,YAAY;AAC3C,mCAAiB,IAAI,MAAM,OAAO,EAAE;AAAA,gBACtC;AAEA,oBAAI,MAAM,cAAc,SAAS,YAAY;AAC3C,wBAAM,oBAAoB,KAAK,IAAI;AACnC,oCAAkB;AAAA,oBAChB,OAAO;AAAA,oBACP,YAAY;AAAA,oBACZ,mBAAmB;AAAA,oBACnB;AAAA;AAAA,kBACF,CAAC;AACD,8BAAY,KAAK,wBAAwB;AAAA,oBACvC,OAAO,MAAM;AAAA,oBACb,WAAW;AAAA,oBACX,WAAW;AAAA,kBACb,CAAC;AAAA,gBACH;AACA;AAAA,cAEF,KAAK;AACH,sBAAM,aAAa,MAAM;AAGzB,oBAAI,CAAC,cAAc,UAAU,GAAG;AAC9B,gCAAc,UAAU,IAAI;AAAA,oBAC1B,MACE,MAAM,MAAM,SAAS,eAAe,SAAS;AAAA,oBAC/C,MAAM,MAAM,MAAM,SAAS,eAAe,KAAK;AAAA,kBACjD;AACA,sBAAI,MAAM,MAAM,SAAS,oBAAoB;AAC3C,qCAAiB,IAAI,YAAY,EAAE;AAAA,kBACrC;AAAA,gBACF;AAEA,oBAAI,MAAM,MAAM,SAAS,cAAc;AACrC,gCAAc,UAAU,EAAE,QAAQ,MAAM,MAAM;AAE9C,wBAAM,eAAe,kBAAkB;AACvC,oCAAkB;AAAA,oBAChB,gBACG,aAAa,iBAAiB,KAC/B,MAAM,MAAM,KAAK;AAAA,kBACrB,CAAC;AAAA,gBACH,WAAW,MAAM,MAAM,SAAS,oBAAoB;AAClD,wBAAM,gBAAgB,iBAAiB,IAAI,UAAU,KAAK;AAC1D,mCAAiB;AAAA,oBACf;AAAA,oBACA,gBAAgB,MAAM,MAAM;AAAA,kBAC9B;AAAA,gBACF,WAES,MAAM,MAAM,SAAS,kBAAkB;AAC9C,wBAAM,kBACJ,cAAc,UAAU,EAAE,YAAY;AACxC,wBAAM,kBACJ,mBAAmB,MAAM,MAAM,YAAY;AAC7C,gCAAc,UAAU,IAAI;AAAA,oBAC1B,GAAG,cAAc,UAAU;AAAA,oBAC3B,UAAU;AAAA,kBACZ;AAGA,oCAAkB;AAAA,oBAChB,OAAO;AAAA,oBACP,YAAY,gBAAgB;AAAA,oBAC5B,mBAAmB;AAAA,kBACrB,CAAC;AAED,8BAAY,MAAM,kBAAkB;AAAA,oBAClC,OAAO;AAAA,oBACP,YAAY,gBAAgB;AAAA,kBAC9B,CAAC;AAAA,gBACH;AACA;AAAA,cAEF,KAAK;AACH,oBAAI,MAAM,MAAM;AACd,+BAAa,MAAM,MAAM;AAC3B,oBAAI,MAAM,MAAM;AACd,iCAAe,MAAM,MAAM;AAC7B,oBAAI,MAAM,OAAO;AACf,0BAAQ,EAAE,GAAG,OAAO,GAAG,MAAM,MAAM;AAGnC,oCAAkB;AAAA,oBAChB,cAAc,MAAM,MAAM;AAAA,kBAC5B,CAAC;AAAA,gBACH;AACA;AAAA,cAEF,KAAK;AACH,sBAAM,YAAY,MAAM;AACxB,sBAAM,QAAQ,cAAc,SAAS;AAGrC,oBAAI,OAAO,SAAS,YAAY;AAC9B,wBAAMA,eAAc,kBAAkB;AACtC,wBAAM,qBAAqBA,aAAY,oBACnC,KAAK,IAAI,IAAIA,aAAY,oBACzB;AACJ,oCAAkB;AAAA,oBAChB,OAAO;AAAA,oBACP;AAAA;AAAA,kBACF,CAAC;AACD,8BAAY,KAAK,2BAA2B;AAAA,oBAC1C,OAAO;AAAA,oBACP,iBAAiB,MAAM,UAAU,UAAU;AAAA,oBAC3C,YAAY;AAAA,kBACd,CAAC;AAAA,gBACH;AAEA,oBACE,OAAO,SAAS,cAChB,iBAAiB,IAAI,SAAS,GAC9B;AACA,wBAAM,UAAU,iBAAiB,IAAI,SAAS;AAC9C,sBAAI,SAAS;AACX,wBAAI;AACF,4BAAM,QAAQ,KAAK,MAAM,OAAO;AAAA,oBAClC,SAAS,OAAO;AACd,4BAAM,WACJ,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AACvD,kCAAY,MAAM,oBAAoB;AAAA,wBACpC,YAAY;AAAA,wBACZ,SACE,QAAQ,SAAS,MACb,QAAQ,MAAM,GAAG,GAAG,IAAI,QACxB;AAAA,wBACN,OAAO;AAAA,sBACT,CAAC;AAGD,4BAAM,QAAQ;AAAA,wBACZ,iBAAiB;AAAA,wBACjB,mBAAmB,sBAAsB,QAAQ;AAAA,wBACjD,sBACE,QAAQ,SAAS,MACb,QAAQ,MAAM,GAAG,GAAG,IAAI,QACxB;AAAA,sBACR;AAAA,oBACF;AACA,qCAAiB,OAAO,SAAS;AAAA,kBACnC;AAAA,gBACF;AACA;AAAA,cAEF,KAAK;AAEH,iCAAiB,MAAM;AACvB;AAAA,YACJ;AAEA,gBAAI,MAAM,SAAS,gBAAgB;AACjC;AAAA,YACF;AAAA,UACF;AAEA,cAAI,CAAC,iBAAiB,CAAC,mBAAmB;AACxC,kBAAM,IAAI,MAAM,+CAA+C;AAAA,UACjE;AAGA,0BAAgB;AAAA,YACd,GAAG,kBAAkB;AAAA,YACrB,SAAS,cAAc,OAAO,OAAO;AAAA,YACrC,aAAa;AAAA,YACb,eAAe;AAAA,YACf,OAAO;AAAA,cACL,GAAG,kBAAkB,QAAQ;AAAA,cAC7B,GAAG;AAAA,YACL;AAAA,UACF;AAEA,iBAAO;AAAA,QACT,OAAO;AAEL,sBAAY,IAAI,0CAA0C;AAAA,YACxD,UAAU,cAAc,WAAW;AAAA,YACnC;AAAA,YACA;AAAA,YACA,kBAAkB,CAAC,CAAC,cAAc;AAAA,YAClC,cAAc,cAAc,SACxB,aAAa,OAAO,UAAU,GAAG,CAAC,IAClC;AAAA,YACJ,WAAW,OAAO;AAAA,YAClB,aAAa;AAAA,YACb,cAAc,OAAO,UAAU,UAAU;AAAA,YACzC,YAAY;AAAA,YACZ,YAAY,YAAY;AAAA,YACxB,gBAAgB;AAAA,YAChB,YAAW,oBAAI,KAAK,GAAE,YAAY;AAAA,YAClC,gBAAgB,cAAc;AAAA,YAC9B,kBAAkB,cAAc;AAAA,UAClC,CAAC;AAED,iBAAO,MAAM,UAAU,KAAK,SAAS,OAAO,QAAQ;AAAA,YAClD;AAAA;AAAA,UACF,CAAC;AAAA,QACH;AAAA,MACF;AAAA,MACA;AAAA,QACE;AAAA;AAAA,QAEA,SAAS,gBAAgB,UACrB,CAAC,EAAE,SAAS,YAAY,OAAO,QAAQ,MAAM;AAC3C,oCAA0B,eAAe,SAAU;AAAA,YACjD;AAAA,YACA;AAAA,YACA,cAAc,MAAM;AAAA,YACpB;AAAA,UACF,CAAC;AAAA,QACH,IACA;AAAA,MACN;AAAA,IACF;AAEA,gBAAY,IAAI,8BAA8B;AAAA,MAC5C,SAAS,SAAS;AAAA,IACpB,CAAC;AAED,UAAM,SAAS,QAAQ,KAAK,IAAI;AAChC,UAAM,aAAa,KAAK,IAAI,IAAI;AAEhC,UAAM,UAAU,SAAS,QAAQ,IAAI,CAAC,UAAwB;AAC5D,UAAI,MAAM,SAAS,QAAQ;AACzB,eAAO;AAAA,UACL,MAAM;AAAA,UACN,MAAM,MAAM;AAAA,QACd;AAAA,MACF,WAAW,MAAM,SAAS,YAAY;AACpC,eAAO;AAAA,UACL,MAAM;AAAA,UACN,IAAI,MAAM;AAAA,UACV,MAAM,MAAM;AAAA,UACZ,OAAO,MAAM;AAAA,QACf;AAAA,MACF;AACA,aAAO;AAAA,IACT,CAAC;AAGD,UAAM,iBAAiB,QAAQ;AAAA,MAC7B,CAAC,UACC,MAAM,SAAS;AAAA,IACnB;AACA,UAAM,cAAc,kBAAkB;AACtC,UAAM,mBACJ,eAAe,SAAS,IACpB;AAAA,MACE,WAAW,eAAe;AAAA,QACxB,CAAC,KAAK,MAAM,OAAO,EAAE,UAAU,UAAU;AAAA,QACzC;AAAA,MACF;AAAA,MACA,YAAY,YAAY,sBAAsB;AAAA,MAC9C,WAAW,YAAY,qBAAqB,KAAK,IAAI;AAAA,MACrD,YAAY;AAAA,IACd,IACA;AAEN,UAAM,mBAAqC;AAAA,MACzC,SAAS;AAAA,QACP,IAAI,SAAS;AAAA,QACb;AAAA,QACA,OAAO,SAAS;AAAA,QAChB,MAAM;AAAA,QACN,aAAa,SAAS;AAAA,QACtB,eAAe,SAAS;AAAA,QACxB,MAAM;AAAA,QACN,OAAO,SAAS;AAAA,MAClB;AAAA,MACA,MAAM;AAAA,MACN,MAAM,OAAO;AAAA,MACb;AAAA,MACA,SAAS;AAAA;AAAA,MACT;AAAA,IACF;AAIA,UAAM,iBAAiB,OAAO,IAAI,YAAU;AAAA,MAC1C,MAAM;AAAA,MACN,SAAS,MAAM;AAAA,IACjB,EAAE;AAEF,sBAAkB;AAAA,MAChB,cAAc,aAAa,KAAK,IAAI;AAAA,MACpC,UAAU,CAAC,GAAG,gBAAgB,GAAG,iBAAiB;AAAA,MAClD;AAAA,MACA,OAAO,SAAS,QACZ;AAAA,QACE,aAAa,SAAS,MAAM;AAAA,QAC5B,cAAc,SAAS,MAAM;AAAA,MAC/B,IACA;AAAA,MACJ,QAAQ;AAAA,QACN;AAAA,QACA,KAAK,KAAK,IAAI;AAAA,MAChB;AAAA,MACA,WAAW;AAAA,IACb,CAAC;AAGD,UAAM,cAAc,SAAS,MAAM;AACnC,UAAM,eAAe,SAAS,MAAM;AACpC,UAAM,2BACJ,SAAS,MAAM,+BAA+B;AAChD,UAAM,uBAAuB,SAAS,MAAM,2BAA2B;AAEvE,UAAM,UACH,cAAc,MAAa,0BAA0B,KAAK,IAC1D,eAAe,MAAa,2BAA2B,KAAK,IAC5D,2BAA2B,MAC1B,0BAA0B,KAAK,IAChC,uBAAuB,OACrB,0BAA0B,KAAK,IAAI;AAExC,qBAAiB,UAAU;AAC3B,mBAAe,SAAS,UAAU;AAGlC;AAAA,MACE;AAAA,QACE;AAAA,QACA;AAAA,QACA,qBAAqB;AAAA,QACrB,iBAAiB;AAAA,MACnB;AAAA,MACA;AAAA,MACA;AAAA,MACA,iBACK;AAAA,QACC,SAAS,eAAe;AAAA,QACxB,WAAW,eAAe;AAAA,QAC1B;AAAA,MACF,IACA;AAAA,IACN;AAEA,WAAO;AAAA,EACT,SAAS,OAAO;AACd,WAAO,6BAA6B,KAAK;AAAA,EAC3C;AACF;AAEA,SAAS,6BAA6B,OAAkC;AACtE,MAAI,iBAAiB,SAAS,MAAM,QAAQ,SAAS,oBAAoB,GAAG;AAC1E,WAAO,+BAA+B,6BAA6B;AAAA,EACrE;AACA,MACE,iBAAiB,SACjB,MAAM,QAAQ,SAAS,gCAAgC,GACvD;AACA,WAAO,+BAA+B,oCAAoC;AAAA,EAC5E;AACA,MACE,iBAAiB,SACjB,MAAM,QAAQ,YAAY,EAAE,SAAS,WAAW,GAChD;AACA,WAAO,+BAA+B,6BAA6B;AAAA,EACrE;AACA,MAAI,iBAAiB,OAAO;AAC1B,QAAI,QAAQ,IAAI,aAAa,eAAe;AAC1C,kBAAY,MAAM,uBAAuB;AAAA,QACvC,SAAS,MAAM;AAAA,QACf,OAAO,MAAM;AAAA,MACf,CAAC;AAAA,IACH;AACA,WAAO;AAAA,MACL,GAAG,wBAAwB,KAAK,MAAM,OAAO;AAAA,IAC/C;AAAA,EACF;AACA,SAAO,+BAA+B,wBAAwB;AAChE;AAEA,SAAS,oBACP,UACgB;AAChB,SAAO,SAAS,IAAI,CAAC,KAAK,UAAU;AAClC,WAAO,IAAI,SAAS,SAChB,0BAA0B,KAAK,QAAQ,SAAS,SAAS,CAAC,IAC1D,+BAA+B,KAAK,QAAQ,SAAS,SAAS,CAAC;AAAA,EACrE,CAAC;AACH;AAEA,eAAe,YACb,UACA,cACA,mBACA,OACA,QACA,SAO2B;AAC3B,QAAM,SAAS,gBAAgB;AAC/B,QAAM,eAAe,gBAAgB;AACrC,QAAM,iBAAiB,SAAS;AAEhC,QAAM,eAAe,SAAS,gBAAgB,aAAa,SAAS,MAAM;AAC1E,MAAI;AAGJ,QAAM,iBAAiB,kBAAkB;AACzC,cAAY,IAAI,uBAAuB;AAAA,IACrC,mBAAmB,CAAC,CAAC;AAAA,IACrB,gBAAgB,cAAc;AAAA,IAC9B,kBAAkB,cAAc;AAAA,IAChC,uBAAuB,cAAc;AAAA,IACrC,sBAAsB,cAAc;AAAA,IACpC,qBAAqB,cAAc;AAAA,IACnC,0BAA0B,CAAC,CAAC,cAAc;AAAA,IAC1C,cAAc,SAAS;AAAA,IACvB,WAAW,kBAAkB,GAAG;AAAA,EAClC,CAAC;AAED,MAAI,cAAc;AAChB,YAAQ,aAAa;AAAA,EACvB,OAAO;AACL,YAAQ,SAAS,SAAS,cAAc,aAAa;AAAA,EACvD;AAEA,MAAI,SAAS,qBAAqB;AAChC,UAAM,CAAC,mBAAmB,IAAI,qBAAqB,YAAY;AAE/D,mBAAe,CAAC,sBAAsB,IAAI,YAAY;AAAA,EACxD;AAEA,QAAM,SAA2B,qBAAqB,YAAY,EAAE;AAAA,IAClE,QAAM;AAAA,MACJ,GAAI,yBACA,EAAE,eAAe,EAAE,MAAM,YAAY,EAAE,IACvC,CAAC;AAAA,MACL,MAAM;AAAA,MACN,MAAM;AAAA,IACR;AAAA,EACF;AAEA,QAAM,cAAc,MAAM,QAAQ;AAAA,IAChC,MAAM;AAAA,MACJ,OAAM,OACH;AAAA,QACC,MAAM;AAAA,QACN,UAAU;AAAA,UACR,MAAM,EAAE;AAAA,UACR,aAAa,MAAM,EAAE,OAAO;AAAA,YAC1B,UAAU,SAAS;AAAA,UACrB,CAAC;AAAA;AAAA,UAED,YACE,qBAAqB,KAAK,EAAE,kBACxB,EAAE,kBACF,gBAAgB,EAAE,WAAW;AAAA,QACrC;AAAA,MACF;AAAA,IACJ;AAAA,EACF;AAEA,QAAM,eAAe,OAAO;AAAA,IAC1B,QACG;AAAA,MACC,MAAM;AAAA,MACN,SAAS,EAAE;AAAA,IACb;AAAA,EACJ;AAEA,QAAM,iBAAiB,yCAAyC,QAAQ;AACxE,QAAM,wBAAwB,KAAK,IAAI;AAGvC,8BAA4B;AAAA,IAC1B,YAAY,aAAa,KAAK,IAAI;AAAA,IAClC,cAAc,qBAAqB,KAAK;AAAA,IACxC,WAAW,CAAC;AAAA;AAAA,IACZ,aAAa,aAAa,KAAK,IAAI;AAAA,EACrC,CAAC;AAED,MAAI,QAAQ,KAAK,IAAI;AACrB,MAAI,gBAAgB;AACpB,MAAI;AAEJ,MAAI;AACF,eAAW,MAAM;AAAA,MACf,OAAM,YAAW;AACf,wBAAgB;AAChB,gBAAQ,KAAK,IAAI;AAEjB,cAAM,YAAY,wBAAwB,YAAY;AACtD,cAAM,SAAS,YAAY,KAAK;AAEhC,cAAM,OAA0C;AAAA,UAC9C;AAAA,UAEA,GAAI,SACA,EAAE,uBAAuB,UAAU,IACnC,EAAE,YAAY,UAAU;AAAA,UAC5B,UAAU,CAAC,GAAG,cAAc,GAAG,cAAc;AAAA,UAE7C,aAAa,SAAS,IAAI;AAAA,QAC5B;AACA,YAAI,OAAO,QAAQ;AACjB;AAAC,UAAC,KAA2C,SAAS;AACtD,eAAK,iBAAiB;AAAA,YACpB,eAAe;AAAA,UACjB;AAAA,QACF;AAEA,YAAI,YAAY,SAAS,GAAG;AAC1B,eAAK,QAAQ;AACb,eAAK,cAAc;AAAA,QACrB;AACA,cAAM,kBAAkB,MAAM,mBAAmB,cAAc,QAAQ;AACvE,YAAI,iBAAiB;AACnB,eAAK,mBAAmB;AAAA,QAC1B;AAEA,YAAI,gBAAgB,aAAa,WAAW;AAC1C,sBAAY,IAAI,4BAA4B;AAAA,YAC1C,kBAAkB,aAAa;AAAA,YAC/B,WAAW,aAAa;AAAA,YACxB,UAAU,aAAa;AAAA,YACvB,SAAS,aAAa;AAAA,YACtB,cAAc,CAAC,CAAC,aAAa;AAAA,YAC7B,WAAW,kBAAkB,GAAG;AAAA,UAClC,CAAC;AAGD,gBAAM,yBACJ,QAAQ,IAAI,qBAAqB;AAEnC,cAAI,wBAAwB;AAE1B,kBAAM,UAAU,oBAAoB,cAAc,YAAY;AAG9D,kBAAM,gBAAsC;AAAA,cAC1C,UAAU;AAAA,cACV,cAAc,aAAa,IAAI,OAAK,EAAE,OAAiB;AAAA,cACvD;AAAA,cACA,WAAW,wBAAwB,YAAY;AAAA,cAC/C,QAAQ,OAAO;AAAA,cACf;AAAA,cACA,aAAa,YAAY,KAAK,IAAI,IAAI;AAAA,cACtC,oBACE,gBAAgB,eAAe;AAAA,cACjC,WAAW;AAAA;AAAA,YACb;AAGA,kBAAM,UAAU,QAAQ,cAAc,aAAa;AAGnD,gBAAI,oBAAoB,sBAAsB,YAAY,GAAG;AAE3D,oBAAM,EAAE,qBAAqB,IAAI,MAAM,OAAO,UAAU;AACxD,oBAAMC,YAAW,MAAM;AAAA,gBACrB;AAAA,gBACA;AAAA,gBACA;AAAA,cACF;AACA,oBAAM,kBAAkB,QAAQ,cAAcA,SAAQ;AAGtD,oBAAM,aAAa;AAAA,gBACjB,MAAM;AAAA,gBACN,SAAS,gBAAgB;AAAA,gBACzB,YAAY,gBAAgB;AAAA,gBAC5B,OAAO;AAAA,kBACL,eAAe,gBAAgB,MAAM;AAAA,kBACrC,mBAAmB,gBAAgB,MAAM;AAAA,gBAC3C;AAAA,cACF;AACA,oBAAM,eAAiC;AAAA,gBACrC,MAAM;AAAA,gBACN,SAAS;AAAA,gBACT,SAAS;AAAA;AAAA,gBACT,YAAY,KAAK,IAAI,IAAI;AAAA,gBACzB,MAAM,GAAG,KAAK,IAAI,CAAC,IAAI,KAAK,OAAO,EAAE,SAAS,EAAE,EAAE,OAAO,GAAG,CAAC,CAAC;AAAA,gBAC9D,YAAY,gBAAgB;AAAA;AAAA,cAC9B;AACA,qBAAO;AAAA,YACT,OAAO;AAEL,oBAAM,IAAI,MAAM;AAAA,gBACd;AAAA,gBACA;AAAA,gBACA;AAAA,gBACA;AAAA,gBACA;AAAA,cACF;AACA,kBAAI;AACJ,kBAAI,OAAO,QAAQ;AACjB,gCAAgB,MAAM;AAAA,kBACpB;AAAA,kBACA;AAAA,gBACF;AAAA,cACF,OAAO;AACL,gCAAgB;AAAA,cAClB;AACA,oBAAM,IAAI,iCAAiC,eAAe,KAAK;AAC/D,qBAAO;AAAA,YACT;AAAA,UACF,OAAO;AAEL,kBAAM,qBAAqB,YAAY,aAAa,SAAS,IACzD,+BACA;AACJ,kBAAM,IAAI,MAAM;AAAA,cACd;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,cACA;AAAA,YACF;AACA,gBAAI;AACJ,gBAAI,KAAK,QAAQ;AACf,8BAAgB,MAAM;AAAA,gBACpB;AAAA,gBACA;AAAA,cACF;AAAA,YACF,OAAO;AACL,8BAAgB;AAAA,YAClB;AACA,kBAAM,IAAI,iCAAiC,eAAe,KAAK;AAC/D,mBAAO;AAAA,UACT;AAAA,QACF,OAAO;AAEL,sBAAY,IAAI,qBAAqB;AAAA,YACnC,oBAAoB,CAAC,CAAC;AAAA,YACtB,gBAAgB,cAAc;AAAA,YAC9B,iBAAiB,CAAC,CAAC,cAAc;AAAA,YACjC,eAAe;AAAA,YACf,aAAa;AAAA,YACb,WAAW,kBAAkB,GAAG;AAAA,UAClC,CAAC;AAGD,gBAAM,eAAe;AAAA,YACnB,oBAAoB,CAAC,CAAC;AAAA,YACtB,gBAAgB,cAAc;AAAA,YAC9B,iBAAiB,CAAC,CAAC,cAAc;AAAA,YACjC,gBAAgB;AAAA,YAChB,WAAW,kBAAkB,GAAG;AAAA,UAClC;AACA,sBAAY,MAAM,0BAA0B,YAAY;AACxD,gBAAM,IAAI;AAAA,YACR,8CAA8C,KAAK,2DAA2D,KAAK,UAAU,YAAY,CAAC;AAAA,UAC5I;AAAA,QACF;AAAA,MACF;AAAA,MACA;AAAA,QACE;AAAA;AAAA,QAEA,SAAS,gBAAgB,UACrB,CAAC,EAAE,SAAS,YAAY,OAAO,QAAQ,MAAM;AAC3C,oCAA0B,eAAe,SAAU;AAAA,YACjD;AAAA,YACA;AAAA,YACA,cAAc,MAAM;AAAA,YACpB;AAAA,UACF,CAAC;AAAA,QACH,IACA;AAAA,MACN;AAAA,IACF;AAAA,EACF,SAAS,OAAO;AACd,aAAS,KAAK;AACd,WAAO,6BAA6B,KAAK;AAAA,EAC3C;AACA,QAAM,aAAa,KAAK,IAAI,IAAI;AAChC,QAAM,6BAA6B,KAAK,IAAI,IAAI;AAEhD,QAAM,cAAc,SAAS,OAAO,iBAAiB;AACrD,QAAM,eAAe,SAAS,OAAO,qBAAqB;AAC1D,QAAM,uBACJ,SAAS,OAAO,sBAAsB,iBAAiB;AACzD,QAAM,2BACJ,SAAS,OAAO,sBAAsB,iBAAiB;AACzD,QAAM,UACH,cAAc,MAAa,uCAC3B,eAAe,MAAa,wCAC5B,uBAAuB,MACtB,mDACD,2BAA2B,MAC1B;AAEJ,iBAAe,SAAS,0BAA0B;AAGlD;AAAA,IACE;AAAA,MACE;AAAA,MACA;AAAA,MACA,qBAAqB;AAAA,MACrB,iBAAiB;AAAA,IACnB;AAAA,IACA;AAAA,IACA;AAAA,IACA,iBACK;AAAA,MACC,SAAS,eAAe;AAAA,MACxB,WAAW,eAAe;AAAA,MAC1B;AAAA,IACF,IACA;AAAA,EACN;AAGA,oBAAkB;AAAA,IAChB,cAAc,aAAa,KAAK,IAAI;AAAA,IACpC,UAAU,CAAC,GAAG,cAAc,GAAG,cAAc;AAAA,IAC7C;AAAA,IACA,OAAO;AAAA,MACL;AAAA,MACA;AAAA,IACF;AAAA,IACA,QAAQ;AAAA,MACN;AAAA,MACA,KAAK,KAAK,IAAI;AAAA,IAChB;AAAA,IACA,WAAW;AAAA,EACb,CAAC;AAED,SAAO;AAAA,IACL,SAAS;AAAA,MACP,GAAG;AAAA,MACH,SAAS,wBAAwB,SAAS,OAAO;AAAA,MACjD,OAAO;AAAA,QACL,cAAc;AAAA,QACd,eAAe;AAAA,QACf,yBAAyB;AAAA,QACzB,6BAA6B;AAAA,MAC/B;AAAA,IACF;AAAA,IACA;AAAA,IACA;AAAA,IACA,MAAM;AAAA,IACN,MAAM,WAAW;AAAA,EACnB;AACF;AAEA,SAAS,wBAAwB,cAA2B;AAE1D,SAAO,cAAc,aAAa;AACpC;AAEA,SAAS,0BAA0B,OAAuB;AAExD,aAAW,kBAAkB,OAAO,OAAO,MAAM,GAAG;AAClD,UAAM,YAAY,eAAe,KAAK,CAAC,MAAW,EAAE,UAAU,KAAK;AACnE,QAAI,WAAW;AACb,aAAO,UAAU,wBAAwB;AAAA,IAC3C;AAAA,EACF;AAEA,SAAO;AACT;AAEA,SAAS,2BAA2B,OAAuB;AAEzD,aAAW,kBAAkB,OAAO,OAAO,MAAM,GAAG;AAClD,UAAM,YAAY,eAAe,KAAK,CAAC,MAAW,EAAE,UAAU,KAAK;AACnE,QAAI,WAAW;AACb,aAAO,UAAU,yBAAyB;AAAA,IAC5C;AAAA,EACF;AAEA,SAAO;AACT;AAGA,eAAsB,WACpB,cACA,UACA,eAAyB,CAAC,GAC1B,QAC2B;AAE3B,SAAO;AAAA,IACL;AAAA,IACA;AAAA,IACA;AAAA;AAAA,IACA,CAAC;AAAA;AAAA,IACD,UAAU,IAAI,gBAAgB,EAAE;AAAA,IAChC;AAAA,MACE,UAAU;AAAA,MACV,OAAO;AAAA,MACP,qBAAqB;AAAA,IACvB;AAAA,EACF;AACF;AAKA,eAAsB,WAAW;AAAA,EAC/B,eAAe,CAAC;AAAA,EAChB;AAAA,EACA;AAAA,EACA,sBAAsB;AAAA,EACtB;AACF,GAM8B;AAC5B,QAAM,WAAW;AAAA,IACf;AAAA,MACE,SAAS,EAAE,MAAM,QAAQ,SAAS,WAAW;AAAA,MAC7C,MAAM;AAAA,MACN,MAAM,WAAW;AAAA,IACnB;AAAA,EACF;AAEA,SAAO,WAAW,SAAS,UAAU,cAAc,MAAM;AAC3D;",
6
- "names": ["streamState", "response"]
4
+ "sourcesContent": ["/**\n * LLM service facade \u2014 thin re-export from decomposed modules in ./llm/.\n *\n * All public API is preserved for backward compatibility with existing importers.\n * New code should import from '@services/llm' (or the specific submodule) directly.\n */\nexport {\n // Constants\n API_ERROR_MESSAGE_PREFIX,\n PROMPT_TOO_LONG_ERROR_MESSAGE,\n CREDIT_BALANCE_TOO_LOW_ERROR_MESSAGE,\n INVALID_API_KEY_ERROR_MESSAGE,\n NO_CONTENT_MESSAGE,\n MAIN_QUERY_TEMPERATURE,\n\n // MintoContext\n generateMintoContext,\n refreshMintoContext,\n\n // Anthropic provider\n fetchAnthropicModels,\n verifyApiKey,\n getAnthropicClient,\n resetAnthropicClient,\n userMessageToMessageParam,\n assistantMessageToMessageParam,\n\n // Dispatch\n queryLLM,\n formatSystemPromptWithContext,\n queryModel,\n queryQuick,\n} from './llm'\n"],
5
+ "mappings": "AAMA;AAAA,EAEE;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EAGA;AAAA,EACA;AAAA,EAGA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,EAGA;AAAA,EACA;AAAA,EACA;AAAA,EACA;AAAA,OACK;",
6
+ "names": []
7
7
  }
@@ -61,7 +61,8 @@ async function testResponsesAPI(config, baseURL, startTime) {
61
61
  const response = await fetch(testURL, {
62
62
  method: "POST",
63
63
  headers,
64
- body: JSON.stringify(testPayload)
64
+ body: JSON.stringify(testPayload),
65
+ signal: AbortSignal.timeout(3e4)
65
66
  });
66
67
  const responseTime = Date.now() - startTime;
67
68
  if (response.ok) {
@@ -165,7 +166,8 @@ async function testChatCompletionsAPI(config, baseURL, startTime) {
165
166
  const response = await fetch(testURL, {
166
167
  method: "POST",
167
168
  headers,
168
- body: JSON.stringify(testPayload)
169
+ body: JSON.stringify(testPayload),
170
+ signal: AbortSignal.timeout(3e4)
169
171
  });
170
172
  const responseTime = Date.now() - startTime;
171
173
  if (response.ok) {
@@ -1,7 +1,7 @@
1
1
  {
2
2
  "version": 3,
3
3
  "sources": ["../../src/services/gpt5ConnectionTest.ts"],
4
- "sourcesContent": ["/**\n * \uD83D\uDD25 GPT-5 Connection Test Service\n *\n * Specialized connection testing for GPT-5 models that supports both\n * Responses API and Chat Completions API with proper fallback handling.\n */\n\nimport { getModelFeatures } from './openai'\n\nexport interface ConnectionTestResult {\n success: boolean\n message: string\n endpoint?: string\n details?: string\n apiUsed?: 'responses' | 'chat_completions'\n responseTime?: number\n}\n\nexport interface GPT5TestConfig {\n model: string\n apiKey: string\n baseURL?: string\n maxTokens?: number\n provider?: string\n}\n\n/**\n * Test GPT-5 model connection with intelligent API selection\n */\nexport async function testGPT5Connection(\n config: GPT5TestConfig,\n): Promise<ConnectionTestResult> {\n const startTime = Date.now()\n\n // Validate configuration\n if (!config.model || !config.apiKey) {\n return {\n success: false,\n message: 'Invalid configuration',\n details: 'Model name and API key are required',\n }\n }\n\n const isGPT5 = config.model.toLowerCase().includes('gpt-5')\n const modelFeatures = getModelFeatures(config.model)\n const baseURL = config.baseURL || 'https://api.openai.com/v1'\n const isOfficialOpenAI =\n !config.baseURL || config.baseURL.includes('api.openai.com')\n\n console.log(`\uD83D\uDD27 Testing GPT-5 connection for model: ${config.model}`)\n console.log(`\uD83D\uDD27 Base URL: ${baseURL}`)\n console.log(`\uD83D\uDD27 Official OpenAI: ${isOfficialOpenAI}`)\n console.log(\n `\uD83D\uDD27 Supports Responses API: ${modelFeatures.supportsResponsesAPI}`,\n )\n\n // Try Responses API first for official GPT-5 models\n if (isGPT5 && modelFeatures.supportsResponsesAPI && isOfficialOpenAI) {\n console.log(`\uD83D\uDE80 Attempting Responses API for ${config.model}`)\n const responsesResult = await testResponsesAPI(config, baseURL, startTime)\n\n if (responsesResult.success) {\n console.log(`\u2705 Responses API test successful for ${config.model}`)\n return responsesResult\n } else {\n console.log(\n `\u26A0\uFE0F Responses API failed, falling back to Chat Completions: ${responsesResult.details}`,\n )\n }\n }\n\n // Fallback to Chat Completions API\n console.log(`\uD83D\uDD04 Using Chat Completions API for ${config.model}`)\n return await testChatCompletionsAPI(config, baseURL, startTime)\n}\n\n/**\n * Test using GPT-5 Responses API\n */\nasync function testResponsesAPI(\n config: GPT5TestConfig,\n baseURL: string,\n startTime: number,\n): Promise<ConnectionTestResult> {\n const testURL = `${baseURL.replace(/\\/+$/, '')}/responses`\n\n const testPayload = {\n model: config.model,\n input: [\n {\n role: 'user',\n content:\n 'Please respond with exactly \"YES\" (in capital letters) to confirm this connection is working.',\n },\n ],\n max_completion_tokens: Math.max(config.maxTokens || 8192, 8192),\n temperature: 1, // GPT-5 requirement\n reasoning: {\n effort: 'low', // Fast response for connection test\n },\n }\n\n const headers = {\n 'Content-Type': 'application/json',\n Authorization: `Bearer ${config.apiKey}`,\n }\n\n console.log(`\uD83D\uDD27 Responses API URL: ${testURL}`)\n console.log(`\uD83D\uDD27 Responses API payload:`, JSON.stringify(testPayload, null, 2))\n\n try {\n const response = await fetch(testURL, {\n method: 'POST',\n headers,\n body: JSON.stringify(testPayload),\n })\n\n const responseTime = Date.now() - startTime\n\n if (response.ok) {\n const data = await response.json()\n console.log(`\u2705 Responses API successful response:`, data)\n\n // Extract content from Responses API format\n let responseContent = ''\n if (data.output_text) {\n responseContent = data.output_text\n } else if (data.output && Array.isArray(data.output)) {\n // Extract from structured output format\n const messageOutput = data.output.find(item => item.type === 'message')\n if (messageOutput && messageOutput.content) {\n const textContent = messageOutput.content.find(\n c => c.type === 'output_text',\n )\n responseContent = textContent?.text || ''\n }\n }\n\n const containsYes = responseContent.toLowerCase().includes('yes')\n\n if (containsYes) {\n return {\n success: true,\n message: '\u2705 GPT-5 Responses API connection successful',\n endpoint: '/responses',\n details: `Model responded correctly: \"${responseContent.trim()}\"`,\n apiUsed: 'responses',\n responseTime,\n }\n } else {\n return {\n success: false,\n message: '\u26A0\uFE0F Responses API connected but unexpected response',\n endpoint: '/responses',\n details: `Expected \"YES\" but got: \"${responseContent.trim() || '(empty response)'}\"`,\n apiUsed: 'responses',\n responseTime,\n }\n }\n } else {\n const errorData = await response.json().catch(() => null)\n const errorMessage =\n errorData?.error?.message || errorData?.message || response.statusText\n\n console.log(`\u274C Responses API error (${response.status}):`, errorData)\n\n return {\n success: false,\n message: `\u274C Responses API failed (${response.status})`,\n endpoint: '/responses',\n details: `Error: ${errorMessage}`,\n apiUsed: 'responses',\n responseTime: Date.now() - startTime,\n }\n }\n } catch (error) {\n console.log(`\u274C Responses API connection error:`, error)\n\n return {\n success: false,\n message: '\u274C Responses API connection failed',\n endpoint: '/responses',\n details: error instanceof Error ? error.message : String(error),\n apiUsed: 'responses',\n responseTime: Date.now() - startTime,\n }\n }\n}\n\n/**\n * Test using Chat Completions API with GPT-5 compatibility\n */\nasync function testChatCompletionsAPI(\n config: GPT5TestConfig,\n baseURL: string,\n startTime: number,\n): Promise<ConnectionTestResult> {\n const testURL = `${baseURL.replace(/\\/+$/, '')}/chat/completions`\n\n const isGPT5 = config.model.toLowerCase().includes('gpt-5')\n\n // Create test payload with GPT-5 compatibility\n const testPayload: any = {\n model: config.model,\n messages: [\n {\n role: 'user',\n content:\n 'Please respond with exactly \"YES\" (in capital letters) to confirm this connection is working.',\n },\n ],\n temperature: isGPT5 ? 1 : 0, // GPT-5 requires temperature=1\n stream: false,\n }\n\n // \uD83D\uDD27 Apply GPT-5 parameter transformations\n if (isGPT5) {\n testPayload.max_completion_tokens = Math.max(config.maxTokens || 8192, 8192)\n delete testPayload.max_tokens // \uD83D\uDD25 CRITICAL: Remove max_tokens for GPT-5\n console.log(\n `\uD83D\uDD27 GPT-5 mode: Using max_completion_tokens = ${testPayload.max_completion_tokens}`,\n )\n } else {\n testPayload.max_tokens = Math.max(config.maxTokens || 8192, 8192)\n }\n\n const headers = {\n 'Content-Type': 'application/json',\n }\n\n // Add provider-specific headers\n if (config.provider === 'azure') {\n headers['api-key'] = config.apiKey\n } else {\n headers['Authorization'] = `Bearer ${config.apiKey}`\n }\n\n console.log(`\uD83D\uDD27 Chat Completions URL: ${testURL}`)\n console.log(\n `\uD83D\uDD27 Chat Completions payload:`,\n JSON.stringify(testPayload, null, 2),\n )\n\n try {\n const response = await fetch(testURL, {\n method: 'POST',\n headers,\n body: JSON.stringify(testPayload),\n })\n\n const responseTime = Date.now() - startTime\n\n if (response.ok) {\n const data = await response.json()\n console.log(`\u2705 Chat Completions successful response:`, data)\n\n const responseContent = data.choices?.[0]?.message?.content || ''\n const containsYes = responseContent.toLowerCase().includes('yes')\n\n if (containsYes) {\n return {\n success: true,\n message: `\u2705 ${isGPT5 ? 'GPT-5' : 'Model'} Chat Completions connection successful`,\n endpoint: '/chat/completions',\n details: `Model responded correctly: \"${responseContent.trim()}\"`,\n apiUsed: 'chat_completions',\n responseTime,\n }\n } else {\n return {\n success: false,\n message: '\u26A0\uFE0F Chat Completions connected but unexpected response',\n endpoint: '/chat/completions',\n details: `Expected \"YES\" but got: \"${responseContent.trim() || '(empty response)'}\"`,\n apiUsed: 'chat_completions',\n responseTime,\n }\n }\n } else {\n const errorData = await response.json().catch(() => null)\n const errorMessage =\n errorData?.error?.message || errorData?.message || response.statusText\n\n console.log(`\u274C Chat Completions error (${response.status}):`, errorData)\n\n // \uD83D\uDD27 Provide specific guidance for GPT-5 errors\n let details = `Error: ${errorMessage}`\n if (\n response.status === 400 &&\n errorMessage.includes('max_tokens') &&\n isGPT5\n ) {\n details +=\n '\\n\\n\uD83D\uDD27 GPT-5 Fix Applied: This error suggests a parameter compatibility issue. Please check if the provider supports GPT-5 with max_completion_tokens.'\n }\n\n return {\n success: false,\n message: `\u274C Chat Completions failed (${response.status})`,\n endpoint: '/chat/completions',\n details: details,\n apiUsed: 'chat_completions',\n responseTime: Date.now() - startTime,\n }\n }\n } catch (error) {\n console.log(`\u274C Chat Completions connection error:`, error)\n\n return {\n success: false,\n message: '\u274C Chat Completions connection failed',\n endpoint: '/chat/completions',\n details: error instanceof Error ? error.message : String(error),\n apiUsed: 'chat_completions',\n responseTime: Date.now() - startTime,\n }\n }\n}\n\n/**\n * Quick validation for GPT-5 configuration\n */\nexport function validateGPT5Config(config: GPT5TestConfig): {\n valid: boolean\n errors: string[]\n} {\n console.log(`\uD83D\uDD27 validateGPT5Config called with:`, {\n model: config.model,\n hasApiKey: !!config.apiKey,\n baseURL: config.baseURL,\n provider: config.provider,\n })\n\n const errors: string[] = []\n\n if (!config.model) {\n errors.push('Model name is required')\n }\n\n if (!config.apiKey) {\n errors.push('API key is required')\n }\n\n if (config.apiKey && config.apiKey.length < 10) {\n errors.push('API key appears to be invalid (too short)')\n }\n\n const isGPT5 = config.model?.toLowerCase().includes('gpt-5')\n if (isGPT5) {\n console.log(\n `\uD83D\uDD27 GPT-5 validation: model=${config.model}, maxTokens=${config.maxTokens}`,\n )\n\n if (config.maxTokens && config.maxTokens < 1000) {\n errors.push('GPT-5 models typically require at least 1000 max tokens')\n }\n\n // \u5B8C\u5168\u79FB\u9664\u7B2C\u4E09\u65B9provider\u9650\u5236\uFF0C\u5141\u8BB8\u6240\u6709\u4EE3\u7406\u4E2D\u8F6C\u7AD9\u4F7F\u7528GPT-5\n console.log(`\uD83D\uDD27 No third-party restrictions applied for GPT-5`)\n }\n\n console.log(`\uD83D\uDD27 Validation result:`, { valid: errors.length === 0, errors })\n\n return {\n valid: errors.length === 0,\n errors,\n }\n}\n"],
5
- "mappings": "AAOA,SAAS,wBAAwB;AAsBjC,eAAsB,mBACpB,QAC+B;AAC/B,QAAM,YAAY,KAAK,IAAI;AAG3B,MAAI,CAAC,OAAO,SAAS,CAAC,OAAO,QAAQ;AACnC,WAAO;AAAA,MACL,SAAS;AAAA,MACT,SAAS;AAAA,MACT,SAAS;AAAA,IACX;AAAA,EACF;AAEA,QAAM,SAAS,OAAO,MAAM,YAAY,EAAE,SAAS,OAAO;AAC1D,QAAM,gBAAgB,iBAAiB,OAAO,KAAK;AACnD,QAAM,UAAU,OAAO,WAAW;AAClC,QAAM,mBACJ,CAAC,OAAO,WAAW,OAAO,QAAQ,SAAS,gBAAgB;AAE7D,UAAQ,IAAI,iDAA0C,OAAO,KAAK,EAAE;AACpE,UAAQ,IAAI,uBAAgB,OAAO,EAAE;AACrC,UAAQ,IAAI,8BAAuB,gBAAgB,EAAE;AACrD,UAAQ;AAAA,IACN,qCAA8B,cAAc,oBAAoB;AAAA,EAClE;AAGA,MAAI,UAAU,cAAc,wBAAwB,kBAAkB;AACpE,YAAQ,IAAI,0CAAmC,OAAO,KAAK,EAAE;AAC7D,UAAM,kBAAkB,MAAM,iBAAiB,QAAQ,SAAS,SAAS;AAEzE,QAAI,gBAAgB,SAAS;AAC3B,cAAQ,IAAI,4CAAuC,OAAO,KAAK,EAAE;AACjE,aAAO;AAAA,IACT,OAAO;AACL,cAAQ;AAAA,QACN,wEAA8D,gBAAgB,OAAO;AAAA,MACvF;AAAA,IACF;AAAA,EACF;AAGA,UAAQ,IAAI,4CAAqC,OAAO,KAAK,EAAE;AAC/D,SAAO,MAAM,uBAAuB,QAAQ,SAAS,SAAS;AAChE;AAKA,eAAe,iBACb,QACA,SACA,WAC+B;AAC/B,QAAM,UAAU,GAAG,QAAQ,QAAQ,QAAQ,EAAE,CAAC;AAE9C,QAAM,cAAc;AAAA,IAClB,OAAO,OAAO;AAAA,IACd,OAAO;AAAA,MACL;AAAA,QACE,MAAM;AAAA,QACN,SACE;AAAA,MACJ;AAAA,IACF;AAAA,IACA,uBAAuB,KAAK,IAAI,OAAO,aAAa,MAAM,IAAI;AAAA,IAC9D,aAAa;AAAA;AAAA,IACb,WAAW;AAAA,MACT,QAAQ;AAAA;AAAA,IACV;AAAA,EACF;AAEA,QAAM,UAAU;AAAA,IACd,gBAAgB;AAAA,IAChB,eAAe,UAAU,OAAO,MAAM;AAAA,EACxC;AAEA,UAAQ,IAAI,gCAAyB,OAAO,EAAE;AAC9C,UAAQ,IAAI,oCAA6B,KAAK,UAAU,aAAa,MAAM,CAAC,CAAC;AAE7E,MAAI;AACF,UAAM,WAAW,MAAM,MAAM,SAAS;AAAA,MACpC,QAAQ;AAAA,MACR;AAAA,MACA,MAAM,KAAK,UAAU,WAAW;AAAA,IAClC,CAAC;AAED,UAAM,eAAe,KAAK,IAAI,IAAI;AAElC,QAAI,SAAS,IAAI;AACf,YAAM,OAAO,MAAM,SAAS,KAAK;AACjC,cAAQ,IAAI,6CAAwC,IAAI;AAGxD,UAAI,kBAAkB;AACtB,UAAI,KAAK,aAAa;AACpB,0BAAkB,KAAK;AAAA,MACzB,WAAW,KAAK,UAAU,MAAM,QAAQ,KAAK,MAAM,GAAG;AAEpD,cAAM,gBAAgB,KAAK,OAAO,KAAK,UAAQ,KAAK,SAAS,SAAS;AACtE,YAAI,iBAAiB,cAAc,SAAS;AAC1C,gBAAM,cAAc,cAAc,QAAQ;AAAA,YACxC,OAAK,EAAE,SAAS;AAAA,UAClB;AACA,4BAAkB,aAAa,QAAQ;AAAA,QACzC;AAAA,MACF;AAEA,YAAM,cAAc,gBAAgB,YAAY,EAAE,SAAS,KAAK;AAEhE,UAAI,aAAa;AACf,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS;AAAA,UACT,UAAU;AAAA,UACV,SAAS,+BAA+B,gBAAgB,KAAK,CAAC;AAAA,UAC9D,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF,OAAO;AACL,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS;AAAA,UACT,UAAU;AAAA,UACV,SAAS,4BAA4B,gBAAgB,KAAK,KAAK,kBAAkB;AAAA,UACjF,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,YAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,IAAI;AACxD,YAAM,eACJ,WAAW,OAAO,WAAW,WAAW,WAAW,SAAS;AAE9D,cAAQ,IAAI,+BAA0B,SAAS,MAAM,MAAM,SAAS;AAEpE,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS,gCAA2B,SAAS,MAAM;AAAA,QACnD,UAAU;AAAA,QACV,SAAS,UAAU,YAAY;AAAA,QAC/B,SAAS;AAAA,QACT,cAAc,KAAK,IAAI,IAAI;AAAA,MAC7B;AAAA,IACF;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,IAAI,0CAAqC,KAAK;AAEtD,WAAO;AAAA,MACL,SAAS;AAAA,MACT,SAAS;AAAA,MACT,UAAU;AAAA,MACV,SAAS,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D,SAAS;AAAA,MACT,cAAc,KAAK,IAAI,IAAI;AAAA,IAC7B;AAAA,EACF;AACF;AAKA,eAAe,uBACb,QACA,SACA,WAC+B;AAC/B,QAAM,UAAU,GAAG,QAAQ,QAAQ,QAAQ,EAAE,CAAC;AAE9C,QAAM,SAAS,OAAO,MAAM,YAAY,EAAE,SAAS,OAAO;AAG1D,QAAM,cAAmB;AAAA,IACvB,OAAO,OAAO;AAAA,IACd,UAAU;AAAA,MACR;AAAA,QACE,MAAM;AAAA,QACN,SACE;AAAA,MACJ;AAAA,IACF;AAAA,IACA,aAAa,SAAS,IAAI;AAAA;AAAA,IAC1B,QAAQ;AAAA,EACV;AAGA,MAAI,QAAQ;AACV,gBAAY,wBAAwB,KAAK,IAAI,OAAO,aAAa,MAAM,IAAI;AAC3E,WAAO,YAAY;AACnB,YAAQ;AAAA,MACN,uDAAgD,YAAY,qBAAqB;AAAA,IACnF;AAAA,EACF,OAAO;AACL,gBAAY,aAAa,KAAK,IAAI,OAAO,aAAa,MAAM,IAAI;AAAA,EAClE;AAEA,QAAM,UAAU;AAAA,IACd,gBAAgB;AAAA,EAClB;AAGA,MAAI,OAAO,aAAa,SAAS;AAC/B,YAAQ,SAAS,IAAI,OAAO;AAAA,EAC9B,OAAO;AACL,YAAQ,eAAe,IAAI,UAAU,OAAO,MAAM;AAAA,EACpD;AAEA,UAAQ,IAAI,mCAA4B,OAAO,EAAE;AACjD,UAAQ;AAAA,IACN;AAAA,IACA,KAAK,UAAU,aAAa,MAAM,CAAC;AAAA,EACrC;AAEA,MAAI;AACF,UAAM,WAAW,MAAM,MAAM,SAAS;AAAA,MACpC,QAAQ;AAAA,MACR;AAAA,MACA,MAAM,KAAK,UAAU,WAAW;AAAA,IAClC,CAAC;AAED,UAAM,eAAe,KAAK,IAAI,IAAI;AAElC,QAAI,SAAS,IAAI;AACf,YAAM,OAAO,MAAM,SAAS,KAAK;AACjC,cAAQ,IAAI,gDAA2C,IAAI;AAE3D,YAAM,kBAAkB,KAAK,UAAU,CAAC,GAAG,SAAS,WAAW;AAC/D,YAAM,cAAc,gBAAgB,YAAY,EAAE,SAAS,KAAK;AAEhE,UAAI,aAAa;AACf,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS,UAAK,SAAS,UAAU,OAAO;AAAA,UACxC,UAAU;AAAA,UACV,SAAS,+BAA+B,gBAAgB,KAAK,CAAC;AAAA,UAC9D,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF,OAAO;AACL,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS;AAAA,UACT,UAAU;AAAA,UACV,SAAS,4BAA4B,gBAAgB,KAAK,KAAK,kBAAkB;AAAA,UACjF,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,YAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,IAAI;AACxD,YAAM,eACJ,WAAW,OAAO,WAAW,WAAW,WAAW,SAAS;AAE9D,cAAQ,IAAI,kCAA6B,SAAS,MAAM,MAAM,SAAS;AAGvE,UAAI,UAAU,UAAU,YAAY;AACpC,UACE,SAAS,WAAW,OACpB,aAAa,SAAS,YAAY,KAClC,QACA;AACA,mBACE;AAAA,MACJ;AAEA,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS,mCAA8B,SAAS,MAAM;AAAA,QACtD,UAAU;AAAA,QACV;AAAA,QACA,SAAS;AAAA,QACT,cAAc,KAAK,IAAI,IAAI;AAAA,MAC7B;AAAA,IACF;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,IAAI,6CAAwC,KAAK;AAEzD,WAAO;AAAA,MACL,SAAS;AAAA,MACT,SAAS;AAAA,MACT,UAAU;AAAA,MACV,SAAS,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D,SAAS;AAAA,MACT,cAAc,KAAK,IAAI,IAAI;AAAA,IAC7B;AAAA,EACF;AACF;AAKO,SAAS,mBAAmB,QAGjC;AACA,UAAQ,IAAI,6CAAsC;AAAA,IAChD,OAAO,OAAO;AAAA,IACd,WAAW,CAAC,CAAC,OAAO;AAAA,IACpB,SAAS,OAAO;AAAA,IAChB,UAAU,OAAO;AAAA,EACnB,CAAC;AAED,QAAM,SAAmB,CAAC;AAE1B,MAAI,CAAC,OAAO,OAAO;AACjB,WAAO,KAAK,wBAAwB;AAAA,EACtC;AAEA,MAAI,CAAC,OAAO,QAAQ;AAClB,WAAO,KAAK,qBAAqB;AAAA,EACnC;AAEA,MAAI,OAAO,UAAU,OAAO,OAAO,SAAS,IAAI;AAC9C,WAAO,KAAK,2CAA2C;AAAA,EACzD;AAEA,QAAM,SAAS,OAAO,OAAO,YAAY,EAAE,SAAS,OAAO;AAC3D,MAAI,QAAQ;AACV,YAAQ;AAAA,MACN,qCAA8B,OAAO,KAAK,eAAe,OAAO,SAAS;AAAA,IAC3E;AAEA,QAAI,OAAO,aAAa,OAAO,YAAY,KAAM;AAC/C,aAAO,KAAK,yDAAyD;AAAA,IACvE;AAGA,YAAQ,IAAI,yDAAkD;AAAA,EAChE;AAEA,UAAQ,IAAI,gCAAyB,EAAE,OAAO,OAAO,WAAW,GAAG,OAAO,CAAC;AAE3E,SAAO;AAAA,IACL,OAAO,OAAO,WAAW;AAAA,IACzB;AAAA,EACF;AACF;",
4
+ "sourcesContent": ["/**\n * \uD83D\uDD25 GPT-5 Connection Test Service\n *\n * Specialized connection testing for GPT-5 models that supports both\n * Responses API and Chat Completions API with proper fallback handling.\n */\n\nimport { getModelFeatures } from './openai'\n\nexport interface ConnectionTestResult {\n success: boolean\n message: string\n endpoint?: string\n details?: string\n apiUsed?: 'responses' | 'chat_completions'\n responseTime?: number\n}\n\nexport interface GPT5TestConfig {\n model: string\n apiKey: string\n baseURL?: string\n maxTokens?: number\n provider?: string\n}\n\n/**\n * Test GPT-5 model connection with intelligent API selection\n */\nexport async function testGPT5Connection(\n config: GPT5TestConfig,\n): Promise<ConnectionTestResult> {\n const startTime = Date.now()\n\n // Validate configuration\n if (!config.model || !config.apiKey) {\n return {\n success: false,\n message: 'Invalid configuration',\n details: 'Model name and API key are required',\n }\n }\n\n const isGPT5 = config.model.toLowerCase().includes('gpt-5')\n const modelFeatures = getModelFeatures(config.model)\n const baseURL = config.baseURL || 'https://api.openai.com/v1'\n const isOfficialOpenAI =\n !config.baseURL || config.baseURL.includes('api.openai.com')\n\n console.log(`\uD83D\uDD27 Testing GPT-5 connection for model: ${config.model}`)\n console.log(`\uD83D\uDD27 Base URL: ${baseURL}`)\n console.log(`\uD83D\uDD27 Official OpenAI: ${isOfficialOpenAI}`)\n console.log(\n `\uD83D\uDD27 Supports Responses API: ${modelFeatures.supportsResponsesAPI}`,\n )\n\n // Try Responses API first for official GPT-5 models\n if (isGPT5 && modelFeatures.supportsResponsesAPI && isOfficialOpenAI) {\n console.log(`\uD83D\uDE80 Attempting Responses API for ${config.model}`)\n const responsesResult = await testResponsesAPI(config, baseURL, startTime)\n\n if (responsesResult.success) {\n console.log(`\u2705 Responses API test successful for ${config.model}`)\n return responsesResult\n } else {\n console.log(\n `\u26A0\uFE0F Responses API failed, falling back to Chat Completions: ${responsesResult.details}`,\n )\n }\n }\n\n // Fallback to Chat Completions API\n console.log(`\uD83D\uDD04 Using Chat Completions API for ${config.model}`)\n return await testChatCompletionsAPI(config, baseURL, startTime)\n}\n\n/**\n * Test using GPT-5 Responses API\n */\nasync function testResponsesAPI(\n config: GPT5TestConfig,\n baseURL: string,\n startTime: number,\n): Promise<ConnectionTestResult> {\n const testURL = `${baseURL.replace(/\\/+$/, '')}/responses`\n\n const testPayload = {\n model: config.model,\n input: [\n {\n role: 'user',\n content:\n 'Please respond with exactly \"YES\" (in capital letters) to confirm this connection is working.',\n },\n ],\n max_completion_tokens: Math.max(config.maxTokens || 8192, 8192),\n temperature: 1, // GPT-5 requirement\n reasoning: {\n effort: 'low', // Fast response for connection test\n },\n }\n\n const headers = {\n 'Content-Type': 'application/json',\n Authorization: `Bearer ${config.apiKey}`,\n }\n\n console.log(`\uD83D\uDD27 Responses API URL: ${testURL}`)\n console.log(`\uD83D\uDD27 Responses API payload:`, JSON.stringify(testPayload, null, 2))\n\n try {\n const response = await fetch(testURL, {\n method: 'POST',\n headers,\n body: JSON.stringify(testPayload),\n signal: AbortSignal.timeout(30_000),\n })\n\n const responseTime = Date.now() - startTime\n\n if (response.ok) {\n const data = await response.json()\n console.log(`\u2705 Responses API successful response:`, data)\n\n // Extract content from Responses API format\n let responseContent = ''\n if (data.output_text) {\n responseContent = data.output_text\n } else if (data.output && Array.isArray(data.output)) {\n // Extract from structured output format\n const messageOutput = data.output.find(item => item.type === 'message')\n if (messageOutput && messageOutput.content) {\n const textContent = messageOutput.content.find(\n c => c.type === 'output_text',\n )\n responseContent = textContent?.text || ''\n }\n }\n\n const containsYes = responseContent.toLowerCase().includes('yes')\n\n if (containsYes) {\n return {\n success: true,\n message: '\u2705 GPT-5 Responses API connection successful',\n endpoint: '/responses',\n details: `Model responded correctly: \"${responseContent.trim()}\"`,\n apiUsed: 'responses',\n responseTime,\n }\n } else {\n return {\n success: false,\n message: '\u26A0\uFE0F Responses API connected but unexpected response',\n endpoint: '/responses',\n details: `Expected \"YES\" but got: \"${responseContent.trim() || '(empty response)'}\"`,\n apiUsed: 'responses',\n responseTime,\n }\n }\n } else {\n const errorData = await response.json().catch(() => null)\n const errorMessage =\n errorData?.error?.message || errorData?.message || response.statusText\n\n console.log(`\u274C Responses API error (${response.status}):`, errorData)\n\n return {\n success: false,\n message: `\u274C Responses API failed (${response.status})`,\n endpoint: '/responses',\n details: `Error: ${errorMessage}`,\n apiUsed: 'responses',\n responseTime: Date.now() - startTime,\n }\n }\n } catch (error) {\n console.log(`\u274C Responses API connection error:`, error)\n\n return {\n success: false,\n message: '\u274C Responses API connection failed',\n endpoint: '/responses',\n details: error instanceof Error ? error.message : String(error),\n apiUsed: 'responses',\n responseTime: Date.now() - startTime,\n }\n }\n}\n\n/**\n * Test using Chat Completions API with GPT-5 compatibility\n */\nasync function testChatCompletionsAPI(\n config: GPT5TestConfig,\n baseURL: string,\n startTime: number,\n): Promise<ConnectionTestResult> {\n const testURL = `${baseURL.replace(/\\/+$/, '')}/chat/completions`\n\n const isGPT5 = config.model.toLowerCase().includes('gpt-5')\n\n // Create test payload with GPT-5 compatibility\n const testPayload: any = {\n model: config.model,\n messages: [\n {\n role: 'user',\n content:\n 'Please respond with exactly \"YES\" (in capital letters) to confirm this connection is working.',\n },\n ],\n temperature: isGPT5 ? 1 : 0, // GPT-5 requires temperature=1\n stream: false,\n }\n\n // \uD83D\uDD27 Apply GPT-5 parameter transformations\n if (isGPT5) {\n testPayload.max_completion_tokens = Math.max(config.maxTokens || 8192, 8192)\n delete testPayload.max_tokens // \uD83D\uDD25 CRITICAL: Remove max_tokens for GPT-5\n console.log(\n `\uD83D\uDD27 GPT-5 mode: Using max_completion_tokens = ${testPayload.max_completion_tokens}`,\n )\n } else {\n testPayload.max_tokens = Math.max(config.maxTokens || 8192, 8192)\n }\n\n const headers = {\n 'Content-Type': 'application/json',\n }\n\n // Add provider-specific headers\n if (config.provider === 'azure') {\n headers['api-key'] = config.apiKey\n } else {\n headers['Authorization'] = `Bearer ${config.apiKey}`\n }\n\n console.log(`\uD83D\uDD27 Chat Completions URL: ${testURL}`)\n console.log(\n `\uD83D\uDD27 Chat Completions payload:`,\n JSON.stringify(testPayload, null, 2),\n )\n\n try {\n const response = await fetch(testURL, {\n method: 'POST',\n headers,\n body: JSON.stringify(testPayload),\n signal: AbortSignal.timeout(30_000),\n })\n\n const responseTime = Date.now() - startTime\n\n if (response.ok) {\n const data = await response.json()\n console.log(`\u2705 Chat Completions successful response:`, data)\n\n const responseContent = data.choices?.[0]?.message?.content || ''\n const containsYes = responseContent.toLowerCase().includes('yes')\n\n if (containsYes) {\n return {\n success: true,\n message: `\u2705 ${isGPT5 ? 'GPT-5' : 'Model'} Chat Completions connection successful`,\n endpoint: '/chat/completions',\n details: `Model responded correctly: \"${responseContent.trim()}\"`,\n apiUsed: 'chat_completions',\n responseTime,\n }\n } else {\n return {\n success: false,\n message: '\u26A0\uFE0F Chat Completions connected but unexpected response',\n endpoint: '/chat/completions',\n details: `Expected \"YES\" but got: \"${responseContent.trim() || '(empty response)'}\"`,\n apiUsed: 'chat_completions',\n responseTime,\n }\n }\n } else {\n const errorData = await response.json().catch(() => null)\n const errorMessage =\n errorData?.error?.message || errorData?.message || response.statusText\n\n console.log(`\u274C Chat Completions error (${response.status}):`, errorData)\n\n // \uD83D\uDD27 Provide specific guidance for GPT-5 errors\n let details = `Error: ${errorMessage}`\n if (\n response.status === 400 &&\n errorMessage.includes('max_tokens') &&\n isGPT5\n ) {\n details +=\n '\\n\\n\uD83D\uDD27 GPT-5 Fix Applied: This error suggests a parameter compatibility issue. Please check if the provider supports GPT-5 with max_completion_tokens.'\n }\n\n return {\n success: false,\n message: `\u274C Chat Completions failed (${response.status})`,\n endpoint: '/chat/completions',\n details: details,\n apiUsed: 'chat_completions',\n responseTime: Date.now() - startTime,\n }\n }\n } catch (error) {\n console.log(`\u274C Chat Completions connection error:`, error)\n\n return {\n success: false,\n message: '\u274C Chat Completions connection failed',\n endpoint: '/chat/completions',\n details: error instanceof Error ? error.message : String(error),\n apiUsed: 'chat_completions',\n responseTime: Date.now() - startTime,\n }\n }\n}\n\n/**\n * Quick validation for GPT-5 configuration\n */\nexport function validateGPT5Config(config: GPT5TestConfig): {\n valid: boolean\n errors: string[]\n} {\n console.log(`\uD83D\uDD27 validateGPT5Config called with:`, {\n model: config.model,\n hasApiKey: !!config.apiKey,\n baseURL: config.baseURL,\n provider: config.provider,\n })\n\n const errors: string[] = []\n\n if (!config.model) {\n errors.push('Model name is required')\n }\n\n if (!config.apiKey) {\n errors.push('API key is required')\n }\n\n if (config.apiKey && config.apiKey.length < 10) {\n errors.push('API key appears to be invalid (too short)')\n }\n\n const isGPT5 = config.model?.toLowerCase().includes('gpt-5')\n if (isGPT5) {\n console.log(\n `\uD83D\uDD27 GPT-5 validation: model=${config.model}, maxTokens=${config.maxTokens}`,\n )\n\n if (config.maxTokens && config.maxTokens < 1000) {\n errors.push('GPT-5 models typically require at least 1000 max tokens')\n }\n\n // \u5B8C\u5168\u79FB\u9664\u7B2C\u4E09\u65B9provider\u9650\u5236\uFF0C\u5141\u8BB8\u6240\u6709\u4EE3\u7406\u4E2D\u8F6C\u7AD9\u4F7F\u7528GPT-5\n console.log(`\uD83D\uDD27 No third-party restrictions applied for GPT-5`)\n }\n\n console.log(`\uD83D\uDD27 Validation result:`, { valid: errors.length === 0, errors })\n\n return {\n valid: errors.length === 0,\n errors,\n }\n}\n"],
5
+ "mappings": "AAOA,SAAS,wBAAwB;AAsBjC,eAAsB,mBACpB,QAC+B;AAC/B,QAAM,YAAY,KAAK,IAAI;AAG3B,MAAI,CAAC,OAAO,SAAS,CAAC,OAAO,QAAQ;AACnC,WAAO;AAAA,MACL,SAAS;AAAA,MACT,SAAS;AAAA,MACT,SAAS;AAAA,IACX;AAAA,EACF;AAEA,QAAM,SAAS,OAAO,MAAM,YAAY,EAAE,SAAS,OAAO;AAC1D,QAAM,gBAAgB,iBAAiB,OAAO,KAAK;AACnD,QAAM,UAAU,OAAO,WAAW;AAClC,QAAM,mBACJ,CAAC,OAAO,WAAW,OAAO,QAAQ,SAAS,gBAAgB;AAE7D,UAAQ,IAAI,iDAA0C,OAAO,KAAK,EAAE;AACpE,UAAQ,IAAI,uBAAgB,OAAO,EAAE;AACrC,UAAQ,IAAI,8BAAuB,gBAAgB,EAAE;AACrD,UAAQ;AAAA,IACN,qCAA8B,cAAc,oBAAoB;AAAA,EAClE;AAGA,MAAI,UAAU,cAAc,wBAAwB,kBAAkB;AACpE,YAAQ,IAAI,0CAAmC,OAAO,KAAK,EAAE;AAC7D,UAAM,kBAAkB,MAAM,iBAAiB,QAAQ,SAAS,SAAS;AAEzE,QAAI,gBAAgB,SAAS;AAC3B,cAAQ,IAAI,4CAAuC,OAAO,KAAK,EAAE;AACjE,aAAO;AAAA,IACT,OAAO;AACL,cAAQ;AAAA,QACN,wEAA8D,gBAAgB,OAAO;AAAA,MACvF;AAAA,IACF;AAAA,EACF;AAGA,UAAQ,IAAI,4CAAqC,OAAO,KAAK,EAAE;AAC/D,SAAO,MAAM,uBAAuB,QAAQ,SAAS,SAAS;AAChE;AAKA,eAAe,iBACb,QACA,SACA,WAC+B;AAC/B,QAAM,UAAU,GAAG,QAAQ,QAAQ,QAAQ,EAAE,CAAC;AAE9C,QAAM,cAAc;AAAA,IAClB,OAAO,OAAO;AAAA,IACd,OAAO;AAAA,MACL;AAAA,QACE,MAAM;AAAA,QACN,SACE;AAAA,MACJ;AAAA,IACF;AAAA,IACA,uBAAuB,KAAK,IAAI,OAAO,aAAa,MAAM,IAAI;AAAA,IAC9D,aAAa;AAAA;AAAA,IACb,WAAW;AAAA,MACT,QAAQ;AAAA;AAAA,IACV;AAAA,EACF;AAEA,QAAM,UAAU;AAAA,IACd,gBAAgB;AAAA,IAChB,eAAe,UAAU,OAAO,MAAM;AAAA,EACxC;AAEA,UAAQ,IAAI,gCAAyB,OAAO,EAAE;AAC9C,UAAQ,IAAI,oCAA6B,KAAK,UAAU,aAAa,MAAM,CAAC,CAAC;AAE7E,MAAI;AACF,UAAM,WAAW,MAAM,MAAM,SAAS;AAAA,MACpC,QAAQ;AAAA,MACR;AAAA,MACA,MAAM,KAAK,UAAU,WAAW;AAAA,MAChC,QAAQ,YAAY,QAAQ,GAAM;AAAA,IACpC,CAAC;AAED,UAAM,eAAe,KAAK,IAAI,IAAI;AAElC,QAAI,SAAS,IAAI;AACf,YAAM,OAAO,MAAM,SAAS,KAAK;AACjC,cAAQ,IAAI,6CAAwC,IAAI;AAGxD,UAAI,kBAAkB;AACtB,UAAI,KAAK,aAAa;AACpB,0BAAkB,KAAK;AAAA,MACzB,WAAW,KAAK,UAAU,MAAM,QAAQ,KAAK,MAAM,GAAG;AAEpD,cAAM,gBAAgB,KAAK,OAAO,KAAK,UAAQ,KAAK,SAAS,SAAS;AACtE,YAAI,iBAAiB,cAAc,SAAS;AAC1C,gBAAM,cAAc,cAAc,QAAQ;AAAA,YACxC,OAAK,EAAE,SAAS;AAAA,UAClB;AACA,4BAAkB,aAAa,QAAQ;AAAA,QACzC;AAAA,MACF;AAEA,YAAM,cAAc,gBAAgB,YAAY,EAAE,SAAS,KAAK;AAEhE,UAAI,aAAa;AACf,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS;AAAA,UACT,UAAU;AAAA,UACV,SAAS,+BAA+B,gBAAgB,KAAK,CAAC;AAAA,UAC9D,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF,OAAO;AACL,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS;AAAA,UACT,UAAU;AAAA,UACV,SAAS,4BAA4B,gBAAgB,KAAK,KAAK,kBAAkB;AAAA,UACjF,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,YAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,IAAI;AACxD,YAAM,eACJ,WAAW,OAAO,WAAW,WAAW,WAAW,SAAS;AAE9D,cAAQ,IAAI,+BAA0B,SAAS,MAAM,MAAM,SAAS;AAEpE,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS,gCAA2B,SAAS,MAAM;AAAA,QACnD,UAAU;AAAA,QACV,SAAS,UAAU,YAAY;AAAA,QAC/B,SAAS;AAAA,QACT,cAAc,KAAK,IAAI,IAAI;AAAA,MAC7B;AAAA,IACF;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,IAAI,0CAAqC,KAAK;AAEtD,WAAO;AAAA,MACL,SAAS;AAAA,MACT,SAAS;AAAA,MACT,UAAU;AAAA,MACV,SAAS,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D,SAAS;AAAA,MACT,cAAc,KAAK,IAAI,IAAI;AAAA,IAC7B;AAAA,EACF;AACF;AAKA,eAAe,uBACb,QACA,SACA,WAC+B;AAC/B,QAAM,UAAU,GAAG,QAAQ,QAAQ,QAAQ,EAAE,CAAC;AAE9C,QAAM,SAAS,OAAO,MAAM,YAAY,EAAE,SAAS,OAAO;AAG1D,QAAM,cAAmB;AAAA,IACvB,OAAO,OAAO;AAAA,IACd,UAAU;AAAA,MACR;AAAA,QACE,MAAM;AAAA,QACN,SACE;AAAA,MACJ;AAAA,IACF;AAAA,IACA,aAAa,SAAS,IAAI;AAAA;AAAA,IAC1B,QAAQ;AAAA,EACV;AAGA,MAAI,QAAQ;AACV,gBAAY,wBAAwB,KAAK,IAAI,OAAO,aAAa,MAAM,IAAI;AAC3E,WAAO,YAAY;AACnB,YAAQ;AAAA,MACN,uDAAgD,YAAY,qBAAqB;AAAA,IACnF;AAAA,EACF,OAAO;AACL,gBAAY,aAAa,KAAK,IAAI,OAAO,aAAa,MAAM,IAAI;AAAA,EAClE;AAEA,QAAM,UAAU;AAAA,IACd,gBAAgB;AAAA,EAClB;AAGA,MAAI,OAAO,aAAa,SAAS;AAC/B,YAAQ,SAAS,IAAI,OAAO;AAAA,EAC9B,OAAO;AACL,YAAQ,eAAe,IAAI,UAAU,OAAO,MAAM;AAAA,EACpD;AAEA,UAAQ,IAAI,mCAA4B,OAAO,EAAE;AACjD,UAAQ;AAAA,IACN;AAAA,IACA,KAAK,UAAU,aAAa,MAAM,CAAC;AAAA,EACrC;AAEA,MAAI;AACF,UAAM,WAAW,MAAM,MAAM,SAAS;AAAA,MACpC,QAAQ;AAAA,MACR;AAAA,MACA,MAAM,KAAK,UAAU,WAAW;AAAA,MAChC,QAAQ,YAAY,QAAQ,GAAM;AAAA,IACpC,CAAC;AAED,UAAM,eAAe,KAAK,IAAI,IAAI;AAElC,QAAI,SAAS,IAAI;AACf,YAAM,OAAO,MAAM,SAAS,KAAK;AACjC,cAAQ,IAAI,gDAA2C,IAAI;AAE3D,YAAM,kBAAkB,KAAK,UAAU,CAAC,GAAG,SAAS,WAAW;AAC/D,YAAM,cAAc,gBAAgB,YAAY,EAAE,SAAS,KAAK;AAEhE,UAAI,aAAa;AACf,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS,UAAK,SAAS,UAAU,OAAO;AAAA,UACxC,UAAU;AAAA,UACV,SAAS,+BAA+B,gBAAgB,KAAK,CAAC;AAAA,UAC9D,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF,OAAO;AACL,eAAO;AAAA,UACL,SAAS;AAAA,UACT,SAAS;AAAA,UACT,UAAU;AAAA,UACV,SAAS,4BAA4B,gBAAgB,KAAK,KAAK,kBAAkB;AAAA,UACjF,SAAS;AAAA,UACT;AAAA,QACF;AAAA,MACF;AAAA,IACF,OAAO;AACL,YAAM,YAAY,MAAM,SAAS,KAAK,EAAE,MAAM,MAAM,IAAI;AACxD,YAAM,eACJ,WAAW,OAAO,WAAW,WAAW,WAAW,SAAS;AAE9D,cAAQ,IAAI,kCAA6B,SAAS,MAAM,MAAM,SAAS;AAGvE,UAAI,UAAU,UAAU,YAAY;AACpC,UACE,SAAS,WAAW,OACpB,aAAa,SAAS,YAAY,KAClC,QACA;AACA,mBACE;AAAA,MACJ;AAEA,aAAO;AAAA,QACL,SAAS;AAAA,QACT,SAAS,mCAA8B,SAAS,MAAM;AAAA,QACtD,UAAU;AAAA,QACV;AAAA,QACA,SAAS;AAAA,QACT,cAAc,KAAK,IAAI,IAAI;AAAA,MAC7B;AAAA,IACF;AAAA,EACF,SAAS,OAAO;AACd,YAAQ,IAAI,6CAAwC,KAAK;AAEzD,WAAO;AAAA,MACL,SAAS;AAAA,MACT,SAAS;AAAA,MACT,UAAU;AAAA,MACV,SAAS,iBAAiB,QAAQ,MAAM,UAAU,OAAO,KAAK;AAAA,MAC9D,SAAS;AAAA,MACT,cAAc,KAAK,IAAI,IAAI;AAAA,IAC7B;AAAA,EACF;AACF;AAKO,SAAS,mBAAmB,QAGjC;AACA,UAAQ,IAAI,6CAAsC;AAAA,IAChD,OAAO,OAAO;AAAA,IACd,WAAW,CAAC,CAAC,OAAO;AAAA,IACpB,SAAS,OAAO;AAAA,IAChB,UAAU,OAAO;AAAA,EACnB,CAAC;AAED,QAAM,SAAmB,CAAC;AAE1B,MAAI,CAAC,OAAO,OAAO;AACjB,WAAO,KAAK,wBAAwB;AAAA,EACtC;AAEA,MAAI,CAAC,OAAO,QAAQ;AAClB,WAAO,KAAK,qBAAqB;AAAA,EACnC;AAEA,MAAI,OAAO,UAAU,OAAO,OAAO,SAAS,IAAI;AAC9C,WAAO,KAAK,2CAA2C;AAAA,EACzD;AAEA,QAAM,SAAS,OAAO,OAAO,YAAY,EAAE,SAAS,OAAO;AAC3D,MAAI,QAAQ;AACV,YAAQ;AAAA,MACN,qCAA8B,OAAO,KAAK,eAAe,OAAO,SAAS;AAAA,IAC3E;AAEA,QAAI,OAAO,aAAa,OAAO,YAAY,KAAM;AAC/C,aAAO,KAAK,yDAAyD;AAAA,IACvE;AAGA,YAAQ,IAAI,yDAAkD;AAAA,EAChE;AAEA,UAAQ,IAAI,gCAAyB,EAAE,OAAO,OAAO,WAAW,GAAG,OAAO,CAAC;AAE3E,SAAO;AAAA,IACL,OAAO,OAAO,WAAW;AAAA,IACzB;AAAA,EACF;AACF;",
6
6
  "names": []
7
7
  }