@danya-ai/cli 0.1.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (184) hide show
  1. package/LICENSE +201 -0
  2. package/README.md +336 -0
  3. package/cli-acp.js +82 -0
  4. package/cli.js +105 -0
  5. package/dist/REPL-EYUOXCEC.js +42 -0
  6. package/dist/REPL-EYUOXCEC.js.map +7 -0
  7. package/dist/acp-S5WNCLMD.js +1372 -0
  8. package/dist/acp-S5WNCLMD.js.map +7 -0
  9. package/dist/agentsValidate-RQ2QDGNY.js +373 -0
  10. package/dist/agentsValidate-RQ2QDGNY.js.map +7 -0
  11. package/dist/ask-TX526UBD.js +129 -0
  12. package/dist/ask-TX526UBD.js.map +7 -0
  13. package/dist/autoUpdater-63RAZ24N.js +17 -0
  14. package/dist/autoUpdater-63RAZ24N.js.map +7 -0
  15. package/dist/chunk-2VQWLLDU.js +16 -0
  16. package/dist/chunk-2VQWLLDU.js.map +7 -0
  17. package/dist/chunk-4CLHMO4I.js +656 -0
  18. package/dist/chunk-4CLHMO4I.js.map +7 -0
  19. package/dist/chunk-4ZNNWJZU.js +5696 -0
  20. package/dist/chunk-4ZNNWJZU.js.map +7 -0
  21. package/dist/chunk-66EZC7Y7.js +149 -0
  22. package/dist/chunk-66EZC7Y7.js.map +7 -0
  23. package/dist/chunk-6EPQRP3S.js +96 -0
  24. package/dist/chunk-6EPQRP3S.js.map +7 -0
  25. package/dist/chunk-77IRSDFR.js +195 -0
  26. package/dist/chunk-77IRSDFR.js.map +7 -0
  27. package/dist/chunk-7RZNLBEK.js +136 -0
  28. package/dist/chunk-7RZNLBEK.js.map +7 -0
  29. package/dist/chunk-BNBV2FXC.js +19 -0
  30. package/dist/chunk-BNBV2FXC.js.map +7 -0
  31. package/dist/chunk-CQCREBDO.js +248 -0
  32. package/dist/chunk-CQCREBDO.js.map +7 -0
  33. package/dist/chunk-D77XS6TB.js +74 -0
  34. package/dist/chunk-D77XS6TB.js.map +7 -0
  35. package/dist/chunk-DHYBJN3V.js +474 -0
  36. package/dist/chunk-DHYBJN3V.js.map +7 -0
  37. package/dist/chunk-DLSLSLTR.js +842 -0
  38. package/dist/chunk-DLSLSLTR.js.map +7 -0
  39. package/dist/chunk-ELAE6Z4H.js +514 -0
  40. package/dist/chunk-ELAE6Z4H.js.map +7 -0
  41. package/dist/chunk-ELZQD7ZR.js +531 -0
  42. package/dist/chunk-ELZQD7ZR.js.map +7 -0
  43. package/dist/chunk-F6DEGMX6.js +31269 -0
  44. package/dist/chunk-F6DEGMX6.js.map +7 -0
  45. package/dist/chunk-GDF2AON2.js +124 -0
  46. package/dist/chunk-GDF2AON2.js.map +7 -0
  47. package/dist/chunk-H7BGBV4P.js +498 -0
  48. package/dist/chunk-H7BGBV4P.js.map +7 -0
  49. package/dist/chunk-HIIHGKXP.js +24 -0
  50. package/dist/chunk-HIIHGKXP.js.map +7 -0
  51. package/dist/chunk-HJCCXED7.js +17 -0
  52. package/dist/chunk-HJCCXED7.js.map +7 -0
  53. package/dist/chunk-IQ6VZB2Y.js +139 -0
  54. package/dist/chunk-IQ6VZB2Y.js.map +7 -0
  55. package/dist/chunk-J4D7AELD.js +518 -0
  56. package/dist/chunk-J4D7AELD.js.map +7 -0
  57. package/dist/chunk-JVGG2YQR.js +23 -0
  58. package/dist/chunk-JVGG2YQR.js.map +7 -0
  59. package/dist/chunk-LGEK2NV7.js +939 -0
  60. package/dist/chunk-LGEK2NV7.js.map +7 -0
  61. package/dist/chunk-LWXT5RGE.js +95 -0
  62. package/dist/chunk-LWXT5RGE.js.map +7 -0
  63. package/dist/chunk-M3TKNAUR.js +35 -0
  64. package/dist/chunk-M3TKNAUR.js.map +7 -0
  65. package/dist/chunk-MRFO7QO5.js +170 -0
  66. package/dist/chunk-MRFO7QO5.js.map +7 -0
  67. package/dist/chunk-MVN3DHQF.js +95 -0
  68. package/dist/chunk-MVN3DHQF.js.map +7 -0
  69. package/dist/chunk-O25PXGOC.js +772 -0
  70. package/dist/chunk-O25PXGOC.js.map +7 -0
  71. package/dist/chunk-OBGVKM3N.js +1618 -0
  72. package/dist/chunk-OBGVKM3N.js.map +7 -0
  73. package/dist/chunk-OV5HJXXQ.js +198 -0
  74. package/dist/chunk-OV5HJXXQ.js.map +7 -0
  75. package/dist/chunk-P5VWDMRD.js +249 -0
  76. package/dist/chunk-P5VWDMRD.js.map +7 -0
  77. package/dist/chunk-PDSAJX7G.js +49 -0
  78. package/dist/chunk-PDSAJX7G.js.map +7 -0
  79. package/dist/chunk-RHNEZOPO.js +739 -0
  80. package/dist/chunk-RHNEZOPO.js.map +7 -0
  81. package/dist/chunk-SQGAHZPM.js +3004 -0
  82. package/dist/chunk-SQGAHZPM.js.map +7 -0
  83. package/dist/chunk-U7Z4MXY4.js +21 -0
  84. package/dist/chunk-U7Z4MXY4.js.map +7 -0
  85. package/dist/chunk-UNCTVIS7.js +146 -0
  86. package/dist/chunk-UNCTVIS7.js.map +7 -0
  87. package/dist/chunk-VMEOI6MH.js +1103 -0
  88. package/dist/chunk-VMEOI6MH.js.map +7 -0
  89. package/dist/chunk-WAY3DKFO.js +47 -0
  90. package/dist/chunk-WAY3DKFO.js.map +7 -0
  91. package/dist/chunk-XEYEKVFT.js +24 -0
  92. package/dist/chunk-XEYEKVFT.js.map +7 -0
  93. package/dist/chunk-Y4BQ36T4.js +796 -0
  94. package/dist/chunk-Y4BQ36T4.js.map +7 -0
  95. package/dist/chunk-Y5LQPJWK.js +12 -0
  96. package/dist/chunk-Y5LQPJWK.js.map +7 -0
  97. package/dist/chunk-YIJWUNWF.js +1260 -0
  98. package/dist/chunk-YIJWUNWF.js.map +7 -0
  99. package/dist/chunk-YMIWYEZ7.js +34 -0
  100. package/dist/chunk-YMIWYEZ7.js.map +7 -0
  101. package/dist/cli-PQNZWJX4.js +3952 -0
  102. package/dist/cli-PQNZWJX4.js.map +7 -0
  103. package/dist/commands-HOBCZ3VQ.js +46 -0
  104. package/dist/commands-HOBCZ3VQ.js.map +7 -0
  105. package/dist/config-MLH7ZTFA.js +81 -0
  106. package/dist/config-MLH7ZTFA.js.map +7 -0
  107. package/dist/context-FZ6G4J63.js +30 -0
  108. package/dist/context-FZ6G4J63.js.map +7 -0
  109. package/dist/costTracker-5WKZXN5S.js +19 -0
  110. package/dist/costTracker-5WKZXN5S.js.map +7 -0
  111. package/dist/customCommands-EB4MMZSS.js +25 -0
  112. package/dist/customCommands-EB4MMZSS.js.map +7 -0
  113. package/dist/env-VMEIP4EW.js +28 -0
  114. package/dist/env-VMEIP4EW.js.map +7 -0
  115. package/dist/index.js +36 -0
  116. package/dist/index.js.map +7 -0
  117. package/dist/kodeAgentSessionId-WUT74FSH.js +16 -0
  118. package/dist/kodeAgentSessionId-WUT74FSH.js.map +7 -0
  119. package/dist/kodeAgentSessionLoad-KR4JSD6D.js +21 -0
  120. package/dist/kodeAgentSessionLoad-KR4JSD6D.js.map +7 -0
  121. package/dist/kodeAgentSessionResume-BCD6UV74.js +18 -0
  122. package/dist/kodeAgentSessionResume-BCD6UV74.js.map +7 -0
  123. package/dist/kodeAgentStreamJson-EDHHWNNX.js +15 -0
  124. package/dist/kodeAgentStreamJson-EDHHWNNX.js.map +7 -0
  125. package/dist/kodeAgentStreamJsonSession-G4RBNZRN.js +133 -0
  126. package/dist/kodeAgentStreamJsonSession-G4RBNZRN.js.map +7 -0
  127. package/dist/kodeAgentStructuredStdio-UA5P5UNU.js +11 -0
  128. package/dist/kodeAgentStructuredStdio-UA5P5UNU.js.map +7 -0
  129. package/dist/kodeHooks-EHM6GSIQ.js +37 -0
  130. package/dist/kodeHooks-EHM6GSIQ.js.map +7 -0
  131. package/dist/llm-SJXCV7DA.js +3138 -0
  132. package/dist/llm-SJXCV7DA.js.map +7 -0
  133. package/dist/llmLazy-2QYJVD6K.js +15 -0
  134. package/dist/llmLazy-2QYJVD6K.js.map +7 -0
  135. package/dist/loader-LJX77EFL.js +28 -0
  136. package/dist/loader-LJX77EFL.js.map +7 -0
  137. package/dist/mcp-DOROSLPN.js +49 -0
  138. package/dist/mcp-DOROSLPN.js.map +7 -0
  139. package/dist/mentionProcessor-5UZRHCGH.js +215 -0
  140. package/dist/mentionProcessor-5UZRHCGH.js.map +7 -0
  141. package/dist/messages-N5KBI53P.js +65 -0
  142. package/dist/messages-N5KBI53P.js.map +7 -0
  143. package/dist/model-HPLBR53R.js +30 -0
  144. package/dist/model-HPLBR53R.js.map +7 -0
  145. package/dist/openai-YP4OJYKF.js +29 -0
  146. package/dist/openai-YP4OJYKF.js.map +7 -0
  147. package/dist/outputStyles-NNALI5D7.js +28 -0
  148. package/dist/outputStyles-NNALI5D7.js.map +7 -0
  149. package/dist/package.json +4 -0
  150. package/dist/pluginRuntime-JYYI5BSQ.js +220 -0
  151. package/dist/pluginRuntime-JYYI5BSQ.js.map +7 -0
  152. package/dist/pluginValidation-JWUFPZUE.js +17 -0
  153. package/dist/pluginValidation-JWUFPZUE.js.map +7 -0
  154. package/dist/prompts-B2SS7CWI.js +50 -0
  155. package/dist/prompts-B2SS7CWI.js.map +7 -0
  156. package/dist/query-HIK457UU.js +50 -0
  157. package/dist/query-HIK457UU.js.map +7 -0
  158. package/dist/responsesStreaming-L2BSN37C.js +10 -0
  159. package/dist/responsesStreaming-L2BSN37C.js.map +7 -0
  160. package/dist/ripgrep-GCKI4UTL.js +17 -0
  161. package/dist/ripgrep-GCKI4UTL.js.map +7 -0
  162. package/dist/skillMarketplace-PCTUUX46.js +37 -0
  163. package/dist/skillMarketplace-PCTUUX46.js.map +7 -0
  164. package/dist/state-XJICGOUA.js +18 -0
  165. package/dist/state-XJICGOUA.js.map +7 -0
  166. package/dist/theme-DP7O4SGH.js +14 -0
  167. package/dist/theme-DP7O4SGH.js.map +7 -0
  168. package/dist/toolPermissionContext-DHAGUPEW.js +17 -0
  169. package/dist/toolPermissionContext-DHAGUPEW.js.map +7 -0
  170. package/dist/toolPermissionSettings-PT65MQIQ.js +18 -0
  171. package/dist/toolPermissionSettings-PT65MQIQ.js.map +7 -0
  172. package/dist/tools-BHW37PCF.js +47 -0
  173. package/dist/tools-BHW37PCF.js.map +7 -0
  174. package/dist/userInput-XDRYT5TI.js +316 -0
  175. package/dist/userInput-XDRYT5TI.js.map +7 -0
  176. package/dist/uuid-QUYJMIUV.js +9 -0
  177. package/dist/uuid-QUYJMIUV.js.map +7 -0
  178. package/dist/yoga.wasm +0 -0
  179. package/package.json +115 -0
  180. package/scripts/binary-utils.cjs +62 -0
  181. package/scripts/cli-acp-wrapper.cjs +82 -0
  182. package/scripts/cli-wrapper.cjs +105 -0
  183. package/scripts/postinstall.js +144 -0
  184. package/yoga.wasm +0 -0
@@ -0,0 +1,3138 @@
1
+ import { createRequire as __kodeCreateRequire } from "node:module";
2
+ const require = __kodeCreateRequire(import.meta.url);
3
+ import {
4
+ getToolDescription
5
+ } from "./chunk-HJCCXED7.js";
6
+ import {
7
+ processResponsesStream
8
+ } from "./chunk-6EPQRP3S.js";
9
+ import {
10
+ formatSystemPromptWithContext,
11
+ generateDanyaContext,
12
+ getCLISyspromptPrefix,
13
+ getReasoningEffort,
14
+ models_default,
15
+ refreshDanyaContext
16
+ } from "./chunk-F6DEGMX6.js";
17
+ import "./chunk-OV5HJXXQ.js";
18
+ import "./chunk-OBGVKM3N.js";
19
+ import "./chunk-DHYBJN3V.js";
20
+ import "./chunk-VMEOI6MH.js";
21
+ import "./chunk-IQ6VZB2Y.js";
22
+ import "./chunk-CQCREBDO.js";
23
+ import "./chunk-Y5LQPJWK.js";
24
+ import {
25
+ setRequestStatus
26
+ } from "./chunk-JVGG2YQR.js";
27
+ import "./chunk-YIJWUNWF.js";
28
+ import "./chunk-U7Z4MXY4.js";
29
+ import "./chunk-ELAE6Z4H.js";
30
+ import {
31
+ getCompletionWithProfile,
32
+ getGPT5CompletionWithProfile
33
+ } from "./chunk-LGEK2NV7.js";
34
+ import "./chunk-YMIWYEZ7.js";
35
+ import "./chunk-MRFO7QO5.js";
36
+ import "./chunk-O25PXGOC.js";
37
+ import "./chunk-77IRSDFR.js";
38
+ import "./chunk-66EZC7Y7.js";
39
+ import "./chunk-MVN3DHQF.js";
40
+ import "./chunk-J4D7AELD.js";
41
+ import "./chunk-XEYEKVFT.js";
42
+ import "./chunk-RHNEZOPO.js";
43
+ import "./chunk-WAY3DKFO.js";
44
+ import "./chunk-2VQWLLDU.js";
45
+ import {
46
+ API_ERROR_MESSAGE_PREFIX,
47
+ CREDIT_BALANCE_TOO_LOW_ERROR_MESSAGE,
48
+ INVALID_API_KEY_ERROR_MESSAGE,
49
+ MAIN_QUERY_TEMPERATURE,
50
+ NO_CONTENT_MESSAGE,
51
+ PROMPT_TOO_LONG_ERROR_MESSAGE,
52
+ createAssistantAPIErrorMessage,
53
+ normalizeContentFromAPI
54
+ } from "./chunk-H7BGBV4P.js";
55
+ import {
56
+ USE_BEDROCK,
57
+ USE_VERTEX,
58
+ getModelManager,
59
+ getVertexRegionForModel
60
+ } from "./chunk-4CLHMO4I.js";
61
+ import "./chunk-ELZQD7ZR.js";
62
+ import "./chunk-GDF2AON2.js";
63
+ import {
64
+ getAnthropicApiKey,
65
+ getGlobalConfig
66
+ } from "./chunk-DLSLSLTR.js";
67
+ import "./chunk-HIIHGKXP.js";
68
+ import {
69
+ debug,
70
+ getCurrentRequest,
71
+ logErrorWithDiagnosis,
72
+ logLLMInteraction,
73
+ logSystemPromptConstruction,
74
+ markPhase
75
+ } from "./chunk-Y4BQ36T4.js";
76
+ import {
77
+ PRODUCT_COMMAND,
78
+ env,
79
+ getCwd,
80
+ init_env,
81
+ init_log,
82
+ init_product,
83
+ init_state,
84
+ logError
85
+ } from "./chunk-SQGAHZPM.js";
86
+ import {
87
+ MACRO,
88
+ init_macros
89
+ } from "./chunk-UNCTVIS7.js";
90
+ import {
91
+ addToTotalCost
92
+ } from "./chunk-LWXT5RGE.js";
93
+ import "./chunk-M3TKNAUR.js";
94
+
95
+ // src/services/ai/llm.ts
96
+ import "@anthropic-ai/sdk/shims/node";
97
+ import Anthropic, { APIConnectionError, APIError } from "@anthropic-ai/sdk";
98
+ import chalk from "chalk";
99
+ import { randomUUID } from "crypto";
100
+ import "dotenv/config";
101
+ init_log();
102
+
103
+ // src/utils/system/http.ts
104
+ init_macros();
105
+ init_product();
106
+ var USER_AGENT = `${PRODUCT_COMMAND}/${MACRO.VERSION} (${process.env.USER_TYPE})`;
107
+
108
+ // src/services/system/vcr.ts
109
+ init_env();
110
+ init_state();
111
+ import { createHash } from "crypto";
112
+ import { mkdirSync, readFileSync, writeFileSync } from "fs";
113
+ import { dirname } from "path";
114
+ import { existsSync } from "fs";
115
+ import * as path from "path";
116
+ import { mapValues } from "lodash-es";
117
+ async function withVCR(messages, f) {
118
+ if (process.env.NODE_ENV !== "test") {
119
+ return await f();
120
+ }
121
+ const dehydratedInput = mapMessages(
122
+ messages.map((_) => _.message.content),
123
+ dehydrateValue
124
+ );
125
+ const filename = `./fixtures/${dehydratedInput.map((_) => createHash("sha1").update(JSON.stringify(_)).digest("hex").slice(0, 6)).join("-")}.json`;
126
+ if (existsSync(filename)) {
127
+ const cached = JSON.parse(readFileSync(filename, "utf-8"));
128
+ return mapAssistantMessage(cached.output, hydrateValue);
129
+ }
130
+ if (env.isCI) {
131
+ process.stderr.write(
132
+ `Anthropic API fixture missing. Re-run bun test locally, then commit the result. ${JSON.stringify({ input: dehydratedInput }, null, 2)}
133
+ `
134
+ );
135
+ }
136
+ const result = await f();
137
+ if (env.isCI) {
138
+ return result;
139
+ }
140
+ if (!existsSync(dirname(filename))) {
141
+ mkdirSync(dirname(filename), { recursive: true });
142
+ }
143
+ writeFileSync(
144
+ filename,
145
+ JSON.stringify(
146
+ {
147
+ input: dehydratedInput,
148
+ output: mapAssistantMessage(result, dehydrateValue)
149
+ },
150
+ null,
151
+ 2
152
+ )
153
+ );
154
+ return result;
155
+ }
156
+ function mapMessages(messages, f) {
157
+ return messages.map((_) => {
158
+ if (typeof _ === "string") {
159
+ return f(_);
160
+ }
161
+ return _.map((_2) => {
162
+ switch (_2.type) {
163
+ case "tool_result":
164
+ if (typeof _2.content === "string") {
165
+ return { ..._2, content: f(_2.content) };
166
+ }
167
+ if (Array.isArray(_2.content)) {
168
+ return {
169
+ ..._2,
170
+ content: _2.content.map((_3) => {
171
+ switch (_3.type) {
172
+ case "text":
173
+ return { ..._3, text: f(_3.text) };
174
+ case "image":
175
+ return _3;
176
+ }
177
+ })
178
+ };
179
+ }
180
+ return _2;
181
+ case "text":
182
+ return { ..._2, text: f(_2.text) };
183
+ case "tool_use":
184
+ return {
185
+ ..._2,
186
+ input: mapValues(_2.input, f)
187
+ };
188
+ case "image":
189
+ return _2;
190
+ }
191
+ });
192
+ });
193
+ }
194
+ function mapAssistantMessage(message, f) {
195
+ return {
196
+ durationMs: "DURATION",
197
+ costUSD: "COST",
198
+ uuid: "UUID",
199
+ message: {
200
+ ...message.message,
201
+ content: message.message.content.map((_) => {
202
+ switch (_.type) {
203
+ case "text":
204
+ return {
205
+ ..._,
206
+ text: f(_.text),
207
+ citations: _.citations || []
208
+ };
209
+ case "tool_use":
210
+ return {
211
+ ..._,
212
+ input: mapValues(_.input, f)
213
+ };
214
+ default:
215
+ return _;
216
+ }
217
+ }).filter(Boolean)
218
+ },
219
+ type: "assistant"
220
+ };
221
+ }
222
+ function dehydrateValue(s) {
223
+ if (typeof s !== "string") {
224
+ return s;
225
+ }
226
+ const s1 = s.replace(/num_files="\d+"/g, 'num_files="[NUM]"').replace(/duration_ms="\d+"/g, 'duration_ms="[DURATION]"').replace(/cost_usd="\d+"/g, 'cost_usd="[COST]"').replace(/\//g, path.sep).replaceAll(getCwd(), "[CWD]");
227
+ if (s1.includes("Files modified by user:")) {
228
+ return "Files modified by user: [FILES]";
229
+ }
230
+ return s1;
231
+ }
232
+ function hydrateValue(s) {
233
+ if (typeof s !== "string") {
234
+ return s;
235
+ }
236
+ return s.replaceAll("[NUM]", "1").replaceAll("[DURATION]", "100").replaceAll("[CWD]", getCwd());
237
+ }
238
+
239
+ // src/services/ai/llm.ts
240
+ import { zodToJsonSchema as zodToJsonSchema4 } from "zod-to-json-schema";
241
+
242
+ // src/services/ai/adapters/base.ts
243
+ function normalizeTokens(apiResponse) {
244
+ if (!apiResponse || typeof apiResponse !== "object") {
245
+ return { input: 0, output: 0 };
246
+ }
247
+ const input = Number(
248
+ apiResponse.prompt_tokens ?? apiResponse.input_tokens ?? apiResponse.promptTokens
249
+ ) || 0;
250
+ const output = Number(
251
+ apiResponse.completion_tokens ?? apiResponse.output_tokens ?? apiResponse.completionTokens
252
+ ) || 0;
253
+ const total = Number(apiResponse.total_tokens ?? apiResponse.totalTokens) || void 0;
254
+ const reasoning = Number(apiResponse.reasoning_tokens ?? apiResponse.reasoningTokens) || void 0;
255
+ return {
256
+ input,
257
+ output,
258
+ total: total && total > 0 ? total : void 0,
259
+ reasoning: reasoning && reasoning > 0 ? reasoning : void 0
260
+ };
261
+ }
262
+ var ModelAPIAdapter = class {
263
+ constructor(capabilities, modelProfile) {
264
+ this.capabilities = capabilities;
265
+ this.modelProfile = modelProfile;
266
+ }
267
+ cumulativeUsage = { input: 0, output: 0 };
268
+ async *parseStreamingResponse(response, signal) {
269
+ return;
270
+ yield;
271
+ }
272
+ resetCumulativeUsage() {
273
+ this.cumulativeUsage = { input: 0, output: 0 };
274
+ }
275
+ updateCumulativeUsage(usage) {
276
+ this.cumulativeUsage.input += usage.input;
277
+ this.cumulativeUsage.output += usage.output;
278
+ if (usage.total) {
279
+ this.cumulativeUsage.total = (this.cumulativeUsage.total || 0) + usage.total;
280
+ }
281
+ if (usage.reasoning) {
282
+ this.cumulativeUsage.reasoning = (this.cumulativeUsage.reasoning || 0) + usage.reasoning;
283
+ }
284
+ }
285
+ getMaxTokensParam() {
286
+ return this.capabilities.parameters.maxTokensField;
287
+ }
288
+ getTemperature() {
289
+ if (this.capabilities.parameters.temperatureMode === "fixed_one") {
290
+ return 1;
291
+ }
292
+ if (this.capabilities.parameters.temperatureMode === "restricted") {
293
+ return Math.min(1, 0.7);
294
+ }
295
+ return 0.7;
296
+ }
297
+ shouldIncludeReasoningEffort() {
298
+ return this.capabilities.parameters.supportsReasoningEffort;
299
+ }
300
+ shouldIncludeVerbosity() {
301
+ return this.capabilities.parameters.supportsVerbosity;
302
+ }
303
+ };
304
+
305
+ // src/services/ai/adapters/openaiAdapter.ts
306
+ import { zodToJsonSchema } from "zod-to-json-schema";
307
+ init_log();
308
+ var OpenAIAdapter = class extends ModelAPIAdapter {
309
+ constructor(capabilities, modelProfile) {
310
+ super(capabilities, modelProfile);
311
+ }
312
+ async parseResponse(response) {
313
+ if (response?.body instanceof ReadableStream) {
314
+ const { assistantMessage } = await this.parseStreamingOpenAIResponse(response);
315
+ return {
316
+ id: assistantMessage.responseId,
317
+ content: assistantMessage.message.content,
318
+ toolCalls: assistantMessage.message.content.filter((block) => block.type === "tool_use").map((block) => ({
319
+ id: block.id,
320
+ type: "function",
321
+ function: {
322
+ name: block.name,
323
+ arguments: JSON.stringify(block.input)
324
+ }
325
+ })),
326
+ usage: this.normalizeUsageForAdapter(assistantMessage.message.usage),
327
+ responseId: assistantMessage.responseId
328
+ };
329
+ }
330
+ return this.parseNonStreamingResponse(response);
331
+ }
332
+ async *parseStreamingResponse(response) {
333
+ const reader = response.body.getReader();
334
+ const decoder = new TextDecoder();
335
+ let buffer = "";
336
+ let responseId = response.id || `openai_${Date.now()}`;
337
+ let hasStarted = false;
338
+ let accumulatedContent = "";
339
+ const reasoningContext = {
340
+ thinkOpen: false,
341
+ thinkClosed: false,
342
+ sawAnySummary: false,
343
+ pendingSummaryParagraph: false
344
+ };
345
+ try {
346
+ while (true) {
347
+ const { done, value } = await reader.read();
348
+ if (done) break;
349
+ buffer += decoder.decode(value, { stream: true });
350
+ const lines = buffer.split("\n");
351
+ buffer = lines.pop() || "";
352
+ for (const line of lines) {
353
+ if (line.trim()) {
354
+ const parsed = this.parseSSEChunk(line);
355
+ if (parsed) {
356
+ if (parsed.id) {
357
+ responseId = parsed.id;
358
+ }
359
+ yield* this.processStreamingChunk(
360
+ parsed,
361
+ responseId,
362
+ hasStarted,
363
+ accumulatedContent,
364
+ reasoningContext
365
+ );
366
+ const stateUpdate = this.updateStreamingState(
367
+ parsed,
368
+ accumulatedContent
369
+ );
370
+ if (stateUpdate.content) accumulatedContent = stateUpdate.content;
371
+ if (stateUpdate.hasStarted) hasStarted = true;
372
+ }
373
+ }
374
+ }
375
+ }
376
+ } catch (error) {
377
+ logError(error);
378
+ debug.warn("OPENAI_ADAPTER_STREAM_READ_ERROR", {
379
+ error: error instanceof Error ? error.message : String(error)
380
+ });
381
+ yield {
382
+ type: "error",
383
+ error: error instanceof Error ? error.message : String(error)
384
+ };
385
+ } finally {
386
+ reader.releaseLock();
387
+ }
388
+ const finalContent = accumulatedContent ? [{ type: "text", text: accumulatedContent, citations: [] }] : [{ type: "text", text: "", citations: [] }];
389
+ yield {
390
+ type: "message_stop",
391
+ message: {
392
+ id: responseId,
393
+ role: "assistant",
394
+ content: finalContent,
395
+ responseId
396
+ }
397
+ };
398
+ }
399
+ parseSSEChunk(line) {
400
+ if (line.startsWith("data: ")) {
401
+ const data = line.slice(6).trim();
402
+ if (data === "[DONE]") {
403
+ return null;
404
+ }
405
+ if (data) {
406
+ try {
407
+ return JSON.parse(data);
408
+ } catch (error) {
409
+ logError(error);
410
+ debug.warn("OPENAI_ADAPTER_SSE_PARSE_ERROR", {
411
+ error: error instanceof Error ? error.message : String(error)
412
+ });
413
+ return null;
414
+ }
415
+ }
416
+ }
417
+ return null;
418
+ }
419
+ handleTextDelta(delta, responseId, hasStarted) {
420
+ const events = [];
421
+ if (!hasStarted && delta) {
422
+ events.push({
423
+ type: "message_start",
424
+ message: {
425
+ role: "assistant",
426
+ content: []
427
+ },
428
+ responseId
429
+ });
430
+ }
431
+ if (delta) {
432
+ events.push({
433
+ type: "text_delta",
434
+ delta,
435
+ responseId
436
+ });
437
+ }
438
+ return events;
439
+ }
440
+ normalizeUsageForAdapter(usage) {
441
+ if (!usage) {
442
+ return {
443
+ input_tokens: 0,
444
+ output_tokens: 0,
445
+ promptTokens: 0,
446
+ completionTokens: 0,
447
+ totalTokens: 0,
448
+ reasoningTokens: 0
449
+ };
450
+ }
451
+ const inputTokens = usage.input_tokens ?? usage.prompt_tokens ?? usage.promptTokens ?? 0;
452
+ const outputTokens = usage.output_tokens ?? usage.completion_tokens ?? usage.completionTokens ?? 0;
453
+ return {
454
+ ...usage,
455
+ input_tokens: inputTokens,
456
+ output_tokens: outputTokens,
457
+ promptTokens: inputTokens,
458
+ completionTokens: outputTokens,
459
+ totalTokens: usage.totalTokens ?? inputTokens + outputTokens,
460
+ reasoningTokens: usage.reasoningTokens ?? 0
461
+ };
462
+ }
463
+ buildTools(tools) {
464
+ return tools.map((tool) => ({
465
+ type: "function",
466
+ function: {
467
+ name: tool.name,
468
+ description: getToolDescription(tool),
469
+ parameters: zodToJsonSchema(tool.inputSchema)
470
+ }
471
+ }));
472
+ }
473
+ };
474
+
475
+ // src/services/ai/adapters/responsesAPI.ts
476
+ import { zodToJsonSchema as zodToJsonSchema2 } from "zod-to-json-schema";
477
+ init_log();
478
+ var ResponsesAPIAdapter = class extends OpenAIAdapter {
479
+ createRequest(params) {
480
+ const {
481
+ messages,
482
+ systemPrompt,
483
+ tools,
484
+ maxTokens,
485
+ reasoningEffort,
486
+ stopSequences
487
+ } = params;
488
+ const request = {
489
+ model: this.modelProfile.modelName,
490
+ input: this.convertMessagesToInput(messages),
491
+ instructions: this.buildInstructions(systemPrompt)
492
+ };
493
+ const maxTokensField = this.getMaxTokensParam();
494
+ request[maxTokensField] = maxTokens;
495
+ if (stopSequences && stopSequences.length > 0) {
496
+ request.stop = stopSequences;
497
+ }
498
+ request.stream = params.stream !== false && this.capabilities.streaming.supported;
499
+ const temperature = this.getTemperature();
500
+ if (temperature !== void 0) {
501
+ request.temperature = temperature;
502
+ }
503
+ const include = [];
504
+ if (this.capabilities.parameters.supportsReasoningEffort && (this.shouldIncludeReasoningEffort() || reasoningEffort)) {
505
+ include.push("reasoning.encrypted_content");
506
+ request.reasoning = {
507
+ effort: reasoningEffort || this.modelProfile.reasoningEffort || "medium"
508
+ };
509
+ }
510
+ if (this.capabilities.parameters.supportsVerbosity && this.shouldIncludeVerbosity()) {
511
+ let defaultVerbosity = "medium";
512
+ if (params.verbosity) {
513
+ defaultVerbosity = params.verbosity;
514
+ } else {
515
+ const modelNameLower = this.modelProfile.modelName.toLowerCase();
516
+ if (modelNameLower.includes("high")) {
517
+ defaultVerbosity = "high";
518
+ } else if (modelNameLower.includes("low")) {
519
+ defaultVerbosity = "low";
520
+ }
521
+ }
522
+ request.text = {
523
+ verbosity: defaultVerbosity
524
+ };
525
+ }
526
+ if (tools && tools.length > 0) {
527
+ request.tools = this.buildTools(tools);
528
+ }
529
+ request.tool_choice = "auto";
530
+ if (this.capabilities.toolCalling.supportsParallelCalls) {
531
+ request.parallel_tool_calls = true;
532
+ }
533
+ request.store = false;
534
+ if (params.previousResponseId && this.capabilities.stateManagement.supportsPreviousResponseId) {
535
+ request.previous_response_id = params.previousResponseId;
536
+ }
537
+ if (include.length > 0) {
538
+ request.include = include;
539
+ }
540
+ return request;
541
+ }
542
+ buildTools(tools) {
543
+ return tools.map((tool) => {
544
+ let parameters = tool.inputJSONSchema;
545
+ if (!parameters && tool.inputSchema) {
546
+ const isPlainObject = (obj) => {
547
+ return obj !== null && typeof obj === "object" && !Array.isArray(obj);
548
+ };
549
+ if (isPlainObject(tool.inputSchema) && ("type" in tool.inputSchema || "properties" in tool.inputSchema)) {
550
+ parameters = tool.inputSchema;
551
+ } else {
552
+ try {
553
+ parameters = zodToJsonSchema2(tool.inputSchema);
554
+ } catch (error) {
555
+ logError(error);
556
+ debug.warn("RESPONSES_API_TOOL_SCHEMA_CONVERSION_FAILED", {
557
+ toolName: tool.name,
558
+ error: error instanceof Error ? error.message : String(error)
559
+ });
560
+ parameters = { type: "object", properties: {} };
561
+ }
562
+ }
563
+ }
564
+ return {
565
+ type: "function",
566
+ name: tool.name,
567
+ description: getToolDescription(tool),
568
+ parameters: parameters || { type: "object", properties: {} }
569
+ };
570
+ });
571
+ }
572
+ async parseResponse(response) {
573
+ if (response?.body instanceof ReadableStream) {
574
+ const { assistantMessage } = await processResponsesStream(
575
+ this.parseStreamingResponse(response),
576
+ Date.now(),
577
+ response.id ?? `resp_${Date.now()}`
578
+ );
579
+ const hasToolUseBlocks = assistantMessage.message.content.some(
580
+ (block) => block.type === "tool_use"
581
+ );
582
+ return {
583
+ id: assistantMessage.responseId,
584
+ content: assistantMessage.message.content,
585
+ toolCalls: hasToolUseBlocks ? [] : [],
586
+ usage: this.normalizeUsageForAdapter(assistantMessage.message.usage),
587
+ responseId: assistantMessage.responseId
588
+ };
589
+ }
590
+ return this.parseNonStreamingResponse(response);
591
+ }
592
+ parseNonStreamingResponse(response) {
593
+ let content = response.output_text || "";
594
+ let reasoningContent = "";
595
+ if (response.output && Array.isArray(response.output)) {
596
+ const messageItems = response.output.filter(
597
+ (item) => item.type === "message"
598
+ );
599
+ if (messageItems.length > 0) {
600
+ content = messageItems.map((item) => {
601
+ if (item.content && Array.isArray(item.content)) {
602
+ return item.content.filter((c) => c.type === "text").map((c) => c.text).join("\n");
603
+ }
604
+ return item.content || "";
605
+ }).filter(Boolean).join("\n\n");
606
+ }
607
+ const reasoningItems = response.output.filter(
608
+ (item) => item.type === "reasoning"
609
+ );
610
+ if (reasoningItems.length > 0) {
611
+ reasoningContent = reasoningItems.map((item) => item.content || "").filter(Boolean).join("\n\n");
612
+ }
613
+ }
614
+ if (reasoningContent) {
615
+ const thinkBlock = `
616
+
617
+ ${reasoningContent}
618
+
619
+ `;
620
+ content = thinkBlock + content;
621
+ }
622
+ const toolCalls = this.parseToolCalls(response);
623
+ const contentArray = content ? [{ type: "text", text: content, citations: [] }] : [{ type: "text", text: "", citations: [] }];
624
+ const promptTokens = response.usage?.input_tokens || 0;
625
+ const completionTokens = response.usage?.output_tokens || 0;
626
+ const totalTokens = response.usage?.total_tokens ?? promptTokens + completionTokens;
627
+ return {
628
+ id: response.id || `resp_${Date.now()}`,
629
+ content: contentArray,
630
+ toolCalls,
631
+ usage: {
632
+ promptTokens,
633
+ completionTokens,
634
+ reasoningTokens: response.usage?.output_tokens_details?.reasoning_tokens
635
+ },
636
+ responseId: response.id
637
+ };
638
+ }
639
+ async *processStreamingChunk(parsed, responseId, hasStarted, accumulatedContent, reasoningContext) {
640
+ if (parsed.type === "response.reasoning_summary_part.added") {
641
+ const partIndex = parsed.summary_index || 0;
642
+ if (!reasoningContext?.thinkingContent) {
643
+ reasoningContext.thinkingContent = "";
644
+ reasoningContext.currentPartIndex = -1;
645
+ }
646
+ reasoningContext.currentPartIndex = partIndex;
647
+ if (partIndex > 0 && reasoningContext.thinkingContent) {
648
+ reasoningContext.thinkingContent += "\n\n";
649
+ yield {
650
+ type: "text_delta",
651
+ delta: "\n\n",
652
+ responseId
653
+ };
654
+ }
655
+ return;
656
+ }
657
+ if (parsed.type === "response.reasoning_summary_text.delta") {
658
+ const delta = parsed.delta || "";
659
+ if (delta && reasoningContext) {
660
+ reasoningContext.thinkingContent += delta;
661
+ yield {
662
+ type: "text_delta",
663
+ delta,
664
+ responseId
665
+ };
666
+ }
667
+ return;
668
+ }
669
+ if (parsed.type === "response.reasoning_text.delta") {
670
+ const delta = parsed.delta || "";
671
+ if (delta && reasoningContext) {
672
+ reasoningContext.thinkingContent += delta;
673
+ yield {
674
+ type: "text_delta",
675
+ delta,
676
+ responseId
677
+ };
678
+ }
679
+ return;
680
+ }
681
+ if (parsed.type === "response.output_text.delta") {
682
+ const delta = parsed.delta || "";
683
+ if (delta) {
684
+ const textEvents = this.handleTextDelta(delta, responseId, hasStarted);
685
+ for (const event of textEvents) {
686
+ yield event;
687
+ }
688
+ }
689
+ }
690
+ if (parsed.type === "response.output_item.done") {
691
+ const item = parsed.item || {};
692
+ if (item.type === "function_call") {
693
+ const callId = item.call_id || item.id;
694
+ const name = item.name;
695
+ const args = item.arguments;
696
+ if (typeof callId === "string" && typeof name === "string" && typeof args === "string") {
697
+ yield {
698
+ type: "tool_request",
699
+ tool: {
700
+ id: callId,
701
+ name,
702
+ input: args
703
+ }
704
+ };
705
+ }
706
+ }
707
+ }
708
+ if (parsed.usage) {
709
+ const normalizedUsage = normalizeTokens(parsed.usage);
710
+ if (parsed.usage.output_tokens_details?.reasoning_tokens) {
711
+ normalizedUsage.reasoning = parsed.usage.output_tokens_details.reasoning_tokens;
712
+ }
713
+ yield {
714
+ type: "usage",
715
+ usage: normalizedUsage
716
+ };
717
+ }
718
+ }
719
+ updateStreamingState(parsed, accumulatedContent) {
720
+ const state = {};
721
+ if (parsed.type === "response.output_text.delta" && parsed.delta) {
722
+ state.content = accumulatedContent + parsed.delta;
723
+ state.hasStarted = true;
724
+ }
725
+ return state;
726
+ }
727
+ async parseStreamingOpenAIResponse(response) {
728
+ const { processResponsesStream: processResponsesStream2 } = await import("./responsesStreaming-L2BSN37C.js");
729
+ return await processResponsesStream2(
730
+ this.parseStreamingResponse(response),
731
+ Date.now(),
732
+ response.id ?? `resp_${Date.now()}`
733
+ );
734
+ }
735
+ normalizeUsageForAdapter(usage) {
736
+ const baseUsage = super.normalizeUsageForAdapter(usage);
737
+ return {
738
+ ...baseUsage,
739
+ reasoningTokens: usage?.output_tokens_details?.reasoning_tokens ?? 0
740
+ };
741
+ }
742
+ convertMessagesToInput(messages) {
743
+ const inputItems = [];
744
+ for (const message of messages) {
745
+ const role = message.role;
746
+ if (role === "tool") {
747
+ const callId = message.tool_call_id || message.id;
748
+ if (typeof callId === "string" && callId) {
749
+ let content2 = message.content || "";
750
+ if (Array.isArray(content2)) {
751
+ const texts = [];
752
+ for (const part of content2) {
753
+ if (typeof part === "object" && part !== null) {
754
+ const t = part.text || part.content;
755
+ if (typeof t === "string" && t) {
756
+ texts.push(t);
757
+ }
758
+ }
759
+ }
760
+ content2 = texts.join("\n");
761
+ }
762
+ if (typeof content2 === "string") {
763
+ inputItems.push({
764
+ type: "function_call_output",
765
+ call_id: callId,
766
+ output: content2
767
+ });
768
+ }
769
+ }
770
+ continue;
771
+ }
772
+ if (role === "assistant" && Array.isArray(message.tool_calls)) {
773
+ for (const tc of message.tool_calls) {
774
+ if (typeof tc !== "object" || tc === null) {
775
+ continue;
776
+ }
777
+ const tcType = tc.type || "function";
778
+ if (tcType !== "function") {
779
+ continue;
780
+ }
781
+ const callId = tc.id || tc.call_id;
782
+ const fn = tc.function;
783
+ const name = typeof fn === "object" && fn !== null ? fn.name : null;
784
+ const args = typeof fn === "object" && fn !== null ? fn.arguments : null;
785
+ if (typeof callId === "string" && typeof name === "string" && typeof args === "string") {
786
+ inputItems.push({
787
+ type: "function_call",
788
+ name,
789
+ arguments: args,
790
+ call_id: callId
791
+ });
792
+ }
793
+ }
794
+ continue;
795
+ }
796
+ const content = message.content || "";
797
+ const contentItems = [];
798
+ if (Array.isArray(content)) {
799
+ for (const part of content) {
800
+ if (typeof part !== "object" || part === null) continue;
801
+ const ptype = part.type;
802
+ if (ptype === "text") {
803
+ const text = part.text || part.content || "";
804
+ if (typeof text === "string" && text) {
805
+ const kind = role === "assistant" ? "output_text" : "input_text";
806
+ contentItems.push({ type: kind, text });
807
+ }
808
+ } else if (ptype === "image_url") {
809
+ const image = part.image_url;
810
+ const url = typeof image === "object" && image !== null ? image.url : image;
811
+ if (typeof url === "string" && url) {
812
+ contentItems.push({ type: "input_image", image_url: url });
813
+ }
814
+ }
815
+ }
816
+ } else if (typeof content === "string" && content) {
817
+ const kind = role === "assistant" ? "output_text" : "input_text";
818
+ contentItems.push({ type: kind, text: content });
819
+ }
820
+ if (contentItems.length) {
821
+ const roleOut = role === "assistant" ? "assistant" : "user";
822
+ inputItems.push({
823
+ type: "message",
824
+ role: roleOut,
825
+ content: contentItems
826
+ });
827
+ }
828
+ }
829
+ return inputItems;
830
+ }
831
+ buildInstructions(systemPrompt) {
832
+ const systemContent = systemPrompt.filter((content) => content.trim()).join("\n\n");
833
+ return systemContent;
834
+ }
835
+ parseToolCalls(response) {
836
+ if (!response.output || !Array.isArray(response.output)) {
837
+ return [];
838
+ }
839
+ const toolCalls = [];
840
+ for (const item of response.output) {
841
+ if (item.type === "function_call") {
842
+ const callId = item.call_id || item.id;
843
+ const name = item.name || "";
844
+ const args = item.arguments || "{}";
845
+ if (typeof callId === "string" && typeof name === "string" && typeof args === "string") {
846
+ toolCalls.push({
847
+ id: callId,
848
+ type: "function",
849
+ function: {
850
+ name,
851
+ arguments: args
852
+ }
853
+ });
854
+ }
855
+ } else if (item.type === "tool_call") {
856
+ const callId = item.id || `tool_${Math.random().toString(36).substring(2, 15)}`;
857
+ toolCalls.push({
858
+ id: callId,
859
+ type: "tool_call",
860
+ name: item.name,
861
+ arguments: item.arguments
862
+ });
863
+ }
864
+ }
865
+ return toolCalls;
866
+ }
867
+ applyReasoningToMessage(message, reasoningSummaryText, reasoningFullText) {
868
+ const rtxtParts = [];
869
+ if (typeof reasoningSummaryText === "string" && reasoningSummaryText.trim()) {
870
+ rtxtParts.push(reasoningSummaryText);
871
+ }
872
+ if (typeof reasoningFullText === "string" && reasoningFullText.trim()) {
873
+ rtxtParts.push(reasoningFullText);
874
+ }
875
+ const rtxt = rtxtParts.filter((p) => p).join("\n\n");
876
+ if (rtxt) {
877
+ const thinkBlock = `<think>
878
+ ${rtxt}
879
+ </think>
880
+ `;
881
+ const contentText = message.content || "";
882
+ message.content = thinkBlock + (typeof contentText === "string" ? contentText : "");
883
+ }
884
+ return message;
885
+ }
886
+ };
887
+
888
+ // src/services/ai/adapters/chatCompletions.ts
889
+ import { zodToJsonSchema as zodToJsonSchema3 } from "zod-to-json-schema";
890
+ var ChatCompletionsAdapter = class extends OpenAIAdapter {
891
+ createRequest(params) {
892
+ const { messages, systemPrompt, tools, maxTokens, stream } = params;
893
+ const fullMessages = this.buildMessages(systemPrompt, messages);
894
+ const request = {
895
+ model: this.modelProfile.modelName,
896
+ messages: fullMessages,
897
+ [this.getMaxTokensParam()]: maxTokens,
898
+ temperature: this.getTemperature()
899
+ };
900
+ if (tools && tools.length > 0) {
901
+ request.tools = this.buildTools(tools);
902
+ request.tool_choice = "auto";
903
+ }
904
+ if (this.capabilities.parameters.supportsReasoningEffort && params.reasoningEffort) {
905
+ request.reasoning_effort = params.reasoningEffort;
906
+ }
907
+ if (this.capabilities.parameters.supportsVerbosity && params.verbosity) {
908
+ request.verbosity = params.verbosity;
909
+ }
910
+ if (stream && this.capabilities.streaming.supported) {
911
+ request.stream = true;
912
+ if (this.capabilities.streaming.includesUsage) {
913
+ request.stream_options = {
914
+ include_usage: true
915
+ };
916
+ }
917
+ }
918
+ if (this.capabilities.parameters.temperatureMode === "fixed_one") {
919
+ delete request.temperature;
920
+ }
921
+ if (!this.capabilities.streaming.supported) {
922
+ delete request.stream;
923
+ delete request.stream_options;
924
+ }
925
+ return request;
926
+ }
927
+ buildTools(tools) {
928
+ return tools.map((tool) => ({
929
+ type: "function",
930
+ function: {
931
+ name: tool.name,
932
+ description: getToolDescription(tool),
933
+ parameters: tool.inputJSONSchema || zodToJsonSchema3(tool.inputSchema)
934
+ }
935
+ }));
936
+ }
937
+ parseNonStreamingResponse(response) {
938
+ if (!response || typeof response !== "object") {
939
+ throw new Error("Invalid response: response must be an object");
940
+ }
941
+ const choice = response.choices?.[0];
942
+ if (!choice) {
943
+ throw new Error("Invalid response: no choices found in response");
944
+ }
945
+ const message = choice.message || {};
946
+ const content = typeof message.content === "string" ? message.content : "";
947
+ const toolCalls = Array.isArray(message.tool_calls) ? message.tool_calls : [];
948
+ const usage = response.usage || {};
949
+ const promptTokens = Number(usage.prompt_tokens) || 0;
950
+ const completionTokens = Number(usage.completion_tokens) || 0;
951
+ return {
952
+ id: response.id || `chatcmpl_${Date.now()}`,
953
+ content,
954
+ toolCalls,
955
+ usage: {
956
+ promptTokens,
957
+ completionTokens
958
+ }
959
+ };
960
+ }
961
+ buildMessages(systemPrompt, messages) {
962
+ const systemMessages = systemPrompt.map((prompt) => ({
963
+ role: "system",
964
+ content: prompt
965
+ }));
966
+ const normalizedMessages = this.normalizeToolMessages(messages);
967
+ return [...systemMessages, ...normalizedMessages];
968
+ }
969
+ normalizeToolMessages(messages) {
970
+ if (!Array.isArray(messages)) {
971
+ return [];
972
+ }
973
+ return messages.map((msg) => {
974
+ if (!msg || typeof msg !== "object") {
975
+ return msg;
976
+ }
977
+ if (msg.role === "tool") {
978
+ if (Array.isArray(msg.content)) {
979
+ return {
980
+ ...msg,
981
+ content: msg.content.map((c) => c?.text || "").filter(Boolean).join("\n\n") || "(empty content)"
982
+ };
983
+ } else if (typeof msg.content !== "string") {
984
+ return {
985
+ ...msg,
986
+ content: msg.content === null || msg.content === void 0 ? "(empty content)" : JSON.stringify(msg.content)
987
+ };
988
+ }
989
+ }
990
+ return msg;
991
+ });
992
+ }
993
+ async *processStreamingChunk(parsed, responseId, hasStarted, accumulatedContent, reasoningContext) {
994
+ if (!parsed || typeof parsed !== "object") {
995
+ return;
996
+ }
997
+ const choice = parsed.choices?.[0];
998
+ if (choice?.delta && typeof choice.delta === "object") {
999
+ const delta = typeof choice.delta.content === "string" ? choice.delta.content : "";
1000
+ const reasoningDelta = typeof choice.delta.reasoning_content === "string" ? choice.delta.reasoning_content : "";
1001
+ const fullDelta = delta + reasoningDelta;
1002
+ if (fullDelta) {
1003
+ const textEvents = this.handleTextDelta(
1004
+ fullDelta,
1005
+ responseId,
1006
+ hasStarted
1007
+ );
1008
+ for (const event of textEvents) {
1009
+ yield event;
1010
+ }
1011
+ }
1012
+ }
1013
+ if (choice?.delta?.tool_calls && Array.isArray(choice.delta.tool_calls)) {
1014
+ for (const toolCall of choice.delta.tool_calls) {
1015
+ if (toolCall && typeof toolCall === "object") {
1016
+ yield {
1017
+ type: "tool_request",
1018
+ tool: {
1019
+ id: toolCall.id || `tool_${Date.now()}`,
1020
+ name: toolCall.function?.name || "unknown",
1021
+ input: toolCall.function?.arguments || "{}"
1022
+ }
1023
+ };
1024
+ }
1025
+ }
1026
+ }
1027
+ if (parsed.usage && typeof parsed.usage === "object") {
1028
+ const normalizedUsage = normalizeTokens(parsed.usage);
1029
+ this.updateCumulativeUsage(normalizedUsage);
1030
+ yield {
1031
+ type: "usage",
1032
+ usage: { ...this.cumulativeUsage }
1033
+ };
1034
+ }
1035
+ }
1036
+ updateStreamingState(parsed, accumulatedContent) {
1037
+ const state = {};
1038
+ const choice = parsed.choices?.[0];
1039
+ if (choice?.delta) {
1040
+ const delta = choice.delta.content || "";
1041
+ const reasoningDelta = choice.delta.reasoning_content || "";
1042
+ const fullDelta = delta + reasoningDelta;
1043
+ if (fullDelta) {
1044
+ state.content = accumulatedContent + fullDelta;
1045
+ state.hasStarted = true;
1046
+ }
1047
+ }
1048
+ return state;
1049
+ }
1050
+ async parseStreamingOpenAIResponse(response, signal) {
1051
+ const contentBlocks = [];
1052
+ const usage = {
1053
+ prompt_tokens: 0,
1054
+ completion_tokens: 0
1055
+ };
1056
+ let responseId = response.id || `chatcmpl_${Date.now()}`;
1057
+ const pendingToolCalls = [];
1058
+ let hasMarkedStreaming = false;
1059
+ try {
1060
+ this.resetCumulativeUsage();
1061
+ for await (const event of this.parseStreamingResponse(response)) {
1062
+ if (signal?.aborted) {
1063
+ throw new Error("Stream aborted by user");
1064
+ }
1065
+ if (event.type === "message_start") {
1066
+ responseId = event.responseId || responseId;
1067
+ continue;
1068
+ }
1069
+ if (event.type === "text_delta") {
1070
+ if (!hasMarkedStreaming) {
1071
+ setRequestStatus({ kind: "streaming" });
1072
+ hasMarkedStreaming = true;
1073
+ }
1074
+ const last = contentBlocks[contentBlocks.length - 1];
1075
+ if (!last || last.type !== "text") {
1076
+ contentBlocks.push({
1077
+ type: "text",
1078
+ text: event.delta,
1079
+ citations: []
1080
+ });
1081
+ } else {
1082
+ last.text += event.delta;
1083
+ }
1084
+ continue;
1085
+ }
1086
+ if (event.type === "tool_request") {
1087
+ setRequestStatus({ kind: "tool", detail: event.tool?.name });
1088
+ pendingToolCalls.push(event.tool);
1089
+ continue;
1090
+ }
1091
+ if (event.type === "usage") {
1092
+ usage.prompt_tokens = event.usage.input;
1093
+ usage.completion_tokens = event.usage.output;
1094
+ usage.totalTokens = event.usage.total ?? event.usage.input + event.usage.output;
1095
+ usage.promptTokens = event.usage.input;
1096
+ usage.completionTokens = event.usage.output;
1097
+ continue;
1098
+ }
1099
+ }
1100
+ } catch (error) {
1101
+ if (signal?.aborted) {
1102
+ const assistantMessage2 = {
1103
+ type: "assistant",
1104
+ message: {
1105
+ role: "assistant",
1106
+ content: contentBlocks,
1107
+ usage: {
1108
+ input_tokens: usage.prompt_tokens ?? 0,
1109
+ output_tokens: usage.completion_tokens ?? 0,
1110
+ prompt_tokens: usage.prompt_tokens ?? 0,
1111
+ completion_tokens: usage.completion_tokens ?? 0,
1112
+ totalTokens: (usage.prompt_tokens || 0) + (usage.completion_tokens || 0)
1113
+ }
1114
+ },
1115
+ costUSD: 0,
1116
+ durationMs: Date.now() - Date.now(),
1117
+ uuid: `${Date.now()}-${Math.random().toString(36).slice(2, 11)}`,
1118
+ responseId
1119
+ };
1120
+ return {
1121
+ assistantMessage: assistantMessage2,
1122
+ rawResponse: {
1123
+ id: responseId,
1124
+ content: contentBlocks,
1125
+ usage,
1126
+ aborted: true
1127
+ }
1128
+ };
1129
+ }
1130
+ throw error;
1131
+ }
1132
+ for (const toolCall of pendingToolCalls) {
1133
+ let toolArgs = {};
1134
+ try {
1135
+ toolArgs = toolCall.input ? JSON.parse(toolCall.input) : {};
1136
+ } catch {
1137
+ }
1138
+ contentBlocks.push({
1139
+ type: "tool_use",
1140
+ id: toolCall.id,
1141
+ name: toolCall.name,
1142
+ input: toolArgs
1143
+ });
1144
+ }
1145
+ const assistantMessage = {
1146
+ type: "assistant",
1147
+ message: {
1148
+ role: "assistant",
1149
+ content: contentBlocks,
1150
+ usage: {
1151
+ input_tokens: usage.prompt_tokens ?? 0,
1152
+ output_tokens: usage.completion_tokens ?? 0,
1153
+ prompt_tokens: usage.prompt_tokens ?? 0,
1154
+ completion_tokens: usage.completion_tokens ?? 0,
1155
+ totalTokens: usage.totalTokens ?? (usage.prompt_tokens || 0) + (usage.completion_tokens || 0)
1156
+ }
1157
+ },
1158
+ costUSD: 0,
1159
+ durationMs: Date.now() - Date.now(),
1160
+ uuid: `${Date.now()}-${Math.random().toString(36).slice(2, 11)}`,
1161
+ responseId
1162
+ };
1163
+ return {
1164
+ assistantMessage,
1165
+ rawResponse: {
1166
+ id: responseId,
1167
+ content: contentBlocks,
1168
+ usage
1169
+ }
1170
+ };
1171
+ }
1172
+ normalizeUsageForAdapter(usage) {
1173
+ return super.normalizeUsageForAdapter(usage);
1174
+ }
1175
+ };
1176
+
1177
+ // src/constants/modelCapabilities.ts
1178
+ var GPT5_CAPABILITIES = {
1179
+ apiArchitecture: {
1180
+ primary: "responses_api",
1181
+ fallback: "chat_completions"
1182
+ },
1183
+ parameters: {
1184
+ maxTokensField: "max_output_tokens",
1185
+ supportsReasoningEffort: true,
1186
+ supportsVerbosity: true,
1187
+ temperatureMode: "fixed_one"
1188
+ },
1189
+ toolCalling: {
1190
+ mode: "custom_tools",
1191
+ supportsFreeform: true,
1192
+ supportsAllowedTools: true,
1193
+ supportsParallelCalls: true
1194
+ },
1195
+ stateManagement: {
1196
+ supportsResponseId: true,
1197
+ supportsConversationChaining: true,
1198
+ supportsPreviousResponseId: true
1199
+ },
1200
+ streaming: {
1201
+ supported: true,
1202
+ includesUsage: true
1203
+ }
1204
+ };
1205
+ var CHAT_COMPLETIONS_CAPABILITIES = {
1206
+ apiArchitecture: {
1207
+ primary: "chat_completions"
1208
+ },
1209
+ parameters: {
1210
+ maxTokensField: "max_tokens",
1211
+ supportsReasoningEffort: false,
1212
+ supportsVerbosity: false,
1213
+ temperatureMode: "flexible"
1214
+ },
1215
+ toolCalling: {
1216
+ mode: "function_calling",
1217
+ supportsFreeform: false,
1218
+ supportsAllowedTools: false,
1219
+ supportsParallelCalls: true
1220
+ },
1221
+ stateManagement: {
1222
+ supportsResponseId: false,
1223
+ supportsConversationChaining: false,
1224
+ supportsPreviousResponseId: false
1225
+ },
1226
+ streaming: {
1227
+ supported: true,
1228
+ includesUsage: true
1229
+ }
1230
+ };
1231
+ var MODEL_CAPABILITIES_REGISTRY = {
1232
+ "gpt-5": GPT5_CAPABILITIES,
1233
+ "gpt-5-mini": GPT5_CAPABILITIES,
1234
+ "gpt-5-nano": GPT5_CAPABILITIES,
1235
+ "gpt-5-chat-latest": GPT5_CAPABILITIES,
1236
+ "gpt-5-codex": GPT5_CAPABILITIES,
1237
+ "gpt-4o": CHAT_COMPLETIONS_CAPABILITIES,
1238
+ "gpt-4o-mini": CHAT_COMPLETIONS_CAPABILITIES,
1239
+ "gpt-4-turbo": CHAT_COMPLETIONS_CAPABILITIES,
1240
+ "gpt-4": CHAT_COMPLETIONS_CAPABILITIES,
1241
+ "claude-3-5-sonnet-20241022": CHAT_COMPLETIONS_CAPABILITIES,
1242
+ "claude-3-5-haiku-20241022": CHAT_COMPLETIONS_CAPABILITIES,
1243
+ "claude-3-opus-20240229": CHAT_COMPLETIONS_CAPABILITIES,
1244
+ o1: {
1245
+ ...CHAT_COMPLETIONS_CAPABILITIES,
1246
+ parameters: {
1247
+ ...CHAT_COMPLETIONS_CAPABILITIES.parameters,
1248
+ maxTokensField: "max_completion_tokens",
1249
+ temperatureMode: "fixed_one"
1250
+ }
1251
+ },
1252
+ "o1-mini": {
1253
+ ...CHAT_COMPLETIONS_CAPABILITIES,
1254
+ parameters: {
1255
+ ...CHAT_COMPLETIONS_CAPABILITIES.parameters,
1256
+ maxTokensField: "max_completion_tokens",
1257
+ temperatureMode: "fixed_one"
1258
+ }
1259
+ },
1260
+ "o1-preview": {
1261
+ ...CHAT_COMPLETIONS_CAPABILITIES,
1262
+ parameters: {
1263
+ ...CHAT_COMPLETIONS_CAPABILITIES.parameters,
1264
+ maxTokensField: "max_completion_tokens",
1265
+ temperatureMode: "fixed_one"
1266
+ }
1267
+ }
1268
+ };
1269
+ function inferModelCapabilities(modelName) {
1270
+ if (!modelName) return null;
1271
+ const lowerName = modelName.toLowerCase();
1272
+ if (lowerName.includes("gpt-5") || lowerName.includes("gpt5")) {
1273
+ return GPT5_CAPABILITIES;
1274
+ }
1275
+ if (lowerName.includes("gpt-6") || lowerName.includes("gpt6")) {
1276
+ return {
1277
+ ...GPT5_CAPABILITIES,
1278
+ streaming: { supported: true, includesUsage: true }
1279
+ };
1280
+ }
1281
+ if (lowerName.includes("glm-5") || lowerName.includes("glm5")) {
1282
+ return {
1283
+ ...CHAT_COMPLETIONS_CAPABILITIES,
1284
+ toolCalling: {
1285
+ ...CHAT_COMPLETIONS_CAPABILITIES.toolCalling,
1286
+ supportsAllowedTools: false
1287
+ }
1288
+ };
1289
+ }
1290
+ if (lowerName.startsWith("o1") || lowerName.includes("o1-")) {
1291
+ return {
1292
+ ...CHAT_COMPLETIONS_CAPABILITIES,
1293
+ parameters: {
1294
+ ...CHAT_COMPLETIONS_CAPABILITIES.parameters,
1295
+ maxTokensField: "max_completion_tokens",
1296
+ temperatureMode: "fixed_one"
1297
+ }
1298
+ };
1299
+ }
1300
+ return null;
1301
+ }
1302
+ var capabilityCache = /* @__PURE__ */ new Map();
1303
+ function getModelCapabilities(modelName) {
1304
+ if (capabilityCache.has(modelName)) {
1305
+ return capabilityCache.get(modelName);
1306
+ }
1307
+ if (MODEL_CAPABILITIES_REGISTRY[modelName]) {
1308
+ const capabilities = MODEL_CAPABILITIES_REGISTRY[modelName];
1309
+ capabilityCache.set(modelName, capabilities);
1310
+ return capabilities;
1311
+ }
1312
+ const inferred = inferModelCapabilities(modelName);
1313
+ if (inferred) {
1314
+ capabilityCache.set(modelName, inferred);
1315
+ return inferred;
1316
+ }
1317
+ const defaultCapabilities = CHAT_COMPLETIONS_CAPABILITIES;
1318
+ capabilityCache.set(modelName, defaultCapabilities);
1319
+ return defaultCapabilities;
1320
+ }
1321
+
1322
+ // src/services/ai/modelAdapterFactory.ts
1323
+ var ModelAdapterFactory = class {
1324
+ static createAdapter(modelProfile) {
1325
+ const capabilities = getModelCapabilities(modelProfile.modelName);
1326
+ const apiType = this.determineAPIType(modelProfile, capabilities);
1327
+ switch (apiType) {
1328
+ case "responses_api":
1329
+ return new ResponsesAPIAdapter(capabilities, modelProfile);
1330
+ case "chat_completions":
1331
+ default:
1332
+ return new ChatCompletionsAdapter(capabilities, modelProfile);
1333
+ }
1334
+ }
1335
+ static determineAPIType(modelProfile, capabilities) {
1336
+ if (capabilities.apiArchitecture.primary !== "responses_api") {
1337
+ return "chat_completions";
1338
+ }
1339
+ const isOfficialOpenAI = !modelProfile.baseURL || modelProfile.baseURL.includes("api.openai.com");
1340
+ if (!isOfficialOpenAI) {
1341
+ if (capabilities.apiArchitecture.fallback === "chat_completions") {
1342
+ return capabilities.apiArchitecture.primary;
1343
+ }
1344
+ return capabilities.apiArchitecture.primary;
1345
+ }
1346
+ return capabilities.apiArchitecture.primary;
1347
+ }
1348
+ static shouldUseResponsesAPI(modelProfile) {
1349
+ const capabilities = getModelCapabilities(modelProfile.modelName);
1350
+ const apiType = this.determineAPIType(modelProfile, capabilities);
1351
+ return apiType === "responses_api";
1352
+ }
1353
+ };
1354
+
1355
+ // src/services/ai/responseStateManager.ts
1356
+ var ResponseStateManager = class {
1357
+ conversationStates = /* @__PURE__ */ new Map();
1358
+ CLEANUP_INTERVAL = 60 * 60 * 1e3;
1359
+ constructor() {
1360
+ setInterval(() => {
1361
+ this.cleanup();
1362
+ }, this.CLEANUP_INTERVAL);
1363
+ }
1364
+ setPreviousResponseId(conversationId, responseId) {
1365
+ this.conversationStates.set(conversationId, {
1366
+ previousResponseId: responseId,
1367
+ lastUpdate: Date.now()
1368
+ });
1369
+ }
1370
+ getPreviousResponseId(conversationId) {
1371
+ const state = this.conversationStates.get(conversationId);
1372
+ if (state) {
1373
+ state.lastUpdate = Date.now();
1374
+ return state.previousResponseId;
1375
+ }
1376
+ return void 0;
1377
+ }
1378
+ clearConversation(conversationId) {
1379
+ this.conversationStates.delete(conversationId);
1380
+ }
1381
+ clearAll() {
1382
+ this.conversationStates.clear();
1383
+ }
1384
+ cleanup() {
1385
+ const now = Date.now();
1386
+ for (const [conversationId, state] of this.conversationStates.entries()) {
1387
+ if (now - state.lastUpdate > this.CLEANUP_INTERVAL) {
1388
+ this.conversationStates.delete(conversationId);
1389
+ }
1390
+ }
1391
+ }
1392
+ getStateSize() {
1393
+ return this.conversationStates.size;
1394
+ }
1395
+ };
1396
+ var responseStateManager = new ResponseStateManager();
1397
+ function getConversationId(agentId, messageId) {
1398
+ return agentId || messageId || `conv_${Date.now()}_${Math.random().toString(36).substr(2, 9)}`;
1399
+ }
1400
+
1401
+ // src/services/ai/llm.ts
1402
+ import { nanoid } from "nanoid";
1403
+
1404
+ // src/utils/tooling/toolUsePartialJson.ts
1405
+ function tokenizePartialJson(input) {
1406
+ let index = 0;
1407
+ const tokens = [];
1408
+ while (index < input.length) {
1409
+ let ch = input[index];
1410
+ if (ch === "\\") {
1411
+ index++;
1412
+ continue;
1413
+ }
1414
+ if (ch === "{") {
1415
+ tokens.push({ type: "brace", value: "{" });
1416
+ index++;
1417
+ continue;
1418
+ }
1419
+ if (ch === "}") {
1420
+ tokens.push({ type: "brace", value: "}" });
1421
+ index++;
1422
+ continue;
1423
+ }
1424
+ if (ch === "[") {
1425
+ tokens.push({ type: "paren", value: "[" });
1426
+ index++;
1427
+ continue;
1428
+ }
1429
+ if (ch === "]") {
1430
+ tokens.push({ type: "paren", value: "]" });
1431
+ index++;
1432
+ continue;
1433
+ }
1434
+ if (ch === ":") {
1435
+ tokens.push({ type: "separator", value: ":" });
1436
+ index++;
1437
+ continue;
1438
+ }
1439
+ if (ch === ",") {
1440
+ tokens.push({ type: "delimiter", value: "," });
1441
+ index++;
1442
+ continue;
1443
+ }
1444
+ if (ch === '"') {
1445
+ let value = "";
1446
+ let incomplete = false;
1447
+ ch = input[++index];
1448
+ while (ch !== '"') {
1449
+ if (index === input.length) {
1450
+ incomplete = true;
1451
+ break;
1452
+ }
1453
+ if (ch === "\\") {
1454
+ if (++index === input.length) {
1455
+ incomplete = true;
1456
+ break;
1457
+ }
1458
+ value += ch + input[index];
1459
+ ch = input[++index];
1460
+ } else {
1461
+ value += ch;
1462
+ ch = input[++index];
1463
+ }
1464
+ }
1465
+ ch = input[++index];
1466
+ if (!incomplete) tokens.push({ type: "string", value });
1467
+ continue;
1468
+ }
1469
+ if (ch && /\s/.test(ch)) {
1470
+ index++;
1471
+ continue;
1472
+ }
1473
+ const digit = /[0-9]/;
1474
+ if (ch && digit.test(ch) || ch === "-" || ch === ".") {
1475
+ let value = "";
1476
+ if (ch === "-") {
1477
+ value += ch;
1478
+ ch = input[++index];
1479
+ }
1480
+ while (ch && digit.test(ch) || ch === ".") {
1481
+ value += ch;
1482
+ ch = input[++index];
1483
+ }
1484
+ tokens.push({ type: "number", value });
1485
+ continue;
1486
+ }
1487
+ const alpha = /[a-z]/i;
1488
+ if (ch && alpha.test(ch)) {
1489
+ let value = "";
1490
+ while (ch && alpha.test(ch)) {
1491
+ if (index === input.length) break;
1492
+ value += ch;
1493
+ ch = input[++index];
1494
+ }
1495
+ if (value === "true" || value === "false" || value === "null") {
1496
+ tokens.push({ type: "name", value });
1497
+ } else {
1498
+ index++;
1499
+ }
1500
+ continue;
1501
+ }
1502
+ index++;
1503
+ }
1504
+ return tokens;
1505
+ }
1506
+ function trimTrailingIncompleteTokens(tokens) {
1507
+ if (tokens.length === 0) return tokens;
1508
+ const last = tokens[tokens.length - 1];
1509
+ if (last.type === "separator") {
1510
+ return trimTrailingIncompleteTokens(tokens.slice(0, -1));
1511
+ }
1512
+ if (last.type === "number") {
1513
+ const lastChar = last.value[last.value.length - 1];
1514
+ if (lastChar === "." || lastChar === "-") {
1515
+ return trimTrailingIncompleteTokens(tokens.slice(0, -1));
1516
+ }
1517
+ }
1518
+ if (last.type === "string" || last.type === "number") {
1519
+ const previous = tokens[tokens.length - 2];
1520
+ if (previous?.type === "delimiter") {
1521
+ return trimTrailingIncompleteTokens(tokens.slice(0, -1));
1522
+ }
1523
+ if (previous?.type === "brace" && previous.value === "{") {
1524
+ return trimTrailingIncompleteTokens(tokens.slice(0, -1));
1525
+ }
1526
+ }
1527
+ if (last.type === "delimiter") {
1528
+ return trimTrailingIncompleteTokens(tokens.slice(0, -1));
1529
+ }
1530
+ return tokens;
1531
+ }
1532
+ function closeOpenBrackets(tokens) {
1533
+ const missingClosers = [];
1534
+ for (const token of tokens) {
1535
+ if (token.type === "brace") {
1536
+ if (token.value === "{") missingClosers.push("}");
1537
+ else missingClosers.splice(missingClosers.lastIndexOf("}"), 1);
1538
+ continue;
1539
+ }
1540
+ if (token.type === "paren") {
1541
+ if (token.value === "[") missingClosers.push("]");
1542
+ else missingClosers.splice(missingClosers.lastIndexOf("]"), 1);
1543
+ }
1544
+ }
1545
+ if (missingClosers.length > 0) {
1546
+ missingClosers.reverse();
1547
+ for (const closer of missingClosers) {
1548
+ if (closer === "}") tokens.push({ type: "brace", value: "}" });
1549
+ else tokens.push({ type: "paren", value: "]" });
1550
+ }
1551
+ }
1552
+ return tokens;
1553
+ }
1554
+ function tokensToJson(tokens) {
1555
+ let out = "";
1556
+ for (const token of tokens) {
1557
+ if (token.type === "string") out += `"${token.value}"`;
1558
+ else out += token.value;
1559
+ }
1560
+ return out;
1561
+ }
1562
+ function parseToolUsePartialJson(input) {
1563
+ const tokens = tokenizePartialJson(input);
1564
+ const trimmed = trimTrailingIncompleteTokens(tokens);
1565
+ const completed = closeOpenBrackets(trimmed);
1566
+ return JSON.parse(tokensToJson(completed));
1567
+ }
1568
+ function parseToolUsePartialJsonOrThrow(input) {
1569
+ try {
1570
+ return parseToolUsePartialJson(input);
1571
+ } catch (error) {
1572
+ throw new Error(
1573
+ `Unable to parse tool parameter JSON from model. Please retry your request or adjust your prompt. Error: ${String(error)}. JSON: ${input}`
1574
+ );
1575
+ }
1576
+ }
1577
+
1578
+ // src/utils/model/openaiMessageConversion.ts
1579
+ function convertAnthropicMessagesToOpenAIMessages(messages) {
1580
+ const openaiMessages = [];
1581
+ const toolResults = {};
1582
+ for (const message of messages) {
1583
+ const blocks = [];
1584
+ if (typeof message.message.content === "string") {
1585
+ blocks.push({ type: "text", text: message.message.content });
1586
+ } else if (Array.isArray(message.message.content)) {
1587
+ blocks.push(...message.message.content);
1588
+ } else if (message.message.content) {
1589
+ blocks.push(message.message.content);
1590
+ }
1591
+ const role = message.message.role;
1592
+ const userContentParts = [];
1593
+ const assistantTextParts = [];
1594
+ const assistantToolCalls = [];
1595
+ for (const block of blocks) {
1596
+ if (block.type === "text") {
1597
+ const text = typeof block.text === "string" ? block.text : "";
1598
+ if (!text) continue;
1599
+ if (role === "user") {
1600
+ userContentParts.push({ type: "text", text });
1601
+ } else if (role === "assistant") {
1602
+ assistantTextParts.push(text);
1603
+ }
1604
+ continue;
1605
+ }
1606
+ if (block.type === "image" && role === "user") {
1607
+ const source = block.source;
1608
+ if (source?.type === "base64") {
1609
+ userContentParts.push({
1610
+ type: "image_url",
1611
+ image_url: {
1612
+ url: `data:${source.media_type};base64,${source.data}`
1613
+ }
1614
+ });
1615
+ } else if (source?.type === "url") {
1616
+ userContentParts.push({
1617
+ type: "image_url",
1618
+ image_url: { url: source.url }
1619
+ });
1620
+ }
1621
+ continue;
1622
+ }
1623
+ if (block.type === "tool_use") {
1624
+ assistantToolCalls.push({
1625
+ type: "function",
1626
+ function: {
1627
+ name: block.name,
1628
+ arguments: JSON.stringify(block.input)
1629
+ },
1630
+ id: block.id
1631
+ });
1632
+ continue;
1633
+ }
1634
+ if (block.type === "tool_result") {
1635
+ const toolUseId = block.tool_use_id;
1636
+ const rawToolContent = block.content;
1637
+ const toolContent = typeof rawToolContent === "string" ? rawToolContent : JSON.stringify(rawToolContent);
1638
+ toolResults[toolUseId] = {
1639
+ role: "tool",
1640
+ content: toolContent,
1641
+ tool_call_id: toolUseId
1642
+ };
1643
+ continue;
1644
+ }
1645
+ }
1646
+ if (role === "user") {
1647
+ if (userContentParts.length === 1 && userContentParts[0]?.type === "text") {
1648
+ openaiMessages.push({
1649
+ role: "user",
1650
+ content: userContentParts[0].text
1651
+ });
1652
+ } else if (userContentParts.length > 0) {
1653
+ openaiMessages.push({ role: "user", content: userContentParts });
1654
+ }
1655
+ continue;
1656
+ }
1657
+ if (role === "assistant") {
1658
+ const text = assistantTextParts.filter(Boolean).join("\n");
1659
+ if (assistantToolCalls.length > 0) {
1660
+ openaiMessages.push({
1661
+ role: "assistant",
1662
+ content: text ? text : void 0,
1663
+ tool_calls: assistantToolCalls
1664
+ });
1665
+ continue;
1666
+ }
1667
+ if (text) {
1668
+ openaiMessages.push({ role: "assistant", content: text });
1669
+ }
1670
+ }
1671
+ }
1672
+ const finalMessages = [];
1673
+ for (const message of openaiMessages) {
1674
+ finalMessages.push(message);
1675
+ if ("tool_calls" in message && message.tool_calls) {
1676
+ for (const toolCall of message.tool_calls) {
1677
+ if (toolResults[toolCall.id]) {
1678
+ finalMessages.push(toolResults[toolCall.id]);
1679
+ }
1680
+ }
1681
+ }
1682
+ }
1683
+ return finalMessages;
1684
+ }
1685
+
1686
+ // src/services/ai/llm.ts
1687
+ var AnthropicBedrock = class {
1688
+ constructor(_args) {
1689
+ throw new Error("Bedrock SDK not installed. Install @anthropic-ai/bedrock-sdk to use Bedrock.");
1690
+ }
1691
+ };
1692
+ var AnthropicVertex = class {
1693
+ constructor(_args) {
1694
+ throw new Error("Vertex SDK not installed. Install @anthropic-ai/vertex-sdk to use Vertex.");
1695
+ }
1696
+ };
1697
+ function isGPT5Model(modelName) {
1698
+ return modelName.startsWith("gpt-5");
1699
+ }
1700
+ var PROMPT_CACHING_ENABLED = !process.env.DISABLE_PROMPT_CACHING;
1701
+ var SONNET_COST_PER_MILLION_INPUT_TOKENS = 3;
1702
+ var SONNET_COST_PER_MILLION_OUTPUT_TOKENS = 15;
1703
+ var SONNET_COST_PER_MILLION_PROMPT_CACHE_WRITE_TOKENS = 3.75;
1704
+ var SONNET_COST_PER_MILLION_PROMPT_CACHE_READ_TOKENS = 0.3;
1705
+ var MAX_RETRIES = process.env.USER_TYPE === "SWE_BENCH" ? 100 : 10;
1706
+ var BASE_DELAY_MS = 500;
1707
+ function abortableDelay(delayMs, signal) {
1708
+ return new Promise((resolve, reject) => {
1709
+ if (signal?.aborted) {
1710
+ reject(new Error("Request was aborted"));
1711
+ return;
1712
+ }
1713
+ const timeoutId = setTimeout(() => {
1714
+ resolve();
1715
+ }, delayMs);
1716
+ if (signal) {
1717
+ const abortHandler = () => {
1718
+ clearTimeout(timeoutId);
1719
+ reject(new Error("Request was aborted"));
1720
+ };
1721
+ signal.addEventListener("abort", abortHandler, { once: true });
1722
+ }
1723
+ });
1724
+ }
1725
+ function getRetryDelay(attempt, retryAfterHeader) {
1726
+ if (retryAfterHeader) {
1727
+ const seconds = parseInt(retryAfterHeader, 10);
1728
+ if (!isNaN(seconds)) {
1729
+ return seconds * 1e3;
1730
+ }
1731
+ }
1732
+ return Math.min(BASE_DELAY_MS * Math.pow(2, attempt - 1), 32e3);
1733
+ }
1734
+ function shouldRetry(error) {
1735
+ if (error.message?.includes('"type":"overloaded_error"')) {
1736
+ return process.env.USER_TYPE === "SWE_BENCH";
1737
+ }
1738
+ const shouldRetryHeader = error.headers?.["x-should-retry"];
1739
+ if (shouldRetryHeader === "true") return true;
1740
+ if (shouldRetryHeader === "false") return false;
1741
+ if (error instanceof APIConnectionError) {
1742
+ return true;
1743
+ }
1744
+ if (!error.status) return false;
1745
+ if (error.status === 408) return true;
1746
+ if (error.status === 409) return true;
1747
+ if (error.status === 429) return true;
1748
+ if (error.status && error.status >= 500) return true;
1749
+ return false;
1750
+ }
1751
+ async function withRetry(operation, options = {}) {
1752
+ const maxRetries = options.maxRetries ?? MAX_RETRIES;
1753
+ let lastError;
1754
+ for (let attempt = 1; attempt <= maxRetries + 1; attempt++) {
1755
+ try {
1756
+ return await operation(attempt);
1757
+ } catch (error) {
1758
+ lastError = error;
1759
+ if (attempt > maxRetries || !(error instanceof APIError) || !shouldRetry(error)) {
1760
+ throw error;
1761
+ }
1762
+ if (options.signal?.aborted) {
1763
+ throw new Error("Request cancelled by user");
1764
+ }
1765
+ const retryAfter = error.headers?.["retry-after"] ?? null;
1766
+ const delayMs = getRetryDelay(attempt, retryAfter);
1767
+ debug.warn("LLM_API_RETRY", {
1768
+ name: error.name,
1769
+ message: error.message,
1770
+ status: error.status,
1771
+ attempt,
1772
+ maxRetries,
1773
+ delayMs
1774
+ });
1775
+ try {
1776
+ await abortableDelay(delayMs, options.signal);
1777
+ } catch (delayError) {
1778
+ if (delayError.message === "Request was aborted") {
1779
+ throw new Error("Request cancelled by user");
1780
+ }
1781
+ throw delayError;
1782
+ }
1783
+ }
1784
+ }
1785
+ throw lastError;
1786
+ }
1787
+ async function fetchAnthropicModels(baseURL, apiKey) {
1788
+ try {
1789
+ const modelsURL = baseURL ? `${baseURL.replace(/\/+$/, "")}/v1/models` : "https://api.anthropic.com/v1/models";
1790
+ const response = await fetch(modelsURL, {
1791
+ method: "GET",
1792
+ headers: {
1793
+ "x-api-key": apiKey,
1794
+ "anthropic-version": "2023-06-01",
1795
+ "User-Agent": USER_AGENT
1796
+ }
1797
+ });
1798
+ if (!response.ok) {
1799
+ if (response.status === 401) {
1800
+ throw new Error(
1801
+ "Invalid API key. Please check your Anthropic API key and try again."
1802
+ );
1803
+ } else if (response.status === 403) {
1804
+ throw new Error(
1805
+ "API key does not have permission to access models. Please check your API key permissions."
1806
+ );
1807
+ } else if (response.status === 429) {
1808
+ throw new Error(
1809
+ "Too many requests. Please wait a moment and try again."
1810
+ );
1811
+ } else if (response.status >= 500) {
1812
+ throw new Error(
1813
+ "Anthropic service is temporarily unavailable. Please try again later."
1814
+ );
1815
+ } else {
1816
+ throw new Error(
1817
+ `Unable to connect to Anthropic API (${response.status}). Please check your internet connection and API key.`
1818
+ );
1819
+ }
1820
+ }
1821
+ const data = await response.json();
1822
+ return data.data || [];
1823
+ } catch (error) {
1824
+ if (error instanceof Error && error.message.includes("API key") || error instanceof Error && error.message.includes("Anthropic")) {
1825
+ throw error;
1826
+ }
1827
+ logError(error);
1828
+ debug.warn("ANTHROPIC_MODELS_FETCH_FAILED", {
1829
+ error: error instanceof Error ? error.message : String(error)
1830
+ });
1831
+ throw new Error(
1832
+ "Unable to connect to Anthropic API. Please check your internet connection and try again."
1833
+ );
1834
+ }
1835
+ }
1836
+ async function verifyApiKey(apiKey, baseURL, provider) {
1837
+ if (!apiKey) {
1838
+ return false;
1839
+ }
1840
+ if (provider && provider !== "anthropic") {
1841
+ try {
1842
+ const headers = {
1843
+ Authorization: `Bearer ${apiKey}`,
1844
+ "Content-Type": "application/json"
1845
+ };
1846
+ if (!baseURL) {
1847
+ debug.warn("API_VERIFICATION_MISSING_BASE_URL", { provider });
1848
+ return false;
1849
+ }
1850
+ const modelsURL = `${baseURL.replace(/\/+$/, "")}/models`;
1851
+ const response = await fetch(modelsURL, {
1852
+ method: "GET",
1853
+ headers
1854
+ });
1855
+ return response.ok;
1856
+ } catch (error) {
1857
+ logError(error);
1858
+ debug.warn("API_VERIFICATION_FAILED", {
1859
+ provider,
1860
+ error: error instanceof Error ? error.message : String(error)
1861
+ });
1862
+ return false;
1863
+ }
1864
+ }
1865
+ const clientConfig = {
1866
+ apiKey,
1867
+ dangerouslyAllowBrowser: true,
1868
+ maxRetries: 3,
1869
+ defaultHeaders: {
1870
+ "User-Agent": USER_AGENT
1871
+ }
1872
+ };
1873
+ if (baseURL && (provider === "anthropic" || provider === "minimax-coding")) {
1874
+ clientConfig.baseURL = baseURL;
1875
+ }
1876
+ const anthropic = new Anthropic(clientConfig);
1877
+ try {
1878
+ await withRetry(
1879
+ async () => {
1880
+ const model = "claude-sonnet-4-20250514";
1881
+ const messages = [{ role: "user", content: "test" }];
1882
+ await anthropic.messages.create({
1883
+ model,
1884
+ max_tokens: 1e3,
1885
+ messages,
1886
+ temperature: 0
1887
+ });
1888
+ return true;
1889
+ },
1890
+ { maxRetries: 2 }
1891
+ );
1892
+ return true;
1893
+ } catch (error) {
1894
+ logError(error);
1895
+ if (error instanceof Error && error.message.includes(
1896
+ '{"type":"error","error":{"type":"authentication_error","message":"invalid x-api-key"}}'
1897
+ )) {
1898
+ return false;
1899
+ }
1900
+ throw error;
1901
+ }
1902
+ }
1903
+ function convertAnthropicMessagesToOpenAIMessages2(messages) {
1904
+ return convertAnthropicMessagesToOpenAIMessages(messages);
1905
+ }
1906
+ function messageReducer(previous, item) {
1907
+ const reduce = (acc, delta) => {
1908
+ acc = { ...acc };
1909
+ for (const [key, value] of Object.entries(delta)) {
1910
+ if (acc[key] === void 0 || acc[key] === null) {
1911
+ acc[key] = value;
1912
+ if (Array.isArray(acc[key])) {
1913
+ for (const arr of acc[key]) {
1914
+ delete arr.index;
1915
+ }
1916
+ }
1917
+ } else if (typeof acc[key] === "string" && typeof value === "string") {
1918
+ acc[key] += value;
1919
+ } else if (typeof acc[key] === "number" && typeof value === "number") {
1920
+ acc[key] = value;
1921
+ } else if (Array.isArray(acc[key]) && Array.isArray(value)) {
1922
+ const accArray = acc[key];
1923
+ for (let i = 0; i < value.length; i++) {
1924
+ const { index, ...chunkTool } = value[i];
1925
+ if (index - accArray.length > 1) {
1926
+ throw new Error(
1927
+ `Error: An array has an empty value when tool_calls are constructed. tool_calls: ${accArray}; tool: ${value}`
1928
+ );
1929
+ }
1930
+ accArray[index] = reduce(accArray[index], chunkTool);
1931
+ }
1932
+ } else if (typeof acc[key] === "object" && typeof value === "object") {
1933
+ acc[key] = reduce(acc[key], value);
1934
+ }
1935
+ }
1936
+ return acc;
1937
+ };
1938
+ const choice = item.choices?.[0];
1939
+ if (!choice) {
1940
+ return previous;
1941
+ }
1942
+ return reduce(previous, choice.delta);
1943
+ }
1944
+ async function handleMessageStream(stream, signal) {
1945
+ const streamStartTime = Date.now();
1946
+ let ttftMs;
1947
+ let chunkCount = 0;
1948
+ let errorCount = 0;
1949
+ debug.api("OPENAI_STREAM_START", {
1950
+ streamStartTime: String(streamStartTime)
1951
+ });
1952
+ let message = {};
1953
+ let id, model, created, object, usage;
1954
+ try {
1955
+ for await (const chunk of stream) {
1956
+ if (signal?.aborted) {
1957
+ debug.flow("OPENAI_STREAM_ABORTED", {
1958
+ chunkCount,
1959
+ timestamp: Date.now()
1960
+ });
1961
+ throw new Error("Request was cancelled");
1962
+ }
1963
+ chunkCount++;
1964
+ try {
1965
+ if (!id) {
1966
+ id = chunk.id;
1967
+ debug.api("OPENAI_STREAM_ID_RECEIVED", {
1968
+ id,
1969
+ chunkNumber: String(chunkCount)
1970
+ });
1971
+ }
1972
+ if (!model) {
1973
+ model = chunk.model;
1974
+ debug.api("OPENAI_STREAM_MODEL_RECEIVED", {
1975
+ model,
1976
+ chunkNumber: String(chunkCount)
1977
+ });
1978
+ }
1979
+ if (!created) {
1980
+ created = chunk.created;
1981
+ }
1982
+ if (!object) {
1983
+ object = chunk.object;
1984
+ }
1985
+ if (!usage) {
1986
+ usage = chunk.usage;
1987
+ }
1988
+ message = messageReducer(message, chunk);
1989
+ if (chunk?.choices?.[0]?.delta?.content) {
1990
+ if (!ttftMs) {
1991
+ ttftMs = Date.now() - streamStartTime;
1992
+ debug.api("OPENAI_STREAM_FIRST_TOKEN", {
1993
+ ttftMs: String(ttftMs),
1994
+ chunkNumber: String(chunkCount)
1995
+ });
1996
+ }
1997
+ }
1998
+ } catch (chunkError) {
1999
+ errorCount++;
2000
+ debug.error("OPENAI_STREAM_CHUNK_ERROR", {
2001
+ chunkNumber: String(chunkCount),
2002
+ errorMessage: chunkError instanceof Error ? chunkError.message : String(chunkError),
2003
+ errorType: chunkError instanceof Error ? chunkError.constructor.name : typeof chunkError
2004
+ });
2005
+ }
2006
+ }
2007
+ debug.api("OPENAI_STREAM_COMPLETE", {
2008
+ totalChunks: String(chunkCount),
2009
+ errorCount: String(errorCount),
2010
+ totalDuration: String(Date.now() - streamStartTime),
2011
+ ttftMs: String(ttftMs || 0),
2012
+ finalMessageId: id || "undefined"
2013
+ });
2014
+ } catch (streamError) {
2015
+ debug.error("OPENAI_STREAM_FATAL_ERROR", {
2016
+ totalChunks: String(chunkCount),
2017
+ errorCount: String(errorCount),
2018
+ errorMessage: streamError instanceof Error ? streamError.message : String(streamError),
2019
+ errorType: streamError instanceof Error ? streamError.constructor.name : typeof streamError
2020
+ });
2021
+ throw streamError;
2022
+ }
2023
+ return {
2024
+ id,
2025
+ created,
2026
+ model,
2027
+ object,
2028
+ choices: [
2029
+ {
2030
+ index: 0,
2031
+ message,
2032
+ finish_reason: "stop",
2033
+ logprobs: void 0
2034
+ }
2035
+ ],
2036
+ usage
2037
+ };
2038
+ }
2039
+ function convertOpenAIResponseToAnthropic(response, tools) {
2040
+ let contentBlocks = [];
2041
+ const message = response.choices?.[0]?.message;
2042
+ if (!message) {
2043
+ return {
2044
+ role: "assistant",
2045
+ content: [],
2046
+ stop_reason: response.choices?.[0]?.finish_reason,
2047
+ type: "message",
2048
+ usage: response.usage
2049
+ };
2050
+ }
2051
+ if (message?.tool_calls) {
2052
+ for (const toolCall of message.tool_calls) {
2053
+ const tool = toolCall.function;
2054
+ const toolName = tool?.name;
2055
+ let toolArgs = {};
2056
+ try {
2057
+ toolArgs = tool?.arguments ? JSON.parse(tool.arguments) : {};
2058
+ } catch (e) {
2059
+ }
2060
+ contentBlocks.push({
2061
+ type: "tool_use",
2062
+ input: toolArgs,
2063
+ name: toolName,
2064
+ id: toolCall.id?.length > 0 ? toolCall.id : nanoid()
2065
+ });
2066
+ }
2067
+ }
2068
+ if (message.reasoning) {
2069
+ contentBlocks.push({
2070
+ type: "thinking",
2071
+ thinking: message.reasoning,
2072
+ signature: ""
2073
+ });
2074
+ }
2075
+ if (message.reasoning_content) {
2076
+ contentBlocks.push({
2077
+ type: "thinking",
2078
+ thinking: message.reasoning_content,
2079
+ signature: ""
2080
+ });
2081
+ }
2082
+ if (message.content) {
2083
+ contentBlocks.push({
2084
+ type: "text",
2085
+ text: message?.content,
2086
+ citations: []
2087
+ });
2088
+ }
2089
+ const finalMessage = {
2090
+ role: "assistant",
2091
+ content: contentBlocks,
2092
+ stop_reason: response.choices?.[0]?.finish_reason,
2093
+ type: "message",
2094
+ usage: response.usage
2095
+ };
2096
+ return finalMessage;
2097
+ }
2098
+ var anthropicClient = null;
2099
+ function getAnthropicClient(model) {
2100
+ const config = getGlobalConfig();
2101
+ const provider = config.primaryProvider;
2102
+ if (anthropicClient && provider) {
2103
+ anthropicClient = null;
2104
+ }
2105
+ if (anthropicClient) {
2106
+ return anthropicClient;
2107
+ }
2108
+ const region = getVertexRegionForModel(model);
2109
+ const modelManager = getModelManager();
2110
+ const modelProfile = modelManager.getModel("main");
2111
+ const defaultHeaders = {
2112
+ "x-app": "cli",
2113
+ "User-Agent": USER_AGENT
2114
+ };
2115
+ if (process.env.ANTHROPIC_AUTH_TOKEN) {
2116
+ defaultHeaders["Authorization"] = `Bearer ${process.env.ANTHROPIC_AUTH_TOKEN}`;
2117
+ }
2118
+ const ARGS = {
2119
+ defaultHeaders,
2120
+ maxRetries: 0,
2121
+ timeout: parseInt(process.env.API_TIMEOUT_MS || String(60 * 1e3), 10)
2122
+ };
2123
+ if (USE_BEDROCK) {
2124
+ const client = new AnthropicBedrock(ARGS);
2125
+ anthropicClient = client;
2126
+ return client;
2127
+ }
2128
+ if (USE_VERTEX) {
2129
+ const vertexArgs = {
2130
+ ...ARGS,
2131
+ region: region || process.env.CLOUD_ML_REGION || "us-east5"
2132
+ };
2133
+ const client = new AnthropicVertex(vertexArgs);
2134
+ anthropicClient = client;
2135
+ return client;
2136
+ }
2137
+ let apiKey;
2138
+ let baseURL;
2139
+ if (modelProfile) {
2140
+ apiKey = modelProfile.apiKey || "";
2141
+ baseURL = modelProfile.baseURL;
2142
+ } else {
2143
+ apiKey = getAnthropicApiKey();
2144
+ baseURL = void 0;
2145
+ }
2146
+ if (process.env.USER_TYPE === "ant" && !apiKey && provider === "anthropic") {
2147
+ console.error(
2148
+ chalk.red(
2149
+ "[ANT-ONLY] Missing API key. Configure an API key in your model profile or environment variables."
2150
+ )
2151
+ );
2152
+ }
2153
+ const clientConfig = {
2154
+ apiKey,
2155
+ dangerouslyAllowBrowser: true,
2156
+ ...ARGS,
2157
+ ...baseURL && { baseURL }
2158
+ };
2159
+ anthropicClient = new Anthropic(clientConfig);
2160
+ return anthropicClient;
2161
+ }
2162
+ function resetAnthropicClient() {
2163
+ anthropicClient = null;
2164
+ }
2165
+ function applyCacheControlWithLimits(systemBlocks, messageParams) {
2166
+ if (!PROMPT_CACHING_ENABLED) {
2167
+ return { systemBlocks, messageParams };
2168
+ }
2169
+ const maxCacheBlocks = 4;
2170
+ let usedCacheBlocks = 0;
2171
+ const processedSystemBlocks = systemBlocks.map((block, index) => {
2172
+ if (usedCacheBlocks < maxCacheBlocks && block.text.length > 1e3) {
2173
+ usedCacheBlocks++;
2174
+ return {
2175
+ ...block,
2176
+ cache_control: { type: "ephemeral" }
2177
+ };
2178
+ }
2179
+ const { cache_control, ...blockWithoutCache } = block;
2180
+ return blockWithoutCache;
2181
+ });
2182
+ const processedMessageParams = messageParams.map((message, messageIndex) => {
2183
+ if (Array.isArray(message.content)) {
2184
+ const processedContent = message.content.map(
2185
+ (contentBlock, blockIndex) => {
2186
+ const shouldCache = usedCacheBlocks < maxCacheBlocks && contentBlock.type === "text" && typeof contentBlock.text === "string" && (contentBlock.text.length > 2e3 || messageIndex === messageParams.length - 1 && blockIndex === message.content.length - 1 && contentBlock.text.length > 500);
2187
+ if (shouldCache) {
2188
+ usedCacheBlocks++;
2189
+ return {
2190
+ ...contentBlock,
2191
+ cache_control: { type: "ephemeral" }
2192
+ };
2193
+ }
2194
+ const { cache_control, ...blockWithoutCache } = contentBlock;
2195
+ return blockWithoutCache;
2196
+ }
2197
+ );
2198
+ return {
2199
+ ...message,
2200
+ content: processedContent
2201
+ };
2202
+ }
2203
+ return message;
2204
+ });
2205
+ return {
2206
+ systemBlocks: processedSystemBlocks,
2207
+ messageParams: processedMessageParams
2208
+ };
2209
+ }
2210
+ function userMessageToMessageParam(message, addCache = false) {
2211
+ if (addCache) {
2212
+ if (typeof message.message.content === "string") {
2213
+ return {
2214
+ role: "user",
2215
+ content: [
2216
+ {
2217
+ type: "text",
2218
+ text: message.message.content
2219
+ }
2220
+ ]
2221
+ };
2222
+ } else {
2223
+ return {
2224
+ role: "user",
2225
+ content: message.message.content.map((_) => ({ ..._ }))
2226
+ };
2227
+ }
2228
+ }
2229
+ return {
2230
+ role: "user",
2231
+ content: message.message.content
2232
+ };
2233
+ }
2234
+ function assistantMessageToMessageParam(message, addCache = false) {
2235
+ if (addCache) {
2236
+ if (typeof message.message.content === "string") {
2237
+ return {
2238
+ role: "assistant",
2239
+ content: [
2240
+ {
2241
+ type: "text",
2242
+ text: message.message.content
2243
+ }
2244
+ ]
2245
+ };
2246
+ } else {
2247
+ return {
2248
+ role: "assistant",
2249
+ content: message.message.content.map((_) => ({ ..._ }))
2250
+ };
2251
+ }
2252
+ }
2253
+ return {
2254
+ role: "assistant",
2255
+ content: message.message.content
2256
+ };
2257
+ }
2258
+ function splitSysPromptPrefix(systemPrompt) {
2259
+ const systemPromptFirstBlock = systemPrompt[0] || "";
2260
+ const systemPromptRest = systemPrompt.slice(1);
2261
+ return [systemPromptFirstBlock, systemPromptRest.join("\n")].filter(Boolean);
2262
+ }
2263
+ async function queryLLM(messages, systemPrompt, maxThinkingTokens, tools, signal, options) {
2264
+ const modelManager = options.__testModelManager ?? getModelManager();
2265
+ const modelResolution = modelManager.resolveModelWithInfo(options.model);
2266
+ if (!modelResolution.success || !modelResolution.profile) {
2267
+ const fallbackProfile = modelManager.resolveModel(options.model);
2268
+ if (!fallbackProfile) {
2269
+ throw new Error(
2270
+ modelResolution.error || `Failed to resolve model: ${options.model}`
2271
+ );
2272
+ }
2273
+ debug.warn("MODEL_RESOLUTION_FALLBACK", {
2274
+ inputParam: options.model,
2275
+ error: modelResolution.error,
2276
+ fallbackModelName: fallbackProfile.modelName,
2277
+ fallbackProvider: fallbackProfile.provider,
2278
+ requestId: getCurrentRequest()?.id
2279
+ });
2280
+ modelResolution.success = true;
2281
+ modelResolution.profile = fallbackProfile;
2282
+ }
2283
+ const modelProfile = modelResolution.profile;
2284
+ const resolvedModel = modelProfile.modelName;
2285
+ const toolUseContext = options.toolUseContext;
2286
+ if (toolUseContext && !toolUseContext.responseState) {
2287
+ const conversationId = getConversationId(
2288
+ toolUseContext.agentId,
2289
+ toolUseContext.messageId
2290
+ );
2291
+ const previousResponseId = responseStateManager.getPreviousResponseId(conversationId);
2292
+ toolUseContext.responseState = {
2293
+ previousResponseId,
2294
+ conversationId
2295
+ };
2296
+ }
2297
+ debug.api("MODEL_RESOLVED", {
2298
+ inputParam: options.model,
2299
+ resolvedModelName: resolvedModel,
2300
+ provider: modelProfile.provider,
2301
+ isPointer: ["main", "task", "compact", "quick"].includes(options.model),
2302
+ hasResponseState: !!toolUseContext?.responseState,
2303
+ conversationId: toolUseContext?.responseState?.conversationId,
2304
+ requestId: getCurrentRequest()?.id
2305
+ });
2306
+ const currentRequest = getCurrentRequest();
2307
+ debug.api("LLM_REQUEST_START", {
2308
+ messageCount: messages.length,
2309
+ systemPromptLength: systemPrompt.join(" ").length,
2310
+ toolCount: tools.length,
2311
+ model: resolvedModel,
2312
+ originalModelParam: options.model,
2313
+ requestId: getCurrentRequest()?.id
2314
+ });
2315
+ markPhase("LLM_CALL");
2316
+ try {
2317
+ const queryFn = options.__testQueryLLMWithPromptCaching ?? queryLLMWithPromptCaching;
2318
+ const cleanOptions = { ...options };
2319
+ delete cleanOptions.__testModelManager;
2320
+ delete cleanOptions.__testQueryLLMWithPromptCaching;
2321
+ const runQuery = () => queryFn(
2322
+ messages,
2323
+ systemPrompt,
2324
+ maxThinkingTokens,
2325
+ tools,
2326
+ signal,
2327
+ {
2328
+ ...cleanOptions,
2329
+ model: resolvedModel,
2330
+ modelProfile,
2331
+ toolUseContext
2332
+ }
2333
+ );
2334
+ const result = options.__testQueryLLMWithPromptCaching ? await runQuery() : await withVCR(messages, runQuery);
2335
+ debug.api("LLM_REQUEST_SUCCESS", {
2336
+ costUSD: result.costUSD,
2337
+ durationMs: result.durationMs,
2338
+ responseLength: result.message.content?.length || 0,
2339
+ requestId: getCurrentRequest()?.id
2340
+ });
2341
+ if (toolUseContext?.responseState?.conversationId && result.responseId) {
2342
+ responseStateManager.setPreviousResponseId(
2343
+ toolUseContext.responseState.conversationId,
2344
+ result.responseId
2345
+ );
2346
+ debug.api("RESPONSE_STATE_UPDATED", {
2347
+ conversationId: toolUseContext.responseState.conversationId,
2348
+ responseId: result.responseId,
2349
+ requestId: getCurrentRequest()?.id
2350
+ });
2351
+ }
2352
+ return result;
2353
+ } catch (error) {
2354
+ logErrorWithDiagnosis(
2355
+ error,
2356
+ {
2357
+ messageCount: messages.length,
2358
+ systemPromptLength: systemPrompt.join(" ").length,
2359
+ model: options.model,
2360
+ toolCount: tools.length,
2361
+ phase: "LLM_CALL"
2362
+ },
2363
+ currentRequest?.id
2364
+ );
2365
+ throw error;
2366
+ }
2367
+ }
2368
+ async function queryLLMWithPromptCaching(messages, systemPrompt, maxThinkingTokens, tools, signal, options) {
2369
+ const config = getGlobalConfig();
2370
+ const modelManager = getModelManager();
2371
+ const toolUseContext = options.toolUseContext;
2372
+ const modelProfile = options.modelProfile || modelManager.getModel("main");
2373
+ let provider;
2374
+ if (modelProfile) {
2375
+ provider = modelProfile.provider || config.primaryProvider || "anthropic";
2376
+ } else {
2377
+ provider = config.primaryProvider || "anthropic";
2378
+ }
2379
+ if (provider === "anthropic" || provider === "bigdream" || provider === "opendev" || provider === "minimax-coding") {
2380
+ return queryAnthropicNative(
2381
+ messages,
2382
+ systemPrompt,
2383
+ maxThinkingTokens,
2384
+ tools,
2385
+ signal,
2386
+ { ...options, modelProfile, toolUseContext }
2387
+ );
2388
+ }
2389
+ return queryOpenAI(messages, systemPrompt, maxThinkingTokens, tools, signal, {
2390
+ ...options,
2391
+ modelProfile,
2392
+ toolUseContext
2393
+ });
2394
+ }
2395
+ async function queryAnthropicNative(messages, systemPrompt, maxThinkingTokens, tools, signal, options) {
2396
+ const config = getGlobalConfig();
2397
+ const modelManager = getModelManager();
2398
+ const toolUseContext = options?.toolUseContext;
2399
+ const modelProfile = options?.modelProfile || modelManager.getModel("main");
2400
+ let anthropic;
2401
+ let model;
2402
+ let provider;
2403
+ debug.api("MODEL_CONFIG_ANTHROPIC", {
2404
+ modelProfileFound: !!modelProfile,
2405
+ modelProfileId: modelProfile?.modelName,
2406
+ modelProfileName: modelProfile?.name,
2407
+ modelProfileModelName: modelProfile?.modelName,
2408
+ modelProfileProvider: modelProfile?.provider,
2409
+ modelProfileBaseURL: modelProfile?.baseURL,
2410
+ modelProfileApiKeyExists: !!modelProfile?.apiKey,
2411
+ optionsModel: options?.model,
2412
+ requestId: getCurrentRequest()?.id
2413
+ });
2414
+ if (modelProfile) {
2415
+ model = modelProfile.modelName;
2416
+ provider = modelProfile.provider || config.primaryProvider || "anthropic";
2417
+ if (modelProfile.provider === "anthropic" || modelProfile.provider === "minimax-coding") {
2418
+ const clientConfig = {
2419
+ apiKey: modelProfile.apiKey,
2420
+ dangerouslyAllowBrowser: true,
2421
+ maxRetries: 0,
2422
+ timeout: parseInt(process.env.API_TIMEOUT_MS || String(60 * 1e3), 10),
2423
+ defaultHeaders: {
2424
+ "x-app": "cli",
2425
+ "User-Agent": USER_AGENT
2426
+ }
2427
+ };
2428
+ if (modelProfile.baseURL) {
2429
+ clientConfig.baseURL = modelProfile.baseURL;
2430
+ }
2431
+ anthropic = new Anthropic(clientConfig);
2432
+ } else {
2433
+ anthropic = getAnthropicClient(model);
2434
+ }
2435
+ } else {
2436
+ const errorDetails = {
2437
+ modelProfileExists: !!modelProfile,
2438
+ modelProfileModelName: modelProfile?.modelName,
2439
+ requestedModel: options?.model,
2440
+ requestId: getCurrentRequest()?.id
2441
+ };
2442
+ debug.error("ANTHROPIC_FALLBACK_ERROR", errorDetails);
2443
+ throw new Error(
2444
+ `No valid ModelProfile available for Anthropic provider. Please configure model through /model command. Debug: ${JSON.stringify(errorDetails)}`
2445
+ );
2446
+ }
2447
+ if (options?.prependCLISysprompt) {
2448
+ const [firstSyspromptBlock] = splitSysPromptPrefix(systemPrompt);
2449
+ systemPrompt = [getCLISyspromptPrefix(), ...systemPrompt];
2450
+ }
2451
+ const system = splitSysPromptPrefix(systemPrompt).map(
2452
+ (_) => ({
2453
+ text: _,
2454
+ type: "text"
2455
+ })
2456
+ );
2457
+ const toolSchemas = await Promise.all(
2458
+ tools.map(
2459
+ async (tool) => ({
2460
+ name: tool.name,
2461
+ description: getToolDescription(tool),
2462
+ input_schema: "inputJSONSchema" in tool && tool.inputJSONSchema ? tool.inputJSONSchema : zodToJsonSchema4(tool.inputSchema)
2463
+ })
2464
+ )
2465
+ );
2466
+ const anthropicMessages = addCacheBreakpoints(messages);
2467
+ const { systemBlocks: processedSystem, messageParams: processedMessages } = applyCacheControlWithLimits(system, anthropicMessages);
2468
+ const startIncludingRetries = Date.now();
2469
+ logSystemPromptConstruction({
2470
+ basePrompt: systemPrompt.join("\n"),
2471
+ danyaContext: generateDanyaContext() || "",
2472
+ reminders: [],
2473
+ finalPrompt: systemPrompt.join("\n")
2474
+ });
2475
+ let start = Date.now();
2476
+ let attemptNumber = 0;
2477
+ let response;
2478
+ try {
2479
+ response = await withRetry(
2480
+ async (attempt) => {
2481
+ attemptNumber = attempt;
2482
+ start = Date.now();
2483
+ const params = {
2484
+ model,
2485
+ max_tokens: options?.maxTokens ?? getMaxTokensFromProfile(modelProfile),
2486
+ messages: processedMessages,
2487
+ system: processedSystem,
2488
+ tools: toolSchemas.length > 0 ? toolSchemas : void 0,
2489
+ tool_choice: toolSchemas.length > 0 ? { type: "auto" } : void 0,
2490
+ ...options?.temperature !== void 0 ? { temperature: options.temperature } : {},
2491
+ ...options?.stopSequences && options.stopSequences.length > 0 ? { stop_sequences: options.stopSequences } : {}
2492
+ };
2493
+ if (maxThinkingTokens > 0) {
2494
+ ;
2495
+ params.extra_headers = {
2496
+ "anthropic-beta": "max-tokens-3-5-sonnet-2024-07-15"
2497
+ };
2498
+ params.thinking = { max_tokens: maxThinkingTokens };
2499
+ }
2500
+ debug.api("ANTHROPIC_API_CALL_START_STREAMING", {
2501
+ endpoint: modelProfile?.baseURL || "DEFAULT_ANTHROPIC",
2502
+ model,
2503
+ provider,
2504
+ apiKeyConfigured: !!modelProfile?.apiKey,
2505
+ apiKeyPrefix: modelProfile?.apiKey ? modelProfile.apiKey.substring(0, 8) : null,
2506
+ maxTokens: params.max_tokens,
2507
+ temperature: options?.temperature ?? MAIN_QUERY_TEMPERATURE,
2508
+ params,
2509
+ messageCount: params.messages?.length || 0,
2510
+ streamMode: true,
2511
+ toolsCount: toolSchemas.length,
2512
+ thinkingTokens: maxThinkingTokens,
2513
+ timestamp: (/* @__PURE__ */ new Date()).toISOString(),
2514
+ modelProfileId: modelProfile?.modelName,
2515
+ modelProfileName: modelProfile?.name
2516
+ });
2517
+ if (config.stream) {
2518
+ const stream = await anthropic.beta.messages.create(
2519
+ {
2520
+ ...params,
2521
+ stream: true
2522
+ },
2523
+ {
2524
+ signal
2525
+ }
2526
+ );
2527
+ let finalResponse = null;
2528
+ let messageStartEvent = null;
2529
+ const contentBlocks = [];
2530
+ const inputJSONBuffers = /* @__PURE__ */ new Map();
2531
+ let usage = null;
2532
+ let stopReason = null;
2533
+ let stopSequence = null;
2534
+ let hasMarkedStreaming = false;
2535
+ for await (const event of stream) {
2536
+ if (signal.aborted) {
2537
+ debug.flow("STREAM_ABORTED", {
2538
+ eventType: event.type,
2539
+ timestamp: Date.now()
2540
+ });
2541
+ throw new Error("Request was cancelled");
2542
+ }
2543
+ switch (event.type) {
2544
+ case "message_start":
2545
+ messageStartEvent = event;
2546
+ finalResponse = {
2547
+ ...event.message,
2548
+ content: []
2549
+ };
2550
+ break;
2551
+ case "content_block_start":
2552
+ contentBlocks[event.index] = { ...event.content_block };
2553
+ const contentBlockType = event.content_block.type;
2554
+ if (contentBlockType === "tool_use" || contentBlockType === "server_tool_use" || contentBlockType === "mcp_tool_use") {
2555
+ setRequestStatus({
2556
+ kind: "tool",
2557
+ detail: event.content_block.name
2558
+ });
2559
+ inputJSONBuffers.set(event.index, "");
2560
+ }
2561
+ break;
2562
+ case "content_block_delta":
2563
+ const blockIndex = event.index;
2564
+ if (!contentBlocks[blockIndex]) {
2565
+ contentBlocks[blockIndex] = {
2566
+ type: event.delta.type === "text_delta" ? "text" : "tool_use",
2567
+ text: event.delta.type === "text_delta" ? "" : void 0
2568
+ };
2569
+ if (event.delta.type === "input_json_delta") {
2570
+ inputJSONBuffers.set(blockIndex, "");
2571
+ }
2572
+ }
2573
+ if (event.delta.type === "text_delta") {
2574
+ if (!hasMarkedStreaming) {
2575
+ setRequestStatus({ kind: "streaming" });
2576
+ hasMarkedStreaming = true;
2577
+ }
2578
+ contentBlocks[blockIndex].text += event.delta.text;
2579
+ } else if (event.delta.type === "input_json_delta") {
2580
+ const currentBuffer = inputJSONBuffers.get(blockIndex) || "";
2581
+ const nextBuffer = currentBuffer + event.delta.partial_json;
2582
+ inputJSONBuffers.set(blockIndex, nextBuffer);
2583
+ const trimmed = nextBuffer.trim();
2584
+ if (trimmed.length === 0) {
2585
+ contentBlocks[blockIndex].input = {};
2586
+ break;
2587
+ }
2588
+ contentBlocks[blockIndex].input = parseToolUsePartialJsonOrThrow(nextBuffer) ?? {};
2589
+ }
2590
+ break;
2591
+ case "message_delta":
2592
+ if (event.delta.stop_reason)
2593
+ stopReason = event.delta.stop_reason;
2594
+ if (event.delta.stop_sequence)
2595
+ stopSequence = event.delta.stop_sequence;
2596
+ if (event.usage) usage = { ...usage, ...event.usage };
2597
+ break;
2598
+ case "content_block_stop":
2599
+ const stopIndex = event.index;
2600
+ const block = contentBlocks[stopIndex];
2601
+ if ((block?.type === "tool_use" || block?.type === "server_tool_use" || block?.type === "mcp_tool_use") && inputJSONBuffers.has(stopIndex)) {
2602
+ const jsonStr = inputJSONBuffers.get(stopIndex) ?? "";
2603
+ if (block.input === void 0) {
2604
+ const trimmed = jsonStr.trim();
2605
+ if (trimmed.length === 0) {
2606
+ block.input = {};
2607
+ } else {
2608
+ block.input = parseToolUsePartialJsonOrThrow(jsonStr) ?? {};
2609
+ }
2610
+ }
2611
+ inputJSONBuffers.delete(stopIndex);
2612
+ }
2613
+ break;
2614
+ case "message_stop":
2615
+ inputJSONBuffers.clear();
2616
+ break;
2617
+ }
2618
+ if (event.type === "message_stop") {
2619
+ break;
2620
+ }
2621
+ }
2622
+ if (!finalResponse || !messageStartEvent) {
2623
+ throw new Error("Stream ended without proper message structure");
2624
+ }
2625
+ finalResponse = {
2626
+ ...messageStartEvent.message,
2627
+ content: contentBlocks.filter(Boolean),
2628
+ stop_reason: stopReason,
2629
+ stop_sequence: stopSequence,
2630
+ usage: {
2631
+ ...messageStartEvent.message.usage,
2632
+ ...usage
2633
+ }
2634
+ };
2635
+ return finalResponse;
2636
+ } else {
2637
+ debug.api("ANTHROPIC_API_CALL_START_NON_STREAMING", {
2638
+ endpoint: modelProfile?.baseURL || "DEFAULT_ANTHROPIC",
2639
+ model,
2640
+ provider,
2641
+ apiKeyConfigured: !!modelProfile?.apiKey,
2642
+ apiKeyPrefix: modelProfile?.apiKey ? modelProfile.apiKey.substring(0, 8) : null,
2643
+ maxTokens: params.max_tokens,
2644
+ temperature: options?.temperature ?? MAIN_QUERY_TEMPERATURE,
2645
+ messageCount: params.messages?.length || 0,
2646
+ streamMode: false,
2647
+ toolsCount: toolSchemas.length,
2648
+ thinkingTokens: maxThinkingTokens,
2649
+ timestamp: (/* @__PURE__ */ new Date()).toISOString(),
2650
+ modelProfileId: modelProfile?.modelName,
2651
+ modelProfileName: modelProfile?.name
2652
+ });
2653
+ return await anthropic.beta.messages.create(params, {
2654
+ signal
2655
+ });
2656
+ }
2657
+ },
2658
+ { signal }
2659
+ );
2660
+ debug.api("ANTHROPIC_API_CALL_SUCCESS", {
2661
+ content: response.content
2662
+ });
2663
+ const ttftMs = start - Date.now();
2664
+ const durationMs = Date.now() - startIncludingRetries;
2665
+ const content = response.content.map((block) => {
2666
+ if (block.type === "text") {
2667
+ return {
2668
+ type: "text",
2669
+ text: block.text
2670
+ };
2671
+ } else if (block.type === "tool_use") {
2672
+ return {
2673
+ type: "tool_use",
2674
+ id: block.id,
2675
+ name: block.name,
2676
+ input: block.input
2677
+ };
2678
+ }
2679
+ return block;
2680
+ });
2681
+ const assistantMessage = {
2682
+ message: {
2683
+ id: response.id,
2684
+ content,
2685
+ model: response.model,
2686
+ role: "assistant",
2687
+ stop_reason: response.stop_reason,
2688
+ stop_sequence: response.stop_sequence,
2689
+ type: "message",
2690
+ usage: response.usage
2691
+ },
2692
+ type: "assistant",
2693
+ uuid: nanoid(),
2694
+ durationMs,
2695
+ costUSD: 0
2696
+ };
2697
+ const systemMessages = system.map((block) => ({
2698
+ role: "system",
2699
+ content: block.text
2700
+ }));
2701
+ logLLMInteraction({
2702
+ systemPrompt: systemPrompt.join("\n"),
2703
+ messages: [...systemMessages, ...anthropicMessages],
2704
+ response,
2705
+ usage: response.usage ? {
2706
+ inputTokens: response.usage.input_tokens,
2707
+ outputTokens: response.usage.output_tokens
2708
+ } : void 0,
2709
+ timing: {
2710
+ start,
2711
+ end: Date.now()
2712
+ },
2713
+ apiFormat: "anthropic"
2714
+ });
2715
+ const inputTokens = response.usage.input_tokens;
2716
+ const outputTokens = response.usage.output_tokens;
2717
+ const cacheCreationInputTokens = response.usage.cache_creation_input_tokens ?? 0;
2718
+ const cacheReadInputTokens = response.usage.cache_read_input_tokens ?? 0;
2719
+ const costUSD = inputTokens / 1e6 * getModelInputTokenCostUSD(model) + outputTokens / 1e6 * getModelOutputTokenCostUSD(model) + cacheCreationInputTokens / 1e6 * getModelInputTokenCostUSD(model) + cacheReadInputTokens / 1e6 * (getModelInputTokenCostUSD(model) * 0.1);
2720
+ assistantMessage.costUSD = costUSD;
2721
+ addToTotalCost(costUSD, durationMs);
2722
+ return assistantMessage;
2723
+ } catch (error) {
2724
+ return getAssistantMessageFromError(error);
2725
+ }
2726
+ }
2727
+ function getAssistantMessageFromError(error) {
2728
+ if (error instanceof Error && error.message.includes("prompt is too long")) {
2729
+ return createAssistantAPIErrorMessage(PROMPT_TOO_LONG_ERROR_MESSAGE);
2730
+ }
2731
+ if (error instanceof Error && error.message.includes("Your credit balance is too low")) {
2732
+ return createAssistantAPIErrorMessage(CREDIT_BALANCE_TOO_LOW_ERROR_MESSAGE);
2733
+ }
2734
+ if (error instanceof Error && error.message.toLowerCase().includes("x-api-key")) {
2735
+ return createAssistantAPIErrorMessage(INVALID_API_KEY_ERROR_MESSAGE);
2736
+ }
2737
+ if (error instanceof Error) {
2738
+ if (process.env.NODE_ENV === "development") {
2739
+ debug.error("ANTHROPIC_API_ERROR", {
2740
+ message: error.message,
2741
+ stack: error.stack
2742
+ });
2743
+ }
2744
+ return createAssistantAPIErrorMessage(
2745
+ `${API_ERROR_MESSAGE_PREFIX}: ${error.message}`
2746
+ );
2747
+ }
2748
+ return createAssistantAPIErrorMessage(API_ERROR_MESSAGE_PREFIX);
2749
+ }
2750
+ function addCacheBreakpoints(messages) {
2751
+ return messages.map((msg, index) => {
2752
+ return msg.type === "user" ? userMessageToMessageParam(msg, index > messages.length - 3) : assistantMessageToMessageParam(msg, index > messages.length - 3);
2753
+ });
2754
+ }
2755
+ async function queryOpenAI(messages, systemPrompt, maxThinkingTokens, tools, signal, options) {
2756
+ const config = getGlobalConfig();
2757
+ const modelManager = getModelManager();
2758
+ const toolUseContext = options?.toolUseContext;
2759
+ const modelProfile = options?.modelProfile || modelManager.getModel("main");
2760
+ let model;
2761
+ const currentRequest = getCurrentRequest();
2762
+ debug.api("MODEL_CONFIG_OPENAI", {
2763
+ modelProfileFound: !!modelProfile,
2764
+ modelProfileId: modelProfile?.modelName,
2765
+ modelProfileName: modelProfile?.name,
2766
+ modelProfileModelName: modelProfile?.modelName,
2767
+ modelProfileProvider: modelProfile?.provider,
2768
+ modelProfileBaseURL: modelProfile?.baseURL,
2769
+ modelProfileApiKeyExists: !!modelProfile?.apiKey,
2770
+ optionsModel: options?.model,
2771
+ requestId: getCurrentRequest()?.id
2772
+ });
2773
+ if (modelProfile) {
2774
+ model = modelProfile.modelName;
2775
+ } else {
2776
+ model = options?.model || modelProfile?.modelName || "";
2777
+ }
2778
+ if (options?.prependCLISysprompt) {
2779
+ const [firstSyspromptBlock] = splitSysPromptPrefix(systemPrompt);
2780
+ systemPrompt = [getCLISyspromptPrefix() + systemPrompt];
2781
+ }
2782
+ const system = splitSysPromptPrefix(systemPrompt).map(
2783
+ (_) => ({
2784
+ ...PROMPT_CACHING_ENABLED ? { cache_control: { type: "ephemeral" } } : {},
2785
+ text: _,
2786
+ type: "text"
2787
+ })
2788
+ );
2789
+ const toolSchemas = await Promise.all(
2790
+ tools.map(
2791
+ async (_) => ({
2792
+ type: "function",
2793
+ function: {
2794
+ name: _.name,
2795
+ description: await _.prompt({
2796
+ safeMode: options?.safeMode
2797
+ }),
2798
+ parameters: "inputJSONSchema" in _ && _.inputJSONSchema ? _.inputJSONSchema : zodToJsonSchema4(_.inputSchema)
2799
+ }
2800
+ })
2801
+ )
2802
+ );
2803
+ const openaiSystem = system.map(
2804
+ (s) => ({
2805
+ role: "system",
2806
+ content: s.text
2807
+ })
2808
+ );
2809
+ const openaiMessages = convertAnthropicMessagesToOpenAIMessages2(messages);
2810
+ logSystemPromptConstruction({
2811
+ basePrompt: systemPrompt.join("\n"),
2812
+ danyaContext: generateDanyaContext() || "",
2813
+ reminders: [],
2814
+ finalPrompt: systemPrompt.join("\n")
2815
+ });
2816
+ let start = Date.now();
2817
+ let adapterContext = null;
2818
+ if (modelProfile && modelProfile.modelName) {
2819
+ debug.api("CHECKING_ADAPTER_SYSTEM", {
2820
+ modelProfileName: modelProfile.modelName,
2821
+ modelName: modelProfile.modelName,
2822
+ provider: modelProfile.provider,
2823
+ requestId: getCurrentRequest()?.id
2824
+ });
2825
+ const USE_NEW_ADAPTER_SYSTEM = process.env.USE_NEW_ADAPTERS !== "false";
2826
+ if (USE_NEW_ADAPTER_SYSTEM) {
2827
+ const shouldUseResponses = ModelAdapterFactory.shouldUseResponsesAPI(modelProfile);
2828
+ if (shouldUseResponses) {
2829
+ const adapter = ModelAdapterFactory.createAdapter(modelProfile);
2830
+ const reasoningEffort = await getReasoningEffort(modelProfile, messages);
2831
+ let verbosity = "medium";
2832
+ const modelNameLower = modelProfile.modelName.toLowerCase();
2833
+ if (modelNameLower.includes("high")) {
2834
+ verbosity = "high";
2835
+ } else if (modelNameLower.includes("low")) {
2836
+ verbosity = "low";
2837
+ }
2838
+ const unifiedParams = {
2839
+ messages: openaiMessages,
2840
+ systemPrompt: openaiSystem.map((s) => s.content),
2841
+ tools,
2842
+ maxTokens: options?.maxTokens ?? getMaxTokensFromProfile(modelProfile),
2843
+ stream: config.stream,
2844
+ reasoningEffort,
2845
+ temperature: options?.temperature ?? (isGPT5Model(model) ? 1 : MAIN_QUERY_TEMPERATURE),
2846
+ previousResponseId: toolUseContext?.responseState?.previousResponseId,
2847
+ verbosity,
2848
+ ...options?.stopSequences && options.stopSequences.length > 0 ? { stopSequences: options.stopSequences } : {}
2849
+ };
2850
+ adapterContext = {
2851
+ adapter,
2852
+ request: adapter.createRequest(unifiedParams),
2853
+ shouldUseResponses: true
2854
+ };
2855
+ }
2856
+ }
2857
+ }
2858
+ let queryResult;
2859
+ let startIncludingRetries = Date.now();
2860
+ try {
2861
+ queryResult = await withRetry(
2862
+ async () => {
2863
+ start = Date.now();
2864
+ if (adapterContext) {
2865
+ if (adapterContext.shouldUseResponses) {
2866
+ const { callGPT5ResponsesAPI } = await import("./openai-YP4OJYKF.js");
2867
+ const response = await callGPT5ResponsesAPI(
2868
+ modelProfile,
2869
+ adapterContext.request,
2870
+ signal
2871
+ );
2872
+ const unifiedResponse = await adapterContext.adapter.parseResponse(response);
2873
+ const assistantMessage2 = buildAssistantMessageFromUnifiedResponse(
2874
+ unifiedResponse,
2875
+ start
2876
+ );
2877
+ assistantMessage2.message.usage = normalizeUsage(
2878
+ assistantMessage2.message.usage
2879
+ );
2880
+ return {
2881
+ assistantMessage: assistantMessage2,
2882
+ rawResponse: unifiedResponse,
2883
+ apiFormat: "openai"
2884
+ };
2885
+ }
2886
+ const s2 = await getCompletionWithProfile(
2887
+ modelProfile,
2888
+ adapterContext.request,
2889
+ 0,
2890
+ 10,
2891
+ signal
2892
+ );
2893
+ let finalResponse2;
2894
+ if (config.stream) {
2895
+ finalResponse2 = await handleMessageStream(
2896
+ s2,
2897
+ signal
2898
+ );
2899
+ } else {
2900
+ finalResponse2 = s2;
2901
+ }
2902
+ const message2 = convertOpenAIResponseToAnthropic(finalResponse2, tools);
2903
+ const assistantMsg2 = {
2904
+ type: "assistant",
2905
+ message: message2,
2906
+ costUSD: 0,
2907
+ durationMs: Date.now() - start,
2908
+ uuid: `${Date.now()}-${Math.random().toString(36).substr(2, 9)}`
2909
+ };
2910
+ return {
2911
+ assistantMessage: assistantMsg2,
2912
+ rawResponse: finalResponse2,
2913
+ apiFormat: "openai"
2914
+ };
2915
+ }
2916
+ const maxTokens = options?.maxTokens ?? getMaxTokensFromProfile(modelProfile);
2917
+ const isGPT5 = isGPT5Model(model);
2918
+ const opts = {
2919
+ model,
2920
+ ...isGPT5 ? { max_completion_tokens: maxTokens } : { max_tokens: maxTokens },
2921
+ messages: [...openaiSystem, ...openaiMessages],
2922
+ temperature: options?.temperature ?? (isGPT5 ? 1 : MAIN_QUERY_TEMPERATURE)
2923
+ };
2924
+ if (options?.stopSequences && options.stopSequences.length > 0) {
2925
+ ;
2926
+ opts.stop = options.stopSequences;
2927
+ }
2928
+ if (config.stream) {
2929
+ ;
2930
+ opts.stream = true;
2931
+ opts.stream_options = {
2932
+ include_usage: true
2933
+ };
2934
+ }
2935
+ if (toolSchemas.length > 0) {
2936
+ opts.tools = toolSchemas;
2937
+ opts.tool_choice = "auto";
2938
+ }
2939
+ const reasoningEffort = await getReasoningEffort(modelProfile, messages);
2940
+ if (reasoningEffort) {
2941
+ opts.reasoning_effort = reasoningEffort;
2942
+ }
2943
+ const completionFunction = isGPT5Model(modelProfile?.modelName || "") ? getGPT5CompletionWithProfile : getCompletionWithProfile;
2944
+ const s = await completionFunction(modelProfile, opts, 0, 10, signal);
2945
+ let finalResponse;
2946
+ if (opts.stream) {
2947
+ finalResponse = await handleMessageStream(
2948
+ s,
2949
+ signal
2950
+ );
2951
+ } else {
2952
+ finalResponse = s;
2953
+ }
2954
+ const message = convertOpenAIResponseToAnthropic(finalResponse, tools);
2955
+ const assistantMsg = {
2956
+ type: "assistant",
2957
+ message,
2958
+ costUSD: 0,
2959
+ durationMs: Date.now() - start,
2960
+ uuid: `${Date.now()}-${Math.random().toString(36).substr(2, 9)}`
2961
+ };
2962
+ return {
2963
+ assistantMessage: assistantMsg,
2964
+ rawResponse: finalResponse,
2965
+ apiFormat: "openai"
2966
+ };
2967
+ },
2968
+ { signal }
2969
+ );
2970
+ } catch (error) {
2971
+ logError(error);
2972
+ return getAssistantMessageFromError(error);
2973
+ }
2974
+ const durationMs = Date.now() - start;
2975
+ const durationMsIncludingRetries = Date.now() - startIncludingRetries;
2976
+ const assistantMessage = queryResult.assistantMessage;
2977
+ assistantMessage.message.content = normalizeContentFromAPI(
2978
+ assistantMessage.message.content || []
2979
+ );
2980
+ const normalizedUsage = normalizeUsage(assistantMessage.message.usage);
2981
+ assistantMessage.message.usage = normalizedUsage;
2982
+ const inputTokens = normalizedUsage.input_tokens ?? 0;
2983
+ const outputTokens = normalizedUsage.output_tokens ?? 0;
2984
+ const cacheReadInputTokens = normalizedUsage.cache_read_input_tokens ?? 0;
2985
+ const cacheCreationInputTokens = normalizedUsage.cache_creation_input_tokens ?? 0;
2986
+ const costUSD = inputTokens / 1e6 * SONNET_COST_PER_MILLION_INPUT_TOKENS + outputTokens / 1e6 * SONNET_COST_PER_MILLION_OUTPUT_TOKENS + cacheReadInputTokens / 1e6 * SONNET_COST_PER_MILLION_PROMPT_CACHE_READ_TOKENS + cacheCreationInputTokens / 1e6 * SONNET_COST_PER_MILLION_PROMPT_CACHE_WRITE_TOKENS;
2987
+ addToTotalCost(costUSD, durationMsIncludingRetries);
2988
+ logLLMInteraction({
2989
+ systemPrompt: systemPrompt.join("\n"),
2990
+ messages: [...openaiSystem, ...openaiMessages],
2991
+ response: assistantMessage.message || queryResult.rawResponse,
2992
+ usage: {
2993
+ inputTokens,
2994
+ outputTokens
2995
+ },
2996
+ timing: {
2997
+ start,
2998
+ end: Date.now()
2999
+ },
3000
+ apiFormat: queryResult.apiFormat
3001
+ });
3002
+ assistantMessage.costUSD = costUSD;
3003
+ assistantMessage.durationMs = durationMs;
3004
+ assistantMessage.uuid = assistantMessage.uuid || randomUUID();
3005
+ return assistantMessage;
3006
+ }
3007
+ function getMaxTokensFromProfile(modelProfile) {
3008
+ return modelProfile?.maxTokens || 8e3;
3009
+ }
3010
+ function buildAssistantMessageFromUnifiedResponse(unifiedResponse, startTime) {
3011
+ const contentBlocks = [...unifiedResponse.content || []];
3012
+ if (unifiedResponse.toolCalls && unifiedResponse.toolCalls.length > 0) {
3013
+ for (const toolCall of unifiedResponse.toolCalls) {
3014
+ const tool = toolCall.function;
3015
+ const toolName = tool?.name;
3016
+ let toolArgs = {};
3017
+ try {
3018
+ toolArgs = tool?.arguments ? JSON.parse(tool.arguments) : {};
3019
+ } catch (e) {
3020
+ }
3021
+ contentBlocks.push({
3022
+ type: "tool_use",
3023
+ input: toolArgs,
3024
+ name: toolName,
3025
+ id: toolCall.id?.length > 0 ? toolCall.id : nanoid()
3026
+ });
3027
+ }
3028
+ }
3029
+ return {
3030
+ type: "assistant",
3031
+ message: {
3032
+ role: "assistant",
3033
+ content: contentBlocks,
3034
+ usage: {
3035
+ input_tokens: unifiedResponse.usage?.promptTokens ?? unifiedResponse.usage?.input_tokens ?? 0,
3036
+ output_tokens: unifiedResponse.usage?.completionTokens ?? unifiedResponse.usage?.output_tokens ?? 0,
3037
+ prompt_tokens: unifiedResponse.usage?.promptTokens ?? unifiedResponse.usage?.input_tokens ?? 0,
3038
+ completion_tokens: unifiedResponse.usage?.completionTokens ?? unifiedResponse.usage?.output_tokens ?? 0,
3039
+ promptTokens: unifiedResponse.usage?.promptTokens ?? unifiedResponse.usage?.input_tokens ?? 0,
3040
+ completionTokens: unifiedResponse.usage?.completionTokens ?? unifiedResponse.usage?.output_tokens ?? 0,
3041
+ totalTokens: unifiedResponse.usage?.totalTokens ?? (unifiedResponse.usage?.promptTokens ?? unifiedResponse.usage?.input_tokens ?? 0) + (unifiedResponse.usage?.completionTokens ?? unifiedResponse.usage?.output_tokens ?? 0)
3042
+ }
3043
+ },
3044
+ costUSD: 0,
3045
+ durationMs: Date.now() - startTime,
3046
+ uuid: `${Date.now()}-${Math.random().toString(36).substr(2, 9)}`,
3047
+ responseId: unifiedResponse.responseId
3048
+ };
3049
+ }
3050
+ function normalizeUsage(usage) {
3051
+ if (!usage) {
3052
+ return {
3053
+ input_tokens: 0,
3054
+ output_tokens: 0,
3055
+ cache_read_input_tokens: 0,
3056
+ cache_creation_input_tokens: 0
3057
+ };
3058
+ }
3059
+ const inputTokens = usage.input_tokens ?? usage.prompt_tokens ?? usage.inputTokens ?? 0;
3060
+ const outputTokens = usage.output_tokens ?? usage.completion_tokens ?? usage.outputTokens ?? 0;
3061
+ const cacheReadInputTokens = usage.cache_read_input_tokens ?? usage.prompt_token_details?.cached_tokens ?? usage.cacheReadInputTokens ?? 0;
3062
+ const cacheCreationInputTokens = usage.cache_creation_input_tokens ?? usage.cacheCreatedInputTokens ?? 0;
3063
+ return {
3064
+ ...usage,
3065
+ input_tokens: inputTokens,
3066
+ output_tokens: outputTokens,
3067
+ cache_read_input_tokens: cacheReadInputTokens,
3068
+ cache_creation_input_tokens: cacheCreationInputTokens
3069
+ };
3070
+ }
3071
+ function getModelInputTokenCostUSD(model) {
3072
+ for (const providerModels of Object.values(models_default)) {
3073
+ const modelInfo = providerModels.find((m) => m.model === model);
3074
+ if (modelInfo) {
3075
+ return modelInfo.input_cost_per_token || 0;
3076
+ }
3077
+ }
3078
+ return 3e-6;
3079
+ }
3080
+ function getModelOutputTokenCostUSD(model) {
3081
+ for (const providerModels of Object.values(models_default)) {
3082
+ const modelInfo = providerModels.find((m) => m.model === model);
3083
+ if (modelInfo) {
3084
+ return modelInfo.output_cost_per_token || 0;
3085
+ }
3086
+ }
3087
+ return 15e-6;
3088
+ }
3089
+ async function queryModel(modelPointer, messages, systemPrompt = [], signal) {
3090
+ return queryLLM(
3091
+ messages,
3092
+ systemPrompt,
3093
+ 0,
3094
+ [],
3095
+ signal || new AbortController().signal,
3096
+ {
3097
+ safeMode: false,
3098
+ model: modelPointer,
3099
+ prependCLISysprompt: true
3100
+ }
3101
+ );
3102
+ }
3103
+ async function queryQuick({
3104
+ systemPrompt = [],
3105
+ userPrompt,
3106
+ assistantPrompt,
3107
+ enablePromptCaching = false,
3108
+ signal
3109
+ }) {
3110
+ const messages = [
3111
+ {
3112
+ message: { role: "user", content: userPrompt },
3113
+ type: "user",
3114
+ uuid: randomUUID()
3115
+ }
3116
+ ];
3117
+ return queryModel("quick", messages, systemPrompt, signal);
3118
+ }
3119
+ export {
3120
+ API_ERROR_MESSAGE_PREFIX,
3121
+ CREDIT_BALANCE_TOO_LOW_ERROR_MESSAGE,
3122
+ INVALID_API_KEY_ERROR_MESSAGE,
3123
+ MAIN_QUERY_TEMPERATURE,
3124
+ NO_CONTENT_MESSAGE,
3125
+ PROMPT_TOO_LONG_ERROR_MESSAGE,
3126
+ assistantMessageToMessageParam,
3127
+ fetchAnthropicModels,
3128
+ formatSystemPromptWithContext,
3129
+ generateDanyaContext,
3130
+ getAnthropicClient,
3131
+ queryLLM,
3132
+ queryModel,
3133
+ queryQuick,
3134
+ refreshDanyaContext,
3135
+ resetAnthropicClient,
3136
+ userMessageToMessageParam,
3137
+ verifyApiKey
3138
+ };