@roo-code/types 1.43.0 → 1.45.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -41,6 +41,7 @@ __export(index_exports, {
41
41
  GLAMA_DEFAULT_TEMPERATURE: () => GLAMA_DEFAULT_TEMPERATURE,
42
42
  GLOBAL_SETTINGS_KEYS: () => GLOBAL_SETTINGS_KEYS,
43
43
  GLOBAL_STATE_KEYS: () => GLOBAL_STATE_KEYS,
44
+ GPT5_DEFAULT_TEMPERATURE: () => GPT5_DEFAULT_TEMPERATURE,
44
45
  HUGGINGFACE_API_URL: () => HUGGINGFACE_API_URL,
45
46
  HUGGINGFACE_CACHE_DURATION: () => HUGGINGFACE_CACHE_DURATION,
46
47
  HUGGINGFACE_DEFAULT_CONTEXT_WINDOW: () => HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
@@ -71,6 +72,7 @@ __export(index_exports, {
71
72
  TaskCommandName: () => TaskCommandName,
72
73
  TelemetryEventName: () => TelemetryEventName,
73
74
  VERTEX_REGIONS: () => VERTEX_REGIONS,
75
+ ZAI_DEFAULT_TEMPERATURE: () => ZAI_DEFAULT_TEMPERATURE,
74
76
  ackSchema: () => ackSchema,
75
77
  anthropicDefaultModelId: () => anthropicDefaultModelId,
76
78
  anthropicModels: () => anthropicModels,
@@ -111,6 +113,9 @@ __export(index_exports, {
111
113
  experimentIds: () => experimentIds,
112
114
  experimentIdsSchema: () => experimentIdsSchema,
113
115
  experimentsSchema: () => experimentsSchema,
116
+ extendedReasoningEffortsSchema: () => extendedReasoningEffortsSchema,
117
+ fireworksDefaultModelId: () => fireworksDefaultModelId,
118
+ fireworksModels: () => fireworksModels,
114
119
  followUpDataSchema: () => followUpDataSchema,
115
120
  geminiDefaultModelId: () => geminiDefaultModelId,
116
121
  geminiModels: () => geminiModels,
@@ -127,6 +132,8 @@ __export(index_exports, {
127
132
  groupOptionsSchema: () => groupOptionsSchema,
128
133
  historyItemSchema: () => historyItemSchema,
129
134
  installMarketplaceItemOptionsSchema: () => installMarketplaceItemOptionsSchema,
135
+ internationalZAiDefaultModelId: () => internationalZAiDefaultModelId,
136
+ internationalZAiModels: () => internationalZAiModels,
130
137
  ipcMessageSchema: () => ipcMessageSchema,
131
138
  isBlockingAsk: () => isBlockingAsk,
132
139
  isGlobalStateKey: () => isGlobalStateKey,
@@ -139,6 +146,8 @@ __export(index_exports, {
139
146
  languagesSchema: () => languagesSchema,
140
147
  litellmDefaultModelId: () => litellmDefaultModelId,
141
148
  litellmDefaultModelInfo: () => litellmDefaultModelInfo,
149
+ mainlandZAiDefaultModelId: () => mainlandZAiDefaultModelId,
150
+ mainlandZAiModels: () => mainlandZAiModels,
142
151
  marketplaceItemSchema: () => marketplaceItemSchema,
143
152
  marketplaceItemTypeSchema: () => marketplaceItemTypeSchema,
144
153
  mcpExecutionStatusSchema: () => mcpExecutionStatusSchema,
@@ -201,6 +210,8 @@ __export(index_exports, {
201
210
  toolUsageSchema: () => toolUsageSchema,
202
211
  unboundDefaultModelId: () => unboundDefaultModelId,
203
212
  unboundDefaultModelInfo: () => unboundDefaultModelInfo,
213
+ verbosityLevels: () => verbosityLevels,
214
+ verbosityLevelsSchema: () => verbosityLevelsSchema,
204
215
  vertexDefaultModelId: () => vertexDefaultModelId,
205
216
  vertexModels: () => vertexModels,
206
217
  vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
@@ -223,6 +234,8 @@ var import_zod3 = require("zod");
223
234
  var import_zod = require("zod");
224
235
  var reasoningEfforts = ["low", "medium", "high"];
225
236
  var reasoningEffortsSchema = import_zod.z.enum(reasoningEfforts);
237
+ var verbosityLevels = ["low", "medium", "high"];
238
+ var verbosityLevelsSchema = import_zod.z.enum(verbosityLevels);
226
239
  var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
227
240
  var modelParametersSchema = import_zod.z.enum(modelParameters);
228
241
  var isModelParameter = (value) => modelParameters.includes(value);
@@ -233,6 +246,8 @@ var modelInfoSchema = import_zod.z.object({
233
246
  supportsImages: import_zod.z.boolean().optional(),
234
247
  supportsComputerUse: import_zod.z.boolean().optional(),
235
248
  supportsPromptCache: import_zod.z.boolean(),
249
+ // Capability flag to indicate whether the model supports an output verbosity parameter
250
+ supportsVerbosity: import_zod.z.boolean().optional(),
236
251
  supportsReasoningBudget: import_zod.z.boolean().optional(),
237
252
  requiredReasoningBudget: import_zod.z.boolean().optional(),
238
253
  supportsReasoningEffort: import_zod.z.boolean().optional(),
@@ -300,6 +315,7 @@ var codebaseIndexProviderSchema = import_zod2.z.object({
300
315
  });
301
316
 
302
317
  // src/provider-settings.ts
318
+ var extendedReasoningEffortsSchema = import_zod3.z.union([reasoningEffortsSchema, import_zod3.z.literal("minimal")]);
303
319
  var providerNames = [
304
320
  "anthropic",
305
321
  "claude-code",
@@ -328,7 +344,9 @@ var providerNames = [
328
344
  "litellm",
329
345
  "huggingface",
330
346
  "cerebras",
331
- "sambanova"
347
+ "sambanova",
348
+ "zai",
349
+ "fireworks"
332
350
  ];
333
351
  var providerNamesSchema = import_zod3.z.enum(providerNames);
334
352
  var providerSettingsEntrySchema = import_zod3.z.object({
@@ -347,9 +365,11 @@ var baseProviderSettingsSchema = import_zod3.z.object({
347
365
  consecutiveMistakeLimit: import_zod3.z.number().min(0).optional(),
348
366
  // Model reasoning.
349
367
  enableReasoningEffort: import_zod3.z.boolean().optional(),
350
- reasoningEffort: reasoningEffortsSchema.optional(),
368
+ reasoningEffort: extendedReasoningEffortsSchema.optional(),
351
369
  modelMaxTokens: import_zod3.z.number().optional(),
352
- modelMaxThinkingTokens: import_zod3.z.number().optional()
370
+ modelMaxThinkingTokens: import_zod3.z.number().optional(),
371
+ // Model verbosity.
372
+ verbosity: verbosityLevelsSchema.optional()
353
373
  });
354
374
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
355
375
  apiModelId: import_zod3.z.string().optional()
@@ -496,6 +516,13 @@ var cerebrasSchema = apiModelIdProviderModelSchema.extend({
496
516
  var sambaNovaSchema = apiModelIdProviderModelSchema.extend({
497
517
  sambaNovaApiKey: import_zod3.z.string().optional()
498
518
  });
519
+ var zaiSchema = apiModelIdProviderModelSchema.extend({
520
+ zaiApiKey: import_zod3.z.string().optional(),
521
+ zaiApiLine: import_zod3.z.union([import_zod3.z.literal("china"), import_zod3.z.literal("international")]).optional()
522
+ });
523
+ var fireworksSchema = apiModelIdProviderModelSchema.extend({
524
+ fireworksApiKey: import_zod3.z.string().optional()
525
+ });
499
526
  var defaultSchema = import_zod3.z.object({
500
527
  apiProvider: import_zod3.z.undefined()
501
528
  });
@@ -528,6 +555,8 @@ var providerSettingsSchemaDiscriminated = import_zod3.z.discriminatedUnion("apiP
528
555
  litellmSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("litellm") })),
529
556
  cerebrasSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("cerebras") })),
530
557
  sambaNovaSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("sambanova") })),
558
+ zaiSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("zai") })),
559
+ fireworksSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("fireworks") })),
531
560
  defaultSchema
532
561
  ]);
533
562
  var providerSettingsSchema = import_zod3.z.object({
@@ -560,6 +589,8 @@ var providerSettingsSchema = import_zod3.z.object({
560
589
  ...litellmSchema.shape,
561
590
  ...cerebrasSchema.shape,
562
591
  ...sambaNovaSchema.shape,
592
+ ...zaiSchema.shape,
593
+ ...fireworksSchema.shape,
563
594
  ...codebaseIndexProviderSchema.shape
564
595
  });
565
596
  var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: import_zod3.z.string().optional() });
@@ -614,12 +645,13 @@ var historyItemSchema = import_zod4.z.object({
614
645
 
615
646
  // src/experiment.ts
616
647
  var import_zod5 = require("zod");
617
- var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption"];
648
+ var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption", "assistantMessageParser"];
618
649
  var experimentIdsSchema = import_zod5.z.enum(experimentIds);
619
650
  var experimentsSchema = import_zod5.z.object({
620
651
  powerSteering: import_zod5.z.boolean().optional(),
621
652
  multiFileApplyDiff: import_zod5.z.boolean().optional(),
622
- preventFocusDisruption: import_zod5.z.boolean().optional()
653
+ preventFocusDisruption: import_zod5.z.boolean().optional(),
654
+ assistantMessageParser: import_zod5.z.boolean().optional()
623
655
  });
624
656
 
625
657
  // src/telemetry.ts
@@ -706,7 +738,14 @@ var clineMessageSchema = import_zod6.z.object({
706
738
  progressStatus: toolProgressStatusSchema.optional(),
707
739
  contextCondense: contextCondenseSchema.optional(),
708
740
  isProtected: import_zod6.z.boolean().optional(),
709
- apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional()
741
+ apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional(),
742
+ metadata: import_zod6.z.object({
743
+ gpt5: import_zod6.z.object({
744
+ previous_response_id: import_zod6.z.string().optional(),
745
+ instructions: import_zod6.z.string().optional(),
746
+ reasoning_summary: import_zod6.z.string().optional()
747
+ }).optional()
748
+ }).optional()
710
749
  });
711
750
  var tokenUsageSchema = import_zod6.z.object({
712
751
  totalTokensIn: import_zod6.z.number(),
@@ -1146,6 +1185,7 @@ var globalSettingsSchema = import_zod11.z.object({
1146
1185
  telemetrySetting: telemetrySettingsSchema.optional(),
1147
1186
  mcpEnabled: import_zod11.z.boolean().optional(),
1148
1187
  enableMcpServerCreation: import_zod11.z.boolean().optional(),
1188
+ remoteControlEnabled: import_zod11.z.boolean().optional(),
1149
1189
  mode: import_zod11.z.string().optional(),
1150
1190
  modeApiConfigs: import_zod11.z.record(import_zod11.z.string(), import_zod11.z.string()).optional(),
1151
1191
  customModes: import_zod11.z.array(modeConfigSchema).optional(),
@@ -1188,7 +1228,8 @@ var SECRET_STATE_KEYS = [
1188
1228
  "codebaseIndexGeminiApiKey",
1189
1229
  "codebaseIndexMistralApiKey",
1190
1230
  "huggingFaceApiKey",
1191
- "sambaNovaApiKey"
1231
+ "sambaNovaApiKey",
1232
+ "fireworksApiKey"
1192
1233
  ];
1193
1234
  var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
1194
1235
  var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
@@ -1255,6 +1296,7 @@ var EVALS_SETTINGS = {
1255
1296
  language: "en",
1256
1297
  telemetrySetting: "enabled",
1257
1298
  mcpEnabled: false,
1299
+ remoteControlEnabled: false,
1258
1300
  mode: "code",
1259
1301
  // "architect",
1260
1302
  customModes: []
@@ -1696,6 +1738,22 @@ var anthropicModels = {
1696
1738
  // $0.30 per million tokens
1697
1739
  supportsReasoningBudget: true
1698
1740
  },
1741
+ "claude-opus-4-1-20250805": {
1742
+ maxTokens: 8192,
1743
+ contextWindow: 2e5,
1744
+ supportsImages: true,
1745
+ supportsComputerUse: true,
1746
+ supportsPromptCache: true,
1747
+ inputPrice: 15,
1748
+ // $15 per million input tokens
1749
+ outputPrice: 75,
1750
+ // $75 per million output tokens
1751
+ cacheWritesPrice: 18.75,
1752
+ // $18.75 per million tokens
1753
+ cacheReadsPrice: 1.5,
1754
+ // $1.50 per million tokens
1755
+ supportsReasoningBudget: true
1756
+ },
1699
1757
  "claude-opus-4-20250514": {
1700
1758
  maxTokens: 32e3,
1701
1759
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -1876,6 +1934,21 @@ var bedrockModels = {
1876
1934
  maxCachePoints: 4,
1877
1935
  cachableFields: ["system", "messages", "tools"]
1878
1936
  },
1937
+ "anthropic.claude-opus-4-1-20250805-v1:0": {
1938
+ maxTokens: 8192,
1939
+ contextWindow: 2e5,
1940
+ supportsImages: true,
1941
+ supportsComputerUse: true,
1942
+ supportsPromptCache: true,
1943
+ supportsReasoningBudget: true,
1944
+ inputPrice: 15,
1945
+ outputPrice: 75,
1946
+ cacheWritesPrice: 18.75,
1947
+ cacheReadsPrice: 1.5,
1948
+ minTokensPerCachePoint: 1024,
1949
+ maxCachePoints: 4,
1950
+ cachableFields: ["system", "messages", "tools"]
1951
+ },
1879
1952
  "anthropic.claude-opus-4-20250514-v1:0": {
1880
1953
  maxTokens: 8192,
1881
1954
  contextWindow: 2e5,
@@ -2248,6 +2321,15 @@ var cerebrasModels = {
2248
2321
  outputPrice: 0,
2249
2322
  description: "SOTA performance with ~1500 tokens/s",
2250
2323
  supportsReasoningEffort: true
2324
+ },
2325
+ "gpt-oss-120b": {
2326
+ maxTokens: 8e3,
2327
+ contextWindow: 64e3,
2328
+ supportsImages: false,
2329
+ supportsPromptCache: false,
2330
+ inputPrice: 0,
2331
+ outputPrice: 0,
2332
+ description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
2251
2333
  }
2252
2334
  };
2253
2335
 
@@ -2503,6 +2585,15 @@ var claudeCodeModels = {
2503
2585
  supportsReasoningBudget: false,
2504
2586
  requiredReasoningBudget: false
2505
2587
  },
2588
+ "claude-opus-4-1-20250805": {
2589
+ ...anthropicModels["claude-opus-4-1-20250805"],
2590
+ supportsImages: false,
2591
+ supportsPromptCache: true,
2592
+ // Claude Code does report cache tokens
2593
+ supportsReasoningEffort: false,
2594
+ supportsReasoningBudget: false,
2595
+ requiredReasoningBudget: false
2596
+ },
2506
2597
  "claude-opus-4-20250514": {
2507
2598
  ...anthropicModels["claude-opus-4-20250514"],
2508
2599
  supportsImages: false,
@@ -2976,6 +3067,24 @@ var groqModels = {
2976
3067
  inputPrice: 1,
2977
3068
  outputPrice: 3,
2978
3069
  description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
3070
+ },
3071
+ "openai/gpt-oss-120b": {
3072
+ maxTokens: 32766,
3073
+ contextWindow: 131072,
3074
+ supportsImages: false,
3075
+ supportsPromptCache: false,
3076
+ inputPrice: 0.15,
3077
+ outputPrice: 0.75,
3078
+ description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
3079
+ },
3080
+ "openai/gpt-oss-20b": {
3081
+ maxTokens: 32768,
3082
+ contextWindow: 131072,
3083
+ supportsImages: false,
3084
+ supportsPromptCache: false,
3085
+ inputPrice: 0.1,
3086
+ outputPrice: 0.5,
3087
+ description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
2979
3088
  }
2980
3089
  };
2981
3090
 
@@ -3004,6 +3113,7 @@ var litellmDefaultModelInfo = {
3004
3113
  };
3005
3114
  var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3006
3115
  "claude-3-5-sonnet-latest",
3116
+ "claude-opus-4-1-20250805",
3007
3117
  "claude-opus-4-20250514",
3008
3118
  "claude-sonnet-4-20250514",
3009
3119
  "claude-3-7-sonnet-latest",
@@ -3013,22 +3123,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3013
3123
  "vertex_ai/claude-3-5-sonnet-v2",
3014
3124
  "vertex_ai/claude-3-5-sonnet-v2@20241022",
3015
3125
  "vertex_ai/claude-3-7-sonnet@20250219",
3126
+ "vertex_ai/claude-opus-4-1@20250805",
3016
3127
  "vertex_ai/claude-opus-4@20250514",
3017
3128
  "vertex_ai/claude-sonnet-4@20250514",
3018
3129
  "openrouter/anthropic/claude-3.5-sonnet",
3019
3130
  "openrouter/anthropic/claude-3.5-sonnet:beta",
3020
3131
  "openrouter/anthropic/claude-3.7-sonnet",
3021
3132
  "openrouter/anthropic/claude-3.7-sonnet:beta",
3133
+ "anthropic.claude-opus-4-1-20250805-v1:0",
3022
3134
  "anthropic.claude-opus-4-20250514-v1:0",
3023
3135
  "anthropic.claude-sonnet-4-20250514-v1:0",
3024
3136
  "anthropic.claude-3-7-sonnet-20250219-v1:0",
3025
3137
  "anthropic.claude-3-5-sonnet-20241022-v2:0",
3026
3138
  "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
3027
3139
  "us.anthropic.claude-3-7-sonnet-20250219-v1:0",
3140
+ "us.anthropic.claude-opus-4-1-20250805-v1:0",
3028
3141
  "us.anthropic.claude-opus-4-20250514-v1:0",
3029
3142
  "us.anthropic.claude-sonnet-4-20250514-v1:0",
3030
3143
  "eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
3031
3144
  "eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
3145
+ "eu.anthropic.claude-opus-4-1-20250805-v1:0",
3032
3146
  "eu.anthropic.claude-opus-4-20250514-v1:0",
3033
3147
  "eu.anthropic.claude-sonnet-4-20250514-v1:0",
3034
3148
  "snowflake/claude-3-5-sonnet"
@@ -3165,8 +3279,48 @@ var ollamaDefaultModelInfo = {
3165
3279
  };
3166
3280
 
3167
3281
  // src/providers/openai.ts
3168
- var openAiNativeDefaultModelId = "gpt-4.1";
3282
+ var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
3169
3283
  var openAiNativeModels = {
3284
+ "gpt-5-2025-08-07": {
3285
+ maxTokens: 128e3,
3286
+ contextWindow: 4e5,
3287
+ supportsImages: true,
3288
+ supportsPromptCache: true,
3289
+ supportsReasoningEffort: true,
3290
+ reasoningEffort: "medium",
3291
+ inputPrice: 1.25,
3292
+ outputPrice: 10,
3293
+ cacheReadsPrice: 0.13,
3294
+ description: "GPT-5: The best model for coding and agentic tasks across domains",
3295
+ // supportsVerbosity is a new capability; ensure ModelInfo includes it
3296
+ supportsVerbosity: true
3297
+ },
3298
+ "gpt-5-mini-2025-08-07": {
3299
+ maxTokens: 128e3,
3300
+ contextWindow: 4e5,
3301
+ supportsImages: true,
3302
+ supportsPromptCache: true,
3303
+ supportsReasoningEffort: true,
3304
+ reasoningEffort: "medium",
3305
+ inputPrice: 0.25,
3306
+ outputPrice: 2,
3307
+ cacheReadsPrice: 0.03,
3308
+ description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
3309
+ supportsVerbosity: true
3310
+ },
3311
+ "gpt-5-nano-2025-08-07": {
3312
+ maxTokens: 128e3,
3313
+ contextWindow: 4e5,
3314
+ supportsImages: true,
3315
+ supportsPromptCache: true,
3316
+ supportsReasoningEffort: true,
3317
+ reasoningEffort: "medium",
3318
+ inputPrice: 0.05,
3319
+ outputPrice: 0.4,
3320
+ cacheReadsPrice: 0.01,
3321
+ description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
3322
+ supportsVerbosity: true
3323
+ },
3170
3324
  "gpt-4.1": {
3171
3325
  maxTokens: 32768,
3172
3326
  contextWindow: 1047576,
@@ -3352,6 +3506,7 @@ var openAiModelInfoSaneDefaults = {
3352
3506
  };
3353
3507
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3354
3508
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3509
+ var GPT5_DEFAULT_TEMPERATURE = 1;
3355
3510
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3356
3511
 
3357
3512
  // src/providers/openrouter.ts
@@ -3389,6 +3544,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3389
3544
  "anthropic/claude-3.7-sonnet:thinking",
3390
3545
  "anthropic/claude-sonnet-4",
3391
3546
  "anthropic/claude-opus-4",
3547
+ "anthropic/claude-opus-4.1",
3392
3548
  "google/gemini-2.5-flash-preview",
3393
3549
  "google/gemini-2.5-flash-preview:thinking",
3394
3550
  "google/gemini-2.5-flash-preview-05-20",
@@ -3406,7 +3562,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3406
3562
  "anthropic/claude-3.7-sonnet:beta",
3407
3563
  "anthropic/claude-3.7-sonnet:thinking",
3408
3564
  "anthropic/claude-sonnet-4",
3409
- "anthropic/claude-opus-4"
3565
+ "anthropic/claude-opus-4",
3566
+ "anthropic/claude-opus-4.1"
3410
3567
  ]);
3411
3568
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3412
3569
  "anthropic/claude-3.7-sonnet:thinking",
@@ -3416,6 +3573,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3416
3573
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3417
3574
  "anthropic/claude-3.7-sonnet:beta",
3418
3575
  "anthropic/claude-opus-4",
3576
+ "anthropic/claude-opus-4.1",
3419
3577
  "anthropic/claude-sonnet-4",
3420
3578
  "google/gemini-2.5-pro-preview",
3421
3579
  "google/gemini-2.5-pro",
@@ -3706,6 +3864,18 @@ var vertexModels = {
3706
3864
  cacheReadsPrice: 0.3,
3707
3865
  supportsReasoningBudget: true
3708
3866
  },
3867
+ "claude-opus-4-1@20250805": {
3868
+ maxTokens: 8192,
3869
+ contextWindow: 2e5,
3870
+ supportsImages: true,
3871
+ supportsComputerUse: true,
3872
+ supportsPromptCache: true,
3873
+ inputPrice: 15,
3874
+ outputPrice: 75,
3875
+ cacheWritesPrice: 18.75,
3876
+ cacheReadsPrice: 1.5,
3877
+ supportsReasoningBudget: true
3878
+ },
3709
3879
  "claude-opus-4@20250514": {
3710
3880
  maxTokens: 8192,
3711
3881
  contextWindow: 2e5,
@@ -4130,6 +4300,185 @@ var doubaoModels = {
4130
4300
  var doubaoDefaultModelInfo = doubaoModels[doubaoDefaultModelId];
4131
4301
  var DOUBAO_API_BASE_URL = "https://ark.cn-beijing.volces.com/api/v3";
4132
4302
  var DOUBAO_API_CHAT_PATH = "/chat/completions";
4303
+
4304
+ // src/providers/zai.ts
4305
+ var internationalZAiDefaultModelId = "glm-4.5";
4306
+ var internationalZAiModels = {
4307
+ "glm-4.5": {
4308
+ maxTokens: 98304,
4309
+ contextWindow: 131072,
4310
+ supportsImages: false,
4311
+ supportsPromptCache: true,
4312
+ inputPrice: 0.6,
4313
+ outputPrice: 2.2,
4314
+ cacheWritesPrice: 0,
4315
+ cacheReadsPrice: 0.11,
4316
+ description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
4317
+ },
4318
+ "glm-4.5-air": {
4319
+ maxTokens: 98304,
4320
+ contextWindow: 131072,
4321
+ supportsImages: false,
4322
+ supportsPromptCache: true,
4323
+ inputPrice: 0.2,
4324
+ outputPrice: 1.1,
4325
+ cacheWritesPrice: 0,
4326
+ cacheReadsPrice: 0.03,
4327
+ description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
4328
+ }
4329
+ };
4330
+ var mainlandZAiDefaultModelId = "glm-4.5";
4331
+ var mainlandZAiModels = {
4332
+ "glm-4.5": {
4333
+ maxTokens: 98304,
4334
+ contextWindow: 131072,
4335
+ supportsImages: false,
4336
+ supportsPromptCache: true,
4337
+ inputPrice: 0.29,
4338
+ outputPrice: 1.14,
4339
+ cacheWritesPrice: 0,
4340
+ cacheReadsPrice: 0.057,
4341
+ description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k.",
4342
+ tiers: [
4343
+ {
4344
+ contextWindow: 32e3,
4345
+ inputPrice: 0.21,
4346
+ outputPrice: 1,
4347
+ cacheReadsPrice: 0.043
4348
+ },
4349
+ {
4350
+ contextWindow: 128e3,
4351
+ inputPrice: 0.29,
4352
+ outputPrice: 1.14,
4353
+ cacheReadsPrice: 0.057
4354
+ },
4355
+ {
4356
+ contextWindow: Infinity,
4357
+ inputPrice: 0.29,
4358
+ outputPrice: 1.14,
4359
+ cacheReadsPrice: 0.057
4360
+ }
4361
+ ]
4362
+ },
4363
+ "glm-4.5-air": {
4364
+ maxTokens: 98304,
4365
+ contextWindow: 131072,
4366
+ supportsImages: false,
4367
+ supportsPromptCache: true,
4368
+ inputPrice: 0.1,
4369
+ outputPrice: 0.6,
4370
+ cacheWritesPrice: 0,
4371
+ cacheReadsPrice: 0.02,
4372
+ description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models.",
4373
+ tiers: [
4374
+ {
4375
+ contextWindow: 32e3,
4376
+ inputPrice: 0.07,
4377
+ outputPrice: 0.4,
4378
+ cacheReadsPrice: 0.014
4379
+ },
4380
+ {
4381
+ contextWindow: 128e3,
4382
+ inputPrice: 0.1,
4383
+ outputPrice: 0.6,
4384
+ cacheReadsPrice: 0.02
4385
+ },
4386
+ {
4387
+ contextWindow: Infinity,
4388
+ inputPrice: 0.1,
4389
+ outputPrice: 0.6,
4390
+ cacheReadsPrice: 0.02
4391
+ }
4392
+ ]
4393
+ }
4394
+ };
4395
+ var ZAI_DEFAULT_TEMPERATURE = 0;
4396
+
4397
+ // src/providers/fireworks.ts
4398
+ var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
4399
+ var fireworksModels = {
4400
+ "accounts/fireworks/models/kimi-k2-instruct": {
4401
+ maxTokens: 16384,
4402
+ contextWindow: 128e3,
4403
+ supportsImages: false,
4404
+ supportsPromptCache: false,
4405
+ inputPrice: 0.6,
4406
+ outputPrice: 2.5,
4407
+ description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
4408
+ },
4409
+ "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
4410
+ maxTokens: 32768,
4411
+ contextWindow: 256e3,
4412
+ supportsImages: false,
4413
+ supportsPromptCache: false,
4414
+ inputPrice: 0.22,
4415
+ outputPrice: 0.88,
4416
+ description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
4417
+ },
4418
+ "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
4419
+ maxTokens: 32768,
4420
+ contextWindow: 256e3,
4421
+ supportsImages: false,
4422
+ supportsPromptCache: false,
4423
+ inputPrice: 0.45,
4424
+ outputPrice: 1.8,
4425
+ description: "Qwen3's most agentic code model to date."
4426
+ },
4427
+ "accounts/fireworks/models/deepseek-r1-0528": {
4428
+ maxTokens: 20480,
4429
+ contextWindow: 16e4,
4430
+ supportsImages: false,
4431
+ supportsPromptCache: false,
4432
+ inputPrice: 3,
4433
+ outputPrice: 8,
4434
+ description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4435
+ },
4436
+ "accounts/fireworks/models/deepseek-v3": {
4437
+ maxTokens: 16384,
4438
+ contextWindow: 128e3,
4439
+ supportsImages: false,
4440
+ supportsPromptCache: false,
4441
+ inputPrice: 0.9,
4442
+ outputPrice: 0.9,
4443
+ description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4444
+ },
4445
+ "accounts/fireworks/models/glm-4p5": {
4446
+ maxTokens: 16384,
4447
+ contextWindow: 128e3,
4448
+ supportsImages: false,
4449
+ supportsPromptCache: false,
4450
+ inputPrice: 0.55,
4451
+ outputPrice: 2.19,
4452
+ description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4453
+ },
4454
+ "accounts/fireworks/models/glm-4p5-air": {
4455
+ maxTokens: 16384,
4456
+ contextWindow: 128e3,
4457
+ supportsImages: false,
4458
+ supportsPromptCache: false,
4459
+ inputPrice: 0.55,
4460
+ outputPrice: 2.19,
4461
+ description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4462
+ },
4463
+ "accounts/fireworks/models/gpt-oss-20b": {
4464
+ maxTokens: 16384,
4465
+ contextWindow: 128e3,
4466
+ supportsImages: false,
4467
+ supportsPromptCache: false,
4468
+ inputPrice: 0.07,
4469
+ outputPrice: 0.3,
4470
+ description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
4471
+ },
4472
+ "accounts/fireworks/models/gpt-oss-120b": {
4473
+ maxTokens: 16384,
4474
+ contextWindow: 128e3,
4475
+ supportsImages: false,
4476
+ supportsPromptCache: false,
4477
+ inputPrice: 0.15,
4478
+ outputPrice: 0.6,
4479
+ description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
4480
+ }
4481
+ };
4133
4482
  // Annotate the CommonJS export names for ESM import in node:
4134
4483
  0 && (module.exports = {
4135
4484
  ANTHROPIC_DEFAULT_MAX_TOKENS,
@@ -4153,6 +4502,7 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
4153
4502
  GLAMA_DEFAULT_TEMPERATURE,
4154
4503
  GLOBAL_SETTINGS_KEYS,
4155
4504
  GLOBAL_STATE_KEYS,
4505
+ GPT5_DEFAULT_TEMPERATURE,
4156
4506
  HUGGINGFACE_API_URL,
4157
4507
  HUGGINGFACE_CACHE_DURATION,
4158
4508
  HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
@@ -4183,6 +4533,7 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
4183
4533
  TaskCommandName,
4184
4534
  TelemetryEventName,
4185
4535
  VERTEX_REGIONS,
4536
+ ZAI_DEFAULT_TEMPERATURE,
4186
4537
  ackSchema,
4187
4538
  anthropicDefaultModelId,
4188
4539
  anthropicModels,
@@ -4223,6 +4574,9 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
4223
4574
  experimentIds,
4224
4575
  experimentIdsSchema,
4225
4576
  experimentsSchema,
4577
+ extendedReasoningEffortsSchema,
4578
+ fireworksDefaultModelId,
4579
+ fireworksModels,
4226
4580
  followUpDataSchema,
4227
4581
  geminiDefaultModelId,
4228
4582
  geminiModels,
@@ -4239,6 +4593,8 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
4239
4593
  groupOptionsSchema,
4240
4594
  historyItemSchema,
4241
4595
  installMarketplaceItemOptionsSchema,
4596
+ internationalZAiDefaultModelId,
4597
+ internationalZAiModels,
4242
4598
  ipcMessageSchema,
4243
4599
  isBlockingAsk,
4244
4600
  isGlobalStateKey,
@@ -4251,6 +4607,8 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
4251
4607
  languagesSchema,
4252
4608
  litellmDefaultModelId,
4253
4609
  litellmDefaultModelInfo,
4610
+ mainlandZAiDefaultModelId,
4611
+ mainlandZAiModels,
4254
4612
  marketplaceItemSchema,
4255
4613
  marketplaceItemTypeSchema,
4256
4614
  mcpExecutionStatusSchema,
@@ -4313,6 +4671,8 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
4313
4671
  toolUsageSchema,
4314
4672
  unboundDefaultModelId,
4315
4673
  unboundDefaultModelInfo,
4674
+ verbosityLevels,
4675
+ verbosityLevelsSchema,
4316
4676
  vertexDefaultModelId,
4317
4677
  vertexModels,
4318
4678
  vscodeLlmDefaultModelId,