@roo-code/types 1.43.0 → 1.45.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -11,6 +11,8 @@ import { z as z3 } from "zod";
11
11
  import { z } from "zod";
12
12
  var reasoningEfforts = ["low", "medium", "high"];
13
13
  var reasoningEffortsSchema = z.enum(reasoningEfforts);
14
+ var verbosityLevels = ["low", "medium", "high"];
15
+ var verbosityLevelsSchema = z.enum(verbosityLevels);
14
16
  var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
15
17
  var modelParametersSchema = z.enum(modelParameters);
16
18
  var isModelParameter = (value) => modelParameters.includes(value);
@@ -21,6 +23,8 @@ var modelInfoSchema = z.object({
21
23
  supportsImages: z.boolean().optional(),
22
24
  supportsComputerUse: z.boolean().optional(),
23
25
  supportsPromptCache: z.boolean(),
26
+ // Capability flag to indicate whether the model supports an output verbosity parameter
27
+ supportsVerbosity: z.boolean().optional(),
24
28
  supportsReasoningBudget: z.boolean().optional(),
25
29
  requiredReasoningBudget: z.boolean().optional(),
26
30
  supportsReasoningEffort: z.boolean().optional(),
@@ -88,6 +92,7 @@ var codebaseIndexProviderSchema = z2.object({
88
92
  });
89
93
 
90
94
  // src/provider-settings.ts
95
+ var extendedReasoningEffortsSchema = z3.union([reasoningEffortsSchema, z3.literal("minimal")]);
91
96
  var providerNames = [
92
97
  "anthropic",
93
98
  "claude-code",
@@ -116,7 +121,9 @@ var providerNames = [
116
121
  "litellm",
117
122
  "huggingface",
118
123
  "cerebras",
119
- "sambanova"
124
+ "sambanova",
125
+ "zai",
126
+ "fireworks"
120
127
  ];
121
128
  var providerNamesSchema = z3.enum(providerNames);
122
129
  var providerSettingsEntrySchema = z3.object({
@@ -135,9 +142,11 @@ var baseProviderSettingsSchema = z3.object({
135
142
  consecutiveMistakeLimit: z3.number().min(0).optional(),
136
143
  // Model reasoning.
137
144
  enableReasoningEffort: z3.boolean().optional(),
138
- reasoningEffort: reasoningEffortsSchema.optional(),
145
+ reasoningEffort: extendedReasoningEffortsSchema.optional(),
139
146
  modelMaxTokens: z3.number().optional(),
140
- modelMaxThinkingTokens: z3.number().optional()
147
+ modelMaxThinkingTokens: z3.number().optional(),
148
+ // Model verbosity.
149
+ verbosity: verbosityLevelsSchema.optional()
141
150
  });
142
151
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
143
152
  apiModelId: z3.string().optional()
@@ -284,6 +293,13 @@ var cerebrasSchema = apiModelIdProviderModelSchema.extend({
284
293
  var sambaNovaSchema = apiModelIdProviderModelSchema.extend({
285
294
  sambaNovaApiKey: z3.string().optional()
286
295
  });
296
+ var zaiSchema = apiModelIdProviderModelSchema.extend({
297
+ zaiApiKey: z3.string().optional(),
298
+ zaiApiLine: z3.union([z3.literal("china"), z3.literal("international")]).optional()
299
+ });
300
+ var fireworksSchema = apiModelIdProviderModelSchema.extend({
301
+ fireworksApiKey: z3.string().optional()
302
+ });
287
303
  var defaultSchema = z3.object({
288
304
  apiProvider: z3.undefined()
289
305
  });
@@ -316,6 +332,8 @@ var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
316
332
  litellmSchema.merge(z3.object({ apiProvider: z3.literal("litellm") })),
317
333
  cerebrasSchema.merge(z3.object({ apiProvider: z3.literal("cerebras") })),
318
334
  sambaNovaSchema.merge(z3.object({ apiProvider: z3.literal("sambanova") })),
335
+ zaiSchema.merge(z3.object({ apiProvider: z3.literal("zai") })),
336
+ fireworksSchema.merge(z3.object({ apiProvider: z3.literal("fireworks") })),
319
337
  defaultSchema
320
338
  ]);
321
339
  var providerSettingsSchema = z3.object({
@@ -348,6 +366,8 @@ var providerSettingsSchema = z3.object({
348
366
  ...litellmSchema.shape,
349
367
  ...cerebrasSchema.shape,
350
368
  ...sambaNovaSchema.shape,
369
+ ...zaiSchema.shape,
370
+ ...fireworksSchema.shape,
351
371
  ...codebaseIndexProviderSchema.shape
352
372
  });
353
373
  var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: z3.string().optional() });
@@ -402,12 +422,13 @@ var historyItemSchema = z4.object({
402
422
 
403
423
  // src/experiment.ts
404
424
  import { z as z5 } from "zod";
405
- var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption"];
425
+ var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption", "assistantMessageParser"];
406
426
  var experimentIdsSchema = z5.enum(experimentIds);
407
427
  var experimentsSchema = z5.object({
408
428
  powerSteering: z5.boolean().optional(),
409
429
  multiFileApplyDiff: z5.boolean().optional(),
410
- preventFocusDisruption: z5.boolean().optional()
430
+ preventFocusDisruption: z5.boolean().optional(),
431
+ assistantMessageParser: z5.boolean().optional()
411
432
  });
412
433
 
413
434
  // src/telemetry.ts
@@ -494,7 +515,14 @@ var clineMessageSchema = z6.object({
494
515
  progressStatus: toolProgressStatusSchema.optional(),
495
516
  contextCondense: contextCondenseSchema.optional(),
496
517
  isProtected: z6.boolean().optional(),
497
- apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional()
518
+ apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional(),
519
+ metadata: z6.object({
520
+ gpt5: z6.object({
521
+ previous_response_id: z6.string().optional(),
522
+ instructions: z6.string().optional(),
523
+ reasoning_summary: z6.string().optional()
524
+ }).optional()
525
+ }).optional()
498
526
  });
499
527
  var tokenUsageSchema = z6.object({
500
528
  totalTokensIn: z6.number(),
@@ -934,6 +962,7 @@ var globalSettingsSchema = z11.object({
934
962
  telemetrySetting: telemetrySettingsSchema.optional(),
935
963
  mcpEnabled: z11.boolean().optional(),
936
964
  enableMcpServerCreation: z11.boolean().optional(),
965
+ remoteControlEnabled: z11.boolean().optional(),
937
966
  mode: z11.string().optional(),
938
967
  modeApiConfigs: z11.record(z11.string(), z11.string()).optional(),
939
968
  customModes: z11.array(modeConfigSchema).optional(),
@@ -976,7 +1005,8 @@ var SECRET_STATE_KEYS = [
976
1005
  "codebaseIndexGeminiApiKey",
977
1006
  "codebaseIndexMistralApiKey",
978
1007
  "huggingFaceApiKey",
979
- "sambaNovaApiKey"
1008
+ "sambaNovaApiKey",
1009
+ "fireworksApiKey"
980
1010
  ];
981
1011
  var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
982
1012
  var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
@@ -1043,6 +1073,7 @@ var EVALS_SETTINGS = {
1043
1073
  language: "en",
1044
1074
  telemetrySetting: "enabled",
1045
1075
  mcpEnabled: false,
1076
+ remoteControlEnabled: false,
1046
1077
  mode: "code",
1047
1078
  // "architect",
1048
1079
  customModes: []
@@ -1484,6 +1515,22 @@ var anthropicModels = {
1484
1515
  // $0.30 per million tokens
1485
1516
  supportsReasoningBudget: true
1486
1517
  },
1518
+ "claude-opus-4-1-20250805": {
1519
+ maxTokens: 8192,
1520
+ contextWindow: 2e5,
1521
+ supportsImages: true,
1522
+ supportsComputerUse: true,
1523
+ supportsPromptCache: true,
1524
+ inputPrice: 15,
1525
+ // $15 per million input tokens
1526
+ outputPrice: 75,
1527
+ // $75 per million output tokens
1528
+ cacheWritesPrice: 18.75,
1529
+ // $18.75 per million tokens
1530
+ cacheReadsPrice: 1.5,
1531
+ // $1.50 per million tokens
1532
+ supportsReasoningBudget: true
1533
+ },
1487
1534
  "claude-opus-4-20250514": {
1488
1535
  maxTokens: 32e3,
1489
1536
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -1664,6 +1711,21 @@ var bedrockModels = {
1664
1711
  maxCachePoints: 4,
1665
1712
  cachableFields: ["system", "messages", "tools"]
1666
1713
  },
1714
+ "anthropic.claude-opus-4-1-20250805-v1:0": {
1715
+ maxTokens: 8192,
1716
+ contextWindow: 2e5,
1717
+ supportsImages: true,
1718
+ supportsComputerUse: true,
1719
+ supportsPromptCache: true,
1720
+ supportsReasoningBudget: true,
1721
+ inputPrice: 15,
1722
+ outputPrice: 75,
1723
+ cacheWritesPrice: 18.75,
1724
+ cacheReadsPrice: 1.5,
1725
+ minTokensPerCachePoint: 1024,
1726
+ maxCachePoints: 4,
1727
+ cachableFields: ["system", "messages", "tools"]
1728
+ },
1667
1729
  "anthropic.claude-opus-4-20250514-v1:0": {
1668
1730
  maxTokens: 8192,
1669
1731
  contextWindow: 2e5,
@@ -2036,6 +2098,15 @@ var cerebrasModels = {
2036
2098
  outputPrice: 0,
2037
2099
  description: "SOTA performance with ~1500 tokens/s",
2038
2100
  supportsReasoningEffort: true
2101
+ },
2102
+ "gpt-oss-120b": {
2103
+ maxTokens: 8e3,
2104
+ contextWindow: 64e3,
2105
+ supportsImages: false,
2106
+ supportsPromptCache: false,
2107
+ inputPrice: 0,
2108
+ outputPrice: 0,
2109
+ description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
2039
2110
  }
2040
2111
  };
2041
2112
 
@@ -2291,6 +2362,15 @@ var claudeCodeModels = {
2291
2362
  supportsReasoningBudget: false,
2292
2363
  requiredReasoningBudget: false
2293
2364
  },
2365
+ "claude-opus-4-1-20250805": {
2366
+ ...anthropicModels["claude-opus-4-1-20250805"],
2367
+ supportsImages: false,
2368
+ supportsPromptCache: true,
2369
+ // Claude Code does report cache tokens
2370
+ supportsReasoningEffort: false,
2371
+ supportsReasoningBudget: false,
2372
+ requiredReasoningBudget: false
2373
+ },
2294
2374
  "claude-opus-4-20250514": {
2295
2375
  ...anthropicModels["claude-opus-4-20250514"],
2296
2376
  supportsImages: false,
@@ -2764,6 +2844,24 @@ var groqModels = {
2764
2844
  inputPrice: 1,
2765
2845
  outputPrice: 3,
2766
2846
  description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
2847
+ },
2848
+ "openai/gpt-oss-120b": {
2849
+ maxTokens: 32766,
2850
+ contextWindow: 131072,
2851
+ supportsImages: false,
2852
+ supportsPromptCache: false,
2853
+ inputPrice: 0.15,
2854
+ outputPrice: 0.75,
2855
+ description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
2856
+ },
2857
+ "openai/gpt-oss-20b": {
2858
+ maxTokens: 32768,
2859
+ contextWindow: 131072,
2860
+ supportsImages: false,
2861
+ supportsPromptCache: false,
2862
+ inputPrice: 0.1,
2863
+ outputPrice: 0.5,
2864
+ description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
2767
2865
  }
2768
2866
  };
2769
2867
 
@@ -2792,6 +2890,7 @@ var litellmDefaultModelInfo = {
2792
2890
  };
2793
2891
  var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2794
2892
  "claude-3-5-sonnet-latest",
2893
+ "claude-opus-4-1-20250805",
2795
2894
  "claude-opus-4-20250514",
2796
2895
  "claude-sonnet-4-20250514",
2797
2896
  "claude-3-7-sonnet-latest",
@@ -2801,22 +2900,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2801
2900
  "vertex_ai/claude-3-5-sonnet-v2",
2802
2901
  "vertex_ai/claude-3-5-sonnet-v2@20241022",
2803
2902
  "vertex_ai/claude-3-7-sonnet@20250219",
2903
+ "vertex_ai/claude-opus-4-1@20250805",
2804
2904
  "vertex_ai/claude-opus-4@20250514",
2805
2905
  "vertex_ai/claude-sonnet-4@20250514",
2806
2906
  "openrouter/anthropic/claude-3.5-sonnet",
2807
2907
  "openrouter/anthropic/claude-3.5-sonnet:beta",
2808
2908
  "openrouter/anthropic/claude-3.7-sonnet",
2809
2909
  "openrouter/anthropic/claude-3.7-sonnet:beta",
2910
+ "anthropic.claude-opus-4-1-20250805-v1:0",
2810
2911
  "anthropic.claude-opus-4-20250514-v1:0",
2811
2912
  "anthropic.claude-sonnet-4-20250514-v1:0",
2812
2913
  "anthropic.claude-3-7-sonnet-20250219-v1:0",
2813
2914
  "anthropic.claude-3-5-sonnet-20241022-v2:0",
2814
2915
  "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
2815
2916
  "us.anthropic.claude-3-7-sonnet-20250219-v1:0",
2917
+ "us.anthropic.claude-opus-4-1-20250805-v1:0",
2816
2918
  "us.anthropic.claude-opus-4-20250514-v1:0",
2817
2919
  "us.anthropic.claude-sonnet-4-20250514-v1:0",
2818
2920
  "eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
2819
2921
  "eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
2922
+ "eu.anthropic.claude-opus-4-1-20250805-v1:0",
2820
2923
  "eu.anthropic.claude-opus-4-20250514-v1:0",
2821
2924
  "eu.anthropic.claude-sonnet-4-20250514-v1:0",
2822
2925
  "snowflake/claude-3-5-sonnet"
@@ -2953,8 +3056,48 @@ var ollamaDefaultModelInfo = {
2953
3056
  };
2954
3057
 
2955
3058
  // src/providers/openai.ts
2956
- var openAiNativeDefaultModelId = "gpt-4.1";
3059
+ var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
2957
3060
  var openAiNativeModels = {
3061
+ "gpt-5-2025-08-07": {
3062
+ maxTokens: 128e3,
3063
+ contextWindow: 4e5,
3064
+ supportsImages: true,
3065
+ supportsPromptCache: true,
3066
+ supportsReasoningEffort: true,
3067
+ reasoningEffort: "medium",
3068
+ inputPrice: 1.25,
3069
+ outputPrice: 10,
3070
+ cacheReadsPrice: 0.13,
3071
+ description: "GPT-5: The best model for coding and agentic tasks across domains",
3072
+ // supportsVerbosity is a new capability; ensure ModelInfo includes it
3073
+ supportsVerbosity: true
3074
+ },
3075
+ "gpt-5-mini-2025-08-07": {
3076
+ maxTokens: 128e3,
3077
+ contextWindow: 4e5,
3078
+ supportsImages: true,
3079
+ supportsPromptCache: true,
3080
+ supportsReasoningEffort: true,
3081
+ reasoningEffort: "medium",
3082
+ inputPrice: 0.25,
3083
+ outputPrice: 2,
3084
+ cacheReadsPrice: 0.03,
3085
+ description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
3086
+ supportsVerbosity: true
3087
+ },
3088
+ "gpt-5-nano-2025-08-07": {
3089
+ maxTokens: 128e3,
3090
+ contextWindow: 4e5,
3091
+ supportsImages: true,
3092
+ supportsPromptCache: true,
3093
+ supportsReasoningEffort: true,
3094
+ reasoningEffort: "medium",
3095
+ inputPrice: 0.05,
3096
+ outputPrice: 0.4,
3097
+ cacheReadsPrice: 0.01,
3098
+ description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
3099
+ supportsVerbosity: true
3100
+ },
2958
3101
  "gpt-4.1": {
2959
3102
  maxTokens: 32768,
2960
3103
  contextWindow: 1047576,
@@ -3140,6 +3283,7 @@ var openAiModelInfoSaneDefaults = {
3140
3283
  };
3141
3284
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3142
3285
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3286
+ var GPT5_DEFAULT_TEMPERATURE = 1;
3143
3287
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3144
3288
 
3145
3289
  // src/providers/openrouter.ts
@@ -3177,6 +3321,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3177
3321
  "anthropic/claude-3.7-sonnet:thinking",
3178
3322
  "anthropic/claude-sonnet-4",
3179
3323
  "anthropic/claude-opus-4",
3324
+ "anthropic/claude-opus-4.1",
3180
3325
  "google/gemini-2.5-flash-preview",
3181
3326
  "google/gemini-2.5-flash-preview:thinking",
3182
3327
  "google/gemini-2.5-flash-preview-05-20",
@@ -3194,7 +3339,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3194
3339
  "anthropic/claude-3.7-sonnet:beta",
3195
3340
  "anthropic/claude-3.7-sonnet:thinking",
3196
3341
  "anthropic/claude-sonnet-4",
3197
- "anthropic/claude-opus-4"
3342
+ "anthropic/claude-opus-4",
3343
+ "anthropic/claude-opus-4.1"
3198
3344
  ]);
3199
3345
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3200
3346
  "anthropic/claude-3.7-sonnet:thinking",
@@ -3204,6 +3350,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3204
3350
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3205
3351
  "anthropic/claude-3.7-sonnet:beta",
3206
3352
  "anthropic/claude-opus-4",
3353
+ "anthropic/claude-opus-4.1",
3207
3354
  "anthropic/claude-sonnet-4",
3208
3355
  "google/gemini-2.5-pro-preview",
3209
3356
  "google/gemini-2.5-pro",
@@ -3494,6 +3641,18 @@ var vertexModels = {
3494
3641
  cacheReadsPrice: 0.3,
3495
3642
  supportsReasoningBudget: true
3496
3643
  },
3644
+ "claude-opus-4-1@20250805": {
3645
+ maxTokens: 8192,
3646
+ contextWindow: 2e5,
3647
+ supportsImages: true,
3648
+ supportsComputerUse: true,
3649
+ supportsPromptCache: true,
3650
+ inputPrice: 15,
3651
+ outputPrice: 75,
3652
+ cacheWritesPrice: 18.75,
3653
+ cacheReadsPrice: 1.5,
3654
+ supportsReasoningBudget: true
3655
+ },
3497
3656
  "claude-opus-4@20250514": {
3498
3657
  maxTokens: 8192,
3499
3658
  contextWindow: 2e5,
@@ -3918,6 +4077,185 @@ var doubaoModels = {
3918
4077
  var doubaoDefaultModelInfo = doubaoModels[doubaoDefaultModelId];
3919
4078
  var DOUBAO_API_BASE_URL = "https://ark.cn-beijing.volces.com/api/v3";
3920
4079
  var DOUBAO_API_CHAT_PATH = "/chat/completions";
4080
+
4081
+ // src/providers/zai.ts
4082
+ var internationalZAiDefaultModelId = "glm-4.5";
4083
+ var internationalZAiModels = {
4084
+ "glm-4.5": {
4085
+ maxTokens: 98304,
4086
+ contextWindow: 131072,
4087
+ supportsImages: false,
4088
+ supportsPromptCache: true,
4089
+ inputPrice: 0.6,
4090
+ outputPrice: 2.2,
4091
+ cacheWritesPrice: 0,
4092
+ cacheReadsPrice: 0.11,
4093
+ description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
4094
+ },
4095
+ "glm-4.5-air": {
4096
+ maxTokens: 98304,
4097
+ contextWindow: 131072,
4098
+ supportsImages: false,
4099
+ supportsPromptCache: true,
4100
+ inputPrice: 0.2,
4101
+ outputPrice: 1.1,
4102
+ cacheWritesPrice: 0,
4103
+ cacheReadsPrice: 0.03,
4104
+ description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
4105
+ }
4106
+ };
4107
+ var mainlandZAiDefaultModelId = "glm-4.5";
4108
+ var mainlandZAiModels = {
4109
+ "glm-4.5": {
4110
+ maxTokens: 98304,
4111
+ contextWindow: 131072,
4112
+ supportsImages: false,
4113
+ supportsPromptCache: true,
4114
+ inputPrice: 0.29,
4115
+ outputPrice: 1.14,
4116
+ cacheWritesPrice: 0,
4117
+ cacheReadsPrice: 0.057,
4118
+ description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k.",
4119
+ tiers: [
4120
+ {
4121
+ contextWindow: 32e3,
4122
+ inputPrice: 0.21,
4123
+ outputPrice: 1,
4124
+ cacheReadsPrice: 0.043
4125
+ },
4126
+ {
4127
+ contextWindow: 128e3,
4128
+ inputPrice: 0.29,
4129
+ outputPrice: 1.14,
4130
+ cacheReadsPrice: 0.057
4131
+ },
4132
+ {
4133
+ contextWindow: Infinity,
4134
+ inputPrice: 0.29,
4135
+ outputPrice: 1.14,
4136
+ cacheReadsPrice: 0.057
4137
+ }
4138
+ ]
4139
+ },
4140
+ "glm-4.5-air": {
4141
+ maxTokens: 98304,
4142
+ contextWindow: 131072,
4143
+ supportsImages: false,
4144
+ supportsPromptCache: true,
4145
+ inputPrice: 0.1,
4146
+ outputPrice: 0.6,
4147
+ cacheWritesPrice: 0,
4148
+ cacheReadsPrice: 0.02,
4149
+ description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models.",
4150
+ tiers: [
4151
+ {
4152
+ contextWindow: 32e3,
4153
+ inputPrice: 0.07,
4154
+ outputPrice: 0.4,
4155
+ cacheReadsPrice: 0.014
4156
+ },
4157
+ {
4158
+ contextWindow: 128e3,
4159
+ inputPrice: 0.1,
4160
+ outputPrice: 0.6,
4161
+ cacheReadsPrice: 0.02
4162
+ },
4163
+ {
4164
+ contextWindow: Infinity,
4165
+ inputPrice: 0.1,
4166
+ outputPrice: 0.6,
4167
+ cacheReadsPrice: 0.02
4168
+ }
4169
+ ]
4170
+ }
4171
+ };
4172
+ var ZAI_DEFAULT_TEMPERATURE = 0;
4173
+
4174
+ // src/providers/fireworks.ts
4175
+ var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
4176
+ var fireworksModels = {
4177
+ "accounts/fireworks/models/kimi-k2-instruct": {
4178
+ maxTokens: 16384,
4179
+ contextWindow: 128e3,
4180
+ supportsImages: false,
4181
+ supportsPromptCache: false,
4182
+ inputPrice: 0.6,
4183
+ outputPrice: 2.5,
4184
+ description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
4185
+ },
4186
+ "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
4187
+ maxTokens: 32768,
4188
+ contextWindow: 256e3,
4189
+ supportsImages: false,
4190
+ supportsPromptCache: false,
4191
+ inputPrice: 0.22,
4192
+ outputPrice: 0.88,
4193
+ description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
4194
+ },
4195
+ "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
4196
+ maxTokens: 32768,
4197
+ contextWindow: 256e3,
4198
+ supportsImages: false,
4199
+ supportsPromptCache: false,
4200
+ inputPrice: 0.45,
4201
+ outputPrice: 1.8,
4202
+ description: "Qwen3's most agentic code model to date."
4203
+ },
4204
+ "accounts/fireworks/models/deepseek-r1-0528": {
4205
+ maxTokens: 20480,
4206
+ contextWindow: 16e4,
4207
+ supportsImages: false,
4208
+ supportsPromptCache: false,
4209
+ inputPrice: 3,
4210
+ outputPrice: 8,
4211
+ description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4212
+ },
4213
+ "accounts/fireworks/models/deepseek-v3": {
4214
+ maxTokens: 16384,
4215
+ contextWindow: 128e3,
4216
+ supportsImages: false,
4217
+ supportsPromptCache: false,
4218
+ inputPrice: 0.9,
4219
+ outputPrice: 0.9,
4220
+ description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4221
+ },
4222
+ "accounts/fireworks/models/glm-4p5": {
4223
+ maxTokens: 16384,
4224
+ contextWindow: 128e3,
4225
+ supportsImages: false,
4226
+ supportsPromptCache: false,
4227
+ inputPrice: 0.55,
4228
+ outputPrice: 2.19,
4229
+ description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4230
+ },
4231
+ "accounts/fireworks/models/glm-4p5-air": {
4232
+ maxTokens: 16384,
4233
+ contextWindow: 128e3,
4234
+ supportsImages: false,
4235
+ supportsPromptCache: false,
4236
+ inputPrice: 0.55,
4237
+ outputPrice: 2.19,
4238
+ description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4239
+ },
4240
+ "accounts/fireworks/models/gpt-oss-20b": {
4241
+ maxTokens: 16384,
4242
+ contextWindow: 128e3,
4243
+ supportsImages: false,
4244
+ supportsPromptCache: false,
4245
+ inputPrice: 0.07,
4246
+ outputPrice: 0.3,
4247
+ description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
4248
+ },
4249
+ "accounts/fireworks/models/gpt-oss-120b": {
4250
+ maxTokens: 16384,
4251
+ contextWindow: 128e3,
4252
+ supportsImages: false,
4253
+ supportsPromptCache: false,
4254
+ inputPrice: 0.15,
4255
+ outputPrice: 0.6,
4256
+ description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
4257
+ }
4258
+ };
3921
4259
  export {
3922
4260
  ANTHROPIC_DEFAULT_MAX_TOKENS,
3923
4261
  ANTHROPIC_STYLE_PROVIDERS,
@@ -3940,6 +4278,7 @@ export {
3940
4278
  GLAMA_DEFAULT_TEMPERATURE,
3941
4279
  GLOBAL_SETTINGS_KEYS,
3942
4280
  GLOBAL_STATE_KEYS,
4281
+ GPT5_DEFAULT_TEMPERATURE,
3943
4282
  HUGGINGFACE_API_URL,
3944
4283
  HUGGINGFACE_CACHE_DURATION,
3945
4284
  HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
@@ -3970,6 +4309,7 @@ export {
3970
4309
  TaskCommandName,
3971
4310
  TelemetryEventName,
3972
4311
  VERTEX_REGIONS,
4312
+ ZAI_DEFAULT_TEMPERATURE,
3973
4313
  ackSchema,
3974
4314
  anthropicDefaultModelId,
3975
4315
  anthropicModels,
@@ -4010,6 +4350,9 @@ export {
4010
4350
  experimentIds,
4011
4351
  experimentIdsSchema,
4012
4352
  experimentsSchema,
4353
+ extendedReasoningEffortsSchema,
4354
+ fireworksDefaultModelId,
4355
+ fireworksModels,
4013
4356
  followUpDataSchema,
4014
4357
  geminiDefaultModelId,
4015
4358
  geminiModels,
@@ -4026,6 +4369,8 @@ export {
4026
4369
  groupOptionsSchema,
4027
4370
  historyItemSchema,
4028
4371
  installMarketplaceItemOptionsSchema,
4372
+ internationalZAiDefaultModelId,
4373
+ internationalZAiModels,
4029
4374
  ipcMessageSchema,
4030
4375
  isBlockingAsk,
4031
4376
  isGlobalStateKey,
@@ -4038,6 +4383,8 @@ export {
4038
4383
  languagesSchema,
4039
4384
  litellmDefaultModelId,
4040
4385
  litellmDefaultModelInfo,
4386
+ mainlandZAiDefaultModelId,
4387
+ mainlandZAiModels,
4041
4388
  marketplaceItemSchema,
4042
4389
  marketplaceItemTypeSchema,
4043
4390
  mcpExecutionStatusSchema,
@@ -4100,6 +4447,8 @@ export {
4100
4447
  toolUsageSchema,
4101
4448
  unboundDefaultModelId,
4102
4449
  unboundDefaultModelInfo,
4450
+ verbosityLevels,
4451
+ verbosityLevelsSchema,
4103
4452
  vertexDefaultModelId,
4104
4453
  vertexModels,
4105
4454
  vscodeLlmDefaultModelId,