@roo-code/types 1.43.0 → 1.45.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +369 -9
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1964 -432
- package/dist/index.d.ts +1964 -432
- package/dist/index.js +358 -9
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.js
CHANGED
|
@@ -11,6 +11,8 @@ import { z as z3 } from "zod";
|
|
|
11
11
|
import { z } from "zod";
|
|
12
12
|
var reasoningEfforts = ["low", "medium", "high"];
|
|
13
13
|
var reasoningEffortsSchema = z.enum(reasoningEfforts);
|
|
14
|
+
var verbosityLevels = ["low", "medium", "high"];
|
|
15
|
+
var verbosityLevelsSchema = z.enum(verbosityLevels);
|
|
14
16
|
var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
|
|
15
17
|
var modelParametersSchema = z.enum(modelParameters);
|
|
16
18
|
var isModelParameter = (value) => modelParameters.includes(value);
|
|
@@ -21,6 +23,8 @@ var modelInfoSchema = z.object({
|
|
|
21
23
|
supportsImages: z.boolean().optional(),
|
|
22
24
|
supportsComputerUse: z.boolean().optional(),
|
|
23
25
|
supportsPromptCache: z.boolean(),
|
|
26
|
+
// Capability flag to indicate whether the model supports an output verbosity parameter
|
|
27
|
+
supportsVerbosity: z.boolean().optional(),
|
|
24
28
|
supportsReasoningBudget: z.boolean().optional(),
|
|
25
29
|
requiredReasoningBudget: z.boolean().optional(),
|
|
26
30
|
supportsReasoningEffort: z.boolean().optional(),
|
|
@@ -88,6 +92,7 @@ var codebaseIndexProviderSchema = z2.object({
|
|
|
88
92
|
});
|
|
89
93
|
|
|
90
94
|
// src/provider-settings.ts
|
|
95
|
+
var extendedReasoningEffortsSchema = z3.union([reasoningEffortsSchema, z3.literal("minimal")]);
|
|
91
96
|
var providerNames = [
|
|
92
97
|
"anthropic",
|
|
93
98
|
"claude-code",
|
|
@@ -116,7 +121,9 @@ var providerNames = [
|
|
|
116
121
|
"litellm",
|
|
117
122
|
"huggingface",
|
|
118
123
|
"cerebras",
|
|
119
|
-
"sambanova"
|
|
124
|
+
"sambanova",
|
|
125
|
+
"zai",
|
|
126
|
+
"fireworks"
|
|
120
127
|
];
|
|
121
128
|
var providerNamesSchema = z3.enum(providerNames);
|
|
122
129
|
var providerSettingsEntrySchema = z3.object({
|
|
@@ -135,9 +142,11 @@ var baseProviderSettingsSchema = z3.object({
|
|
|
135
142
|
consecutiveMistakeLimit: z3.number().min(0).optional(),
|
|
136
143
|
// Model reasoning.
|
|
137
144
|
enableReasoningEffort: z3.boolean().optional(),
|
|
138
|
-
reasoningEffort:
|
|
145
|
+
reasoningEffort: extendedReasoningEffortsSchema.optional(),
|
|
139
146
|
modelMaxTokens: z3.number().optional(),
|
|
140
|
-
modelMaxThinkingTokens: z3.number().optional()
|
|
147
|
+
modelMaxThinkingTokens: z3.number().optional(),
|
|
148
|
+
// Model verbosity.
|
|
149
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
141
150
|
});
|
|
142
151
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
143
152
|
apiModelId: z3.string().optional()
|
|
@@ -284,6 +293,13 @@ var cerebrasSchema = apiModelIdProviderModelSchema.extend({
|
|
|
284
293
|
var sambaNovaSchema = apiModelIdProviderModelSchema.extend({
|
|
285
294
|
sambaNovaApiKey: z3.string().optional()
|
|
286
295
|
});
|
|
296
|
+
var zaiSchema = apiModelIdProviderModelSchema.extend({
|
|
297
|
+
zaiApiKey: z3.string().optional(),
|
|
298
|
+
zaiApiLine: z3.union([z3.literal("china"), z3.literal("international")]).optional()
|
|
299
|
+
});
|
|
300
|
+
var fireworksSchema = apiModelIdProviderModelSchema.extend({
|
|
301
|
+
fireworksApiKey: z3.string().optional()
|
|
302
|
+
});
|
|
287
303
|
var defaultSchema = z3.object({
|
|
288
304
|
apiProvider: z3.undefined()
|
|
289
305
|
});
|
|
@@ -316,6 +332,8 @@ var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
|
|
|
316
332
|
litellmSchema.merge(z3.object({ apiProvider: z3.literal("litellm") })),
|
|
317
333
|
cerebrasSchema.merge(z3.object({ apiProvider: z3.literal("cerebras") })),
|
|
318
334
|
sambaNovaSchema.merge(z3.object({ apiProvider: z3.literal("sambanova") })),
|
|
335
|
+
zaiSchema.merge(z3.object({ apiProvider: z3.literal("zai") })),
|
|
336
|
+
fireworksSchema.merge(z3.object({ apiProvider: z3.literal("fireworks") })),
|
|
319
337
|
defaultSchema
|
|
320
338
|
]);
|
|
321
339
|
var providerSettingsSchema = z3.object({
|
|
@@ -348,6 +366,8 @@ var providerSettingsSchema = z3.object({
|
|
|
348
366
|
...litellmSchema.shape,
|
|
349
367
|
...cerebrasSchema.shape,
|
|
350
368
|
...sambaNovaSchema.shape,
|
|
369
|
+
...zaiSchema.shape,
|
|
370
|
+
...fireworksSchema.shape,
|
|
351
371
|
...codebaseIndexProviderSchema.shape
|
|
352
372
|
});
|
|
353
373
|
var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: z3.string().optional() });
|
|
@@ -402,12 +422,13 @@ var historyItemSchema = z4.object({
|
|
|
402
422
|
|
|
403
423
|
// src/experiment.ts
|
|
404
424
|
import { z as z5 } from "zod";
|
|
405
|
-
var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption"];
|
|
425
|
+
var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption", "assistantMessageParser"];
|
|
406
426
|
var experimentIdsSchema = z5.enum(experimentIds);
|
|
407
427
|
var experimentsSchema = z5.object({
|
|
408
428
|
powerSteering: z5.boolean().optional(),
|
|
409
429
|
multiFileApplyDiff: z5.boolean().optional(),
|
|
410
|
-
preventFocusDisruption: z5.boolean().optional()
|
|
430
|
+
preventFocusDisruption: z5.boolean().optional(),
|
|
431
|
+
assistantMessageParser: z5.boolean().optional()
|
|
411
432
|
});
|
|
412
433
|
|
|
413
434
|
// src/telemetry.ts
|
|
@@ -494,7 +515,14 @@ var clineMessageSchema = z6.object({
|
|
|
494
515
|
progressStatus: toolProgressStatusSchema.optional(),
|
|
495
516
|
contextCondense: contextCondenseSchema.optional(),
|
|
496
517
|
isProtected: z6.boolean().optional(),
|
|
497
|
-
apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional()
|
|
518
|
+
apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional(),
|
|
519
|
+
metadata: z6.object({
|
|
520
|
+
gpt5: z6.object({
|
|
521
|
+
previous_response_id: z6.string().optional(),
|
|
522
|
+
instructions: z6.string().optional(),
|
|
523
|
+
reasoning_summary: z6.string().optional()
|
|
524
|
+
}).optional()
|
|
525
|
+
}).optional()
|
|
498
526
|
});
|
|
499
527
|
var tokenUsageSchema = z6.object({
|
|
500
528
|
totalTokensIn: z6.number(),
|
|
@@ -934,6 +962,7 @@ var globalSettingsSchema = z11.object({
|
|
|
934
962
|
telemetrySetting: telemetrySettingsSchema.optional(),
|
|
935
963
|
mcpEnabled: z11.boolean().optional(),
|
|
936
964
|
enableMcpServerCreation: z11.boolean().optional(),
|
|
965
|
+
remoteControlEnabled: z11.boolean().optional(),
|
|
937
966
|
mode: z11.string().optional(),
|
|
938
967
|
modeApiConfigs: z11.record(z11.string(), z11.string()).optional(),
|
|
939
968
|
customModes: z11.array(modeConfigSchema).optional(),
|
|
@@ -976,7 +1005,8 @@ var SECRET_STATE_KEYS = [
|
|
|
976
1005
|
"codebaseIndexGeminiApiKey",
|
|
977
1006
|
"codebaseIndexMistralApiKey",
|
|
978
1007
|
"huggingFaceApiKey",
|
|
979
|
-
"sambaNovaApiKey"
|
|
1008
|
+
"sambaNovaApiKey",
|
|
1009
|
+
"fireworksApiKey"
|
|
980
1010
|
];
|
|
981
1011
|
var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
|
|
982
1012
|
var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
|
|
@@ -1043,6 +1073,7 @@ var EVALS_SETTINGS = {
|
|
|
1043
1073
|
language: "en",
|
|
1044
1074
|
telemetrySetting: "enabled",
|
|
1045
1075
|
mcpEnabled: false,
|
|
1076
|
+
remoteControlEnabled: false,
|
|
1046
1077
|
mode: "code",
|
|
1047
1078
|
// "architect",
|
|
1048
1079
|
customModes: []
|
|
@@ -1484,6 +1515,22 @@ var anthropicModels = {
|
|
|
1484
1515
|
// $0.30 per million tokens
|
|
1485
1516
|
supportsReasoningBudget: true
|
|
1486
1517
|
},
|
|
1518
|
+
"claude-opus-4-1-20250805": {
|
|
1519
|
+
maxTokens: 8192,
|
|
1520
|
+
contextWindow: 2e5,
|
|
1521
|
+
supportsImages: true,
|
|
1522
|
+
supportsComputerUse: true,
|
|
1523
|
+
supportsPromptCache: true,
|
|
1524
|
+
inputPrice: 15,
|
|
1525
|
+
// $15 per million input tokens
|
|
1526
|
+
outputPrice: 75,
|
|
1527
|
+
// $75 per million output tokens
|
|
1528
|
+
cacheWritesPrice: 18.75,
|
|
1529
|
+
// $18.75 per million tokens
|
|
1530
|
+
cacheReadsPrice: 1.5,
|
|
1531
|
+
// $1.50 per million tokens
|
|
1532
|
+
supportsReasoningBudget: true
|
|
1533
|
+
},
|
|
1487
1534
|
"claude-opus-4-20250514": {
|
|
1488
1535
|
maxTokens: 32e3,
|
|
1489
1536
|
// Overridden to 8k if `enableReasoningEffort` is false.
|
|
@@ -1664,6 +1711,21 @@ var bedrockModels = {
|
|
|
1664
1711
|
maxCachePoints: 4,
|
|
1665
1712
|
cachableFields: ["system", "messages", "tools"]
|
|
1666
1713
|
},
|
|
1714
|
+
"anthropic.claude-opus-4-1-20250805-v1:0": {
|
|
1715
|
+
maxTokens: 8192,
|
|
1716
|
+
contextWindow: 2e5,
|
|
1717
|
+
supportsImages: true,
|
|
1718
|
+
supportsComputerUse: true,
|
|
1719
|
+
supportsPromptCache: true,
|
|
1720
|
+
supportsReasoningBudget: true,
|
|
1721
|
+
inputPrice: 15,
|
|
1722
|
+
outputPrice: 75,
|
|
1723
|
+
cacheWritesPrice: 18.75,
|
|
1724
|
+
cacheReadsPrice: 1.5,
|
|
1725
|
+
minTokensPerCachePoint: 1024,
|
|
1726
|
+
maxCachePoints: 4,
|
|
1727
|
+
cachableFields: ["system", "messages", "tools"]
|
|
1728
|
+
},
|
|
1667
1729
|
"anthropic.claude-opus-4-20250514-v1:0": {
|
|
1668
1730
|
maxTokens: 8192,
|
|
1669
1731
|
contextWindow: 2e5,
|
|
@@ -2036,6 +2098,15 @@ var cerebrasModels = {
|
|
|
2036
2098
|
outputPrice: 0,
|
|
2037
2099
|
description: "SOTA performance with ~1500 tokens/s",
|
|
2038
2100
|
supportsReasoningEffort: true
|
|
2101
|
+
},
|
|
2102
|
+
"gpt-oss-120b": {
|
|
2103
|
+
maxTokens: 8e3,
|
|
2104
|
+
contextWindow: 64e3,
|
|
2105
|
+
supportsImages: false,
|
|
2106
|
+
supportsPromptCache: false,
|
|
2107
|
+
inputPrice: 0,
|
|
2108
|
+
outputPrice: 0,
|
|
2109
|
+
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
2039
2110
|
}
|
|
2040
2111
|
};
|
|
2041
2112
|
|
|
@@ -2291,6 +2362,15 @@ var claudeCodeModels = {
|
|
|
2291
2362
|
supportsReasoningBudget: false,
|
|
2292
2363
|
requiredReasoningBudget: false
|
|
2293
2364
|
},
|
|
2365
|
+
"claude-opus-4-1-20250805": {
|
|
2366
|
+
...anthropicModels["claude-opus-4-1-20250805"],
|
|
2367
|
+
supportsImages: false,
|
|
2368
|
+
supportsPromptCache: true,
|
|
2369
|
+
// Claude Code does report cache tokens
|
|
2370
|
+
supportsReasoningEffort: false,
|
|
2371
|
+
supportsReasoningBudget: false,
|
|
2372
|
+
requiredReasoningBudget: false
|
|
2373
|
+
},
|
|
2294
2374
|
"claude-opus-4-20250514": {
|
|
2295
2375
|
...anthropicModels["claude-opus-4-20250514"],
|
|
2296
2376
|
supportsImages: false,
|
|
@@ -2764,6 +2844,24 @@ var groqModels = {
|
|
|
2764
2844
|
inputPrice: 1,
|
|
2765
2845
|
outputPrice: 3,
|
|
2766
2846
|
description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
|
|
2847
|
+
},
|
|
2848
|
+
"openai/gpt-oss-120b": {
|
|
2849
|
+
maxTokens: 32766,
|
|
2850
|
+
contextWindow: 131072,
|
|
2851
|
+
supportsImages: false,
|
|
2852
|
+
supportsPromptCache: false,
|
|
2853
|
+
inputPrice: 0.15,
|
|
2854
|
+
outputPrice: 0.75,
|
|
2855
|
+
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
2856
|
+
},
|
|
2857
|
+
"openai/gpt-oss-20b": {
|
|
2858
|
+
maxTokens: 32768,
|
|
2859
|
+
contextWindow: 131072,
|
|
2860
|
+
supportsImages: false,
|
|
2861
|
+
supportsPromptCache: false,
|
|
2862
|
+
inputPrice: 0.1,
|
|
2863
|
+
outputPrice: 0.5,
|
|
2864
|
+
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
2767
2865
|
}
|
|
2768
2866
|
};
|
|
2769
2867
|
|
|
@@ -2792,6 +2890,7 @@ var litellmDefaultModelInfo = {
|
|
|
2792
2890
|
};
|
|
2793
2891
|
var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
2794
2892
|
"claude-3-5-sonnet-latest",
|
|
2893
|
+
"claude-opus-4-1-20250805",
|
|
2795
2894
|
"claude-opus-4-20250514",
|
|
2796
2895
|
"claude-sonnet-4-20250514",
|
|
2797
2896
|
"claude-3-7-sonnet-latest",
|
|
@@ -2801,22 +2900,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
2801
2900
|
"vertex_ai/claude-3-5-sonnet-v2",
|
|
2802
2901
|
"vertex_ai/claude-3-5-sonnet-v2@20241022",
|
|
2803
2902
|
"vertex_ai/claude-3-7-sonnet@20250219",
|
|
2903
|
+
"vertex_ai/claude-opus-4-1@20250805",
|
|
2804
2904
|
"vertex_ai/claude-opus-4@20250514",
|
|
2805
2905
|
"vertex_ai/claude-sonnet-4@20250514",
|
|
2806
2906
|
"openrouter/anthropic/claude-3.5-sonnet",
|
|
2807
2907
|
"openrouter/anthropic/claude-3.5-sonnet:beta",
|
|
2808
2908
|
"openrouter/anthropic/claude-3.7-sonnet",
|
|
2809
2909
|
"openrouter/anthropic/claude-3.7-sonnet:beta",
|
|
2910
|
+
"anthropic.claude-opus-4-1-20250805-v1:0",
|
|
2810
2911
|
"anthropic.claude-opus-4-20250514-v1:0",
|
|
2811
2912
|
"anthropic.claude-sonnet-4-20250514-v1:0",
|
|
2812
2913
|
"anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
2813
2914
|
"anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
2814
2915
|
"us.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
2815
2916
|
"us.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
2917
|
+
"us.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
2816
2918
|
"us.anthropic.claude-opus-4-20250514-v1:0",
|
|
2817
2919
|
"us.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
2818
2920
|
"eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
2819
2921
|
"eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
2922
|
+
"eu.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
2820
2923
|
"eu.anthropic.claude-opus-4-20250514-v1:0",
|
|
2821
2924
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
2822
2925
|
"snowflake/claude-3-5-sonnet"
|
|
@@ -2953,8 +3056,48 @@ var ollamaDefaultModelInfo = {
|
|
|
2953
3056
|
};
|
|
2954
3057
|
|
|
2955
3058
|
// src/providers/openai.ts
|
|
2956
|
-
var openAiNativeDefaultModelId = "gpt-
|
|
3059
|
+
var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
|
|
2957
3060
|
var openAiNativeModels = {
|
|
3061
|
+
"gpt-5-2025-08-07": {
|
|
3062
|
+
maxTokens: 128e3,
|
|
3063
|
+
contextWindow: 4e5,
|
|
3064
|
+
supportsImages: true,
|
|
3065
|
+
supportsPromptCache: true,
|
|
3066
|
+
supportsReasoningEffort: true,
|
|
3067
|
+
reasoningEffort: "medium",
|
|
3068
|
+
inputPrice: 1.25,
|
|
3069
|
+
outputPrice: 10,
|
|
3070
|
+
cacheReadsPrice: 0.13,
|
|
3071
|
+
description: "GPT-5: The best model for coding and agentic tasks across domains",
|
|
3072
|
+
// supportsVerbosity is a new capability; ensure ModelInfo includes it
|
|
3073
|
+
supportsVerbosity: true
|
|
3074
|
+
},
|
|
3075
|
+
"gpt-5-mini-2025-08-07": {
|
|
3076
|
+
maxTokens: 128e3,
|
|
3077
|
+
contextWindow: 4e5,
|
|
3078
|
+
supportsImages: true,
|
|
3079
|
+
supportsPromptCache: true,
|
|
3080
|
+
supportsReasoningEffort: true,
|
|
3081
|
+
reasoningEffort: "medium",
|
|
3082
|
+
inputPrice: 0.25,
|
|
3083
|
+
outputPrice: 2,
|
|
3084
|
+
cacheReadsPrice: 0.03,
|
|
3085
|
+
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
|
|
3086
|
+
supportsVerbosity: true
|
|
3087
|
+
},
|
|
3088
|
+
"gpt-5-nano-2025-08-07": {
|
|
3089
|
+
maxTokens: 128e3,
|
|
3090
|
+
contextWindow: 4e5,
|
|
3091
|
+
supportsImages: true,
|
|
3092
|
+
supportsPromptCache: true,
|
|
3093
|
+
supportsReasoningEffort: true,
|
|
3094
|
+
reasoningEffort: "medium",
|
|
3095
|
+
inputPrice: 0.05,
|
|
3096
|
+
outputPrice: 0.4,
|
|
3097
|
+
cacheReadsPrice: 0.01,
|
|
3098
|
+
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
|
|
3099
|
+
supportsVerbosity: true
|
|
3100
|
+
},
|
|
2958
3101
|
"gpt-4.1": {
|
|
2959
3102
|
maxTokens: 32768,
|
|
2960
3103
|
contextWindow: 1047576,
|
|
@@ -3140,6 +3283,7 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3140
3283
|
};
|
|
3141
3284
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3142
3285
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
3286
|
+
var GPT5_DEFAULT_TEMPERATURE = 1;
|
|
3143
3287
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
3144
3288
|
|
|
3145
3289
|
// src/providers/openrouter.ts
|
|
@@ -3177,6 +3321,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
|
|
|
3177
3321
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3178
3322
|
"anthropic/claude-sonnet-4",
|
|
3179
3323
|
"anthropic/claude-opus-4",
|
|
3324
|
+
"anthropic/claude-opus-4.1",
|
|
3180
3325
|
"google/gemini-2.5-flash-preview",
|
|
3181
3326
|
"google/gemini-2.5-flash-preview:thinking",
|
|
3182
3327
|
"google/gemini-2.5-flash-preview-05-20",
|
|
@@ -3194,7 +3339,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
3194
3339
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3195
3340
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3196
3341
|
"anthropic/claude-sonnet-4",
|
|
3197
|
-
"anthropic/claude-opus-4"
|
|
3342
|
+
"anthropic/claude-opus-4",
|
|
3343
|
+
"anthropic/claude-opus-4.1"
|
|
3198
3344
|
]);
|
|
3199
3345
|
var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3200
3346
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
@@ -3204,6 +3350,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
|
3204
3350
|
var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3205
3351
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3206
3352
|
"anthropic/claude-opus-4",
|
|
3353
|
+
"anthropic/claude-opus-4.1",
|
|
3207
3354
|
"anthropic/claude-sonnet-4",
|
|
3208
3355
|
"google/gemini-2.5-pro-preview",
|
|
3209
3356
|
"google/gemini-2.5-pro",
|
|
@@ -3494,6 +3641,18 @@ var vertexModels = {
|
|
|
3494
3641
|
cacheReadsPrice: 0.3,
|
|
3495
3642
|
supportsReasoningBudget: true
|
|
3496
3643
|
},
|
|
3644
|
+
"claude-opus-4-1@20250805": {
|
|
3645
|
+
maxTokens: 8192,
|
|
3646
|
+
contextWindow: 2e5,
|
|
3647
|
+
supportsImages: true,
|
|
3648
|
+
supportsComputerUse: true,
|
|
3649
|
+
supportsPromptCache: true,
|
|
3650
|
+
inputPrice: 15,
|
|
3651
|
+
outputPrice: 75,
|
|
3652
|
+
cacheWritesPrice: 18.75,
|
|
3653
|
+
cacheReadsPrice: 1.5,
|
|
3654
|
+
supportsReasoningBudget: true
|
|
3655
|
+
},
|
|
3497
3656
|
"claude-opus-4@20250514": {
|
|
3498
3657
|
maxTokens: 8192,
|
|
3499
3658
|
contextWindow: 2e5,
|
|
@@ -3918,6 +4077,185 @@ var doubaoModels = {
|
|
|
3918
4077
|
var doubaoDefaultModelInfo = doubaoModels[doubaoDefaultModelId];
|
|
3919
4078
|
var DOUBAO_API_BASE_URL = "https://ark.cn-beijing.volces.com/api/v3";
|
|
3920
4079
|
var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
4080
|
+
|
|
4081
|
+
// src/providers/zai.ts
|
|
4082
|
+
var internationalZAiDefaultModelId = "glm-4.5";
|
|
4083
|
+
var internationalZAiModels = {
|
|
4084
|
+
"glm-4.5": {
|
|
4085
|
+
maxTokens: 98304,
|
|
4086
|
+
contextWindow: 131072,
|
|
4087
|
+
supportsImages: false,
|
|
4088
|
+
supportsPromptCache: true,
|
|
4089
|
+
inputPrice: 0.6,
|
|
4090
|
+
outputPrice: 2.2,
|
|
4091
|
+
cacheWritesPrice: 0,
|
|
4092
|
+
cacheReadsPrice: 0.11,
|
|
4093
|
+
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
|
|
4094
|
+
},
|
|
4095
|
+
"glm-4.5-air": {
|
|
4096
|
+
maxTokens: 98304,
|
|
4097
|
+
contextWindow: 131072,
|
|
4098
|
+
supportsImages: false,
|
|
4099
|
+
supportsPromptCache: true,
|
|
4100
|
+
inputPrice: 0.2,
|
|
4101
|
+
outputPrice: 1.1,
|
|
4102
|
+
cacheWritesPrice: 0,
|
|
4103
|
+
cacheReadsPrice: 0.03,
|
|
4104
|
+
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
|
|
4105
|
+
}
|
|
4106
|
+
};
|
|
4107
|
+
var mainlandZAiDefaultModelId = "glm-4.5";
|
|
4108
|
+
var mainlandZAiModels = {
|
|
4109
|
+
"glm-4.5": {
|
|
4110
|
+
maxTokens: 98304,
|
|
4111
|
+
contextWindow: 131072,
|
|
4112
|
+
supportsImages: false,
|
|
4113
|
+
supportsPromptCache: true,
|
|
4114
|
+
inputPrice: 0.29,
|
|
4115
|
+
outputPrice: 1.14,
|
|
4116
|
+
cacheWritesPrice: 0,
|
|
4117
|
+
cacheReadsPrice: 0.057,
|
|
4118
|
+
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k.",
|
|
4119
|
+
tiers: [
|
|
4120
|
+
{
|
|
4121
|
+
contextWindow: 32e3,
|
|
4122
|
+
inputPrice: 0.21,
|
|
4123
|
+
outputPrice: 1,
|
|
4124
|
+
cacheReadsPrice: 0.043
|
|
4125
|
+
},
|
|
4126
|
+
{
|
|
4127
|
+
contextWindow: 128e3,
|
|
4128
|
+
inputPrice: 0.29,
|
|
4129
|
+
outputPrice: 1.14,
|
|
4130
|
+
cacheReadsPrice: 0.057
|
|
4131
|
+
},
|
|
4132
|
+
{
|
|
4133
|
+
contextWindow: Infinity,
|
|
4134
|
+
inputPrice: 0.29,
|
|
4135
|
+
outputPrice: 1.14,
|
|
4136
|
+
cacheReadsPrice: 0.057
|
|
4137
|
+
}
|
|
4138
|
+
]
|
|
4139
|
+
},
|
|
4140
|
+
"glm-4.5-air": {
|
|
4141
|
+
maxTokens: 98304,
|
|
4142
|
+
contextWindow: 131072,
|
|
4143
|
+
supportsImages: false,
|
|
4144
|
+
supportsPromptCache: true,
|
|
4145
|
+
inputPrice: 0.1,
|
|
4146
|
+
outputPrice: 0.6,
|
|
4147
|
+
cacheWritesPrice: 0,
|
|
4148
|
+
cacheReadsPrice: 0.02,
|
|
4149
|
+
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models.",
|
|
4150
|
+
tiers: [
|
|
4151
|
+
{
|
|
4152
|
+
contextWindow: 32e3,
|
|
4153
|
+
inputPrice: 0.07,
|
|
4154
|
+
outputPrice: 0.4,
|
|
4155
|
+
cacheReadsPrice: 0.014
|
|
4156
|
+
},
|
|
4157
|
+
{
|
|
4158
|
+
contextWindow: 128e3,
|
|
4159
|
+
inputPrice: 0.1,
|
|
4160
|
+
outputPrice: 0.6,
|
|
4161
|
+
cacheReadsPrice: 0.02
|
|
4162
|
+
},
|
|
4163
|
+
{
|
|
4164
|
+
contextWindow: Infinity,
|
|
4165
|
+
inputPrice: 0.1,
|
|
4166
|
+
outputPrice: 0.6,
|
|
4167
|
+
cacheReadsPrice: 0.02
|
|
4168
|
+
}
|
|
4169
|
+
]
|
|
4170
|
+
}
|
|
4171
|
+
};
|
|
4172
|
+
var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
4173
|
+
|
|
4174
|
+
// src/providers/fireworks.ts
|
|
4175
|
+
var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
|
|
4176
|
+
var fireworksModels = {
|
|
4177
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
4178
|
+
maxTokens: 16384,
|
|
4179
|
+
contextWindow: 128e3,
|
|
4180
|
+
supportsImages: false,
|
|
4181
|
+
supportsPromptCache: false,
|
|
4182
|
+
inputPrice: 0.6,
|
|
4183
|
+
outputPrice: 2.5,
|
|
4184
|
+
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
4185
|
+
},
|
|
4186
|
+
"accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
|
|
4187
|
+
maxTokens: 32768,
|
|
4188
|
+
contextWindow: 256e3,
|
|
4189
|
+
supportsImages: false,
|
|
4190
|
+
supportsPromptCache: false,
|
|
4191
|
+
inputPrice: 0.22,
|
|
4192
|
+
outputPrice: 0.88,
|
|
4193
|
+
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
4194
|
+
},
|
|
4195
|
+
"accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
|
|
4196
|
+
maxTokens: 32768,
|
|
4197
|
+
contextWindow: 256e3,
|
|
4198
|
+
supportsImages: false,
|
|
4199
|
+
supportsPromptCache: false,
|
|
4200
|
+
inputPrice: 0.45,
|
|
4201
|
+
outputPrice: 1.8,
|
|
4202
|
+
description: "Qwen3's most agentic code model to date."
|
|
4203
|
+
},
|
|
4204
|
+
"accounts/fireworks/models/deepseek-r1-0528": {
|
|
4205
|
+
maxTokens: 20480,
|
|
4206
|
+
contextWindow: 16e4,
|
|
4207
|
+
supportsImages: false,
|
|
4208
|
+
supportsPromptCache: false,
|
|
4209
|
+
inputPrice: 3,
|
|
4210
|
+
outputPrice: 8,
|
|
4211
|
+
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4212
|
+
},
|
|
4213
|
+
"accounts/fireworks/models/deepseek-v3": {
|
|
4214
|
+
maxTokens: 16384,
|
|
4215
|
+
contextWindow: 128e3,
|
|
4216
|
+
supportsImages: false,
|
|
4217
|
+
supportsPromptCache: false,
|
|
4218
|
+
inputPrice: 0.9,
|
|
4219
|
+
outputPrice: 0.9,
|
|
4220
|
+
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4221
|
+
},
|
|
4222
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
4223
|
+
maxTokens: 16384,
|
|
4224
|
+
contextWindow: 128e3,
|
|
4225
|
+
supportsImages: false,
|
|
4226
|
+
supportsPromptCache: false,
|
|
4227
|
+
inputPrice: 0.55,
|
|
4228
|
+
outputPrice: 2.19,
|
|
4229
|
+
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4230
|
+
},
|
|
4231
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
4232
|
+
maxTokens: 16384,
|
|
4233
|
+
contextWindow: 128e3,
|
|
4234
|
+
supportsImages: false,
|
|
4235
|
+
supportsPromptCache: false,
|
|
4236
|
+
inputPrice: 0.55,
|
|
4237
|
+
outputPrice: 2.19,
|
|
4238
|
+
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4239
|
+
},
|
|
4240
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
4241
|
+
maxTokens: 16384,
|
|
4242
|
+
contextWindow: 128e3,
|
|
4243
|
+
supportsImages: false,
|
|
4244
|
+
supportsPromptCache: false,
|
|
4245
|
+
inputPrice: 0.07,
|
|
4246
|
+
outputPrice: 0.3,
|
|
4247
|
+
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
4248
|
+
},
|
|
4249
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
4250
|
+
maxTokens: 16384,
|
|
4251
|
+
contextWindow: 128e3,
|
|
4252
|
+
supportsImages: false,
|
|
4253
|
+
supportsPromptCache: false,
|
|
4254
|
+
inputPrice: 0.15,
|
|
4255
|
+
outputPrice: 0.6,
|
|
4256
|
+
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
4257
|
+
}
|
|
4258
|
+
};
|
|
3921
4259
|
export {
|
|
3922
4260
|
ANTHROPIC_DEFAULT_MAX_TOKENS,
|
|
3923
4261
|
ANTHROPIC_STYLE_PROVIDERS,
|
|
@@ -3940,6 +4278,7 @@ export {
|
|
|
3940
4278
|
GLAMA_DEFAULT_TEMPERATURE,
|
|
3941
4279
|
GLOBAL_SETTINGS_KEYS,
|
|
3942
4280
|
GLOBAL_STATE_KEYS,
|
|
4281
|
+
GPT5_DEFAULT_TEMPERATURE,
|
|
3943
4282
|
HUGGINGFACE_API_URL,
|
|
3944
4283
|
HUGGINGFACE_CACHE_DURATION,
|
|
3945
4284
|
HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
|
|
@@ -3970,6 +4309,7 @@ export {
|
|
|
3970
4309
|
TaskCommandName,
|
|
3971
4310
|
TelemetryEventName,
|
|
3972
4311
|
VERTEX_REGIONS,
|
|
4312
|
+
ZAI_DEFAULT_TEMPERATURE,
|
|
3973
4313
|
ackSchema,
|
|
3974
4314
|
anthropicDefaultModelId,
|
|
3975
4315
|
anthropicModels,
|
|
@@ -4010,6 +4350,9 @@ export {
|
|
|
4010
4350
|
experimentIds,
|
|
4011
4351
|
experimentIdsSchema,
|
|
4012
4352
|
experimentsSchema,
|
|
4353
|
+
extendedReasoningEffortsSchema,
|
|
4354
|
+
fireworksDefaultModelId,
|
|
4355
|
+
fireworksModels,
|
|
4013
4356
|
followUpDataSchema,
|
|
4014
4357
|
geminiDefaultModelId,
|
|
4015
4358
|
geminiModels,
|
|
@@ -4026,6 +4369,8 @@ export {
|
|
|
4026
4369
|
groupOptionsSchema,
|
|
4027
4370
|
historyItemSchema,
|
|
4028
4371
|
installMarketplaceItemOptionsSchema,
|
|
4372
|
+
internationalZAiDefaultModelId,
|
|
4373
|
+
internationalZAiModels,
|
|
4029
4374
|
ipcMessageSchema,
|
|
4030
4375
|
isBlockingAsk,
|
|
4031
4376
|
isGlobalStateKey,
|
|
@@ -4038,6 +4383,8 @@ export {
|
|
|
4038
4383
|
languagesSchema,
|
|
4039
4384
|
litellmDefaultModelId,
|
|
4040
4385
|
litellmDefaultModelInfo,
|
|
4386
|
+
mainlandZAiDefaultModelId,
|
|
4387
|
+
mainlandZAiModels,
|
|
4041
4388
|
marketplaceItemSchema,
|
|
4042
4389
|
marketplaceItemTypeSchema,
|
|
4043
4390
|
mcpExecutionStatusSchema,
|
|
@@ -4100,6 +4447,8 @@ export {
|
|
|
4100
4447
|
toolUsageSchema,
|
|
4101
4448
|
unboundDefaultModelId,
|
|
4102
4449
|
unboundDefaultModelInfo,
|
|
4450
|
+
verbosityLevels,
|
|
4451
|
+
verbosityLevelsSchema,
|
|
4103
4452
|
vertexDefaultModelId,
|
|
4104
4453
|
vertexModels,
|
|
4105
4454
|
vscodeLlmDefaultModelId,
|