@roo-code/types 1.43.0 → 1.45.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +369 -9
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1964 -432
- package/dist/index.d.ts +1964 -432
- package/dist/index.js +358 -9
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.cjs
CHANGED
|
@@ -41,6 +41,7 @@ __export(index_exports, {
|
|
|
41
41
|
GLAMA_DEFAULT_TEMPERATURE: () => GLAMA_DEFAULT_TEMPERATURE,
|
|
42
42
|
GLOBAL_SETTINGS_KEYS: () => GLOBAL_SETTINGS_KEYS,
|
|
43
43
|
GLOBAL_STATE_KEYS: () => GLOBAL_STATE_KEYS,
|
|
44
|
+
GPT5_DEFAULT_TEMPERATURE: () => GPT5_DEFAULT_TEMPERATURE,
|
|
44
45
|
HUGGINGFACE_API_URL: () => HUGGINGFACE_API_URL,
|
|
45
46
|
HUGGINGFACE_CACHE_DURATION: () => HUGGINGFACE_CACHE_DURATION,
|
|
46
47
|
HUGGINGFACE_DEFAULT_CONTEXT_WINDOW: () => HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
|
|
@@ -71,6 +72,7 @@ __export(index_exports, {
|
|
|
71
72
|
TaskCommandName: () => TaskCommandName,
|
|
72
73
|
TelemetryEventName: () => TelemetryEventName,
|
|
73
74
|
VERTEX_REGIONS: () => VERTEX_REGIONS,
|
|
75
|
+
ZAI_DEFAULT_TEMPERATURE: () => ZAI_DEFAULT_TEMPERATURE,
|
|
74
76
|
ackSchema: () => ackSchema,
|
|
75
77
|
anthropicDefaultModelId: () => anthropicDefaultModelId,
|
|
76
78
|
anthropicModels: () => anthropicModels,
|
|
@@ -111,6 +113,9 @@ __export(index_exports, {
|
|
|
111
113
|
experimentIds: () => experimentIds,
|
|
112
114
|
experimentIdsSchema: () => experimentIdsSchema,
|
|
113
115
|
experimentsSchema: () => experimentsSchema,
|
|
116
|
+
extendedReasoningEffortsSchema: () => extendedReasoningEffortsSchema,
|
|
117
|
+
fireworksDefaultModelId: () => fireworksDefaultModelId,
|
|
118
|
+
fireworksModels: () => fireworksModels,
|
|
114
119
|
followUpDataSchema: () => followUpDataSchema,
|
|
115
120
|
geminiDefaultModelId: () => geminiDefaultModelId,
|
|
116
121
|
geminiModels: () => geminiModels,
|
|
@@ -127,6 +132,8 @@ __export(index_exports, {
|
|
|
127
132
|
groupOptionsSchema: () => groupOptionsSchema,
|
|
128
133
|
historyItemSchema: () => historyItemSchema,
|
|
129
134
|
installMarketplaceItemOptionsSchema: () => installMarketplaceItemOptionsSchema,
|
|
135
|
+
internationalZAiDefaultModelId: () => internationalZAiDefaultModelId,
|
|
136
|
+
internationalZAiModels: () => internationalZAiModels,
|
|
130
137
|
ipcMessageSchema: () => ipcMessageSchema,
|
|
131
138
|
isBlockingAsk: () => isBlockingAsk,
|
|
132
139
|
isGlobalStateKey: () => isGlobalStateKey,
|
|
@@ -139,6 +146,8 @@ __export(index_exports, {
|
|
|
139
146
|
languagesSchema: () => languagesSchema,
|
|
140
147
|
litellmDefaultModelId: () => litellmDefaultModelId,
|
|
141
148
|
litellmDefaultModelInfo: () => litellmDefaultModelInfo,
|
|
149
|
+
mainlandZAiDefaultModelId: () => mainlandZAiDefaultModelId,
|
|
150
|
+
mainlandZAiModels: () => mainlandZAiModels,
|
|
142
151
|
marketplaceItemSchema: () => marketplaceItemSchema,
|
|
143
152
|
marketplaceItemTypeSchema: () => marketplaceItemTypeSchema,
|
|
144
153
|
mcpExecutionStatusSchema: () => mcpExecutionStatusSchema,
|
|
@@ -201,6 +210,8 @@ __export(index_exports, {
|
|
|
201
210
|
toolUsageSchema: () => toolUsageSchema,
|
|
202
211
|
unboundDefaultModelId: () => unboundDefaultModelId,
|
|
203
212
|
unboundDefaultModelInfo: () => unboundDefaultModelInfo,
|
|
213
|
+
verbosityLevels: () => verbosityLevels,
|
|
214
|
+
verbosityLevelsSchema: () => verbosityLevelsSchema,
|
|
204
215
|
vertexDefaultModelId: () => vertexDefaultModelId,
|
|
205
216
|
vertexModels: () => vertexModels,
|
|
206
217
|
vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
|
|
@@ -223,6 +234,8 @@ var import_zod3 = require("zod");
|
|
|
223
234
|
var import_zod = require("zod");
|
|
224
235
|
var reasoningEfforts = ["low", "medium", "high"];
|
|
225
236
|
var reasoningEffortsSchema = import_zod.z.enum(reasoningEfforts);
|
|
237
|
+
var verbosityLevels = ["low", "medium", "high"];
|
|
238
|
+
var verbosityLevelsSchema = import_zod.z.enum(verbosityLevels);
|
|
226
239
|
var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
|
|
227
240
|
var modelParametersSchema = import_zod.z.enum(modelParameters);
|
|
228
241
|
var isModelParameter = (value) => modelParameters.includes(value);
|
|
@@ -233,6 +246,8 @@ var modelInfoSchema = import_zod.z.object({
|
|
|
233
246
|
supportsImages: import_zod.z.boolean().optional(),
|
|
234
247
|
supportsComputerUse: import_zod.z.boolean().optional(),
|
|
235
248
|
supportsPromptCache: import_zod.z.boolean(),
|
|
249
|
+
// Capability flag to indicate whether the model supports an output verbosity parameter
|
|
250
|
+
supportsVerbosity: import_zod.z.boolean().optional(),
|
|
236
251
|
supportsReasoningBudget: import_zod.z.boolean().optional(),
|
|
237
252
|
requiredReasoningBudget: import_zod.z.boolean().optional(),
|
|
238
253
|
supportsReasoningEffort: import_zod.z.boolean().optional(),
|
|
@@ -300,6 +315,7 @@ var codebaseIndexProviderSchema = import_zod2.z.object({
|
|
|
300
315
|
});
|
|
301
316
|
|
|
302
317
|
// src/provider-settings.ts
|
|
318
|
+
var extendedReasoningEffortsSchema = import_zod3.z.union([reasoningEffortsSchema, import_zod3.z.literal("minimal")]);
|
|
303
319
|
var providerNames = [
|
|
304
320
|
"anthropic",
|
|
305
321
|
"claude-code",
|
|
@@ -328,7 +344,9 @@ var providerNames = [
|
|
|
328
344
|
"litellm",
|
|
329
345
|
"huggingface",
|
|
330
346
|
"cerebras",
|
|
331
|
-
"sambanova"
|
|
347
|
+
"sambanova",
|
|
348
|
+
"zai",
|
|
349
|
+
"fireworks"
|
|
332
350
|
];
|
|
333
351
|
var providerNamesSchema = import_zod3.z.enum(providerNames);
|
|
334
352
|
var providerSettingsEntrySchema = import_zod3.z.object({
|
|
@@ -347,9 +365,11 @@ var baseProviderSettingsSchema = import_zod3.z.object({
|
|
|
347
365
|
consecutiveMistakeLimit: import_zod3.z.number().min(0).optional(),
|
|
348
366
|
// Model reasoning.
|
|
349
367
|
enableReasoningEffort: import_zod3.z.boolean().optional(),
|
|
350
|
-
reasoningEffort:
|
|
368
|
+
reasoningEffort: extendedReasoningEffortsSchema.optional(),
|
|
351
369
|
modelMaxTokens: import_zod3.z.number().optional(),
|
|
352
|
-
modelMaxThinkingTokens: import_zod3.z.number().optional()
|
|
370
|
+
modelMaxThinkingTokens: import_zod3.z.number().optional(),
|
|
371
|
+
// Model verbosity.
|
|
372
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
353
373
|
});
|
|
354
374
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
355
375
|
apiModelId: import_zod3.z.string().optional()
|
|
@@ -496,6 +516,13 @@ var cerebrasSchema = apiModelIdProviderModelSchema.extend({
|
|
|
496
516
|
var sambaNovaSchema = apiModelIdProviderModelSchema.extend({
|
|
497
517
|
sambaNovaApiKey: import_zod3.z.string().optional()
|
|
498
518
|
});
|
|
519
|
+
var zaiSchema = apiModelIdProviderModelSchema.extend({
|
|
520
|
+
zaiApiKey: import_zod3.z.string().optional(),
|
|
521
|
+
zaiApiLine: import_zod3.z.union([import_zod3.z.literal("china"), import_zod3.z.literal("international")]).optional()
|
|
522
|
+
});
|
|
523
|
+
var fireworksSchema = apiModelIdProviderModelSchema.extend({
|
|
524
|
+
fireworksApiKey: import_zod3.z.string().optional()
|
|
525
|
+
});
|
|
499
526
|
var defaultSchema = import_zod3.z.object({
|
|
500
527
|
apiProvider: import_zod3.z.undefined()
|
|
501
528
|
});
|
|
@@ -528,6 +555,8 @@ var providerSettingsSchemaDiscriminated = import_zod3.z.discriminatedUnion("apiP
|
|
|
528
555
|
litellmSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("litellm") })),
|
|
529
556
|
cerebrasSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("cerebras") })),
|
|
530
557
|
sambaNovaSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("sambanova") })),
|
|
558
|
+
zaiSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("zai") })),
|
|
559
|
+
fireworksSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("fireworks") })),
|
|
531
560
|
defaultSchema
|
|
532
561
|
]);
|
|
533
562
|
var providerSettingsSchema = import_zod3.z.object({
|
|
@@ -560,6 +589,8 @@ var providerSettingsSchema = import_zod3.z.object({
|
|
|
560
589
|
...litellmSchema.shape,
|
|
561
590
|
...cerebrasSchema.shape,
|
|
562
591
|
...sambaNovaSchema.shape,
|
|
592
|
+
...zaiSchema.shape,
|
|
593
|
+
...fireworksSchema.shape,
|
|
563
594
|
...codebaseIndexProviderSchema.shape
|
|
564
595
|
});
|
|
565
596
|
var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: import_zod3.z.string().optional() });
|
|
@@ -614,12 +645,13 @@ var historyItemSchema = import_zod4.z.object({
|
|
|
614
645
|
|
|
615
646
|
// src/experiment.ts
|
|
616
647
|
var import_zod5 = require("zod");
|
|
617
|
-
var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption"];
|
|
648
|
+
var experimentIds = ["powerSteering", "multiFileApplyDiff", "preventFocusDisruption", "assistantMessageParser"];
|
|
618
649
|
var experimentIdsSchema = import_zod5.z.enum(experimentIds);
|
|
619
650
|
var experimentsSchema = import_zod5.z.object({
|
|
620
651
|
powerSteering: import_zod5.z.boolean().optional(),
|
|
621
652
|
multiFileApplyDiff: import_zod5.z.boolean().optional(),
|
|
622
|
-
preventFocusDisruption: import_zod5.z.boolean().optional()
|
|
653
|
+
preventFocusDisruption: import_zod5.z.boolean().optional(),
|
|
654
|
+
assistantMessageParser: import_zod5.z.boolean().optional()
|
|
623
655
|
});
|
|
624
656
|
|
|
625
657
|
// src/telemetry.ts
|
|
@@ -706,7 +738,14 @@ var clineMessageSchema = import_zod6.z.object({
|
|
|
706
738
|
progressStatus: toolProgressStatusSchema.optional(),
|
|
707
739
|
contextCondense: contextCondenseSchema.optional(),
|
|
708
740
|
isProtected: import_zod6.z.boolean().optional(),
|
|
709
|
-
apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional()
|
|
741
|
+
apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional(),
|
|
742
|
+
metadata: import_zod6.z.object({
|
|
743
|
+
gpt5: import_zod6.z.object({
|
|
744
|
+
previous_response_id: import_zod6.z.string().optional(),
|
|
745
|
+
instructions: import_zod6.z.string().optional(),
|
|
746
|
+
reasoning_summary: import_zod6.z.string().optional()
|
|
747
|
+
}).optional()
|
|
748
|
+
}).optional()
|
|
710
749
|
});
|
|
711
750
|
var tokenUsageSchema = import_zod6.z.object({
|
|
712
751
|
totalTokensIn: import_zod6.z.number(),
|
|
@@ -1146,6 +1185,7 @@ var globalSettingsSchema = import_zod11.z.object({
|
|
|
1146
1185
|
telemetrySetting: telemetrySettingsSchema.optional(),
|
|
1147
1186
|
mcpEnabled: import_zod11.z.boolean().optional(),
|
|
1148
1187
|
enableMcpServerCreation: import_zod11.z.boolean().optional(),
|
|
1188
|
+
remoteControlEnabled: import_zod11.z.boolean().optional(),
|
|
1149
1189
|
mode: import_zod11.z.string().optional(),
|
|
1150
1190
|
modeApiConfigs: import_zod11.z.record(import_zod11.z.string(), import_zod11.z.string()).optional(),
|
|
1151
1191
|
customModes: import_zod11.z.array(modeConfigSchema).optional(),
|
|
@@ -1188,7 +1228,8 @@ var SECRET_STATE_KEYS = [
|
|
|
1188
1228
|
"codebaseIndexGeminiApiKey",
|
|
1189
1229
|
"codebaseIndexMistralApiKey",
|
|
1190
1230
|
"huggingFaceApiKey",
|
|
1191
|
-
"sambaNovaApiKey"
|
|
1231
|
+
"sambaNovaApiKey",
|
|
1232
|
+
"fireworksApiKey"
|
|
1192
1233
|
];
|
|
1193
1234
|
var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
|
|
1194
1235
|
var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
|
|
@@ -1255,6 +1296,7 @@ var EVALS_SETTINGS = {
|
|
|
1255
1296
|
language: "en",
|
|
1256
1297
|
telemetrySetting: "enabled",
|
|
1257
1298
|
mcpEnabled: false,
|
|
1299
|
+
remoteControlEnabled: false,
|
|
1258
1300
|
mode: "code",
|
|
1259
1301
|
// "architect",
|
|
1260
1302
|
customModes: []
|
|
@@ -1696,6 +1738,22 @@ var anthropicModels = {
|
|
|
1696
1738
|
// $0.30 per million tokens
|
|
1697
1739
|
supportsReasoningBudget: true
|
|
1698
1740
|
},
|
|
1741
|
+
"claude-opus-4-1-20250805": {
|
|
1742
|
+
maxTokens: 8192,
|
|
1743
|
+
contextWindow: 2e5,
|
|
1744
|
+
supportsImages: true,
|
|
1745
|
+
supportsComputerUse: true,
|
|
1746
|
+
supportsPromptCache: true,
|
|
1747
|
+
inputPrice: 15,
|
|
1748
|
+
// $15 per million input tokens
|
|
1749
|
+
outputPrice: 75,
|
|
1750
|
+
// $75 per million output tokens
|
|
1751
|
+
cacheWritesPrice: 18.75,
|
|
1752
|
+
// $18.75 per million tokens
|
|
1753
|
+
cacheReadsPrice: 1.5,
|
|
1754
|
+
// $1.50 per million tokens
|
|
1755
|
+
supportsReasoningBudget: true
|
|
1756
|
+
},
|
|
1699
1757
|
"claude-opus-4-20250514": {
|
|
1700
1758
|
maxTokens: 32e3,
|
|
1701
1759
|
// Overridden to 8k if `enableReasoningEffort` is false.
|
|
@@ -1876,6 +1934,21 @@ var bedrockModels = {
|
|
|
1876
1934
|
maxCachePoints: 4,
|
|
1877
1935
|
cachableFields: ["system", "messages", "tools"]
|
|
1878
1936
|
},
|
|
1937
|
+
"anthropic.claude-opus-4-1-20250805-v1:0": {
|
|
1938
|
+
maxTokens: 8192,
|
|
1939
|
+
contextWindow: 2e5,
|
|
1940
|
+
supportsImages: true,
|
|
1941
|
+
supportsComputerUse: true,
|
|
1942
|
+
supportsPromptCache: true,
|
|
1943
|
+
supportsReasoningBudget: true,
|
|
1944
|
+
inputPrice: 15,
|
|
1945
|
+
outputPrice: 75,
|
|
1946
|
+
cacheWritesPrice: 18.75,
|
|
1947
|
+
cacheReadsPrice: 1.5,
|
|
1948
|
+
minTokensPerCachePoint: 1024,
|
|
1949
|
+
maxCachePoints: 4,
|
|
1950
|
+
cachableFields: ["system", "messages", "tools"]
|
|
1951
|
+
},
|
|
1879
1952
|
"anthropic.claude-opus-4-20250514-v1:0": {
|
|
1880
1953
|
maxTokens: 8192,
|
|
1881
1954
|
contextWindow: 2e5,
|
|
@@ -2248,6 +2321,15 @@ var cerebrasModels = {
|
|
|
2248
2321
|
outputPrice: 0,
|
|
2249
2322
|
description: "SOTA performance with ~1500 tokens/s",
|
|
2250
2323
|
supportsReasoningEffort: true
|
|
2324
|
+
},
|
|
2325
|
+
"gpt-oss-120b": {
|
|
2326
|
+
maxTokens: 8e3,
|
|
2327
|
+
contextWindow: 64e3,
|
|
2328
|
+
supportsImages: false,
|
|
2329
|
+
supportsPromptCache: false,
|
|
2330
|
+
inputPrice: 0,
|
|
2331
|
+
outputPrice: 0,
|
|
2332
|
+
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
2251
2333
|
}
|
|
2252
2334
|
};
|
|
2253
2335
|
|
|
@@ -2503,6 +2585,15 @@ var claudeCodeModels = {
|
|
|
2503
2585
|
supportsReasoningBudget: false,
|
|
2504
2586
|
requiredReasoningBudget: false
|
|
2505
2587
|
},
|
|
2588
|
+
"claude-opus-4-1-20250805": {
|
|
2589
|
+
...anthropicModels["claude-opus-4-1-20250805"],
|
|
2590
|
+
supportsImages: false,
|
|
2591
|
+
supportsPromptCache: true,
|
|
2592
|
+
// Claude Code does report cache tokens
|
|
2593
|
+
supportsReasoningEffort: false,
|
|
2594
|
+
supportsReasoningBudget: false,
|
|
2595
|
+
requiredReasoningBudget: false
|
|
2596
|
+
},
|
|
2506
2597
|
"claude-opus-4-20250514": {
|
|
2507
2598
|
...anthropicModels["claude-opus-4-20250514"],
|
|
2508
2599
|
supportsImages: false,
|
|
@@ -2976,6 +3067,24 @@ var groqModels = {
|
|
|
2976
3067
|
inputPrice: 1,
|
|
2977
3068
|
outputPrice: 3,
|
|
2978
3069
|
description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
|
|
3070
|
+
},
|
|
3071
|
+
"openai/gpt-oss-120b": {
|
|
3072
|
+
maxTokens: 32766,
|
|
3073
|
+
contextWindow: 131072,
|
|
3074
|
+
supportsImages: false,
|
|
3075
|
+
supportsPromptCache: false,
|
|
3076
|
+
inputPrice: 0.15,
|
|
3077
|
+
outputPrice: 0.75,
|
|
3078
|
+
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
3079
|
+
},
|
|
3080
|
+
"openai/gpt-oss-20b": {
|
|
3081
|
+
maxTokens: 32768,
|
|
3082
|
+
contextWindow: 131072,
|
|
3083
|
+
supportsImages: false,
|
|
3084
|
+
supportsPromptCache: false,
|
|
3085
|
+
inputPrice: 0.1,
|
|
3086
|
+
outputPrice: 0.5,
|
|
3087
|
+
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
2979
3088
|
}
|
|
2980
3089
|
};
|
|
2981
3090
|
|
|
@@ -3004,6 +3113,7 @@ var litellmDefaultModelInfo = {
|
|
|
3004
3113
|
};
|
|
3005
3114
|
var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
3006
3115
|
"claude-3-5-sonnet-latest",
|
|
3116
|
+
"claude-opus-4-1-20250805",
|
|
3007
3117
|
"claude-opus-4-20250514",
|
|
3008
3118
|
"claude-sonnet-4-20250514",
|
|
3009
3119
|
"claude-3-7-sonnet-latest",
|
|
@@ -3013,22 +3123,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
3013
3123
|
"vertex_ai/claude-3-5-sonnet-v2",
|
|
3014
3124
|
"vertex_ai/claude-3-5-sonnet-v2@20241022",
|
|
3015
3125
|
"vertex_ai/claude-3-7-sonnet@20250219",
|
|
3126
|
+
"vertex_ai/claude-opus-4-1@20250805",
|
|
3016
3127
|
"vertex_ai/claude-opus-4@20250514",
|
|
3017
3128
|
"vertex_ai/claude-sonnet-4@20250514",
|
|
3018
3129
|
"openrouter/anthropic/claude-3.5-sonnet",
|
|
3019
3130
|
"openrouter/anthropic/claude-3.5-sonnet:beta",
|
|
3020
3131
|
"openrouter/anthropic/claude-3.7-sonnet",
|
|
3021
3132
|
"openrouter/anthropic/claude-3.7-sonnet:beta",
|
|
3133
|
+
"anthropic.claude-opus-4-1-20250805-v1:0",
|
|
3022
3134
|
"anthropic.claude-opus-4-20250514-v1:0",
|
|
3023
3135
|
"anthropic.claude-sonnet-4-20250514-v1:0",
|
|
3024
3136
|
"anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
3025
3137
|
"anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
3026
3138
|
"us.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
3027
3139
|
"us.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
3140
|
+
"us.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
3028
3141
|
"us.anthropic.claude-opus-4-20250514-v1:0",
|
|
3029
3142
|
"us.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
3030
3143
|
"eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
3031
3144
|
"eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
3145
|
+
"eu.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
3032
3146
|
"eu.anthropic.claude-opus-4-20250514-v1:0",
|
|
3033
3147
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
3034
3148
|
"snowflake/claude-3-5-sonnet"
|
|
@@ -3165,8 +3279,48 @@ var ollamaDefaultModelInfo = {
|
|
|
3165
3279
|
};
|
|
3166
3280
|
|
|
3167
3281
|
// src/providers/openai.ts
|
|
3168
|
-
var openAiNativeDefaultModelId = "gpt-
|
|
3282
|
+
var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
|
|
3169
3283
|
var openAiNativeModels = {
|
|
3284
|
+
"gpt-5-2025-08-07": {
|
|
3285
|
+
maxTokens: 128e3,
|
|
3286
|
+
contextWindow: 4e5,
|
|
3287
|
+
supportsImages: true,
|
|
3288
|
+
supportsPromptCache: true,
|
|
3289
|
+
supportsReasoningEffort: true,
|
|
3290
|
+
reasoningEffort: "medium",
|
|
3291
|
+
inputPrice: 1.25,
|
|
3292
|
+
outputPrice: 10,
|
|
3293
|
+
cacheReadsPrice: 0.13,
|
|
3294
|
+
description: "GPT-5: The best model for coding and agentic tasks across domains",
|
|
3295
|
+
// supportsVerbosity is a new capability; ensure ModelInfo includes it
|
|
3296
|
+
supportsVerbosity: true
|
|
3297
|
+
},
|
|
3298
|
+
"gpt-5-mini-2025-08-07": {
|
|
3299
|
+
maxTokens: 128e3,
|
|
3300
|
+
contextWindow: 4e5,
|
|
3301
|
+
supportsImages: true,
|
|
3302
|
+
supportsPromptCache: true,
|
|
3303
|
+
supportsReasoningEffort: true,
|
|
3304
|
+
reasoningEffort: "medium",
|
|
3305
|
+
inputPrice: 0.25,
|
|
3306
|
+
outputPrice: 2,
|
|
3307
|
+
cacheReadsPrice: 0.03,
|
|
3308
|
+
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
|
|
3309
|
+
supportsVerbosity: true
|
|
3310
|
+
},
|
|
3311
|
+
"gpt-5-nano-2025-08-07": {
|
|
3312
|
+
maxTokens: 128e3,
|
|
3313
|
+
contextWindow: 4e5,
|
|
3314
|
+
supportsImages: true,
|
|
3315
|
+
supportsPromptCache: true,
|
|
3316
|
+
supportsReasoningEffort: true,
|
|
3317
|
+
reasoningEffort: "medium",
|
|
3318
|
+
inputPrice: 0.05,
|
|
3319
|
+
outputPrice: 0.4,
|
|
3320
|
+
cacheReadsPrice: 0.01,
|
|
3321
|
+
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
|
|
3322
|
+
supportsVerbosity: true
|
|
3323
|
+
},
|
|
3170
3324
|
"gpt-4.1": {
|
|
3171
3325
|
maxTokens: 32768,
|
|
3172
3326
|
contextWindow: 1047576,
|
|
@@ -3352,6 +3506,7 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3352
3506
|
};
|
|
3353
3507
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3354
3508
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
3509
|
+
var GPT5_DEFAULT_TEMPERATURE = 1;
|
|
3355
3510
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
3356
3511
|
|
|
3357
3512
|
// src/providers/openrouter.ts
|
|
@@ -3389,6 +3544,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
|
|
|
3389
3544
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3390
3545
|
"anthropic/claude-sonnet-4",
|
|
3391
3546
|
"anthropic/claude-opus-4",
|
|
3547
|
+
"anthropic/claude-opus-4.1",
|
|
3392
3548
|
"google/gemini-2.5-flash-preview",
|
|
3393
3549
|
"google/gemini-2.5-flash-preview:thinking",
|
|
3394
3550
|
"google/gemini-2.5-flash-preview-05-20",
|
|
@@ -3406,7 +3562,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
3406
3562
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3407
3563
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3408
3564
|
"anthropic/claude-sonnet-4",
|
|
3409
|
-
"anthropic/claude-opus-4"
|
|
3565
|
+
"anthropic/claude-opus-4",
|
|
3566
|
+
"anthropic/claude-opus-4.1"
|
|
3410
3567
|
]);
|
|
3411
3568
|
var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3412
3569
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
@@ -3416,6 +3573,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
|
3416
3573
|
var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3417
3574
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3418
3575
|
"anthropic/claude-opus-4",
|
|
3576
|
+
"anthropic/claude-opus-4.1",
|
|
3419
3577
|
"anthropic/claude-sonnet-4",
|
|
3420
3578
|
"google/gemini-2.5-pro-preview",
|
|
3421
3579
|
"google/gemini-2.5-pro",
|
|
@@ -3706,6 +3864,18 @@ var vertexModels = {
|
|
|
3706
3864
|
cacheReadsPrice: 0.3,
|
|
3707
3865
|
supportsReasoningBudget: true
|
|
3708
3866
|
},
|
|
3867
|
+
"claude-opus-4-1@20250805": {
|
|
3868
|
+
maxTokens: 8192,
|
|
3869
|
+
contextWindow: 2e5,
|
|
3870
|
+
supportsImages: true,
|
|
3871
|
+
supportsComputerUse: true,
|
|
3872
|
+
supportsPromptCache: true,
|
|
3873
|
+
inputPrice: 15,
|
|
3874
|
+
outputPrice: 75,
|
|
3875
|
+
cacheWritesPrice: 18.75,
|
|
3876
|
+
cacheReadsPrice: 1.5,
|
|
3877
|
+
supportsReasoningBudget: true
|
|
3878
|
+
},
|
|
3709
3879
|
"claude-opus-4@20250514": {
|
|
3710
3880
|
maxTokens: 8192,
|
|
3711
3881
|
contextWindow: 2e5,
|
|
@@ -4130,6 +4300,185 @@ var doubaoModels = {
|
|
|
4130
4300
|
var doubaoDefaultModelInfo = doubaoModels[doubaoDefaultModelId];
|
|
4131
4301
|
var DOUBAO_API_BASE_URL = "https://ark.cn-beijing.volces.com/api/v3";
|
|
4132
4302
|
var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
4303
|
+
|
|
4304
|
+
// src/providers/zai.ts
|
|
4305
|
+
var internationalZAiDefaultModelId = "glm-4.5";
|
|
4306
|
+
var internationalZAiModels = {
|
|
4307
|
+
"glm-4.5": {
|
|
4308
|
+
maxTokens: 98304,
|
|
4309
|
+
contextWindow: 131072,
|
|
4310
|
+
supportsImages: false,
|
|
4311
|
+
supportsPromptCache: true,
|
|
4312
|
+
inputPrice: 0.6,
|
|
4313
|
+
outputPrice: 2.2,
|
|
4314
|
+
cacheWritesPrice: 0,
|
|
4315
|
+
cacheReadsPrice: 0.11,
|
|
4316
|
+
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k."
|
|
4317
|
+
},
|
|
4318
|
+
"glm-4.5-air": {
|
|
4319
|
+
maxTokens: 98304,
|
|
4320
|
+
contextWindow: 131072,
|
|
4321
|
+
supportsImages: false,
|
|
4322
|
+
supportsPromptCache: true,
|
|
4323
|
+
inputPrice: 0.2,
|
|
4324
|
+
outputPrice: 1.1,
|
|
4325
|
+
cacheWritesPrice: 0,
|
|
4326
|
+
cacheReadsPrice: 0.03,
|
|
4327
|
+
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
|
|
4328
|
+
}
|
|
4329
|
+
};
|
|
4330
|
+
var mainlandZAiDefaultModelId = "glm-4.5";
|
|
4331
|
+
var mainlandZAiModels = {
|
|
4332
|
+
"glm-4.5": {
|
|
4333
|
+
maxTokens: 98304,
|
|
4334
|
+
contextWindow: 131072,
|
|
4335
|
+
supportsImages: false,
|
|
4336
|
+
supportsPromptCache: true,
|
|
4337
|
+
inputPrice: 0.29,
|
|
4338
|
+
outputPrice: 1.14,
|
|
4339
|
+
cacheWritesPrice: 0,
|
|
4340
|
+
cacheReadsPrice: 0.057,
|
|
4341
|
+
description: "GLM-4.5 is Zhipu's latest featured model. Its comprehensive capabilities in reasoning, coding, and agent reach the state-of-the-art (SOTA) level among open-source models, with a context length of up to 128k.",
|
|
4342
|
+
tiers: [
|
|
4343
|
+
{
|
|
4344
|
+
contextWindow: 32e3,
|
|
4345
|
+
inputPrice: 0.21,
|
|
4346
|
+
outputPrice: 1,
|
|
4347
|
+
cacheReadsPrice: 0.043
|
|
4348
|
+
},
|
|
4349
|
+
{
|
|
4350
|
+
contextWindow: 128e3,
|
|
4351
|
+
inputPrice: 0.29,
|
|
4352
|
+
outputPrice: 1.14,
|
|
4353
|
+
cacheReadsPrice: 0.057
|
|
4354
|
+
},
|
|
4355
|
+
{
|
|
4356
|
+
contextWindow: Infinity,
|
|
4357
|
+
inputPrice: 0.29,
|
|
4358
|
+
outputPrice: 1.14,
|
|
4359
|
+
cacheReadsPrice: 0.057
|
|
4360
|
+
}
|
|
4361
|
+
]
|
|
4362
|
+
},
|
|
4363
|
+
"glm-4.5-air": {
|
|
4364
|
+
maxTokens: 98304,
|
|
4365
|
+
contextWindow: 131072,
|
|
4366
|
+
supportsImages: false,
|
|
4367
|
+
supportsPromptCache: true,
|
|
4368
|
+
inputPrice: 0.1,
|
|
4369
|
+
outputPrice: 0.6,
|
|
4370
|
+
cacheWritesPrice: 0,
|
|
4371
|
+
cacheReadsPrice: 0.02,
|
|
4372
|
+
description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models.",
|
|
4373
|
+
tiers: [
|
|
4374
|
+
{
|
|
4375
|
+
contextWindow: 32e3,
|
|
4376
|
+
inputPrice: 0.07,
|
|
4377
|
+
outputPrice: 0.4,
|
|
4378
|
+
cacheReadsPrice: 0.014
|
|
4379
|
+
},
|
|
4380
|
+
{
|
|
4381
|
+
contextWindow: 128e3,
|
|
4382
|
+
inputPrice: 0.1,
|
|
4383
|
+
outputPrice: 0.6,
|
|
4384
|
+
cacheReadsPrice: 0.02
|
|
4385
|
+
},
|
|
4386
|
+
{
|
|
4387
|
+
contextWindow: Infinity,
|
|
4388
|
+
inputPrice: 0.1,
|
|
4389
|
+
outputPrice: 0.6,
|
|
4390
|
+
cacheReadsPrice: 0.02
|
|
4391
|
+
}
|
|
4392
|
+
]
|
|
4393
|
+
}
|
|
4394
|
+
};
|
|
4395
|
+
var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
4396
|
+
|
|
4397
|
+
// src/providers/fireworks.ts
|
|
4398
|
+
var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
|
|
4399
|
+
var fireworksModels = {
|
|
4400
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
4401
|
+
maxTokens: 16384,
|
|
4402
|
+
contextWindow: 128e3,
|
|
4403
|
+
supportsImages: false,
|
|
4404
|
+
supportsPromptCache: false,
|
|
4405
|
+
inputPrice: 0.6,
|
|
4406
|
+
outputPrice: 2.5,
|
|
4407
|
+
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
4408
|
+
},
|
|
4409
|
+
"accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
|
|
4410
|
+
maxTokens: 32768,
|
|
4411
|
+
contextWindow: 256e3,
|
|
4412
|
+
supportsImages: false,
|
|
4413
|
+
supportsPromptCache: false,
|
|
4414
|
+
inputPrice: 0.22,
|
|
4415
|
+
outputPrice: 0.88,
|
|
4416
|
+
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
4417
|
+
},
|
|
4418
|
+
"accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
|
|
4419
|
+
maxTokens: 32768,
|
|
4420
|
+
contextWindow: 256e3,
|
|
4421
|
+
supportsImages: false,
|
|
4422
|
+
supportsPromptCache: false,
|
|
4423
|
+
inputPrice: 0.45,
|
|
4424
|
+
outputPrice: 1.8,
|
|
4425
|
+
description: "Qwen3's most agentic code model to date."
|
|
4426
|
+
},
|
|
4427
|
+
"accounts/fireworks/models/deepseek-r1-0528": {
|
|
4428
|
+
maxTokens: 20480,
|
|
4429
|
+
contextWindow: 16e4,
|
|
4430
|
+
supportsImages: false,
|
|
4431
|
+
supportsPromptCache: false,
|
|
4432
|
+
inputPrice: 3,
|
|
4433
|
+
outputPrice: 8,
|
|
4434
|
+
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4435
|
+
},
|
|
4436
|
+
"accounts/fireworks/models/deepseek-v3": {
|
|
4437
|
+
maxTokens: 16384,
|
|
4438
|
+
contextWindow: 128e3,
|
|
4439
|
+
supportsImages: false,
|
|
4440
|
+
supportsPromptCache: false,
|
|
4441
|
+
inputPrice: 0.9,
|
|
4442
|
+
outputPrice: 0.9,
|
|
4443
|
+
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4444
|
+
},
|
|
4445
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
4446
|
+
maxTokens: 16384,
|
|
4447
|
+
contextWindow: 128e3,
|
|
4448
|
+
supportsImages: false,
|
|
4449
|
+
supportsPromptCache: false,
|
|
4450
|
+
inputPrice: 0.55,
|
|
4451
|
+
outputPrice: 2.19,
|
|
4452
|
+
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4453
|
+
},
|
|
4454
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
4455
|
+
maxTokens: 16384,
|
|
4456
|
+
contextWindow: 128e3,
|
|
4457
|
+
supportsImages: false,
|
|
4458
|
+
supportsPromptCache: false,
|
|
4459
|
+
inputPrice: 0.55,
|
|
4460
|
+
outputPrice: 2.19,
|
|
4461
|
+
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4462
|
+
},
|
|
4463
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
4464
|
+
maxTokens: 16384,
|
|
4465
|
+
contextWindow: 128e3,
|
|
4466
|
+
supportsImages: false,
|
|
4467
|
+
supportsPromptCache: false,
|
|
4468
|
+
inputPrice: 0.07,
|
|
4469
|
+
outputPrice: 0.3,
|
|
4470
|
+
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
4471
|
+
},
|
|
4472
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
4473
|
+
maxTokens: 16384,
|
|
4474
|
+
contextWindow: 128e3,
|
|
4475
|
+
supportsImages: false,
|
|
4476
|
+
supportsPromptCache: false,
|
|
4477
|
+
inputPrice: 0.15,
|
|
4478
|
+
outputPrice: 0.6,
|
|
4479
|
+
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
4480
|
+
}
|
|
4481
|
+
};
|
|
4133
4482
|
// Annotate the CommonJS export names for ESM import in node:
|
|
4134
4483
|
0 && (module.exports = {
|
|
4135
4484
|
ANTHROPIC_DEFAULT_MAX_TOKENS,
|
|
@@ -4153,6 +4502,7 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
|
4153
4502
|
GLAMA_DEFAULT_TEMPERATURE,
|
|
4154
4503
|
GLOBAL_SETTINGS_KEYS,
|
|
4155
4504
|
GLOBAL_STATE_KEYS,
|
|
4505
|
+
GPT5_DEFAULT_TEMPERATURE,
|
|
4156
4506
|
HUGGINGFACE_API_URL,
|
|
4157
4507
|
HUGGINGFACE_CACHE_DURATION,
|
|
4158
4508
|
HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
|
|
@@ -4183,6 +4533,7 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
|
4183
4533
|
TaskCommandName,
|
|
4184
4534
|
TelemetryEventName,
|
|
4185
4535
|
VERTEX_REGIONS,
|
|
4536
|
+
ZAI_DEFAULT_TEMPERATURE,
|
|
4186
4537
|
ackSchema,
|
|
4187
4538
|
anthropicDefaultModelId,
|
|
4188
4539
|
anthropicModels,
|
|
@@ -4223,6 +4574,9 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
|
4223
4574
|
experimentIds,
|
|
4224
4575
|
experimentIdsSchema,
|
|
4225
4576
|
experimentsSchema,
|
|
4577
|
+
extendedReasoningEffortsSchema,
|
|
4578
|
+
fireworksDefaultModelId,
|
|
4579
|
+
fireworksModels,
|
|
4226
4580
|
followUpDataSchema,
|
|
4227
4581
|
geminiDefaultModelId,
|
|
4228
4582
|
geminiModels,
|
|
@@ -4239,6 +4593,8 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
|
4239
4593
|
groupOptionsSchema,
|
|
4240
4594
|
historyItemSchema,
|
|
4241
4595
|
installMarketplaceItemOptionsSchema,
|
|
4596
|
+
internationalZAiDefaultModelId,
|
|
4597
|
+
internationalZAiModels,
|
|
4242
4598
|
ipcMessageSchema,
|
|
4243
4599
|
isBlockingAsk,
|
|
4244
4600
|
isGlobalStateKey,
|
|
@@ -4251,6 +4607,8 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
|
4251
4607
|
languagesSchema,
|
|
4252
4608
|
litellmDefaultModelId,
|
|
4253
4609
|
litellmDefaultModelInfo,
|
|
4610
|
+
mainlandZAiDefaultModelId,
|
|
4611
|
+
mainlandZAiModels,
|
|
4254
4612
|
marketplaceItemSchema,
|
|
4255
4613
|
marketplaceItemTypeSchema,
|
|
4256
4614
|
mcpExecutionStatusSchema,
|
|
@@ -4313,6 +4671,8 @@ var DOUBAO_API_CHAT_PATH = "/chat/completions";
|
|
|
4313
4671
|
toolUsageSchema,
|
|
4314
4672
|
unboundDefaultModelId,
|
|
4315
4673
|
unboundDefaultModelInfo,
|
|
4674
|
+
verbosityLevels,
|
|
4675
|
+
verbosityLevelsSchema,
|
|
4316
4676
|
vertexDefaultModelId,
|
|
4317
4677
|
vertexModels,
|
|
4318
4678
|
vscodeLlmDefaultModelId,
|