@roo-code/types 1.44.0 → 1.45.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +254 -7
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1584 -429
- package/dist/index.d.ts +1584 -429
- package/dist/index.js +248 -7
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.cjs
CHANGED
|
@@ -41,6 +41,7 @@ __export(index_exports, {
|
|
|
41
41
|
GLAMA_DEFAULT_TEMPERATURE: () => GLAMA_DEFAULT_TEMPERATURE,
|
|
42
42
|
GLOBAL_SETTINGS_KEYS: () => GLOBAL_SETTINGS_KEYS,
|
|
43
43
|
GLOBAL_STATE_KEYS: () => GLOBAL_STATE_KEYS,
|
|
44
|
+
GPT5_DEFAULT_TEMPERATURE: () => GPT5_DEFAULT_TEMPERATURE,
|
|
44
45
|
HUGGINGFACE_API_URL: () => HUGGINGFACE_API_URL,
|
|
45
46
|
HUGGINGFACE_CACHE_DURATION: () => HUGGINGFACE_CACHE_DURATION,
|
|
46
47
|
HUGGINGFACE_DEFAULT_CONTEXT_WINDOW: () => HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
|
|
@@ -112,6 +113,9 @@ __export(index_exports, {
|
|
|
112
113
|
experimentIds: () => experimentIds,
|
|
113
114
|
experimentIdsSchema: () => experimentIdsSchema,
|
|
114
115
|
experimentsSchema: () => experimentsSchema,
|
|
116
|
+
extendedReasoningEffortsSchema: () => extendedReasoningEffortsSchema,
|
|
117
|
+
fireworksDefaultModelId: () => fireworksDefaultModelId,
|
|
118
|
+
fireworksModels: () => fireworksModels,
|
|
115
119
|
followUpDataSchema: () => followUpDataSchema,
|
|
116
120
|
geminiDefaultModelId: () => geminiDefaultModelId,
|
|
117
121
|
geminiModels: () => geminiModels,
|
|
@@ -206,6 +210,8 @@ __export(index_exports, {
|
|
|
206
210
|
toolUsageSchema: () => toolUsageSchema,
|
|
207
211
|
unboundDefaultModelId: () => unboundDefaultModelId,
|
|
208
212
|
unboundDefaultModelInfo: () => unboundDefaultModelInfo,
|
|
213
|
+
verbosityLevels: () => verbosityLevels,
|
|
214
|
+
verbosityLevelsSchema: () => verbosityLevelsSchema,
|
|
209
215
|
vertexDefaultModelId: () => vertexDefaultModelId,
|
|
210
216
|
vertexModels: () => vertexModels,
|
|
211
217
|
vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
|
|
@@ -228,6 +234,8 @@ var import_zod3 = require("zod");
|
|
|
228
234
|
var import_zod = require("zod");
|
|
229
235
|
var reasoningEfforts = ["low", "medium", "high"];
|
|
230
236
|
var reasoningEffortsSchema = import_zod.z.enum(reasoningEfforts);
|
|
237
|
+
var verbosityLevels = ["low", "medium", "high"];
|
|
238
|
+
var verbosityLevelsSchema = import_zod.z.enum(verbosityLevels);
|
|
231
239
|
var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
|
|
232
240
|
var modelParametersSchema = import_zod.z.enum(modelParameters);
|
|
233
241
|
var isModelParameter = (value) => modelParameters.includes(value);
|
|
@@ -238,6 +246,8 @@ var modelInfoSchema = import_zod.z.object({
|
|
|
238
246
|
supportsImages: import_zod.z.boolean().optional(),
|
|
239
247
|
supportsComputerUse: import_zod.z.boolean().optional(),
|
|
240
248
|
supportsPromptCache: import_zod.z.boolean(),
|
|
249
|
+
// Capability flag to indicate whether the model supports an output verbosity parameter
|
|
250
|
+
supportsVerbosity: import_zod.z.boolean().optional(),
|
|
241
251
|
supportsReasoningBudget: import_zod.z.boolean().optional(),
|
|
242
252
|
requiredReasoningBudget: import_zod.z.boolean().optional(),
|
|
243
253
|
supportsReasoningEffort: import_zod.z.boolean().optional(),
|
|
@@ -305,6 +315,7 @@ var codebaseIndexProviderSchema = import_zod2.z.object({
|
|
|
305
315
|
});
|
|
306
316
|
|
|
307
317
|
// src/provider-settings.ts
|
|
318
|
+
var extendedReasoningEffortsSchema = import_zod3.z.union([reasoningEffortsSchema, import_zod3.z.literal("minimal")]);
|
|
308
319
|
var providerNames = [
|
|
309
320
|
"anthropic",
|
|
310
321
|
"claude-code",
|
|
@@ -334,7 +345,8 @@ var providerNames = [
|
|
|
334
345
|
"huggingface",
|
|
335
346
|
"cerebras",
|
|
336
347
|
"sambanova",
|
|
337
|
-
"zai"
|
|
348
|
+
"zai",
|
|
349
|
+
"fireworks"
|
|
338
350
|
];
|
|
339
351
|
var providerNamesSchema = import_zod3.z.enum(providerNames);
|
|
340
352
|
var providerSettingsEntrySchema = import_zod3.z.object({
|
|
@@ -353,9 +365,11 @@ var baseProviderSettingsSchema = import_zod3.z.object({
|
|
|
353
365
|
consecutiveMistakeLimit: import_zod3.z.number().min(0).optional(),
|
|
354
366
|
// Model reasoning.
|
|
355
367
|
enableReasoningEffort: import_zod3.z.boolean().optional(),
|
|
356
|
-
reasoningEffort:
|
|
368
|
+
reasoningEffort: extendedReasoningEffortsSchema.optional(),
|
|
357
369
|
modelMaxTokens: import_zod3.z.number().optional(),
|
|
358
|
-
modelMaxThinkingTokens: import_zod3.z.number().optional()
|
|
370
|
+
modelMaxThinkingTokens: import_zod3.z.number().optional(),
|
|
371
|
+
// Model verbosity.
|
|
372
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
359
373
|
});
|
|
360
374
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
361
375
|
apiModelId: import_zod3.z.string().optional()
|
|
@@ -506,6 +520,9 @@ var zaiSchema = apiModelIdProviderModelSchema.extend({
|
|
|
506
520
|
zaiApiKey: import_zod3.z.string().optional(),
|
|
507
521
|
zaiApiLine: import_zod3.z.union([import_zod3.z.literal("china"), import_zod3.z.literal("international")]).optional()
|
|
508
522
|
});
|
|
523
|
+
var fireworksSchema = apiModelIdProviderModelSchema.extend({
|
|
524
|
+
fireworksApiKey: import_zod3.z.string().optional()
|
|
525
|
+
});
|
|
509
526
|
var defaultSchema = import_zod3.z.object({
|
|
510
527
|
apiProvider: import_zod3.z.undefined()
|
|
511
528
|
});
|
|
@@ -539,6 +556,7 @@ var providerSettingsSchemaDiscriminated = import_zod3.z.discriminatedUnion("apiP
|
|
|
539
556
|
cerebrasSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("cerebras") })),
|
|
540
557
|
sambaNovaSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("sambanova") })),
|
|
541
558
|
zaiSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("zai") })),
|
|
559
|
+
fireworksSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("fireworks") })),
|
|
542
560
|
defaultSchema
|
|
543
561
|
]);
|
|
544
562
|
var providerSettingsSchema = import_zod3.z.object({
|
|
@@ -572,6 +590,7 @@ var providerSettingsSchema = import_zod3.z.object({
|
|
|
572
590
|
...cerebrasSchema.shape,
|
|
573
591
|
...sambaNovaSchema.shape,
|
|
574
592
|
...zaiSchema.shape,
|
|
593
|
+
...fireworksSchema.shape,
|
|
575
594
|
...codebaseIndexProviderSchema.shape
|
|
576
595
|
});
|
|
577
596
|
var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: import_zod3.z.string().optional() });
|
|
@@ -719,7 +738,14 @@ var clineMessageSchema = import_zod6.z.object({
|
|
|
719
738
|
progressStatus: toolProgressStatusSchema.optional(),
|
|
720
739
|
contextCondense: contextCondenseSchema.optional(),
|
|
721
740
|
isProtected: import_zod6.z.boolean().optional(),
|
|
722
|
-
apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional()
|
|
741
|
+
apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional(),
|
|
742
|
+
metadata: import_zod6.z.object({
|
|
743
|
+
gpt5: import_zod6.z.object({
|
|
744
|
+
previous_response_id: import_zod6.z.string().optional(),
|
|
745
|
+
instructions: import_zod6.z.string().optional(),
|
|
746
|
+
reasoning_summary: import_zod6.z.string().optional()
|
|
747
|
+
}).optional()
|
|
748
|
+
}).optional()
|
|
723
749
|
});
|
|
724
750
|
var tokenUsageSchema = import_zod6.z.object({
|
|
725
751
|
totalTokensIn: import_zod6.z.number(),
|
|
@@ -1202,7 +1228,8 @@ var SECRET_STATE_KEYS = [
|
|
|
1202
1228
|
"codebaseIndexGeminiApiKey",
|
|
1203
1229
|
"codebaseIndexMistralApiKey",
|
|
1204
1230
|
"huggingFaceApiKey",
|
|
1205
|
-
"sambaNovaApiKey"
|
|
1231
|
+
"sambaNovaApiKey",
|
|
1232
|
+
"fireworksApiKey"
|
|
1206
1233
|
];
|
|
1207
1234
|
var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
|
|
1208
1235
|
var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
|
|
@@ -1711,6 +1738,22 @@ var anthropicModels = {
|
|
|
1711
1738
|
// $0.30 per million tokens
|
|
1712
1739
|
supportsReasoningBudget: true
|
|
1713
1740
|
},
|
|
1741
|
+
"claude-opus-4-1-20250805": {
|
|
1742
|
+
maxTokens: 8192,
|
|
1743
|
+
contextWindow: 2e5,
|
|
1744
|
+
supportsImages: true,
|
|
1745
|
+
supportsComputerUse: true,
|
|
1746
|
+
supportsPromptCache: true,
|
|
1747
|
+
inputPrice: 15,
|
|
1748
|
+
// $15 per million input tokens
|
|
1749
|
+
outputPrice: 75,
|
|
1750
|
+
// $75 per million output tokens
|
|
1751
|
+
cacheWritesPrice: 18.75,
|
|
1752
|
+
// $18.75 per million tokens
|
|
1753
|
+
cacheReadsPrice: 1.5,
|
|
1754
|
+
// $1.50 per million tokens
|
|
1755
|
+
supportsReasoningBudget: true
|
|
1756
|
+
},
|
|
1714
1757
|
"claude-opus-4-20250514": {
|
|
1715
1758
|
maxTokens: 32e3,
|
|
1716
1759
|
// Overridden to 8k if `enableReasoningEffort` is false.
|
|
@@ -1891,6 +1934,21 @@ var bedrockModels = {
|
|
|
1891
1934
|
maxCachePoints: 4,
|
|
1892
1935
|
cachableFields: ["system", "messages", "tools"]
|
|
1893
1936
|
},
|
|
1937
|
+
"anthropic.claude-opus-4-1-20250805-v1:0": {
|
|
1938
|
+
maxTokens: 8192,
|
|
1939
|
+
contextWindow: 2e5,
|
|
1940
|
+
supportsImages: true,
|
|
1941
|
+
supportsComputerUse: true,
|
|
1942
|
+
supportsPromptCache: true,
|
|
1943
|
+
supportsReasoningBudget: true,
|
|
1944
|
+
inputPrice: 15,
|
|
1945
|
+
outputPrice: 75,
|
|
1946
|
+
cacheWritesPrice: 18.75,
|
|
1947
|
+
cacheReadsPrice: 1.5,
|
|
1948
|
+
minTokensPerCachePoint: 1024,
|
|
1949
|
+
maxCachePoints: 4,
|
|
1950
|
+
cachableFields: ["system", "messages", "tools"]
|
|
1951
|
+
},
|
|
1894
1952
|
"anthropic.claude-opus-4-20250514-v1:0": {
|
|
1895
1953
|
maxTokens: 8192,
|
|
1896
1954
|
contextWindow: 2e5,
|
|
@@ -2263,6 +2321,15 @@ var cerebrasModels = {
|
|
|
2263
2321
|
outputPrice: 0,
|
|
2264
2322
|
description: "SOTA performance with ~1500 tokens/s",
|
|
2265
2323
|
supportsReasoningEffort: true
|
|
2324
|
+
},
|
|
2325
|
+
"gpt-oss-120b": {
|
|
2326
|
+
maxTokens: 8e3,
|
|
2327
|
+
contextWindow: 64e3,
|
|
2328
|
+
supportsImages: false,
|
|
2329
|
+
supportsPromptCache: false,
|
|
2330
|
+
inputPrice: 0,
|
|
2331
|
+
outputPrice: 0,
|
|
2332
|
+
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
2266
2333
|
}
|
|
2267
2334
|
};
|
|
2268
2335
|
|
|
@@ -2518,6 +2585,15 @@ var claudeCodeModels = {
|
|
|
2518
2585
|
supportsReasoningBudget: false,
|
|
2519
2586
|
requiredReasoningBudget: false
|
|
2520
2587
|
},
|
|
2588
|
+
"claude-opus-4-1-20250805": {
|
|
2589
|
+
...anthropicModels["claude-opus-4-1-20250805"],
|
|
2590
|
+
supportsImages: false,
|
|
2591
|
+
supportsPromptCache: true,
|
|
2592
|
+
// Claude Code does report cache tokens
|
|
2593
|
+
supportsReasoningEffort: false,
|
|
2594
|
+
supportsReasoningBudget: false,
|
|
2595
|
+
requiredReasoningBudget: false
|
|
2596
|
+
},
|
|
2521
2597
|
"claude-opus-4-20250514": {
|
|
2522
2598
|
...anthropicModels["claude-opus-4-20250514"],
|
|
2523
2599
|
supportsImages: false,
|
|
@@ -2991,6 +3067,24 @@ var groqModels = {
|
|
|
2991
3067
|
inputPrice: 1,
|
|
2992
3068
|
outputPrice: 3,
|
|
2993
3069
|
description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
|
|
3070
|
+
},
|
|
3071
|
+
"openai/gpt-oss-120b": {
|
|
3072
|
+
maxTokens: 32766,
|
|
3073
|
+
contextWindow: 131072,
|
|
3074
|
+
supportsImages: false,
|
|
3075
|
+
supportsPromptCache: false,
|
|
3076
|
+
inputPrice: 0.15,
|
|
3077
|
+
outputPrice: 0.75,
|
|
3078
|
+
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
3079
|
+
},
|
|
3080
|
+
"openai/gpt-oss-20b": {
|
|
3081
|
+
maxTokens: 32768,
|
|
3082
|
+
contextWindow: 131072,
|
|
3083
|
+
supportsImages: false,
|
|
3084
|
+
supportsPromptCache: false,
|
|
3085
|
+
inputPrice: 0.1,
|
|
3086
|
+
outputPrice: 0.5,
|
|
3087
|
+
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
2994
3088
|
}
|
|
2995
3089
|
};
|
|
2996
3090
|
|
|
@@ -3019,6 +3113,7 @@ var litellmDefaultModelInfo = {
|
|
|
3019
3113
|
};
|
|
3020
3114
|
var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
3021
3115
|
"claude-3-5-sonnet-latest",
|
|
3116
|
+
"claude-opus-4-1-20250805",
|
|
3022
3117
|
"claude-opus-4-20250514",
|
|
3023
3118
|
"claude-sonnet-4-20250514",
|
|
3024
3119
|
"claude-3-7-sonnet-latest",
|
|
@@ -3028,22 +3123,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
3028
3123
|
"vertex_ai/claude-3-5-sonnet-v2",
|
|
3029
3124
|
"vertex_ai/claude-3-5-sonnet-v2@20241022",
|
|
3030
3125
|
"vertex_ai/claude-3-7-sonnet@20250219",
|
|
3126
|
+
"vertex_ai/claude-opus-4-1@20250805",
|
|
3031
3127
|
"vertex_ai/claude-opus-4@20250514",
|
|
3032
3128
|
"vertex_ai/claude-sonnet-4@20250514",
|
|
3033
3129
|
"openrouter/anthropic/claude-3.5-sonnet",
|
|
3034
3130
|
"openrouter/anthropic/claude-3.5-sonnet:beta",
|
|
3035
3131
|
"openrouter/anthropic/claude-3.7-sonnet",
|
|
3036
3132
|
"openrouter/anthropic/claude-3.7-sonnet:beta",
|
|
3133
|
+
"anthropic.claude-opus-4-1-20250805-v1:0",
|
|
3037
3134
|
"anthropic.claude-opus-4-20250514-v1:0",
|
|
3038
3135
|
"anthropic.claude-sonnet-4-20250514-v1:0",
|
|
3039
3136
|
"anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
3040
3137
|
"anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
3041
3138
|
"us.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
3042
3139
|
"us.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
3140
|
+
"us.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
3043
3141
|
"us.anthropic.claude-opus-4-20250514-v1:0",
|
|
3044
3142
|
"us.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
3045
3143
|
"eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
3046
3144
|
"eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
3145
|
+
"eu.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
3047
3146
|
"eu.anthropic.claude-opus-4-20250514-v1:0",
|
|
3048
3147
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
3049
3148
|
"snowflake/claude-3-5-sonnet"
|
|
@@ -3180,8 +3279,48 @@ var ollamaDefaultModelInfo = {
|
|
|
3180
3279
|
};
|
|
3181
3280
|
|
|
3182
3281
|
// src/providers/openai.ts
|
|
3183
|
-
var openAiNativeDefaultModelId = "gpt-
|
|
3282
|
+
var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
|
|
3184
3283
|
var openAiNativeModels = {
|
|
3284
|
+
"gpt-5-2025-08-07": {
|
|
3285
|
+
maxTokens: 128e3,
|
|
3286
|
+
contextWindow: 4e5,
|
|
3287
|
+
supportsImages: true,
|
|
3288
|
+
supportsPromptCache: true,
|
|
3289
|
+
supportsReasoningEffort: true,
|
|
3290
|
+
reasoningEffort: "medium",
|
|
3291
|
+
inputPrice: 1.25,
|
|
3292
|
+
outputPrice: 10,
|
|
3293
|
+
cacheReadsPrice: 0.13,
|
|
3294
|
+
description: "GPT-5: The best model for coding and agentic tasks across domains",
|
|
3295
|
+
// supportsVerbosity is a new capability; ensure ModelInfo includes it
|
|
3296
|
+
supportsVerbosity: true
|
|
3297
|
+
},
|
|
3298
|
+
"gpt-5-mini-2025-08-07": {
|
|
3299
|
+
maxTokens: 128e3,
|
|
3300
|
+
contextWindow: 4e5,
|
|
3301
|
+
supportsImages: true,
|
|
3302
|
+
supportsPromptCache: true,
|
|
3303
|
+
supportsReasoningEffort: true,
|
|
3304
|
+
reasoningEffort: "medium",
|
|
3305
|
+
inputPrice: 0.25,
|
|
3306
|
+
outputPrice: 2,
|
|
3307
|
+
cacheReadsPrice: 0.03,
|
|
3308
|
+
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
|
|
3309
|
+
supportsVerbosity: true
|
|
3310
|
+
},
|
|
3311
|
+
"gpt-5-nano-2025-08-07": {
|
|
3312
|
+
maxTokens: 128e3,
|
|
3313
|
+
contextWindow: 4e5,
|
|
3314
|
+
supportsImages: true,
|
|
3315
|
+
supportsPromptCache: true,
|
|
3316
|
+
supportsReasoningEffort: true,
|
|
3317
|
+
reasoningEffort: "medium",
|
|
3318
|
+
inputPrice: 0.05,
|
|
3319
|
+
outputPrice: 0.4,
|
|
3320
|
+
cacheReadsPrice: 0.01,
|
|
3321
|
+
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
|
|
3322
|
+
supportsVerbosity: true
|
|
3323
|
+
},
|
|
3185
3324
|
"gpt-4.1": {
|
|
3186
3325
|
maxTokens: 32768,
|
|
3187
3326
|
contextWindow: 1047576,
|
|
@@ -3367,6 +3506,7 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3367
3506
|
};
|
|
3368
3507
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3369
3508
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
3509
|
+
var GPT5_DEFAULT_TEMPERATURE = 1;
|
|
3370
3510
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
3371
3511
|
|
|
3372
3512
|
// src/providers/openrouter.ts
|
|
@@ -3404,6 +3544,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
|
|
|
3404
3544
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3405
3545
|
"anthropic/claude-sonnet-4",
|
|
3406
3546
|
"anthropic/claude-opus-4",
|
|
3547
|
+
"anthropic/claude-opus-4.1",
|
|
3407
3548
|
"google/gemini-2.5-flash-preview",
|
|
3408
3549
|
"google/gemini-2.5-flash-preview:thinking",
|
|
3409
3550
|
"google/gemini-2.5-flash-preview-05-20",
|
|
@@ -3421,7 +3562,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
3421
3562
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3422
3563
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3423
3564
|
"anthropic/claude-sonnet-4",
|
|
3424
|
-
"anthropic/claude-opus-4"
|
|
3565
|
+
"anthropic/claude-opus-4",
|
|
3566
|
+
"anthropic/claude-opus-4.1"
|
|
3425
3567
|
]);
|
|
3426
3568
|
var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3427
3569
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
@@ -3431,6 +3573,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
|
3431
3573
|
var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3432
3574
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3433
3575
|
"anthropic/claude-opus-4",
|
|
3576
|
+
"anthropic/claude-opus-4.1",
|
|
3434
3577
|
"anthropic/claude-sonnet-4",
|
|
3435
3578
|
"google/gemini-2.5-pro-preview",
|
|
3436
3579
|
"google/gemini-2.5-pro",
|
|
@@ -3721,6 +3864,18 @@ var vertexModels = {
|
|
|
3721
3864
|
cacheReadsPrice: 0.3,
|
|
3722
3865
|
supportsReasoningBudget: true
|
|
3723
3866
|
},
|
|
3867
|
+
"claude-opus-4-1@20250805": {
|
|
3868
|
+
maxTokens: 8192,
|
|
3869
|
+
contextWindow: 2e5,
|
|
3870
|
+
supportsImages: true,
|
|
3871
|
+
supportsComputerUse: true,
|
|
3872
|
+
supportsPromptCache: true,
|
|
3873
|
+
inputPrice: 15,
|
|
3874
|
+
outputPrice: 75,
|
|
3875
|
+
cacheWritesPrice: 18.75,
|
|
3876
|
+
cacheReadsPrice: 1.5,
|
|
3877
|
+
supportsReasoningBudget: true
|
|
3878
|
+
},
|
|
3724
3879
|
"claude-opus-4@20250514": {
|
|
3725
3880
|
maxTokens: 8192,
|
|
3726
3881
|
contextWindow: 2e5,
|
|
@@ -4238,6 +4393,92 @@ var mainlandZAiModels = {
|
|
|
4238
4393
|
}
|
|
4239
4394
|
};
|
|
4240
4395
|
var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
4396
|
+
|
|
4397
|
+
// src/providers/fireworks.ts
|
|
4398
|
+
var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
|
|
4399
|
+
var fireworksModels = {
|
|
4400
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
4401
|
+
maxTokens: 16384,
|
|
4402
|
+
contextWindow: 128e3,
|
|
4403
|
+
supportsImages: false,
|
|
4404
|
+
supportsPromptCache: false,
|
|
4405
|
+
inputPrice: 0.6,
|
|
4406
|
+
outputPrice: 2.5,
|
|
4407
|
+
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
4408
|
+
},
|
|
4409
|
+
"accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
|
|
4410
|
+
maxTokens: 32768,
|
|
4411
|
+
contextWindow: 256e3,
|
|
4412
|
+
supportsImages: false,
|
|
4413
|
+
supportsPromptCache: false,
|
|
4414
|
+
inputPrice: 0.22,
|
|
4415
|
+
outputPrice: 0.88,
|
|
4416
|
+
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
4417
|
+
},
|
|
4418
|
+
"accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
|
|
4419
|
+
maxTokens: 32768,
|
|
4420
|
+
contextWindow: 256e3,
|
|
4421
|
+
supportsImages: false,
|
|
4422
|
+
supportsPromptCache: false,
|
|
4423
|
+
inputPrice: 0.45,
|
|
4424
|
+
outputPrice: 1.8,
|
|
4425
|
+
description: "Qwen3's most agentic code model to date."
|
|
4426
|
+
},
|
|
4427
|
+
"accounts/fireworks/models/deepseek-r1-0528": {
|
|
4428
|
+
maxTokens: 20480,
|
|
4429
|
+
contextWindow: 16e4,
|
|
4430
|
+
supportsImages: false,
|
|
4431
|
+
supportsPromptCache: false,
|
|
4432
|
+
inputPrice: 3,
|
|
4433
|
+
outputPrice: 8,
|
|
4434
|
+
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4435
|
+
},
|
|
4436
|
+
"accounts/fireworks/models/deepseek-v3": {
|
|
4437
|
+
maxTokens: 16384,
|
|
4438
|
+
contextWindow: 128e3,
|
|
4439
|
+
supportsImages: false,
|
|
4440
|
+
supportsPromptCache: false,
|
|
4441
|
+
inputPrice: 0.9,
|
|
4442
|
+
outputPrice: 0.9,
|
|
4443
|
+
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4444
|
+
},
|
|
4445
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
4446
|
+
maxTokens: 16384,
|
|
4447
|
+
contextWindow: 128e3,
|
|
4448
|
+
supportsImages: false,
|
|
4449
|
+
supportsPromptCache: false,
|
|
4450
|
+
inputPrice: 0.55,
|
|
4451
|
+
outputPrice: 2.19,
|
|
4452
|
+
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4453
|
+
},
|
|
4454
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
4455
|
+
maxTokens: 16384,
|
|
4456
|
+
contextWindow: 128e3,
|
|
4457
|
+
supportsImages: false,
|
|
4458
|
+
supportsPromptCache: false,
|
|
4459
|
+
inputPrice: 0.55,
|
|
4460
|
+
outputPrice: 2.19,
|
|
4461
|
+
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4462
|
+
},
|
|
4463
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
4464
|
+
maxTokens: 16384,
|
|
4465
|
+
contextWindow: 128e3,
|
|
4466
|
+
supportsImages: false,
|
|
4467
|
+
supportsPromptCache: false,
|
|
4468
|
+
inputPrice: 0.07,
|
|
4469
|
+
outputPrice: 0.3,
|
|
4470
|
+
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
4471
|
+
},
|
|
4472
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
4473
|
+
maxTokens: 16384,
|
|
4474
|
+
contextWindow: 128e3,
|
|
4475
|
+
supportsImages: false,
|
|
4476
|
+
supportsPromptCache: false,
|
|
4477
|
+
inputPrice: 0.15,
|
|
4478
|
+
outputPrice: 0.6,
|
|
4479
|
+
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
4480
|
+
}
|
|
4481
|
+
};
|
|
4241
4482
|
// Annotate the CommonJS export names for ESM import in node:
|
|
4242
4483
|
0 && (module.exports = {
|
|
4243
4484
|
ANTHROPIC_DEFAULT_MAX_TOKENS,
|
|
@@ -4261,6 +4502,7 @@ var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
|
4261
4502
|
GLAMA_DEFAULT_TEMPERATURE,
|
|
4262
4503
|
GLOBAL_SETTINGS_KEYS,
|
|
4263
4504
|
GLOBAL_STATE_KEYS,
|
|
4505
|
+
GPT5_DEFAULT_TEMPERATURE,
|
|
4264
4506
|
HUGGINGFACE_API_URL,
|
|
4265
4507
|
HUGGINGFACE_CACHE_DURATION,
|
|
4266
4508
|
HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
|
|
@@ -4332,6 +4574,9 @@ var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
|
4332
4574
|
experimentIds,
|
|
4333
4575
|
experimentIdsSchema,
|
|
4334
4576
|
experimentsSchema,
|
|
4577
|
+
extendedReasoningEffortsSchema,
|
|
4578
|
+
fireworksDefaultModelId,
|
|
4579
|
+
fireworksModels,
|
|
4335
4580
|
followUpDataSchema,
|
|
4336
4581
|
geminiDefaultModelId,
|
|
4337
4582
|
geminiModels,
|
|
@@ -4426,6 +4671,8 @@ var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
|
4426
4671
|
toolUsageSchema,
|
|
4427
4672
|
unboundDefaultModelId,
|
|
4428
4673
|
unboundDefaultModelInfo,
|
|
4674
|
+
verbosityLevels,
|
|
4675
|
+
verbosityLevelsSchema,
|
|
4429
4676
|
vertexDefaultModelId,
|
|
4430
4677
|
vertexModels,
|
|
4431
4678
|
vscodeLlmDefaultModelId,
|