@roo-code/types 1.44.0 → 1.45.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +254 -7
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1584 -429
- package/dist/index.d.ts +1584 -429
- package/dist/index.js +248 -7
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.js
CHANGED
|
@@ -11,6 +11,8 @@ import { z as z3 } from "zod";
|
|
|
11
11
|
import { z } from "zod";
|
|
12
12
|
var reasoningEfforts = ["low", "medium", "high"];
|
|
13
13
|
var reasoningEffortsSchema = z.enum(reasoningEfforts);
|
|
14
|
+
var verbosityLevels = ["low", "medium", "high"];
|
|
15
|
+
var verbosityLevelsSchema = z.enum(verbosityLevels);
|
|
14
16
|
var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
|
|
15
17
|
var modelParametersSchema = z.enum(modelParameters);
|
|
16
18
|
var isModelParameter = (value) => modelParameters.includes(value);
|
|
@@ -21,6 +23,8 @@ var modelInfoSchema = z.object({
|
|
|
21
23
|
supportsImages: z.boolean().optional(),
|
|
22
24
|
supportsComputerUse: z.boolean().optional(),
|
|
23
25
|
supportsPromptCache: z.boolean(),
|
|
26
|
+
// Capability flag to indicate whether the model supports an output verbosity parameter
|
|
27
|
+
supportsVerbosity: z.boolean().optional(),
|
|
24
28
|
supportsReasoningBudget: z.boolean().optional(),
|
|
25
29
|
requiredReasoningBudget: z.boolean().optional(),
|
|
26
30
|
supportsReasoningEffort: z.boolean().optional(),
|
|
@@ -88,6 +92,7 @@ var codebaseIndexProviderSchema = z2.object({
|
|
|
88
92
|
});
|
|
89
93
|
|
|
90
94
|
// src/provider-settings.ts
|
|
95
|
+
var extendedReasoningEffortsSchema = z3.union([reasoningEffortsSchema, z3.literal("minimal")]);
|
|
91
96
|
var providerNames = [
|
|
92
97
|
"anthropic",
|
|
93
98
|
"claude-code",
|
|
@@ -117,7 +122,8 @@ var providerNames = [
|
|
|
117
122
|
"huggingface",
|
|
118
123
|
"cerebras",
|
|
119
124
|
"sambanova",
|
|
120
|
-
"zai"
|
|
125
|
+
"zai",
|
|
126
|
+
"fireworks"
|
|
121
127
|
];
|
|
122
128
|
var providerNamesSchema = z3.enum(providerNames);
|
|
123
129
|
var providerSettingsEntrySchema = z3.object({
|
|
@@ -136,9 +142,11 @@ var baseProviderSettingsSchema = z3.object({
|
|
|
136
142
|
consecutiveMistakeLimit: z3.number().min(0).optional(),
|
|
137
143
|
// Model reasoning.
|
|
138
144
|
enableReasoningEffort: z3.boolean().optional(),
|
|
139
|
-
reasoningEffort:
|
|
145
|
+
reasoningEffort: extendedReasoningEffortsSchema.optional(),
|
|
140
146
|
modelMaxTokens: z3.number().optional(),
|
|
141
|
-
modelMaxThinkingTokens: z3.number().optional()
|
|
147
|
+
modelMaxThinkingTokens: z3.number().optional(),
|
|
148
|
+
// Model verbosity.
|
|
149
|
+
verbosity: verbosityLevelsSchema.optional()
|
|
142
150
|
});
|
|
143
151
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
144
152
|
apiModelId: z3.string().optional()
|
|
@@ -289,6 +297,9 @@ var zaiSchema = apiModelIdProviderModelSchema.extend({
|
|
|
289
297
|
zaiApiKey: z3.string().optional(),
|
|
290
298
|
zaiApiLine: z3.union([z3.literal("china"), z3.literal("international")]).optional()
|
|
291
299
|
});
|
|
300
|
+
var fireworksSchema = apiModelIdProviderModelSchema.extend({
|
|
301
|
+
fireworksApiKey: z3.string().optional()
|
|
302
|
+
});
|
|
292
303
|
var defaultSchema = z3.object({
|
|
293
304
|
apiProvider: z3.undefined()
|
|
294
305
|
});
|
|
@@ -322,6 +333,7 @@ var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
|
|
|
322
333
|
cerebrasSchema.merge(z3.object({ apiProvider: z3.literal("cerebras") })),
|
|
323
334
|
sambaNovaSchema.merge(z3.object({ apiProvider: z3.literal("sambanova") })),
|
|
324
335
|
zaiSchema.merge(z3.object({ apiProvider: z3.literal("zai") })),
|
|
336
|
+
fireworksSchema.merge(z3.object({ apiProvider: z3.literal("fireworks") })),
|
|
325
337
|
defaultSchema
|
|
326
338
|
]);
|
|
327
339
|
var providerSettingsSchema = z3.object({
|
|
@@ -355,6 +367,7 @@ var providerSettingsSchema = z3.object({
|
|
|
355
367
|
...cerebrasSchema.shape,
|
|
356
368
|
...sambaNovaSchema.shape,
|
|
357
369
|
...zaiSchema.shape,
|
|
370
|
+
...fireworksSchema.shape,
|
|
358
371
|
...codebaseIndexProviderSchema.shape
|
|
359
372
|
});
|
|
360
373
|
var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: z3.string().optional() });
|
|
@@ -502,7 +515,14 @@ var clineMessageSchema = z6.object({
|
|
|
502
515
|
progressStatus: toolProgressStatusSchema.optional(),
|
|
503
516
|
contextCondense: contextCondenseSchema.optional(),
|
|
504
517
|
isProtected: z6.boolean().optional(),
|
|
505
|
-
apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional()
|
|
518
|
+
apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional(),
|
|
519
|
+
metadata: z6.object({
|
|
520
|
+
gpt5: z6.object({
|
|
521
|
+
previous_response_id: z6.string().optional(),
|
|
522
|
+
instructions: z6.string().optional(),
|
|
523
|
+
reasoning_summary: z6.string().optional()
|
|
524
|
+
}).optional()
|
|
525
|
+
}).optional()
|
|
506
526
|
});
|
|
507
527
|
var tokenUsageSchema = z6.object({
|
|
508
528
|
totalTokensIn: z6.number(),
|
|
@@ -985,7 +1005,8 @@ var SECRET_STATE_KEYS = [
|
|
|
985
1005
|
"codebaseIndexGeminiApiKey",
|
|
986
1006
|
"codebaseIndexMistralApiKey",
|
|
987
1007
|
"huggingFaceApiKey",
|
|
988
|
-
"sambaNovaApiKey"
|
|
1008
|
+
"sambaNovaApiKey",
|
|
1009
|
+
"fireworksApiKey"
|
|
989
1010
|
];
|
|
990
1011
|
var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
|
|
991
1012
|
var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
|
|
@@ -1494,6 +1515,22 @@ var anthropicModels = {
|
|
|
1494
1515
|
// $0.30 per million tokens
|
|
1495
1516
|
supportsReasoningBudget: true
|
|
1496
1517
|
},
|
|
1518
|
+
"claude-opus-4-1-20250805": {
|
|
1519
|
+
maxTokens: 8192,
|
|
1520
|
+
contextWindow: 2e5,
|
|
1521
|
+
supportsImages: true,
|
|
1522
|
+
supportsComputerUse: true,
|
|
1523
|
+
supportsPromptCache: true,
|
|
1524
|
+
inputPrice: 15,
|
|
1525
|
+
// $15 per million input tokens
|
|
1526
|
+
outputPrice: 75,
|
|
1527
|
+
// $75 per million output tokens
|
|
1528
|
+
cacheWritesPrice: 18.75,
|
|
1529
|
+
// $18.75 per million tokens
|
|
1530
|
+
cacheReadsPrice: 1.5,
|
|
1531
|
+
// $1.50 per million tokens
|
|
1532
|
+
supportsReasoningBudget: true
|
|
1533
|
+
},
|
|
1497
1534
|
"claude-opus-4-20250514": {
|
|
1498
1535
|
maxTokens: 32e3,
|
|
1499
1536
|
// Overridden to 8k if `enableReasoningEffort` is false.
|
|
@@ -1674,6 +1711,21 @@ var bedrockModels = {
|
|
|
1674
1711
|
maxCachePoints: 4,
|
|
1675
1712
|
cachableFields: ["system", "messages", "tools"]
|
|
1676
1713
|
},
|
|
1714
|
+
"anthropic.claude-opus-4-1-20250805-v1:0": {
|
|
1715
|
+
maxTokens: 8192,
|
|
1716
|
+
contextWindow: 2e5,
|
|
1717
|
+
supportsImages: true,
|
|
1718
|
+
supportsComputerUse: true,
|
|
1719
|
+
supportsPromptCache: true,
|
|
1720
|
+
supportsReasoningBudget: true,
|
|
1721
|
+
inputPrice: 15,
|
|
1722
|
+
outputPrice: 75,
|
|
1723
|
+
cacheWritesPrice: 18.75,
|
|
1724
|
+
cacheReadsPrice: 1.5,
|
|
1725
|
+
minTokensPerCachePoint: 1024,
|
|
1726
|
+
maxCachePoints: 4,
|
|
1727
|
+
cachableFields: ["system", "messages", "tools"]
|
|
1728
|
+
},
|
|
1677
1729
|
"anthropic.claude-opus-4-20250514-v1:0": {
|
|
1678
1730
|
maxTokens: 8192,
|
|
1679
1731
|
contextWindow: 2e5,
|
|
@@ -2046,6 +2098,15 @@ var cerebrasModels = {
|
|
|
2046
2098
|
outputPrice: 0,
|
|
2047
2099
|
description: "SOTA performance with ~1500 tokens/s",
|
|
2048
2100
|
supportsReasoningEffort: true
|
|
2101
|
+
},
|
|
2102
|
+
"gpt-oss-120b": {
|
|
2103
|
+
maxTokens: 8e3,
|
|
2104
|
+
contextWindow: 64e3,
|
|
2105
|
+
supportsImages: false,
|
|
2106
|
+
supportsPromptCache: false,
|
|
2107
|
+
inputPrice: 0,
|
|
2108
|
+
outputPrice: 0,
|
|
2109
|
+
description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
|
|
2049
2110
|
}
|
|
2050
2111
|
};
|
|
2051
2112
|
|
|
@@ -2301,6 +2362,15 @@ var claudeCodeModels = {
|
|
|
2301
2362
|
supportsReasoningBudget: false,
|
|
2302
2363
|
requiredReasoningBudget: false
|
|
2303
2364
|
},
|
|
2365
|
+
"claude-opus-4-1-20250805": {
|
|
2366
|
+
...anthropicModels["claude-opus-4-1-20250805"],
|
|
2367
|
+
supportsImages: false,
|
|
2368
|
+
supportsPromptCache: true,
|
|
2369
|
+
// Claude Code does report cache tokens
|
|
2370
|
+
supportsReasoningEffort: false,
|
|
2371
|
+
supportsReasoningBudget: false,
|
|
2372
|
+
requiredReasoningBudget: false
|
|
2373
|
+
},
|
|
2304
2374
|
"claude-opus-4-20250514": {
|
|
2305
2375
|
...anthropicModels["claude-opus-4-20250514"],
|
|
2306
2376
|
supportsImages: false,
|
|
@@ -2774,6 +2844,24 @@ var groqModels = {
|
|
|
2774
2844
|
inputPrice: 1,
|
|
2775
2845
|
outputPrice: 3,
|
|
2776
2846
|
description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
|
|
2847
|
+
},
|
|
2848
|
+
"openai/gpt-oss-120b": {
|
|
2849
|
+
maxTokens: 32766,
|
|
2850
|
+
contextWindow: 131072,
|
|
2851
|
+
supportsImages: false,
|
|
2852
|
+
supportsPromptCache: false,
|
|
2853
|
+
inputPrice: 0.15,
|
|
2854
|
+
outputPrice: 0.75,
|
|
2855
|
+
description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
|
|
2856
|
+
},
|
|
2857
|
+
"openai/gpt-oss-20b": {
|
|
2858
|
+
maxTokens: 32768,
|
|
2859
|
+
contextWindow: 131072,
|
|
2860
|
+
supportsImages: false,
|
|
2861
|
+
supportsPromptCache: false,
|
|
2862
|
+
inputPrice: 0.1,
|
|
2863
|
+
outputPrice: 0.5,
|
|
2864
|
+
description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
|
|
2777
2865
|
}
|
|
2778
2866
|
};
|
|
2779
2867
|
|
|
@@ -2802,6 +2890,7 @@ var litellmDefaultModelInfo = {
|
|
|
2802
2890
|
};
|
|
2803
2891
|
var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
2804
2892
|
"claude-3-5-sonnet-latest",
|
|
2893
|
+
"claude-opus-4-1-20250805",
|
|
2805
2894
|
"claude-opus-4-20250514",
|
|
2806
2895
|
"claude-sonnet-4-20250514",
|
|
2807
2896
|
"claude-3-7-sonnet-latest",
|
|
@@ -2811,22 +2900,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
2811
2900
|
"vertex_ai/claude-3-5-sonnet-v2",
|
|
2812
2901
|
"vertex_ai/claude-3-5-sonnet-v2@20241022",
|
|
2813
2902
|
"vertex_ai/claude-3-7-sonnet@20250219",
|
|
2903
|
+
"vertex_ai/claude-opus-4-1@20250805",
|
|
2814
2904
|
"vertex_ai/claude-opus-4@20250514",
|
|
2815
2905
|
"vertex_ai/claude-sonnet-4@20250514",
|
|
2816
2906
|
"openrouter/anthropic/claude-3.5-sonnet",
|
|
2817
2907
|
"openrouter/anthropic/claude-3.5-sonnet:beta",
|
|
2818
2908
|
"openrouter/anthropic/claude-3.7-sonnet",
|
|
2819
2909
|
"openrouter/anthropic/claude-3.7-sonnet:beta",
|
|
2910
|
+
"anthropic.claude-opus-4-1-20250805-v1:0",
|
|
2820
2911
|
"anthropic.claude-opus-4-20250514-v1:0",
|
|
2821
2912
|
"anthropic.claude-sonnet-4-20250514-v1:0",
|
|
2822
2913
|
"anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
2823
2914
|
"anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
2824
2915
|
"us.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
2825
2916
|
"us.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
2917
|
+
"us.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
2826
2918
|
"us.anthropic.claude-opus-4-20250514-v1:0",
|
|
2827
2919
|
"us.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
2828
2920
|
"eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
|
|
2829
2921
|
"eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
|
|
2922
|
+
"eu.anthropic.claude-opus-4-1-20250805-v1:0",
|
|
2830
2923
|
"eu.anthropic.claude-opus-4-20250514-v1:0",
|
|
2831
2924
|
"eu.anthropic.claude-sonnet-4-20250514-v1:0",
|
|
2832
2925
|
"snowflake/claude-3-5-sonnet"
|
|
@@ -2963,8 +3056,48 @@ var ollamaDefaultModelInfo = {
|
|
|
2963
3056
|
};
|
|
2964
3057
|
|
|
2965
3058
|
// src/providers/openai.ts
|
|
2966
|
-
var openAiNativeDefaultModelId = "gpt-
|
|
3059
|
+
var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
|
|
2967
3060
|
var openAiNativeModels = {
|
|
3061
|
+
"gpt-5-2025-08-07": {
|
|
3062
|
+
maxTokens: 128e3,
|
|
3063
|
+
contextWindow: 4e5,
|
|
3064
|
+
supportsImages: true,
|
|
3065
|
+
supportsPromptCache: true,
|
|
3066
|
+
supportsReasoningEffort: true,
|
|
3067
|
+
reasoningEffort: "medium",
|
|
3068
|
+
inputPrice: 1.25,
|
|
3069
|
+
outputPrice: 10,
|
|
3070
|
+
cacheReadsPrice: 0.13,
|
|
3071
|
+
description: "GPT-5: The best model for coding and agentic tasks across domains",
|
|
3072
|
+
// supportsVerbosity is a new capability; ensure ModelInfo includes it
|
|
3073
|
+
supportsVerbosity: true
|
|
3074
|
+
},
|
|
3075
|
+
"gpt-5-mini-2025-08-07": {
|
|
3076
|
+
maxTokens: 128e3,
|
|
3077
|
+
contextWindow: 4e5,
|
|
3078
|
+
supportsImages: true,
|
|
3079
|
+
supportsPromptCache: true,
|
|
3080
|
+
supportsReasoningEffort: true,
|
|
3081
|
+
reasoningEffort: "medium",
|
|
3082
|
+
inputPrice: 0.25,
|
|
3083
|
+
outputPrice: 2,
|
|
3084
|
+
cacheReadsPrice: 0.03,
|
|
3085
|
+
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
|
|
3086
|
+
supportsVerbosity: true
|
|
3087
|
+
},
|
|
3088
|
+
"gpt-5-nano-2025-08-07": {
|
|
3089
|
+
maxTokens: 128e3,
|
|
3090
|
+
contextWindow: 4e5,
|
|
3091
|
+
supportsImages: true,
|
|
3092
|
+
supportsPromptCache: true,
|
|
3093
|
+
supportsReasoningEffort: true,
|
|
3094
|
+
reasoningEffort: "medium",
|
|
3095
|
+
inputPrice: 0.05,
|
|
3096
|
+
outputPrice: 0.4,
|
|
3097
|
+
cacheReadsPrice: 0.01,
|
|
3098
|
+
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
|
|
3099
|
+
supportsVerbosity: true
|
|
3100
|
+
},
|
|
2968
3101
|
"gpt-4.1": {
|
|
2969
3102
|
maxTokens: 32768,
|
|
2970
3103
|
contextWindow: 1047576,
|
|
@@ -3150,6 +3283,7 @@ var openAiModelInfoSaneDefaults = {
|
|
|
3150
3283
|
};
|
|
3151
3284
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
3152
3285
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
3286
|
+
var GPT5_DEFAULT_TEMPERATURE = 1;
|
|
3153
3287
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
3154
3288
|
|
|
3155
3289
|
// src/providers/openrouter.ts
|
|
@@ -3187,6 +3321,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
|
|
|
3187
3321
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3188
3322
|
"anthropic/claude-sonnet-4",
|
|
3189
3323
|
"anthropic/claude-opus-4",
|
|
3324
|
+
"anthropic/claude-opus-4.1",
|
|
3190
3325
|
"google/gemini-2.5-flash-preview",
|
|
3191
3326
|
"google/gemini-2.5-flash-preview:thinking",
|
|
3192
3327
|
"google/gemini-2.5-flash-preview-05-20",
|
|
@@ -3204,7 +3339,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
|
|
|
3204
3339
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3205
3340
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
3206
3341
|
"anthropic/claude-sonnet-4",
|
|
3207
|
-
"anthropic/claude-opus-4"
|
|
3342
|
+
"anthropic/claude-opus-4",
|
|
3343
|
+
"anthropic/claude-opus-4.1"
|
|
3208
3344
|
]);
|
|
3209
3345
|
var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3210
3346
|
"anthropic/claude-3.7-sonnet:thinking",
|
|
@@ -3214,6 +3350,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
|
3214
3350
|
var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
|
|
3215
3351
|
"anthropic/claude-3.7-sonnet:beta",
|
|
3216
3352
|
"anthropic/claude-opus-4",
|
|
3353
|
+
"anthropic/claude-opus-4.1",
|
|
3217
3354
|
"anthropic/claude-sonnet-4",
|
|
3218
3355
|
"google/gemini-2.5-pro-preview",
|
|
3219
3356
|
"google/gemini-2.5-pro",
|
|
@@ -3504,6 +3641,18 @@ var vertexModels = {
|
|
|
3504
3641
|
cacheReadsPrice: 0.3,
|
|
3505
3642
|
supportsReasoningBudget: true
|
|
3506
3643
|
},
|
|
3644
|
+
"claude-opus-4-1@20250805": {
|
|
3645
|
+
maxTokens: 8192,
|
|
3646
|
+
contextWindow: 2e5,
|
|
3647
|
+
supportsImages: true,
|
|
3648
|
+
supportsComputerUse: true,
|
|
3649
|
+
supportsPromptCache: true,
|
|
3650
|
+
inputPrice: 15,
|
|
3651
|
+
outputPrice: 75,
|
|
3652
|
+
cacheWritesPrice: 18.75,
|
|
3653
|
+
cacheReadsPrice: 1.5,
|
|
3654
|
+
supportsReasoningBudget: true
|
|
3655
|
+
},
|
|
3507
3656
|
"claude-opus-4@20250514": {
|
|
3508
3657
|
maxTokens: 8192,
|
|
3509
3658
|
contextWindow: 2e5,
|
|
@@ -4021,6 +4170,92 @@ var mainlandZAiModels = {
|
|
|
4021
4170
|
}
|
|
4022
4171
|
};
|
|
4023
4172
|
var ZAI_DEFAULT_TEMPERATURE = 0;
|
|
4173
|
+
|
|
4174
|
+
// src/providers/fireworks.ts
|
|
4175
|
+
var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
|
|
4176
|
+
var fireworksModels = {
|
|
4177
|
+
"accounts/fireworks/models/kimi-k2-instruct": {
|
|
4178
|
+
maxTokens: 16384,
|
|
4179
|
+
contextWindow: 128e3,
|
|
4180
|
+
supportsImages: false,
|
|
4181
|
+
supportsPromptCache: false,
|
|
4182
|
+
inputPrice: 0.6,
|
|
4183
|
+
outputPrice: 2.5,
|
|
4184
|
+
description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
|
|
4185
|
+
},
|
|
4186
|
+
"accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
|
|
4187
|
+
maxTokens: 32768,
|
|
4188
|
+
contextWindow: 256e3,
|
|
4189
|
+
supportsImages: false,
|
|
4190
|
+
supportsPromptCache: false,
|
|
4191
|
+
inputPrice: 0.22,
|
|
4192
|
+
outputPrice: 0.88,
|
|
4193
|
+
description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
|
|
4194
|
+
},
|
|
4195
|
+
"accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
|
|
4196
|
+
maxTokens: 32768,
|
|
4197
|
+
contextWindow: 256e3,
|
|
4198
|
+
supportsImages: false,
|
|
4199
|
+
supportsPromptCache: false,
|
|
4200
|
+
inputPrice: 0.45,
|
|
4201
|
+
outputPrice: 1.8,
|
|
4202
|
+
description: "Qwen3's most agentic code model to date."
|
|
4203
|
+
},
|
|
4204
|
+
"accounts/fireworks/models/deepseek-r1-0528": {
|
|
4205
|
+
maxTokens: 20480,
|
|
4206
|
+
contextWindow: 16e4,
|
|
4207
|
+
supportsImages: false,
|
|
4208
|
+
supportsPromptCache: false,
|
|
4209
|
+
inputPrice: 3,
|
|
4210
|
+
outputPrice: 8,
|
|
4211
|
+
description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4212
|
+
},
|
|
4213
|
+
"accounts/fireworks/models/deepseek-v3": {
|
|
4214
|
+
maxTokens: 16384,
|
|
4215
|
+
contextWindow: 128e3,
|
|
4216
|
+
supportsImages: false,
|
|
4217
|
+
supportsPromptCache: false,
|
|
4218
|
+
inputPrice: 0.9,
|
|
4219
|
+
outputPrice: 0.9,
|
|
4220
|
+
description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
|
|
4221
|
+
},
|
|
4222
|
+
"accounts/fireworks/models/glm-4p5": {
|
|
4223
|
+
maxTokens: 16384,
|
|
4224
|
+
contextWindow: 128e3,
|
|
4225
|
+
supportsImages: false,
|
|
4226
|
+
supportsPromptCache: false,
|
|
4227
|
+
inputPrice: 0.55,
|
|
4228
|
+
outputPrice: 2.19,
|
|
4229
|
+
description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4230
|
+
},
|
|
4231
|
+
"accounts/fireworks/models/glm-4p5-air": {
|
|
4232
|
+
maxTokens: 16384,
|
|
4233
|
+
contextWindow: 128e3,
|
|
4234
|
+
supportsImages: false,
|
|
4235
|
+
supportsPromptCache: false,
|
|
4236
|
+
inputPrice: 0.55,
|
|
4237
|
+
outputPrice: 2.19,
|
|
4238
|
+
description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
|
|
4239
|
+
},
|
|
4240
|
+
"accounts/fireworks/models/gpt-oss-20b": {
|
|
4241
|
+
maxTokens: 16384,
|
|
4242
|
+
contextWindow: 128e3,
|
|
4243
|
+
supportsImages: false,
|
|
4244
|
+
supportsPromptCache: false,
|
|
4245
|
+
inputPrice: 0.07,
|
|
4246
|
+
outputPrice: 0.3,
|
|
4247
|
+
description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
|
|
4248
|
+
},
|
|
4249
|
+
"accounts/fireworks/models/gpt-oss-120b": {
|
|
4250
|
+
maxTokens: 16384,
|
|
4251
|
+
contextWindow: 128e3,
|
|
4252
|
+
supportsImages: false,
|
|
4253
|
+
supportsPromptCache: false,
|
|
4254
|
+
inputPrice: 0.15,
|
|
4255
|
+
outputPrice: 0.6,
|
|
4256
|
+
description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
|
|
4257
|
+
}
|
|
4258
|
+
};
|
|
4024
4259
|
export {
|
|
4025
4260
|
ANTHROPIC_DEFAULT_MAX_TOKENS,
|
|
4026
4261
|
ANTHROPIC_STYLE_PROVIDERS,
|
|
@@ -4043,6 +4278,7 @@ export {
|
|
|
4043
4278
|
GLAMA_DEFAULT_TEMPERATURE,
|
|
4044
4279
|
GLOBAL_SETTINGS_KEYS,
|
|
4045
4280
|
GLOBAL_STATE_KEYS,
|
|
4281
|
+
GPT5_DEFAULT_TEMPERATURE,
|
|
4046
4282
|
HUGGINGFACE_API_URL,
|
|
4047
4283
|
HUGGINGFACE_CACHE_DURATION,
|
|
4048
4284
|
HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
|
|
@@ -4114,6 +4350,9 @@ export {
|
|
|
4114
4350
|
experimentIds,
|
|
4115
4351
|
experimentIdsSchema,
|
|
4116
4352
|
experimentsSchema,
|
|
4353
|
+
extendedReasoningEffortsSchema,
|
|
4354
|
+
fireworksDefaultModelId,
|
|
4355
|
+
fireworksModels,
|
|
4117
4356
|
followUpDataSchema,
|
|
4118
4357
|
geminiDefaultModelId,
|
|
4119
4358
|
geminiModels,
|
|
@@ -4208,6 +4447,8 @@ export {
|
|
|
4208
4447
|
toolUsageSchema,
|
|
4209
4448
|
unboundDefaultModelId,
|
|
4210
4449
|
unboundDefaultModelInfo,
|
|
4450
|
+
verbosityLevels,
|
|
4451
|
+
verbosityLevelsSchema,
|
|
4211
4452
|
vertexDefaultModelId,
|
|
4212
4453
|
vertexModels,
|
|
4213
4454
|
vscodeLlmDefaultModelId,
|