@roo-code/types 1.44.0 → 1.45.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -41,6 +41,7 @@ __export(index_exports, {
41
41
  GLAMA_DEFAULT_TEMPERATURE: () => GLAMA_DEFAULT_TEMPERATURE,
42
42
  GLOBAL_SETTINGS_KEYS: () => GLOBAL_SETTINGS_KEYS,
43
43
  GLOBAL_STATE_KEYS: () => GLOBAL_STATE_KEYS,
44
+ GPT5_DEFAULT_TEMPERATURE: () => GPT5_DEFAULT_TEMPERATURE,
44
45
  HUGGINGFACE_API_URL: () => HUGGINGFACE_API_URL,
45
46
  HUGGINGFACE_CACHE_DURATION: () => HUGGINGFACE_CACHE_DURATION,
46
47
  HUGGINGFACE_DEFAULT_CONTEXT_WINDOW: () => HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
@@ -112,6 +113,9 @@ __export(index_exports, {
112
113
  experimentIds: () => experimentIds,
113
114
  experimentIdsSchema: () => experimentIdsSchema,
114
115
  experimentsSchema: () => experimentsSchema,
116
+ extendedReasoningEffortsSchema: () => extendedReasoningEffortsSchema,
117
+ fireworksDefaultModelId: () => fireworksDefaultModelId,
118
+ fireworksModels: () => fireworksModels,
115
119
  followUpDataSchema: () => followUpDataSchema,
116
120
  geminiDefaultModelId: () => geminiDefaultModelId,
117
121
  geminiModels: () => geminiModels,
@@ -206,6 +210,8 @@ __export(index_exports, {
206
210
  toolUsageSchema: () => toolUsageSchema,
207
211
  unboundDefaultModelId: () => unboundDefaultModelId,
208
212
  unboundDefaultModelInfo: () => unboundDefaultModelInfo,
213
+ verbosityLevels: () => verbosityLevels,
214
+ verbosityLevelsSchema: () => verbosityLevelsSchema,
209
215
  vertexDefaultModelId: () => vertexDefaultModelId,
210
216
  vertexModels: () => vertexModels,
211
217
  vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
@@ -228,6 +234,8 @@ var import_zod3 = require("zod");
228
234
  var import_zod = require("zod");
229
235
  var reasoningEfforts = ["low", "medium", "high"];
230
236
  var reasoningEffortsSchema = import_zod.z.enum(reasoningEfforts);
237
+ var verbosityLevels = ["low", "medium", "high"];
238
+ var verbosityLevelsSchema = import_zod.z.enum(verbosityLevels);
231
239
  var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
232
240
  var modelParametersSchema = import_zod.z.enum(modelParameters);
233
241
  var isModelParameter = (value) => modelParameters.includes(value);
@@ -238,6 +246,8 @@ var modelInfoSchema = import_zod.z.object({
238
246
  supportsImages: import_zod.z.boolean().optional(),
239
247
  supportsComputerUse: import_zod.z.boolean().optional(),
240
248
  supportsPromptCache: import_zod.z.boolean(),
249
+ // Capability flag to indicate whether the model supports an output verbosity parameter
250
+ supportsVerbosity: import_zod.z.boolean().optional(),
241
251
  supportsReasoningBudget: import_zod.z.boolean().optional(),
242
252
  requiredReasoningBudget: import_zod.z.boolean().optional(),
243
253
  supportsReasoningEffort: import_zod.z.boolean().optional(),
@@ -305,6 +315,7 @@ var codebaseIndexProviderSchema = import_zod2.z.object({
305
315
  });
306
316
 
307
317
  // src/provider-settings.ts
318
+ var extendedReasoningEffortsSchema = import_zod3.z.union([reasoningEffortsSchema, import_zod3.z.literal("minimal")]);
308
319
  var providerNames = [
309
320
  "anthropic",
310
321
  "claude-code",
@@ -334,7 +345,8 @@ var providerNames = [
334
345
  "huggingface",
335
346
  "cerebras",
336
347
  "sambanova",
337
- "zai"
348
+ "zai",
349
+ "fireworks"
338
350
  ];
339
351
  var providerNamesSchema = import_zod3.z.enum(providerNames);
340
352
  var providerSettingsEntrySchema = import_zod3.z.object({
@@ -353,9 +365,11 @@ var baseProviderSettingsSchema = import_zod3.z.object({
353
365
  consecutiveMistakeLimit: import_zod3.z.number().min(0).optional(),
354
366
  // Model reasoning.
355
367
  enableReasoningEffort: import_zod3.z.boolean().optional(),
356
- reasoningEffort: reasoningEffortsSchema.optional(),
368
+ reasoningEffort: extendedReasoningEffortsSchema.optional(),
357
369
  modelMaxTokens: import_zod3.z.number().optional(),
358
- modelMaxThinkingTokens: import_zod3.z.number().optional()
370
+ modelMaxThinkingTokens: import_zod3.z.number().optional(),
371
+ // Model verbosity.
372
+ verbosity: verbosityLevelsSchema.optional()
359
373
  });
360
374
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
361
375
  apiModelId: import_zod3.z.string().optional()
@@ -506,6 +520,9 @@ var zaiSchema = apiModelIdProviderModelSchema.extend({
506
520
  zaiApiKey: import_zod3.z.string().optional(),
507
521
  zaiApiLine: import_zod3.z.union([import_zod3.z.literal("china"), import_zod3.z.literal("international")]).optional()
508
522
  });
523
+ var fireworksSchema = apiModelIdProviderModelSchema.extend({
524
+ fireworksApiKey: import_zod3.z.string().optional()
525
+ });
509
526
  var defaultSchema = import_zod3.z.object({
510
527
  apiProvider: import_zod3.z.undefined()
511
528
  });
@@ -539,6 +556,7 @@ var providerSettingsSchemaDiscriminated = import_zod3.z.discriminatedUnion("apiP
539
556
  cerebrasSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("cerebras") })),
540
557
  sambaNovaSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("sambanova") })),
541
558
  zaiSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("zai") })),
559
+ fireworksSchema.merge(import_zod3.z.object({ apiProvider: import_zod3.z.literal("fireworks") })),
542
560
  defaultSchema
543
561
  ]);
544
562
  var providerSettingsSchema = import_zod3.z.object({
@@ -572,6 +590,7 @@ var providerSettingsSchema = import_zod3.z.object({
572
590
  ...cerebrasSchema.shape,
573
591
  ...sambaNovaSchema.shape,
574
592
  ...zaiSchema.shape,
593
+ ...fireworksSchema.shape,
575
594
  ...codebaseIndexProviderSchema.shape
576
595
  });
577
596
  var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: import_zod3.z.string().optional() });
@@ -719,7 +738,14 @@ var clineMessageSchema = import_zod6.z.object({
719
738
  progressStatus: toolProgressStatusSchema.optional(),
720
739
  contextCondense: contextCondenseSchema.optional(),
721
740
  isProtected: import_zod6.z.boolean().optional(),
722
- apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional()
741
+ apiProtocol: import_zod6.z.union([import_zod6.z.literal("openai"), import_zod6.z.literal("anthropic")]).optional(),
742
+ metadata: import_zod6.z.object({
743
+ gpt5: import_zod6.z.object({
744
+ previous_response_id: import_zod6.z.string().optional(),
745
+ instructions: import_zod6.z.string().optional(),
746
+ reasoning_summary: import_zod6.z.string().optional()
747
+ }).optional()
748
+ }).optional()
723
749
  });
724
750
  var tokenUsageSchema = import_zod6.z.object({
725
751
  totalTokensIn: import_zod6.z.number(),
@@ -1202,7 +1228,8 @@ var SECRET_STATE_KEYS = [
1202
1228
  "codebaseIndexGeminiApiKey",
1203
1229
  "codebaseIndexMistralApiKey",
1204
1230
  "huggingFaceApiKey",
1205
- "sambaNovaApiKey"
1231
+ "sambaNovaApiKey",
1232
+ "fireworksApiKey"
1206
1233
  ];
1207
1234
  var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
1208
1235
  var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
@@ -1711,6 +1738,22 @@ var anthropicModels = {
1711
1738
  // $0.30 per million tokens
1712
1739
  supportsReasoningBudget: true
1713
1740
  },
1741
+ "claude-opus-4-1-20250805": {
1742
+ maxTokens: 8192,
1743
+ contextWindow: 2e5,
1744
+ supportsImages: true,
1745
+ supportsComputerUse: true,
1746
+ supportsPromptCache: true,
1747
+ inputPrice: 15,
1748
+ // $15 per million input tokens
1749
+ outputPrice: 75,
1750
+ // $75 per million output tokens
1751
+ cacheWritesPrice: 18.75,
1752
+ // $18.75 per million tokens
1753
+ cacheReadsPrice: 1.5,
1754
+ // $1.50 per million tokens
1755
+ supportsReasoningBudget: true
1756
+ },
1714
1757
  "claude-opus-4-20250514": {
1715
1758
  maxTokens: 32e3,
1716
1759
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -1891,6 +1934,21 @@ var bedrockModels = {
1891
1934
  maxCachePoints: 4,
1892
1935
  cachableFields: ["system", "messages", "tools"]
1893
1936
  },
1937
+ "anthropic.claude-opus-4-1-20250805-v1:0": {
1938
+ maxTokens: 8192,
1939
+ contextWindow: 2e5,
1940
+ supportsImages: true,
1941
+ supportsComputerUse: true,
1942
+ supportsPromptCache: true,
1943
+ supportsReasoningBudget: true,
1944
+ inputPrice: 15,
1945
+ outputPrice: 75,
1946
+ cacheWritesPrice: 18.75,
1947
+ cacheReadsPrice: 1.5,
1948
+ minTokensPerCachePoint: 1024,
1949
+ maxCachePoints: 4,
1950
+ cachableFields: ["system", "messages", "tools"]
1951
+ },
1894
1952
  "anthropic.claude-opus-4-20250514-v1:0": {
1895
1953
  maxTokens: 8192,
1896
1954
  contextWindow: 2e5,
@@ -2263,6 +2321,15 @@ var cerebrasModels = {
2263
2321
  outputPrice: 0,
2264
2322
  description: "SOTA performance with ~1500 tokens/s",
2265
2323
  supportsReasoningEffort: true
2324
+ },
2325
+ "gpt-oss-120b": {
2326
+ maxTokens: 8e3,
2327
+ contextWindow: 64e3,
2328
+ supportsImages: false,
2329
+ supportsPromptCache: false,
2330
+ inputPrice: 0,
2331
+ outputPrice: 0,
2332
+ description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
2266
2333
  }
2267
2334
  };
2268
2335
 
@@ -2518,6 +2585,15 @@ var claudeCodeModels = {
2518
2585
  supportsReasoningBudget: false,
2519
2586
  requiredReasoningBudget: false
2520
2587
  },
2588
+ "claude-opus-4-1-20250805": {
2589
+ ...anthropicModels["claude-opus-4-1-20250805"],
2590
+ supportsImages: false,
2591
+ supportsPromptCache: true,
2592
+ // Claude Code does report cache tokens
2593
+ supportsReasoningEffort: false,
2594
+ supportsReasoningBudget: false,
2595
+ requiredReasoningBudget: false
2596
+ },
2521
2597
  "claude-opus-4-20250514": {
2522
2598
  ...anthropicModels["claude-opus-4-20250514"],
2523
2599
  supportsImages: false,
@@ -2991,6 +3067,24 @@ var groqModels = {
2991
3067
  inputPrice: 1,
2992
3068
  outputPrice: 3,
2993
3069
  description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
3070
+ },
3071
+ "openai/gpt-oss-120b": {
3072
+ maxTokens: 32766,
3073
+ contextWindow: 131072,
3074
+ supportsImages: false,
3075
+ supportsPromptCache: false,
3076
+ inputPrice: 0.15,
3077
+ outputPrice: 0.75,
3078
+ description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
3079
+ },
3080
+ "openai/gpt-oss-20b": {
3081
+ maxTokens: 32768,
3082
+ contextWindow: 131072,
3083
+ supportsImages: false,
3084
+ supportsPromptCache: false,
3085
+ inputPrice: 0.1,
3086
+ outputPrice: 0.5,
3087
+ description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
2994
3088
  }
2995
3089
  };
2996
3090
 
@@ -3019,6 +3113,7 @@ var litellmDefaultModelInfo = {
3019
3113
  };
3020
3114
  var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3021
3115
  "claude-3-5-sonnet-latest",
3116
+ "claude-opus-4-1-20250805",
3022
3117
  "claude-opus-4-20250514",
3023
3118
  "claude-sonnet-4-20250514",
3024
3119
  "claude-3-7-sonnet-latest",
@@ -3028,22 +3123,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3028
3123
  "vertex_ai/claude-3-5-sonnet-v2",
3029
3124
  "vertex_ai/claude-3-5-sonnet-v2@20241022",
3030
3125
  "vertex_ai/claude-3-7-sonnet@20250219",
3126
+ "vertex_ai/claude-opus-4-1@20250805",
3031
3127
  "vertex_ai/claude-opus-4@20250514",
3032
3128
  "vertex_ai/claude-sonnet-4@20250514",
3033
3129
  "openrouter/anthropic/claude-3.5-sonnet",
3034
3130
  "openrouter/anthropic/claude-3.5-sonnet:beta",
3035
3131
  "openrouter/anthropic/claude-3.7-sonnet",
3036
3132
  "openrouter/anthropic/claude-3.7-sonnet:beta",
3133
+ "anthropic.claude-opus-4-1-20250805-v1:0",
3037
3134
  "anthropic.claude-opus-4-20250514-v1:0",
3038
3135
  "anthropic.claude-sonnet-4-20250514-v1:0",
3039
3136
  "anthropic.claude-3-7-sonnet-20250219-v1:0",
3040
3137
  "anthropic.claude-3-5-sonnet-20241022-v2:0",
3041
3138
  "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
3042
3139
  "us.anthropic.claude-3-7-sonnet-20250219-v1:0",
3140
+ "us.anthropic.claude-opus-4-1-20250805-v1:0",
3043
3141
  "us.anthropic.claude-opus-4-20250514-v1:0",
3044
3142
  "us.anthropic.claude-sonnet-4-20250514-v1:0",
3045
3143
  "eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
3046
3144
  "eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
3145
+ "eu.anthropic.claude-opus-4-1-20250805-v1:0",
3047
3146
  "eu.anthropic.claude-opus-4-20250514-v1:0",
3048
3147
  "eu.anthropic.claude-sonnet-4-20250514-v1:0",
3049
3148
  "snowflake/claude-3-5-sonnet"
@@ -3180,8 +3279,48 @@ var ollamaDefaultModelInfo = {
3180
3279
  };
3181
3280
 
3182
3281
  // src/providers/openai.ts
3183
- var openAiNativeDefaultModelId = "gpt-4.1";
3282
+ var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
3184
3283
  var openAiNativeModels = {
3284
+ "gpt-5-2025-08-07": {
3285
+ maxTokens: 128e3,
3286
+ contextWindow: 4e5,
3287
+ supportsImages: true,
3288
+ supportsPromptCache: true,
3289
+ supportsReasoningEffort: true,
3290
+ reasoningEffort: "medium",
3291
+ inputPrice: 1.25,
3292
+ outputPrice: 10,
3293
+ cacheReadsPrice: 0.13,
3294
+ description: "GPT-5: The best model for coding and agentic tasks across domains",
3295
+ // supportsVerbosity is a new capability; ensure ModelInfo includes it
3296
+ supportsVerbosity: true
3297
+ },
3298
+ "gpt-5-mini-2025-08-07": {
3299
+ maxTokens: 128e3,
3300
+ contextWindow: 4e5,
3301
+ supportsImages: true,
3302
+ supportsPromptCache: true,
3303
+ supportsReasoningEffort: true,
3304
+ reasoningEffort: "medium",
3305
+ inputPrice: 0.25,
3306
+ outputPrice: 2,
3307
+ cacheReadsPrice: 0.03,
3308
+ description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
3309
+ supportsVerbosity: true
3310
+ },
3311
+ "gpt-5-nano-2025-08-07": {
3312
+ maxTokens: 128e3,
3313
+ contextWindow: 4e5,
3314
+ supportsImages: true,
3315
+ supportsPromptCache: true,
3316
+ supportsReasoningEffort: true,
3317
+ reasoningEffort: "medium",
3318
+ inputPrice: 0.05,
3319
+ outputPrice: 0.4,
3320
+ cacheReadsPrice: 0.01,
3321
+ description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
3322
+ supportsVerbosity: true
3323
+ },
3185
3324
  "gpt-4.1": {
3186
3325
  maxTokens: 32768,
3187
3326
  contextWindow: 1047576,
@@ -3367,6 +3506,7 @@ var openAiModelInfoSaneDefaults = {
3367
3506
  };
3368
3507
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3369
3508
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3509
+ var GPT5_DEFAULT_TEMPERATURE = 1;
3370
3510
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3371
3511
 
3372
3512
  // src/providers/openrouter.ts
@@ -3404,6 +3544,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3404
3544
  "anthropic/claude-3.7-sonnet:thinking",
3405
3545
  "anthropic/claude-sonnet-4",
3406
3546
  "anthropic/claude-opus-4",
3547
+ "anthropic/claude-opus-4.1",
3407
3548
  "google/gemini-2.5-flash-preview",
3408
3549
  "google/gemini-2.5-flash-preview:thinking",
3409
3550
  "google/gemini-2.5-flash-preview-05-20",
@@ -3421,7 +3562,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3421
3562
  "anthropic/claude-3.7-sonnet:beta",
3422
3563
  "anthropic/claude-3.7-sonnet:thinking",
3423
3564
  "anthropic/claude-sonnet-4",
3424
- "anthropic/claude-opus-4"
3565
+ "anthropic/claude-opus-4",
3566
+ "anthropic/claude-opus-4.1"
3425
3567
  ]);
3426
3568
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3427
3569
  "anthropic/claude-3.7-sonnet:thinking",
@@ -3431,6 +3573,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3431
3573
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3432
3574
  "anthropic/claude-3.7-sonnet:beta",
3433
3575
  "anthropic/claude-opus-4",
3576
+ "anthropic/claude-opus-4.1",
3434
3577
  "anthropic/claude-sonnet-4",
3435
3578
  "google/gemini-2.5-pro-preview",
3436
3579
  "google/gemini-2.5-pro",
@@ -3721,6 +3864,18 @@ var vertexModels = {
3721
3864
  cacheReadsPrice: 0.3,
3722
3865
  supportsReasoningBudget: true
3723
3866
  },
3867
+ "claude-opus-4-1@20250805": {
3868
+ maxTokens: 8192,
3869
+ contextWindow: 2e5,
3870
+ supportsImages: true,
3871
+ supportsComputerUse: true,
3872
+ supportsPromptCache: true,
3873
+ inputPrice: 15,
3874
+ outputPrice: 75,
3875
+ cacheWritesPrice: 18.75,
3876
+ cacheReadsPrice: 1.5,
3877
+ supportsReasoningBudget: true
3878
+ },
3724
3879
  "claude-opus-4@20250514": {
3725
3880
  maxTokens: 8192,
3726
3881
  contextWindow: 2e5,
@@ -4238,6 +4393,92 @@ var mainlandZAiModels = {
4238
4393
  }
4239
4394
  };
4240
4395
  var ZAI_DEFAULT_TEMPERATURE = 0;
4396
+
4397
+ // src/providers/fireworks.ts
4398
+ var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
4399
+ var fireworksModels = {
4400
+ "accounts/fireworks/models/kimi-k2-instruct": {
4401
+ maxTokens: 16384,
4402
+ contextWindow: 128e3,
4403
+ supportsImages: false,
4404
+ supportsPromptCache: false,
4405
+ inputPrice: 0.6,
4406
+ outputPrice: 2.5,
4407
+ description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
4408
+ },
4409
+ "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
4410
+ maxTokens: 32768,
4411
+ contextWindow: 256e3,
4412
+ supportsImages: false,
4413
+ supportsPromptCache: false,
4414
+ inputPrice: 0.22,
4415
+ outputPrice: 0.88,
4416
+ description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
4417
+ },
4418
+ "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
4419
+ maxTokens: 32768,
4420
+ contextWindow: 256e3,
4421
+ supportsImages: false,
4422
+ supportsPromptCache: false,
4423
+ inputPrice: 0.45,
4424
+ outputPrice: 1.8,
4425
+ description: "Qwen3's most agentic code model to date."
4426
+ },
4427
+ "accounts/fireworks/models/deepseek-r1-0528": {
4428
+ maxTokens: 20480,
4429
+ contextWindow: 16e4,
4430
+ supportsImages: false,
4431
+ supportsPromptCache: false,
4432
+ inputPrice: 3,
4433
+ outputPrice: 8,
4434
+ description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4435
+ },
4436
+ "accounts/fireworks/models/deepseek-v3": {
4437
+ maxTokens: 16384,
4438
+ contextWindow: 128e3,
4439
+ supportsImages: false,
4440
+ supportsPromptCache: false,
4441
+ inputPrice: 0.9,
4442
+ outputPrice: 0.9,
4443
+ description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4444
+ },
4445
+ "accounts/fireworks/models/glm-4p5": {
4446
+ maxTokens: 16384,
4447
+ contextWindow: 128e3,
4448
+ supportsImages: false,
4449
+ supportsPromptCache: false,
4450
+ inputPrice: 0.55,
4451
+ outputPrice: 2.19,
4452
+ description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4453
+ },
4454
+ "accounts/fireworks/models/glm-4p5-air": {
4455
+ maxTokens: 16384,
4456
+ contextWindow: 128e3,
4457
+ supportsImages: false,
4458
+ supportsPromptCache: false,
4459
+ inputPrice: 0.55,
4460
+ outputPrice: 2.19,
4461
+ description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4462
+ },
4463
+ "accounts/fireworks/models/gpt-oss-20b": {
4464
+ maxTokens: 16384,
4465
+ contextWindow: 128e3,
4466
+ supportsImages: false,
4467
+ supportsPromptCache: false,
4468
+ inputPrice: 0.07,
4469
+ outputPrice: 0.3,
4470
+ description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
4471
+ },
4472
+ "accounts/fireworks/models/gpt-oss-120b": {
4473
+ maxTokens: 16384,
4474
+ contextWindow: 128e3,
4475
+ supportsImages: false,
4476
+ supportsPromptCache: false,
4477
+ inputPrice: 0.15,
4478
+ outputPrice: 0.6,
4479
+ description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
4480
+ }
4481
+ };
4241
4482
  // Annotate the CommonJS export names for ESM import in node:
4242
4483
  0 && (module.exports = {
4243
4484
  ANTHROPIC_DEFAULT_MAX_TOKENS,
@@ -4261,6 +4502,7 @@ var ZAI_DEFAULT_TEMPERATURE = 0;
4261
4502
  GLAMA_DEFAULT_TEMPERATURE,
4262
4503
  GLOBAL_SETTINGS_KEYS,
4263
4504
  GLOBAL_STATE_KEYS,
4505
+ GPT5_DEFAULT_TEMPERATURE,
4264
4506
  HUGGINGFACE_API_URL,
4265
4507
  HUGGINGFACE_CACHE_DURATION,
4266
4508
  HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
@@ -4332,6 +4574,9 @@ var ZAI_DEFAULT_TEMPERATURE = 0;
4332
4574
  experimentIds,
4333
4575
  experimentIdsSchema,
4334
4576
  experimentsSchema,
4577
+ extendedReasoningEffortsSchema,
4578
+ fireworksDefaultModelId,
4579
+ fireworksModels,
4335
4580
  followUpDataSchema,
4336
4581
  geminiDefaultModelId,
4337
4582
  geminiModels,
@@ -4426,6 +4671,8 @@ var ZAI_DEFAULT_TEMPERATURE = 0;
4426
4671
  toolUsageSchema,
4427
4672
  unboundDefaultModelId,
4428
4673
  unboundDefaultModelInfo,
4674
+ verbosityLevels,
4675
+ verbosityLevelsSchema,
4429
4676
  vertexDefaultModelId,
4430
4677
  vertexModels,
4431
4678
  vscodeLlmDefaultModelId,