@roo-code/types 1.44.0 → 1.45.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -11,6 +11,8 @@ import { z as z3 } from "zod";
11
11
  import { z } from "zod";
12
12
  var reasoningEfforts = ["low", "medium", "high"];
13
13
  var reasoningEffortsSchema = z.enum(reasoningEfforts);
14
+ var verbosityLevels = ["low", "medium", "high"];
15
+ var verbosityLevelsSchema = z.enum(verbosityLevels);
14
16
  var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
15
17
  var modelParametersSchema = z.enum(modelParameters);
16
18
  var isModelParameter = (value) => modelParameters.includes(value);
@@ -21,6 +23,8 @@ var modelInfoSchema = z.object({
21
23
  supportsImages: z.boolean().optional(),
22
24
  supportsComputerUse: z.boolean().optional(),
23
25
  supportsPromptCache: z.boolean(),
26
+ // Capability flag to indicate whether the model supports an output verbosity parameter
27
+ supportsVerbosity: z.boolean().optional(),
24
28
  supportsReasoningBudget: z.boolean().optional(),
25
29
  requiredReasoningBudget: z.boolean().optional(),
26
30
  supportsReasoningEffort: z.boolean().optional(),
@@ -88,6 +92,7 @@ var codebaseIndexProviderSchema = z2.object({
88
92
  });
89
93
 
90
94
  // src/provider-settings.ts
95
+ var extendedReasoningEffortsSchema = z3.union([reasoningEffortsSchema, z3.literal("minimal")]);
91
96
  var providerNames = [
92
97
  "anthropic",
93
98
  "claude-code",
@@ -117,7 +122,8 @@ var providerNames = [
117
122
  "huggingface",
118
123
  "cerebras",
119
124
  "sambanova",
120
- "zai"
125
+ "zai",
126
+ "fireworks"
121
127
  ];
122
128
  var providerNamesSchema = z3.enum(providerNames);
123
129
  var providerSettingsEntrySchema = z3.object({
@@ -136,9 +142,11 @@ var baseProviderSettingsSchema = z3.object({
136
142
  consecutiveMistakeLimit: z3.number().min(0).optional(),
137
143
  // Model reasoning.
138
144
  enableReasoningEffort: z3.boolean().optional(),
139
- reasoningEffort: reasoningEffortsSchema.optional(),
145
+ reasoningEffort: extendedReasoningEffortsSchema.optional(),
140
146
  modelMaxTokens: z3.number().optional(),
141
- modelMaxThinkingTokens: z3.number().optional()
147
+ modelMaxThinkingTokens: z3.number().optional(),
148
+ // Model verbosity.
149
+ verbosity: verbosityLevelsSchema.optional()
142
150
  });
143
151
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
144
152
  apiModelId: z3.string().optional()
@@ -289,6 +297,9 @@ var zaiSchema = apiModelIdProviderModelSchema.extend({
289
297
  zaiApiKey: z3.string().optional(),
290
298
  zaiApiLine: z3.union([z3.literal("china"), z3.literal("international")]).optional()
291
299
  });
300
+ var fireworksSchema = apiModelIdProviderModelSchema.extend({
301
+ fireworksApiKey: z3.string().optional()
302
+ });
292
303
  var defaultSchema = z3.object({
293
304
  apiProvider: z3.undefined()
294
305
  });
@@ -322,6 +333,7 @@ var providerSettingsSchemaDiscriminated = z3.discriminatedUnion("apiProvider", [
322
333
  cerebrasSchema.merge(z3.object({ apiProvider: z3.literal("cerebras") })),
323
334
  sambaNovaSchema.merge(z3.object({ apiProvider: z3.literal("sambanova") })),
324
335
  zaiSchema.merge(z3.object({ apiProvider: z3.literal("zai") })),
336
+ fireworksSchema.merge(z3.object({ apiProvider: z3.literal("fireworks") })),
325
337
  defaultSchema
326
338
  ]);
327
339
  var providerSettingsSchema = z3.object({
@@ -355,6 +367,7 @@ var providerSettingsSchema = z3.object({
355
367
  ...cerebrasSchema.shape,
356
368
  ...sambaNovaSchema.shape,
357
369
  ...zaiSchema.shape,
370
+ ...fireworksSchema.shape,
358
371
  ...codebaseIndexProviderSchema.shape
359
372
  });
360
373
  var providerSettingsWithIdSchema = providerSettingsSchema.extend({ id: z3.string().optional() });
@@ -502,7 +515,14 @@ var clineMessageSchema = z6.object({
502
515
  progressStatus: toolProgressStatusSchema.optional(),
503
516
  contextCondense: contextCondenseSchema.optional(),
504
517
  isProtected: z6.boolean().optional(),
505
- apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional()
518
+ apiProtocol: z6.union([z6.literal("openai"), z6.literal("anthropic")]).optional(),
519
+ metadata: z6.object({
520
+ gpt5: z6.object({
521
+ previous_response_id: z6.string().optional(),
522
+ instructions: z6.string().optional(),
523
+ reasoning_summary: z6.string().optional()
524
+ }).optional()
525
+ }).optional()
506
526
  });
507
527
  var tokenUsageSchema = z6.object({
508
528
  totalTokensIn: z6.number(),
@@ -985,7 +1005,8 @@ var SECRET_STATE_KEYS = [
985
1005
  "codebaseIndexGeminiApiKey",
986
1006
  "codebaseIndexMistralApiKey",
987
1007
  "huggingFaceApiKey",
988
- "sambaNovaApiKey"
1008
+ "sambaNovaApiKey",
1009
+ "fireworksApiKey"
989
1010
  ];
990
1011
  var isSecretStateKey = (key) => SECRET_STATE_KEYS.includes(key);
991
1012
  var GLOBAL_STATE_KEYS = [...GLOBAL_SETTINGS_KEYS, ...PROVIDER_SETTINGS_KEYS].filter(
@@ -1494,6 +1515,22 @@ var anthropicModels = {
1494
1515
  // $0.30 per million tokens
1495
1516
  supportsReasoningBudget: true
1496
1517
  },
1518
+ "claude-opus-4-1-20250805": {
1519
+ maxTokens: 8192,
1520
+ contextWindow: 2e5,
1521
+ supportsImages: true,
1522
+ supportsComputerUse: true,
1523
+ supportsPromptCache: true,
1524
+ inputPrice: 15,
1525
+ // $15 per million input tokens
1526
+ outputPrice: 75,
1527
+ // $75 per million output tokens
1528
+ cacheWritesPrice: 18.75,
1529
+ // $18.75 per million tokens
1530
+ cacheReadsPrice: 1.5,
1531
+ // $1.50 per million tokens
1532
+ supportsReasoningBudget: true
1533
+ },
1497
1534
  "claude-opus-4-20250514": {
1498
1535
  maxTokens: 32e3,
1499
1536
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -1674,6 +1711,21 @@ var bedrockModels = {
1674
1711
  maxCachePoints: 4,
1675
1712
  cachableFields: ["system", "messages", "tools"]
1676
1713
  },
1714
+ "anthropic.claude-opus-4-1-20250805-v1:0": {
1715
+ maxTokens: 8192,
1716
+ contextWindow: 2e5,
1717
+ supportsImages: true,
1718
+ supportsComputerUse: true,
1719
+ supportsPromptCache: true,
1720
+ supportsReasoningBudget: true,
1721
+ inputPrice: 15,
1722
+ outputPrice: 75,
1723
+ cacheWritesPrice: 18.75,
1724
+ cacheReadsPrice: 1.5,
1725
+ minTokensPerCachePoint: 1024,
1726
+ maxCachePoints: 4,
1727
+ cachableFields: ["system", "messages", "tools"]
1728
+ },
1677
1729
  "anthropic.claude-opus-4-20250514-v1:0": {
1678
1730
  maxTokens: 8192,
1679
1731
  contextWindow: 2e5,
@@ -2046,6 +2098,15 @@ var cerebrasModels = {
2046
2098
  outputPrice: 0,
2047
2099
  description: "SOTA performance with ~1500 tokens/s",
2048
2100
  supportsReasoningEffort: true
2101
+ },
2102
+ "gpt-oss-120b": {
2103
+ maxTokens: 8e3,
2104
+ contextWindow: 64e3,
2105
+ supportsImages: false,
2106
+ supportsPromptCache: false,
2107
+ inputPrice: 0,
2108
+ outputPrice: 0,
2109
+ description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
2049
2110
  }
2050
2111
  };
2051
2112
 
@@ -2301,6 +2362,15 @@ var claudeCodeModels = {
2301
2362
  supportsReasoningBudget: false,
2302
2363
  requiredReasoningBudget: false
2303
2364
  },
2365
+ "claude-opus-4-1-20250805": {
2366
+ ...anthropicModels["claude-opus-4-1-20250805"],
2367
+ supportsImages: false,
2368
+ supportsPromptCache: true,
2369
+ // Claude Code does report cache tokens
2370
+ supportsReasoningEffort: false,
2371
+ supportsReasoningBudget: false,
2372
+ requiredReasoningBudget: false
2373
+ },
2304
2374
  "claude-opus-4-20250514": {
2305
2375
  ...anthropicModels["claude-opus-4-20250514"],
2306
2376
  supportsImages: false,
@@ -2774,6 +2844,24 @@ var groqModels = {
2774
2844
  inputPrice: 1,
2775
2845
  outputPrice: 3,
2776
2846
  description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
2847
+ },
2848
+ "openai/gpt-oss-120b": {
2849
+ maxTokens: 32766,
2850
+ contextWindow: 131072,
2851
+ supportsImages: false,
2852
+ supportsPromptCache: false,
2853
+ inputPrice: 0.15,
2854
+ outputPrice: 0.75,
2855
+ description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
2856
+ },
2857
+ "openai/gpt-oss-20b": {
2858
+ maxTokens: 32768,
2859
+ contextWindow: 131072,
2860
+ supportsImages: false,
2861
+ supportsPromptCache: false,
2862
+ inputPrice: 0.1,
2863
+ outputPrice: 0.5,
2864
+ description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
2777
2865
  }
2778
2866
  };
2779
2867
 
@@ -2802,6 +2890,7 @@ var litellmDefaultModelInfo = {
2802
2890
  };
2803
2891
  var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2804
2892
  "claude-3-5-sonnet-latest",
2893
+ "claude-opus-4-1-20250805",
2805
2894
  "claude-opus-4-20250514",
2806
2895
  "claude-sonnet-4-20250514",
2807
2896
  "claude-3-7-sonnet-latest",
@@ -2811,22 +2900,26 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2811
2900
  "vertex_ai/claude-3-5-sonnet-v2",
2812
2901
  "vertex_ai/claude-3-5-sonnet-v2@20241022",
2813
2902
  "vertex_ai/claude-3-7-sonnet@20250219",
2903
+ "vertex_ai/claude-opus-4-1@20250805",
2814
2904
  "vertex_ai/claude-opus-4@20250514",
2815
2905
  "vertex_ai/claude-sonnet-4@20250514",
2816
2906
  "openrouter/anthropic/claude-3.5-sonnet",
2817
2907
  "openrouter/anthropic/claude-3.5-sonnet:beta",
2818
2908
  "openrouter/anthropic/claude-3.7-sonnet",
2819
2909
  "openrouter/anthropic/claude-3.7-sonnet:beta",
2910
+ "anthropic.claude-opus-4-1-20250805-v1:0",
2820
2911
  "anthropic.claude-opus-4-20250514-v1:0",
2821
2912
  "anthropic.claude-sonnet-4-20250514-v1:0",
2822
2913
  "anthropic.claude-3-7-sonnet-20250219-v1:0",
2823
2914
  "anthropic.claude-3-5-sonnet-20241022-v2:0",
2824
2915
  "us.anthropic.claude-3-5-sonnet-20241022-v2:0",
2825
2916
  "us.anthropic.claude-3-7-sonnet-20250219-v1:0",
2917
+ "us.anthropic.claude-opus-4-1-20250805-v1:0",
2826
2918
  "us.anthropic.claude-opus-4-20250514-v1:0",
2827
2919
  "us.anthropic.claude-sonnet-4-20250514-v1:0",
2828
2920
  "eu.anthropic.claude-3-5-sonnet-20241022-v2:0",
2829
2921
  "eu.anthropic.claude-3-7-sonnet-20250219-v1:0",
2922
+ "eu.anthropic.claude-opus-4-1-20250805-v1:0",
2830
2923
  "eu.anthropic.claude-opus-4-20250514-v1:0",
2831
2924
  "eu.anthropic.claude-sonnet-4-20250514-v1:0",
2832
2925
  "snowflake/claude-3-5-sonnet"
@@ -2963,8 +3056,48 @@ var ollamaDefaultModelInfo = {
2963
3056
  };
2964
3057
 
2965
3058
  // src/providers/openai.ts
2966
- var openAiNativeDefaultModelId = "gpt-4.1";
3059
+ var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
2967
3060
  var openAiNativeModels = {
3061
+ "gpt-5-2025-08-07": {
3062
+ maxTokens: 128e3,
3063
+ contextWindow: 4e5,
3064
+ supportsImages: true,
3065
+ supportsPromptCache: true,
3066
+ supportsReasoningEffort: true,
3067
+ reasoningEffort: "medium",
3068
+ inputPrice: 1.25,
3069
+ outputPrice: 10,
3070
+ cacheReadsPrice: 0.13,
3071
+ description: "GPT-5: The best model for coding and agentic tasks across domains",
3072
+ // supportsVerbosity is a new capability; ensure ModelInfo includes it
3073
+ supportsVerbosity: true
3074
+ },
3075
+ "gpt-5-mini-2025-08-07": {
3076
+ maxTokens: 128e3,
3077
+ contextWindow: 4e5,
3078
+ supportsImages: true,
3079
+ supportsPromptCache: true,
3080
+ supportsReasoningEffort: true,
3081
+ reasoningEffort: "medium",
3082
+ inputPrice: 0.25,
3083
+ outputPrice: 2,
3084
+ cacheReadsPrice: 0.03,
3085
+ description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
3086
+ supportsVerbosity: true
3087
+ },
3088
+ "gpt-5-nano-2025-08-07": {
3089
+ maxTokens: 128e3,
3090
+ contextWindow: 4e5,
3091
+ supportsImages: true,
3092
+ supportsPromptCache: true,
3093
+ supportsReasoningEffort: true,
3094
+ reasoningEffort: "medium",
3095
+ inputPrice: 0.05,
3096
+ outputPrice: 0.4,
3097
+ cacheReadsPrice: 0.01,
3098
+ description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
3099
+ supportsVerbosity: true
3100
+ },
2968
3101
  "gpt-4.1": {
2969
3102
  maxTokens: 32768,
2970
3103
  contextWindow: 1047576,
@@ -3150,6 +3283,7 @@ var openAiModelInfoSaneDefaults = {
3150
3283
  };
3151
3284
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3152
3285
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3286
+ var GPT5_DEFAULT_TEMPERATURE = 1;
3153
3287
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3154
3288
 
3155
3289
  // src/providers/openrouter.ts
@@ -3187,6 +3321,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3187
3321
  "anthropic/claude-3.7-sonnet:thinking",
3188
3322
  "anthropic/claude-sonnet-4",
3189
3323
  "anthropic/claude-opus-4",
3324
+ "anthropic/claude-opus-4.1",
3190
3325
  "google/gemini-2.5-flash-preview",
3191
3326
  "google/gemini-2.5-flash-preview:thinking",
3192
3327
  "google/gemini-2.5-flash-preview-05-20",
@@ -3204,7 +3339,8 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3204
3339
  "anthropic/claude-3.7-sonnet:beta",
3205
3340
  "anthropic/claude-3.7-sonnet:thinking",
3206
3341
  "anthropic/claude-sonnet-4",
3207
- "anthropic/claude-opus-4"
3342
+ "anthropic/claude-opus-4",
3343
+ "anthropic/claude-opus-4.1"
3208
3344
  ]);
3209
3345
  var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3210
3346
  "anthropic/claude-3.7-sonnet:thinking",
@@ -3214,6 +3350,7 @@ var OPEN_ROUTER_REQUIRED_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3214
3350
  var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3215
3351
  "anthropic/claude-3.7-sonnet:beta",
3216
3352
  "anthropic/claude-opus-4",
3353
+ "anthropic/claude-opus-4.1",
3217
3354
  "anthropic/claude-sonnet-4",
3218
3355
  "google/gemini-2.5-pro-preview",
3219
3356
  "google/gemini-2.5-pro",
@@ -3504,6 +3641,18 @@ var vertexModels = {
3504
3641
  cacheReadsPrice: 0.3,
3505
3642
  supportsReasoningBudget: true
3506
3643
  },
3644
+ "claude-opus-4-1@20250805": {
3645
+ maxTokens: 8192,
3646
+ contextWindow: 2e5,
3647
+ supportsImages: true,
3648
+ supportsComputerUse: true,
3649
+ supportsPromptCache: true,
3650
+ inputPrice: 15,
3651
+ outputPrice: 75,
3652
+ cacheWritesPrice: 18.75,
3653
+ cacheReadsPrice: 1.5,
3654
+ supportsReasoningBudget: true
3655
+ },
3507
3656
  "claude-opus-4@20250514": {
3508
3657
  maxTokens: 8192,
3509
3658
  contextWindow: 2e5,
@@ -4021,6 +4170,92 @@ var mainlandZAiModels = {
4021
4170
  }
4022
4171
  };
4023
4172
  var ZAI_DEFAULT_TEMPERATURE = 0;
4173
+
4174
+ // src/providers/fireworks.ts
4175
+ var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
4176
+ var fireworksModels = {
4177
+ "accounts/fireworks/models/kimi-k2-instruct": {
4178
+ maxTokens: 16384,
4179
+ contextWindow: 128e3,
4180
+ supportsImages: false,
4181
+ supportsPromptCache: false,
4182
+ inputPrice: 0.6,
4183
+ outputPrice: 2.5,
4184
+ description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
4185
+ },
4186
+ "accounts/fireworks/models/qwen3-235b-a22b-instruct-2507": {
4187
+ maxTokens: 32768,
4188
+ contextWindow: 256e3,
4189
+ supportsImages: false,
4190
+ supportsPromptCache: false,
4191
+ inputPrice: 0.22,
4192
+ outputPrice: 0.88,
4193
+ description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
4194
+ },
4195
+ "accounts/fireworks/models/qwen3-coder-480b-a35b-instruct": {
4196
+ maxTokens: 32768,
4197
+ contextWindow: 256e3,
4198
+ supportsImages: false,
4199
+ supportsPromptCache: false,
4200
+ inputPrice: 0.45,
4201
+ outputPrice: 1.8,
4202
+ description: "Qwen3's most agentic code model to date."
4203
+ },
4204
+ "accounts/fireworks/models/deepseek-r1-0528": {
4205
+ maxTokens: 20480,
4206
+ contextWindow: 16e4,
4207
+ supportsImages: false,
4208
+ supportsPromptCache: false,
4209
+ inputPrice: 3,
4210
+ outputPrice: 8,
4211
+ description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4212
+ },
4213
+ "accounts/fireworks/models/deepseek-v3": {
4214
+ maxTokens: 16384,
4215
+ contextWindow: 128e3,
4216
+ supportsImages: false,
4217
+ supportsPromptCache: false,
4218
+ inputPrice: 0.9,
4219
+ outputPrice: 0.9,
4220
+ description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
4221
+ },
4222
+ "accounts/fireworks/models/glm-4p5": {
4223
+ maxTokens: 16384,
4224
+ contextWindow: 128e3,
4225
+ supportsImages: false,
4226
+ supportsPromptCache: false,
4227
+ inputPrice: 0.55,
4228
+ outputPrice: 2.19,
4229
+ description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4230
+ },
4231
+ "accounts/fireworks/models/glm-4p5-air": {
4232
+ maxTokens: 16384,
4233
+ contextWindow: 128e3,
4234
+ supportsImages: false,
4235
+ supportsPromptCache: false,
4236
+ inputPrice: 0.55,
4237
+ outputPrice: 2.19,
4238
+ description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
4239
+ },
4240
+ "accounts/fireworks/models/gpt-oss-20b": {
4241
+ maxTokens: 16384,
4242
+ contextWindow: 128e3,
4243
+ supportsImages: false,
4244
+ supportsPromptCache: false,
4245
+ inputPrice: 0.07,
4246
+ outputPrice: 0.3,
4247
+ description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
4248
+ },
4249
+ "accounts/fireworks/models/gpt-oss-120b": {
4250
+ maxTokens: 16384,
4251
+ contextWindow: 128e3,
4252
+ supportsImages: false,
4253
+ supportsPromptCache: false,
4254
+ inputPrice: 0.15,
4255
+ outputPrice: 0.6,
4256
+ description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
4257
+ }
4258
+ };
4024
4259
  export {
4025
4260
  ANTHROPIC_DEFAULT_MAX_TOKENS,
4026
4261
  ANTHROPIC_STYLE_PROVIDERS,
@@ -4043,6 +4278,7 @@ export {
4043
4278
  GLAMA_DEFAULT_TEMPERATURE,
4044
4279
  GLOBAL_SETTINGS_KEYS,
4045
4280
  GLOBAL_STATE_KEYS,
4281
+ GPT5_DEFAULT_TEMPERATURE,
4046
4282
  HUGGINGFACE_API_URL,
4047
4283
  HUGGINGFACE_CACHE_DURATION,
4048
4284
  HUGGINGFACE_DEFAULT_CONTEXT_WINDOW,
@@ -4114,6 +4350,9 @@ export {
4114
4350
  experimentIds,
4115
4351
  experimentIdsSchema,
4116
4352
  experimentsSchema,
4353
+ extendedReasoningEffortsSchema,
4354
+ fireworksDefaultModelId,
4355
+ fireworksModels,
4117
4356
  followUpDataSchema,
4118
4357
  geminiDefaultModelId,
4119
4358
  geminiModels,
@@ -4208,6 +4447,8 @@ export {
4208
4447
  toolUsageSchema,
4209
4448
  unboundDefaultModelId,
4210
4449
  unboundDefaultModelInfo,
4450
+ verbosityLevels,
4451
+ verbosityLevelsSchema,
4211
4452
  vertexDefaultModelId,
4212
4453
  vertexModels,
4213
4454
  vscodeLlmDefaultModelId,