@roo-code/types 1.85.0 → 1.86.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -21,10 +21,6 @@ var clineAsks = [
21
21
  "auto_approval_max_req_reached"
22
22
  ];
23
23
  var clineAskSchema = z.enum(clineAsks);
24
- var nonBlockingAsks = ["command_output"];
25
- function isNonBlockingAsk(ask) {
26
- return nonBlockingAsks.includes(ask);
27
- }
28
24
  var idleAsks = [
29
25
  "completion_result",
30
26
  "api_req_failed",
@@ -49,6 +45,10 @@ var interactiveAsks = [
49
45
  function isInteractiveAsk(ask) {
50
46
  return interactiveAsks.includes(ask);
51
47
  }
48
+ var nonBlockingAsks = ["command_output"];
49
+ function isNonBlockingAsk(ask) {
50
+ return nonBlockingAsks.includes(ask);
51
+ }
52
52
  var clineSays = [
53
53
  "error",
54
54
  "api_req_started",
@@ -103,12 +103,7 @@ var clineMessageSchema = z.object({
103
103
  contextCondense: contextCondenseSchema.optional(),
104
104
  isProtected: z.boolean().optional(),
105
105
  apiProtocol: z.union([z.literal("openai"), z.literal("anthropic")]).optional(),
106
- isAnswered: z.boolean().optional(),
107
- metadata: z.object({
108
- gpt5: z.object({
109
- previous_response_id: z.string().optional()
110
- }).optional()
111
- }).optional()
106
+ isAnswered: z.boolean().optional()
112
107
  });
113
108
  var tokenUsageSchema = z.object({
114
109
  totalTokensIn: z.number(),
@@ -159,6 +154,16 @@ var toolUsageSchema = z2.record(
159
154
  failures: z2.number()
160
155
  })
161
156
  );
157
+ var TOOL_PROTOCOL = {
158
+ XML: "xml",
159
+ NATIVE: "native"
160
+ };
161
+ function isNativeProtocol(protocol) {
162
+ return protocol === TOOL_PROTOCOL.NATIVE;
163
+ }
164
+ function getEffectiveProtocol(toolProtocol) {
165
+ return toolProtocol || TOOL_PROTOCOL.XML;
166
+ }
162
167
 
163
168
  // src/events.ts
164
169
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
@@ -358,6 +363,10 @@ import { z as z5 } from "zod";
358
363
  var reasoningEfforts = ["low", "medium", "high"];
359
364
  var reasoningEffortsSchema = z5.enum(reasoningEfforts);
360
365
  var reasoningEffortWithMinimalSchema = z5.union([reasoningEffortsSchema, z5.literal("minimal")]);
366
+ var reasoningEffortsExtended = ["none", "minimal", "low", "medium", "high"];
367
+ var reasoningEffortExtendedSchema = z5.enum(reasoningEffortsExtended);
368
+ var reasoningEffortSettingValues = ["disable", "none", "minimal", "low", "medium", "high"];
369
+ var reasoningEffortSettingSchema = z5.enum(reasoningEffortSettingValues);
361
370
  var verbosityLevels = ["low", "medium", "high"];
362
371
  var verbosityLevelsSchema = z5.enum(verbosityLevels);
363
372
  var serviceTiers = ["default", "flex", "priority"];
@@ -371,6 +380,10 @@ var modelInfoSchema = z5.object({
371
380
  contextWindow: z5.number(),
372
381
  supportsImages: z5.boolean().optional(),
373
382
  supportsPromptCache: z5.boolean(),
383
+ // Optional default prompt cache retention policy for providers that support it.
384
+ // When set to "24h", extended prompt caching will be requested; when omitted
385
+ // or set to "in_memory", the default in‑memory cache is used.
386
+ promptCacheRetention: z5.enum(["in_memory", "24h"]).optional(),
374
387
  // Capability flag to indicate whether the model supports an output verbosity parameter
375
388
  supportsVerbosity: z5.boolean().optional(),
376
389
  supportsReasoningBudget: z5.boolean().optional(),
@@ -380,7 +393,7 @@ var modelInfoSchema = z5.object({
380
393
  supportsTemperature: z5.boolean().optional(),
381
394
  defaultTemperature: z5.number().optional(),
382
395
  requiredReasoningBudget: z5.boolean().optional(),
383
- supportsReasoningEffort: z5.boolean().optional(),
396
+ supportsReasoningEffort: z5.union([z5.boolean(), z5.array(z5.enum(["disable", "none", "minimal", "low", "medium", "high"]))]).optional(),
384
397
  requiredReasoningEffort: z5.boolean().optional(),
385
398
  preserveReasoning: z5.boolean().optional(),
386
399
  supportedParameters: z5.array(modelParametersSchema).optional(),
@@ -389,7 +402,8 @@ var modelInfoSchema = z5.object({
389
402
  cacheWritesPrice: z5.number().optional(),
390
403
  cacheReadsPrice: z5.number().optional(),
391
404
  description: z5.string().optional(),
392
- reasoningEffort: reasoningEffortsSchema.optional(),
405
+ // Default effort value for models that support reasoning effort
406
+ reasoningEffort: reasoningEffortExtendedSchema.optional(),
393
407
  minTokensPerCachePoint: z5.number().optional(),
394
408
  maxCachePoints: z5.number().optional(),
395
409
  cachableFields: z5.array(z5.string()).optional(),
@@ -397,6 +411,10 @@ var modelInfoSchema = z5.object({
397
411
  deprecated: z5.boolean().optional(),
398
412
  // Flag to indicate if the model is free (no cost)
399
413
  isFree: z5.boolean().optional(),
414
+ // Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
415
+ supportsNativeTools: z5.boolean().optional(),
416
+ // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
417
+ defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
400
418
  /**
401
419
  * Service tiers with pricing information.
402
420
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -2635,85 +2653,130 @@ var ollamaDefaultModelInfo = {
2635
2653
  };
2636
2654
 
2637
2655
  // src/providers/openai.ts
2638
- var openAiNativeDefaultModelId = "gpt-5-2025-08-07";
2656
+ var openAiNativeDefaultModelId = "gpt-5.1";
2639
2657
  var openAiNativeModels = {
2640
- "gpt-5-chat-latest": {
2658
+ "gpt-5.1": {
2641
2659
  maxTokens: 128e3,
2642
2660
  contextWindow: 4e5,
2643
2661
  supportsImages: true,
2644
2662
  supportsPromptCache: true,
2645
- supportsReasoningEffort: false,
2663
+ promptCacheRetention: "24h",
2664
+ supportsReasoningEffort: ["none", "low", "medium", "high"],
2665
+ reasoningEffort: "medium",
2646
2666
  inputPrice: 1.25,
2647
2667
  outputPrice: 10,
2648
- cacheReadsPrice: 0.13,
2649
- description: "GPT-5 Chat Latest: Optimized for conversational AI and non-reasoning tasks",
2650
- supportsVerbosity: true
2668
+ cacheReadsPrice: 0.125,
2669
+ supportsVerbosity: true,
2670
+ supportsTemperature: false,
2671
+ tiers: [
2672
+ { name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
2673
+ { name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
2674
+ ],
2675
+ description: "GPT-5.1: The best model for coding and agentic tasks across domains"
2651
2676
  },
2652
- "gpt-5-2025-08-07": {
2677
+ "gpt-5.1-codex": {
2678
+ maxTokens: 128e3,
2679
+ contextWindow: 4e5,
2680
+ supportsImages: true,
2681
+ supportsPromptCache: true,
2682
+ promptCacheRetention: "24h",
2683
+ supportsReasoningEffort: ["low", "medium", "high"],
2684
+ reasoningEffort: "medium",
2685
+ inputPrice: 1.25,
2686
+ outputPrice: 10,
2687
+ cacheReadsPrice: 0.125,
2688
+ supportsTemperature: false,
2689
+ tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }],
2690
+ description: "GPT-5.1 Codex: A version of GPT-5.1 optimized for agentic coding in Codex"
2691
+ },
2692
+ "gpt-5.1-codex-mini": {
2693
+ maxTokens: 128e3,
2694
+ contextWindow: 4e5,
2695
+ supportsImages: true,
2696
+ supportsPromptCache: true,
2697
+ promptCacheRetention: "24h",
2698
+ supportsReasoningEffort: ["low", "medium", "high"],
2699
+ reasoningEffort: "medium",
2700
+ inputPrice: 0.25,
2701
+ outputPrice: 2,
2702
+ cacheReadsPrice: 0.025,
2703
+ supportsTemperature: false,
2704
+ description: "GPT-5.1 Codex mini: A version of GPT-5.1 optimized for agentic coding in Codex"
2705
+ },
2706
+ "gpt-5": {
2653
2707
  maxTokens: 128e3,
2654
2708
  contextWindow: 4e5,
2655
2709
  supportsImages: true,
2656
2710
  supportsPromptCache: true,
2657
- supportsReasoningEffort: true,
2711
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2658
2712
  reasoningEffort: "medium",
2659
2713
  inputPrice: 1.25,
2660
2714
  outputPrice: 10,
2661
- cacheReadsPrice: 0.13,
2662
- description: "GPT-5: The best model for coding and agentic tasks across domains",
2663
- // supportsVerbosity is a new capability; ensure ModelInfo includes it
2715
+ cacheReadsPrice: 0.125,
2664
2716
  supportsVerbosity: true,
2665
2717
  supportsTemperature: false,
2666
2718
  tiers: [
2667
2719
  { name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
2668
2720
  { name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
2669
- ]
2721
+ ],
2722
+ description: "GPT-5: The best model for coding and agentic tasks across domains"
2670
2723
  },
2671
- "gpt-5-mini-2025-08-07": {
2724
+ "gpt-5-mini": {
2672
2725
  maxTokens: 128e3,
2673
2726
  contextWindow: 4e5,
2674
2727
  supportsImages: true,
2675
2728
  supportsPromptCache: true,
2676
- supportsReasoningEffort: true,
2729
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2677
2730
  reasoningEffort: "medium",
2678
2731
  inputPrice: 0.25,
2679
2732
  outputPrice: 2,
2680
- cacheReadsPrice: 0.03,
2681
- description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
2733
+ cacheReadsPrice: 0.025,
2682
2734
  supportsVerbosity: true,
2683
2735
  supportsTemperature: false,
2684
2736
  tiers: [
2685
2737
  { name: "flex", contextWindow: 4e5, inputPrice: 0.125, outputPrice: 1, cacheReadsPrice: 0.0125 },
2686
2738
  { name: "priority", contextWindow: 4e5, inputPrice: 0.45, outputPrice: 3.6, cacheReadsPrice: 0.045 }
2687
- ]
2739
+ ],
2740
+ description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks"
2688
2741
  },
2689
- "gpt-5-nano-2025-08-07": {
2742
+ "gpt-5-codex": {
2743
+ maxTokens: 128e3,
2744
+ contextWindow: 4e5,
2745
+ supportsImages: true,
2746
+ supportsPromptCache: true,
2747
+ supportsReasoningEffort: ["low", "medium", "high"],
2748
+ reasoningEffort: "medium",
2749
+ inputPrice: 1.25,
2750
+ outputPrice: 10,
2751
+ cacheReadsPrice: 0.125,
2752
+ supportsTemperature: false,
2753
+ tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }],
2754
+ description: "GPT-5-Codex: A version of GPT-5 optimized for agentic coding in Codex"
2755
+ },
2756
+ "gpt-5-nano": {
2690
2757
  maxTokens: 128e3,
2691
2758
  contextWindow: 4e5,
2692
2759
  supportsImages: true,
2693
2760
  supportsPromptCache: true,
2694
- supportsReasoningEffort: true,
2761
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2695
2762
  reasoningEffort: "medium",
2696
2763
  inputPrice: 0.05,
2697
2764
  outputPrice: 0.4,
2698
- cacheReadsPrice: 0.01,
2699
- description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
2765
+ cacheReadsPrice: 5e-3,
2700
2766
  supportsVerbosity: true,
2701
2767
  supportsTemperature: false,
2702
- tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }]
2768
+ tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }],
2769
+ description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5"
2703
2770
  },
2704
- "gpt-5-codex": {
2771
+ "gpt-5-chat-latest": {
2705
2772
  maxTokens: 128e3,
2706
2773
  contextWindow: 4e5,
2707
2774
  supportsImages: true,
2708
2775
  supportsPromptCache: true,
2709
- supportsReasoningEffort: true,
2710
- reasoningEffort: "medium",
2711
2776
  inputPrice: 1.25,
2712
2777
  outputPrice: 10,
2713
- cacheReadsPrice: 0.13,
2714
- description: "GPT-5-Codex: A version of GPT-5 optimized for agentic coding in Codex",
2715
- supportsVerbosity: true,
2716
- supportsTemperature: false
2778
+ cacheReadsPrice: 0.125,
2779
+ description: "GPT-5 Chat: Optimized for conversational AI and non-reasoning tasks"
2717
2780
  },
2718
2781
  "gpt-4.1": {
2719
2782
  maxTokens: 32768,
@@ -2762,7 +2825,7 @@ var openAiNativeModels = {
2762
2825
  inputPrice: 2,
2763
2826
  outputPrice: 8,
2764
2827
  cacheReadsPrice: 0.5,
2765
- supportsReasoningEffort: true,
2828
+ supportsReasoningEffort: ["low", "medium", "high"],
2766
2829
  reasoningEffort: "medium",
2767
2830
  supportsTemperature: false,
2768
2831
  tiers: [
@@ -2800,7 +2863,7 @@ var openAiNativeModels = {
2800
2863
  inputPrice: 1.1,
2801
2864
  outputPrice: 4.4,
2802
2865
  cacheReadsPrice: 0.275,
2803
- supportsReasoningEffort: true,
2866
+ supportsReasoningEffort: ["low", "medium", "high"],
2804
2867
  reasoningEffort: "medium",
2805
2868
  supportsTemperature: false,
2806
2869
  tiers: [
@@ -2838,7 +2901,7 @@ var openAiNativeModels = {
2838
2901
  inputPrice: 1.1,
2839
2902
  outputPrice: 4.4,
2840
2903
  cacheReadsPrice: 0.55,
2841
- supportsReasoningEffort: true,
2904
+ supportsReasoningEffort: ["low", "medium", "high"],
2842
2905
  reasoningEffort: "medium",
2843
2906
  supportsTemperature: false
2844
2907
  },
@@ -2927,9 +2990,61 @@ var openAiNativeModels = {
2927
2990
  supportsPromptCache: false,
2928
2991
  inputPrice: 1.5,
2929
2992
  outputPrice: 6,
2930
- cacheReadsPrice: 0,
2993
+ cacheReadsPrice: 0.375,
2931
2994
  supportsTemperature: false,
2932
2995
  description: "Codex Mini: Cloud-based software engineering agent powered by codex-1, a version of o3 optimized for coding tasks. Trained with reinforcement learning to generate human-style code, adhere to instructions, and iteratively run tests."
2996
+ },
2997
+ // Dated clones (snapshots) preserved for backward compatibility
2998
+ "gpt-5-2025-08-07": {
2999
+ maxTokens: 128e3,
3000
+ contextWindow: 4e5,
3001
+ supportsImages: true,
3002
+ supportsPromptCache: true,
3003
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3004
+ reasoningEffort: "medium",
3005
+ inputPrice: 1.25,
3006
+ outputPrice: 10,
3007
+ cacheReadsPrice: 0.125,
3008
+ supportsVerbosity: true,
3009
+ supportsTemperature: false,
3010
+ tiers: [
3011
+ { name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
3012
+ { name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
3013
+ ],
3014
+ description: "GPT-5: The best model for coding and agentic tasks across domains"
3015
+ },
3016
+ "gpt-5-mini-2025-08-07": {
3017
+ maxTokens: 128e3,
3018
+ contextWindow: 4e5,
3019
+ supportsImages: true,
3020
+ supportsPromptCache: true,
3021
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3022
+ reasoningEffort: "medium",
3023
+ inputPrice: 0.25,
3024
+ outputPrice: 2,
3025
+ cacheReadsPrice: 0.025,
3026
+ supportsVerbosity: true,
3027
+ supportsTemperature: false,
3028
+ tiers: [
3029
+ { name: "flex", contextWindow: 4e5, inputPrice: 0.125, outputPrice: 1, cacheReadsPrice: 0.0125 },
3030
+ { name: "priority", contextWindow: 4e5, inputPrice: 0.45, outputPrice: 3.6, cacheReadsPrice: 0.045 }
3031
+ ],
3032
+ description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks"
3033
+ },
3034
+ "gpt-5-nano-2025-08-07": {
3035
+ maxTokens: 128e3,
3036
+ contextWindow: 4e5,
3037
+ supportsImages: true,
3038
+ supportsPromptCache: true,
3039
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3040
+ reasoningEffort: "medium",
3041
+ inputPrice: 0.05,
3042
+ outputPrice: 0.4,
3043
+ cacheReadsPrice: 5e-3,
3044
+ supportsVerbosity: true,
3045
+ supportsTemperature: false,
3046
+ tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }],
3047
+ description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5"
2933
3048
  }
2934
3049
  };
2935
3050
  var openAiModelInfoSaneDefaults = {
@@ -2942,7 +3057,6 @@ var openAiModelInfoSaneDefaults = {
2942
3057
  };
2943
3058
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
2944
3059
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
2945
- var GPT5_DEFAULT_TEMPERATURE = 1;
2946
3060
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
2947
3061
 
2948
3062
  // src/providers/openrouter.ts
@@ -2952,6 +3066,7 @@ var openRouterDefaultModelInfo = {
2952
3066
  contextWindow: 2e5,
2953
3067
  supportsImages: true,
2954
3068
  supportsPromptCache: true,
3069
+ supportsNativeTools: true,
2955
3070
  inputPrice: 3,
2956
3071
  outputPrice: 15,
2957
3072
  cacheWritesPrice: 3.75,
@@ -3060,6 +3175,16 @@ var requestyDefaultModelInfo = {
3060
3175
  import { z as z7 } from "zod";
3061
3176
  var rooDefaultModelId = "xai/grok-code-fast-1";
3062
3177
  var rooModels = {};
3178
+ var rooModelDefaults = {
3179
+ // Add model-specific defaults below.
3180
+ // You can configure defaultToolProtocol and other ModelInfo fields for specific model IDs.
3181
+ "anthropic/claude-haiku-4.5": {
3182
+ defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3183
+ },
3184
+ "minimax/minimax-m2:free": {
3185
+ defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3186
+ }
3187
+ };
3063
3188
  var RooPricingSchema = z7.object({
3064
3189
  input: z7.string(),
3065
3190
  output: z7.string(),
@@ -4346,11 +4471,13 @@ var baseProviderSettingsSchema = z8.object({
4346
4471
  consecutiveMistakeLimit: z8.number().min(0).optional(),
4347
4472
  // Model reasoning.
4348
4473
  enableReasoningEffort: z8.boolean().optional(),
4349
- reasoningEffort: reasoningEffortWithMinimalSchema.optional(),
4474
+ reasoningEffort: reasoningEffortSettingSchema.optional(),
4350
4475
  modelMaxTokens: z8.number().optional(),
4351
4476
  modelMaxThinkingTokens: z8.number().optional(),
4352
4477
  // Model verbosity.
4353
- verbosity: verbosityLevelsSchema.optional()
4478
+ verbosity: verbosityLevelsSchema.optional(),
4479
+ // Tool protocol override for this profile.
4480
+ toolProtocol: z8.enum(["xml", "native"]).optional()
4354
4481
  });
4355
4482
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
4356
4483
  apiModelId: z8.string().optional()
@@ -4827,7 +4954,8 @@ var experimentIds = [
4827
4954
  "multiFileApplyDiff",
4828
4955
  "preventFocusDisruption",
4829
4956
  "imageGeneration",
4830
- "runSlashCommand"
4957
+ "runSlashCommand",
4958
+ "nativeToolCalling"
4831
4959
  ];
4832
4960
  var experimentIdsSchema = z10.enum(experimentIds);
4833
4961
  var experimentsSchema = z10.object({
@@ -4835,7 +4963,8 @@ var experimentsSchema = z10.object({
4835
4963
  multiFileApplyDiff: z10.boolean().optional(),
4836
4964
  preventFocusDisruption: z10.boolean().optional(),
4837
4965
  imageGeneration: z10.boolean().optional(),
4838
- runSlashCommand: z10.boolean().optional()
4966
+ runSlashCommand: z10.boolean().optional(),
4967
+ nativeToolCalling: z10.boolean().optional()
4839
4968
  });
4840
4969
 
4841
4970
  // src/telemetry.ts
@@ -5227,6 +5356,12 @@ var globalSettingsSchema = z14.object({
5227
5356
  * @default true
5228
5357
  */
5229
5358
  includeCurrentCost: z14.boolean().optional(),
5359
+ /**
5360
+ * Maximum number of git status file entries to include in the environment details.
5361
+ * Set to 0 to disable git status. The header (branch, commits) is always included when > 0.
5362
+ * @default 0
5363
+ */
5364
+ maxGitStatusFiles: z14.number().optional(),
5230
5365
  /**
5231
5366
  * Whether to include diagnostic messages (errors, warnings) in tool outputs
5232
5367
  * @default true
@@ -5394,6 +5529,7 @@ var EVALS_SETTINGS = {
5394
5529
  rateLimitSeconds: 0,
5395
5530
  maxOpenTabsContext: 20,
5396
5531
  maxWorkspaceFiles: 200,
5532
+ maxGitStatusFiles: 20,
5397
5533
  showRooIgnoredFiles: true,
5398
5534
  maxReadFileLine: -1,
5399
5535
  // -1 to enable full file reading.
@@ -6036,7 +6172,6 @@ export {
6036
6172
  GLOBAL_SECRET_KEYS,
6037
6173
  GLOBAL_SETTINGS_KEYS,
6038
6174
  GLOBAL_STATE_KEYS,
6039
- GPT5_DEFAULT_TEMPERATURE,
6040
6175
  HEARTBEAT_INTERVAL_MS,
6041
6176
  HUGGINGFACE_API_URL,
6042
6177
  HUGGINGFACE_CACHE_DURATION,
@@ -6071,6 +6206,7 @@ export {
6071
6206
  RooModelsResponseSchema,
6072
6207
  RooPricingSchema,
6073
6208
  SECRET_STATE_KEYS,
6209
+ TOOL_PROTOCOL,
6074
6210
  TaskBridgeCommandName,
6075
6211
  TaskBridgeEventName,
6076
6212
  TaskCommandName,
@@ -6142,6 +6278,7 @@ export {
6142
6278
  geminiModels,
6143
6279
  getApiProtocol,
6144
6280
  getClaudeCodeModelId,
6281
+ getEffectiveProtocol,
6145
6282
  getModelId,
6146
6283
  getProviderDefaultModelId,
6147
6284
  gitPropertiesSchema,
@@ -6173,6 +6310,7 @@ export {
6173
6310
  isLanguage,
6174
6311
  isLocalProvider,
6175
6312
  isModelParameter,
6313
+ isNativeProtocol,
6176
6314
  isNonBlockingAsk,
6177
6315
  isProviderName,
6178
6316
  isResumableAsk,
@@ -6229,8 +6367,12 @@ export {
6229
6367
  queuedMessageSchema,
6230
6368
  qwenCodeDefaultModelId,
6231
6369
  qwenCodeModels,
6370
+ reasoningEffortExtendedSchema,
6371
+ reasoningEffortSettingSchema,
6372
+ reasoningEffortSettingValues,
6232
6373
  reasoningEffortWithMinimalSchema,
6233
6374
  reasoningEfforts,
6375
+ reasoningEffortsExtended,
6234
6376
  reasoningEffortsSchema,
6235
6377
  requestyDefaultModelId,
6236
6378
  requestyDefaultModelInfo,
@@ -6239,6 +6381,7 @@ export {
6239
6381
  rooCodeSettingsSchema,
6240
6382
  rooCodeTelemetryEventSchema,
6241
6383
  rooDefaultModelId,
6384
+ rooModelDefaults,
6242
6385
  rooModels,
6243
6386
  sambaNovaDefaultModelId,
6244
6387
  sambaNovaModels,