@roo-code/types 1.85.0 → 1.86.0
This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
- package/dist/index.cjs +202 -52
- package/dist/index.cjs.map +1 -1
- package/dist/index.d.cts +1115 -2867
- package/dist/index.d.ts +1115 -2867
- package/dist/index.js +194 -51
- package/dist/index.js.map +1 -1
- package/package.json +1 -1
package/dist/index.js
CHANGED
|
@@ -21,10 +21,6 @@ var clineAsks = [
|
|
|
21
21
|
"auto_approval_max_req_reached"
|
|
22
22
|
];
|
|
23
23
|
var clineAskSchema = z.enum(clineAsks);
|
|
24
|
-
var nonBlockingAsks = ["command_output"];
|
|
25
|
-
function isNonBlockingAsk(ask) {
|
|
26
|
-
return nonBlockingAsks.includes(ask);
|
|
27
|
-
}
|
|
28
24
|
var idleAsks = [
|
|
29
25
|
"completion_result",
|
|
30
26
|
"api_req_failed",
|
|
@@ -49,6 +45,10 @@ var interactiveAsks = [
|
|
|
49
45
|
function isInteractiveAsk(ask) {
|
|
50
46
|
return interactiveAsks.includes(ask);
|
|
51
47
|
}
|
|
48
|
+
var nonBlockingAsks = ["command_output"];
|
|
49
|
+
function isNonBlockingAsk(ask) {
|
|
50
|
+
return nonBlockingAsks.includes(ask);
|
|
51
|
+
}
|
|
52
52
|
var clineSays = [
|
|
53
53
|
"error",
|
|
54
54
|
"api_req_started",
|
|
@@ -103,12 +103,7 @@ var clineMessageSchema = z.object({
|
|
|
103
103
|
contextCondense: contextCondenseSchema.optional(),
|
|
104
104
|
isProtected: z.boolean().optional(),
|
|
105
105
|
apiProtocol: z.union([z.literal("openai"), z.literal("anthropic")]).optional(),
|
|
106
|
-
isAnswered: z.boolean().optional()
|
|
107
|
-
metadata: z.object({
|
|
108
|
-
gpt5: z.object({
|
|
109
|
-
previous_response_id: z.string().optional()
|
|
110
|
-
}).optional()
|
|
111
|
-
}).optional()
|
|
106
|
+
isAnswered: z.boolean().optional()
|
|
112
107
|
});
|
|
113
108
|
var tokenUsageSchema = z.object({
|
|
114
109
|
totalTokensIn: z.number(),
|
|
@@ -159,6 +154,16 @@ var toolUsageSchema = z2.record(
|
|
|
159
154
|
failures: z2.number()
|
|
160
155
|
})
|
|
161
156
|
);
|
|
157
|
+
var TOOL_PROTOCOL = {
|
|
158
|
+
XML: "xml",
|
|
159
|
+
NATIVE: "native"
|
|
160
|
+
};
|
|
161
|
+
function isNativeProtocol(protocol) {
|
|
162
|
+
return protocol === TOOL_PROTOCOL.NATIVE;
|
|
163
|
+
}
|
|
164
|
+
function getEffectiveProtocol(toolProtocol) {
|
|
165
|
+
return toolProtocol || TOOL_PROTOCOL.XML;
|
|
166
|
+
}
|
|
162
167
|
|
|
163
168
|
// src/events.ts
|
|
164
169
|
var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
|
|
@@ -358,6 +363,10 @@ import { z as z5 } from "zod";
|
|
|
358
363
|
var reasoningEfforts = ["low", "medium", "high"];
|
|
359
364
|
var reasoningEffortsSchema = z5.enum(reasoningEfforts);
|
|
360
365
|
var reasoningEffortWithMinimalSchema = z5.union([reasoningEffortsSchema, z5.literal("minimal")]);
|
|
366
|
+
var reasoningEffortsExtended = ["none", "minimal", "low", "medium", "high"];
|
|
367
|
+
var reasoningEffortExtendedSchema = z5.enum(reasoningEffortsExtended);
|
|
368
|
+
var reasoningEffortSettingValues = ["disable", "none", "minimal", "low", "medium", "high"];
|
|
369
|
+
var reasoningEffortSettingSchema = z5.enum(reasoningEffortSettingValues);
|
|
361
370
|
var verbosityLevels = ["low", "medium", "high"];
|
|
362
371
|
var verbosityLevelsSchema = z5.enum(verbosityLevels);
|
|
363
372
|
var serviceTiers = ["default", "flex", "priority"];
|
|
@@ -371,6 +380,10 @@ var modelInfoSchema = z5.object({
|
|
|
371
380
|
contextWindow: z5.number(),
|
|
372
381
|
supportsImages: z5.boolean().optional(),
|
|
373
382
|
supportsPromptCache: z5.boolean(),
|
|
383
|
+
// Optional default prompt cache retention policy for providers that support it.
|
|
384
|
+
// When set to "24h", extended prompt caching will be requested; when omitted
|
|
385
|
+
// or set to "in_memory", the default in‑memory cache is used.
|
|
386
|
+
promptCacheRetention: z5.enum(["in_memory", "24h"]).optional(),
|
|
374
387
|
// Capability flag to indicate whether the model supports an output verbosity parameter
|
|
375
388
|
supportsVerbosity: z5.boolean().optional(),
|
|
376
389
|
supportsReasoningBudget: z5.boolean().optional(),
|
|
@@ -380,7 +393,7 @@ var modelInfoSchema = z5.object({
|
|
|
380
393
|
supportsTemperature: z5.boolean().optional(),
|
|
381
394
|
defaultTemperature: z5.number().optional(),
|
|
382
395
|
requiredReasoningBudget: z5.boolean().optional(),
|
|
383
|
-
supportsReasoningEffort: z5.boolean().optional(),
|
|
396
|
+
supportsReasoningEffort: z5.union([z5.boolean(), z5.array(z5.enum(["disable", "none", "minimal", "low", "medium", "high"]))]).optional(),
|
|
384
397
|
requiredReasoningEffort: z5.boolean().optional(),
|
|
385
398
|
preserveReasoning: z5.boolean().optional(),
|
|
386
399
|
supportedParameters: z5.array(modelParametersSchema).optional(),
|
|
@@ -389,7 +402,8 @@ var modelInfoSchema = z5.object({
|
|
|
389
402
|
cacheWritesPrice: z5.number().optional(),
|
|
390
403
|
cacheReadsPrice: z5.number().optional(),
|
|
391
404
|
description: z5.string().optional(),
|
|
392
|
-
|
|
405
|
+
// Default effort value for models that support reasoning effort
|
|
406
|
+
reasoningEffort: reasoningEffortExtendedSchema.optional(),
|
|
393
407
|
minTokensPerCachePoint: z5.number().optional(),
|
|
394
408
|
maxCachePoints: z5.number().optional(),
|
|
395
409
|
cachableFields: z5.array(z5.string()).optional(),
|
|
@@ -397,6 +411,10 @@ var modelInfoSchema = z5.object({
|
|
|
397
411
|
deprecated: z5.boolean().optional(),
|
|
398
412
|
// Flag to indicate if the model is free (no cost)
|
|
399
413
|
isFree: z5.boolean().optional(),
|
|
414
|
+
// Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
|
|
415
|
+
supportsNativeTools: z5.boolean().optional(),
|
|
416
|
+
// Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
|
|
417
|
+
defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
|
|
400
418
|
/**
|
|
401
419
|
* Service tiers with pricing information.
|
|
402
420
|
* Each tier can have a name (for OpenAI service tiers) and pricing overrides.
|
|
@@ -2635,85 +2653,130 @@ var ollamaDefaultModelInfo = {
|
|
|
2635
2653
|
};
|
|
2636
2654
|
|
|
2637
2655
|
// src/providers/openai.ts
|
|
2638
|
-
var openAiNativeDefaultModelId = "gpt-5
|
|
2656
|
+
var openAiNativeDefaultModelId = "gpt-5.1";
|
|
2639
2657
|
var openAiNativeModels = {
|
|
2640
|
-
"gpt-5
|
|
2658
|
+
"gpt-5.1": {
|
|
2641
2659
|
maxTokens: 128e3,
|
|
2642
2660
|
contextWindow: 4e5,
|
|
2643
2661
|
supportsImages: true,
|
|
2644
2662
|
supportsPromptCache: true,
|
|
2645
|
-
|
|
2663
|
+
promptCacheRetention: "24h",
|
|
2664
|
+
supportsReasoningEffort: ["none", "low", "medium", "high"],
|
|
2665
|
+
reasoningEffort: "medium",
|
|
2646
2666
|
inputPrice: 1.25,
|
|
2647
2667
|
outputPrice: 10,
|
|
2648
|
-
cacheReadsPrice: 0.
|
|
2649
|
-
|
|
2650
|
-
|
|
2668
|
+
cacheReadsPrice: 0.125,
|
|
2669
|
+
supportsVerbosity: true,
|
|
2670
|
+
supportsTemperature: false,
|
|
2671
|
+
tiers: [
|
|
2672
|
+
{ name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
|
|
2673
|
+
{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
|
|
2674
|
+
],
|
|
2675
|
+
description: "GPT-5.1: The best model for coding and agentic tasks across domains"
|
|
2651
2676
|
},
|
|
2652
|
-
"gpt-5-
|
|
2677
|
+
"gpt-5.1-codex": {
|
|
2678
|
+
maxTokens: 128e3,
|
|
2679
|
+
contextWindow: 4e5,
|
|
2680
|
+
supportsImages: true,
|
|
2681
|
+
supportsPromptCache: true,
|
|
2682
|
+
promptCacheRetention: "24h",
|
|
2683
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
2684
|
+
reasoningEffort: "medium",
|
|
2685
|
+
inputPrice: 1.25,
|
|
2686
|
+
outputPrice: 10,
|
|
2687
|
+
cacheReadsPrice: 0.125,
|
|
2688
|
+
supportsTemperature: false,
|
|
2689
|
+
tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }],
|
|
2690
|
+
description: "GPT-5.1 Codex: A version of GPT-5.1 optimized for agentic coding in Codex"
|
|
2691
|
+
},
|
|
2692
|
+
"gpt-5.1-codex-mini": {
|
|
2693
|
+
maxTokens: 128e3,
|
|
2694
|
+
contextWindow: 4e5,
|
|
2695
|
+
supportsImages: true,
|
|
2696
|
+
supportsPromptCache: true,
|
|
2697
|
+
promptCacheRetention: "24h",
|
|
2698
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
2699
|
+
reasoningEffort: "medium",
|
|
2700
|
+
inputPrice: 0.25,
|
|
2701
|
+
outputPrice: 2,
|
|
2702
|
+
cacheReadsPrice: 0.025,
|
|
2703
|
+
supportsTemperature: false,
|
|
2704
|
+
description: "GPT-5.1 Codex mini: A version of GPT-5.1 optimized for agentic coding in Codex"
|
|
2705
|
+
},
|
|
2706
|
+
"gpt-5": {
|
|
2653
2707
|
maxTokens: 128e3,
|
|
2654
2708
|
contextWindow: 4e5,
|
|
2655
2709
|
supportsImages: true,
|
|
2656
2710
|
supportsPromptCache: true,
|
|
2657
|
-
supportsReasoningEffort:
|
|
2711
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2658
2712
|
reasoningEffort: "medium",
|
|
2659
2713
|
inputPrice: 1.25,
|
|
2660
2714
|
outputPrice: 10,
|
|
2661
|
-
cacheReadsPrice: 0.
|
|
2662
|
-
description: "GPT-5: The best model for coding and agentic tasks across domains",
|
|
2663
|
-
// supportsVerbosity is a new capability; ensure ModelInfo includes it
|
|
2715
|
+
cacheReadsPrice: 0.125,
|
|
2664
2716
|
supportsVerbosity: true,
|
|
2665
2717
|
supportsTemperature: false,
|
|
2666
2718
|
tiers: [
|
|
2667
2719
|
{ name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
|
|
2668
2720
|
{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
|
|
2669
|
-
]
|
|
2721
|
+
],
|
|
2722
|
+
description: "GPT-5: The best model for coding and agentic tasks across domains"
|
|
2670
2723
|
},
|
|
2671
|
-
"gpt-5-mini
|
|
2724
|
+
"gpt-5-mini": {
|
|
2672
2725
|
maxTokens: 128e3,
|
|
2673
2726
|
contextWindow: 4e5,
|
|
2674
2727
|
supportsImages: true,
|
|
2675
2728
|
supportsPromptCache: true,
|
|
2676
|
-
supportsReasoningEffort:
|
|
2729
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2677
2730
|
reasoningEffort: "medium",
|
|
2678
2731
|
inputPrice: 0.25,
|
|
2679
2732
|
outputPrice: 2,
|
|
2680
|
-
cacheReadsPrice: 0.
|
|
2681
|
-
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
|
|
2733
|
+
cacheReadsPrice: 0.025,
|
|
2682
2734
|
supportsVerbosity: true,
|
|
2683
2735
|
supportsTemperature: false,
|
|
2684
2736
|
tiers: [
|
|
2685
2737
|
{ name: "flex", contextWindow: 4e5, inputPrice: 0.125, outputPrice: 1, cacheReadsPrice: 0.0125 },
|
|
2686
2738
|
{ name: "priority", contextWindow: 4e5, inputPrice: 0.45, outputPrice: 3.6, cacheReadsPrice: 0.045 }
|
|
2687
|
-
]
|
|
2739
|
+
],
|
|
2740
|
+
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks"
|
|
2688
2741
|
},
|
|
2689
|
-
"gpt-5-
|
|
2742
|
+
"gpt-5-codex": {
|
|
2743
|
+
maxTokens: 128e3,
|
|
2744
|
+
contextWindow: 4e5,
|
|
2745
|
+
supportsImages: true,
|
|
2746
|
+
supportsPromptCache: true,
|
|
2747
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
2748
|
+
reasoningEffort: "medium",
|
|
2749
|
+
inputPrice: 1.25,
|
|
2750
|
+
outputPrice: 10,
|
|
2751
|
+
cacheReadsPrice: 0.125,
|
|
2752
|
+
supportsTemperature: false,
|
|
2753
|
+
tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }],
|
|
2754
|
+
description: "GPT-5-Codex: A version of GPT-5 optimized for agentic coding in Codex"
|
|
2755
|
+
},
|
|
2756
|
+
"gpt-5-nano": {
|
|
2690
2757
|
maxTokens: 128e3,
|
|
2691
2758
|
contextWindow: 4e5,
|
|
2692
2759
|
supportsImages: true,
|
|
2693
2760
|
supportsPromptCache: true,
|
|
2694
|
-
supportsReasoningEffort:
|
|
2761
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
2695
2762
|
reasoningEffort: "medium",
|
|
2696
2763
|
inputPrice: 0.05,
|
|
2697
2764
|
outputPrice: 0.4,
|
|
2698
|
-
cacheReadsPrice:
|
|
2699
|
-
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
|
|
2765
|
+
cacheReadsPrice: 5e-3,
|
|
2700
2766
|
supportsVerbosity: true,
|
|
2701
2767
|
supportsTemperature: false,
|
|
2702
|
-
tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }]
|
|
2768
|
+
tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }],
|
|
2769
|
+
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5"
|
|
2703
2770
|
},
|
|
2704
|
-
"gpt-5-
|
|
2771
|
+
"gpt-5-chat-latest": {
|
|
2705
2772
|
maxTokens: 128e3,
|
|
2706
2773
|
contextWindow: 4e5,
|
|
2707
2774
|
supportsImages: true,
|
|
2708
2775
|
supportsPromptCache: true,
|
|
2709
|
-
supportsReasoningEffort: true,
|
|
2710
|
-
reasoningEffort: "medium",
|
|
2711
2776
|
inputPrice: 1.25,
|
|
2712
2777
|
outputPrice: 10,
|
|
2713
|
-
cacheReadsPrice: 0.
|
|
2714
|
-
description: "GPT-5
|
|
2715
|
-
supportsVerbosity: true,
|
|
2716
|
-
supportsTemperature: false
|
|
2778
|
+
cacheReadsPrice: 0.125,
|
|
2779
|
+
description: "GPT-5 Chat: Optimized for conversational AI and non-reasoning tasks"
|
|
2717
2780
|
},
|
|
2718
2781
|
"gpt-4.1": {
|
|
2719
2782
|
maxTokens: 32768,
|
|
@@ -2762,7 +2825,7 @@ var openAiNativeModels = {
|
|
|
2762
2825
|
inputPrice: 2,
|
|
2763
2826
|
outputPrice: 8,
|
|
2764
2827
|
cacheReadsPrice: 0.5,
|
|
2765
|
-
supportsReasoningEffort:
|
|
2828
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
2766
2829
|
reasoningEffort: "medium",
|
|
2767
2830
|
supportsTemperature: false,
|
|
2768
2831
|
tiers: [
|
|
@@ -2800,7 +2863,7 @@ var openAiNativeModels = {
|
|
|
2800
2863
|
inputPrice: 1.1,
|
|
2801
2864
|
outputPrice: 4.4,
|
|
2802
2865
|
cacheReadsPrice: 0.275,
|
|
2803
|
-
supportsReasoningEffort:
|
|
2866
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
2804
2867
|
reasoningEffort: "medium",
|
|
2805
2868
|
supportsTemperature: false,
|
|
2806
2869
|
tiers: [
|
|
@@ -2838,7 +2901,7 @@ var openAiNativeModels = {
|
|
|
2838
2901
|
inputPrice: 1.1,
|
|
2839
2902
|
outputPrice: 4.4,
|
|
2840
2903
|
cacheReadsPrice: 0.55,
|
|
2841
|
-
supportsReasoningEffort:
|
|
2904
|
+
supportsReasoningEffort: ["low", "medium", "high"],
|
|
2842
2905
|
reasoningEffort: "medium",
|
|
2843
2906
|
supportsTemperature: false
|
|
2844
2907
|
},
|
|
@@ -2927,9 +2990,61 @@ var openAiNativeModels = {
|
|
|
2927
2990
|
supportsPromptCache: false,
|
|
2928
2991
|
inputPrice: 1.5,
|
|
2929
2992
|
outputPrice: 6,
|
|
2930
|
-
cacheReadsPrice: 0,
|
|
2993
|
+
cacheReadsPrice: 0.375,
|
|
2931
2994
|
supportsTemperature: false,
|
|
2932
2995
|
description: "Codex Mini: Cloud-based software engineering agent powered by codex-1, a version of o3 optimized for coding tasks. Trained with reinforcement learning to generate human-style code, adhere to instructions, and iteratively run tests."
|
|
2996
|
+
},
|
|
2997
|
+
// Dated clones (snapshots) preserved for backward compatibility
|
|
2998
|
+
"gpt-5-2025-08-07": {
|
|
2999
|
+
maxTokens: 128e3,
|
|
3000
|
+
contextWindow: 4e5,
|
|
3001
|
+
supportsImages: true,
|
|
3002
|
+
supportsPromptCache: true,
|
|
3003
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3004
|
+
reasoningEffort: "medium",
|
|
3005
|
+
inputPrice: 1.25,
|
|
3006
|
+
outputPrice: 10,
|
|
3007
|
+
cacheReadsPrice: 0.125,
|
|
3008
|
+
supportsVerbosity: true,
|
|
3009
|
+
supportsTemperature: false,
|
|
3010
|
+
tiers: [
|
|
3011
|
+
{ name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
|
|
3012
|
+
{ name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
|
|
3013
|
+
],
|
|
3014
|
+
description: "GPT-5: The best model for coding and agentic tasks across domains"
|
|
3015
|
+
},
|
|
3016
|
+
"gpt-5-mini-2025-08-07": {
|
|
3017
|
+
maxTokens: 128e3,
|
|
3018
|
+
contextWindow: 4e5,
|
|
3019
|
+
supportsImages: true,
|
|
3020
|
+
supportsPromptCache: true,
|
|
3021
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3022
|
+
reasoningEffort: "medium",
|
|
3023
|
+
inputPrice: 0.25,
|
|
3024
|
+
outputPrice: 2,
|
|
3025
|
+
cacheReadsPrice: 0.025,
|
|
3026
|
+
supportsVerbosity: true,
|
|
3027
|
+
supportsTemperature: false,
|
|
3028
|
+
tiers: [
|
|
3029
|
+
{ name: "flex", contextWindow: 4e5, inputPrice: 0.125, outputPrice: 1, cacheReadsPrice: 0.0125 },
|
|
3030
|
+
{ name: "priority", contextWindow: 4e5, inputPrice: 0.45, outputPrice: 3.6, cacheReadsPrice: 0.045 }
|
|
3031
|
+
],
|
|
3032
|
+
description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks"
|
|
3033
|
+
},
|
|
3034
|
+
"gpt-5-nano-2025-08-07": {
|
|
3035
|
+
maxTokens: 128e3,
|
|
3036
|
+
contextWindow: 4e5,
|
|
3037
|
+
supportsImages: true,
|
|
3038
|
+
supportsPromptCache: true,
|
|
3039
|
+
supportsReasoningEffort: ["minimal", "low", "medium", "high"],
|
|
3040
|
+
reasoningEffort: "medium",
|
|
3041
|
+
inputPrice: 0.05,
|
|
3042
|
+
outputPrice: 0.4,
|
|
3043
|
+
cacheReadsPrice: 5e-3,
|
|
3044
|
+
supportsVerbosity: true,
|
|
3045
|
+
supportsTemperature: false,
|
|
3046
|
+
tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }],
|
|
3047
|
+
description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5"
|
|
2933
3048
|
}
|
|
2934
3049
|
};
|
|
2935
3050
|
var openAiModelInfoSaneDefaults = {
|
|
@@ -2942,7 +3057,6 @@ var openAiModelInfoSaneDefaults = {
|
|
|
2942
3057
|
};
|
|
2943
3058
|
var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
|
|
2944
3059
|
var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
|
|
2945
|
-
var GPT5_DEFAULT_TEMPERATURE = 1;
|
|
2946
3060
|
var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
|
|
2947
3061
|
|
|
2948
3062
|
// src/providers/openrouter.ts
|
|
@@ -2952,6 +3066,7 @@ var openRouterDefaultModelInfo = {
|
|
|
2952
3066
|
contextWindow: 2e5,
|
|
2953
3067
|
supportsImages: true,
|
|
2954
3068
|
supportsPromptCache: true,
|
|
3069
|
+
supportsNativeTools: true,
|
|
2955
3070
|
inputPrice: 3,
|
|
2956
3071
|
outputPrice: 15,
|
|
2957
3072
|
cacheWritesPrice: 3.75,
|
|
@@ -3060,6 +3175,16 @@ var requestyDefaultModelInfo = {
|
|
|
3060
3175
|
import { z as z7 } from "zod";
|
|
3061
3176
|
var rooDefaultModelId = "xai/grok-code-fast-1";
|
|
3062
3177
|
var rooModels = {};
|
|
3178
|
+
var rooModelDefaults = {
|
|
3179
|
+
// Add model-specific defaults below.
|
|
3180
|
+
// You can configure defaultToolProtocol and other ModelInfo fields for specific model IDs.
|
|
3181
|
+
"anthropic/claude-haiku-4.5": {
|
|
3182
|
+
defaultToolProtocol: TOOL_PROTOCOL.NATIVE
|
|
3183
|
+
},
|
|
3184
|
+
"minimax/minimax-m2:free": {
|
|
3185
|
+
defaultToolProtocol: TOOL_PROTOCOL.NATIVE
|
|
3186
|
+
}
|
|
3187
|
+
};
|
|
3063
3188
|
var RooPricingSchema = z7.object({
|
|
3064
3189
|
input: z7.string(),
|
|
3065
3190
|
output: z7.string(),
|
|
@@ -4346,11 +4471,13 @@ var baseProviderSettingsSchema = z8.object({
|
|
|
4346
4471
|
consecutiveMistakeLimit: z8.number().min(0).optional(),
|
|
4347
4472
|
// Model reasoning.
|
|
4348
4473
|
enableReasoningEffort: z8.boolean().optional(),
|
|
4349
|
-
reasoningEffort:
|
|
4474
|
+
reasoningEffort: reasoningEffortSettingSchema.optional(),
|
|
4350
4475
|
modelMaxTokens: z8.number().optional(),
|
|
4351
4476
|
modelMaxThinkingTokens: z8.number().optional(),
|
|
4352
4477
|
// Model verbosity.
|
|
4353
|
-
verbosity: verbosityLevelsSchema.optional()
|
|
4478
|
+
verbosity: verbosityLevelsSchema.optional(),
|
|
4479
|
+
// Tool protocol override for this profile.
|
|
4480
|
+
toolProtocol: z8.enum(["xml", "native"]).optional()
|
|
4354
4481
|
});
|
|
4355
4482
|
var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
|
|
4356
4483
|
apiModelId: z8.string().optional()
|
|
@@ -4827,7 +4954,8 @@ var experimentIds = [
|
|
|
4827
4954
|
"multiFileApplyDiff",
|
|
4828
4955
|
"preventFocusDisruption",
|
|
4829
4956
|
"imageGeneration",
|
|
4830
|
-
"runSlashCommand"
|
|
4957
|
+
"runSlashCommand",
|
|
4958
|
+
"nativeToolCalling"
|
|
4831
4959
|
];
|
|
4832
4960
|
var experimentIdsSchema = z10.enum(experimentIds);
|
|
4833
4961
|
var experimentsSchema = z10.object({
|
|
@@ -4835,7 +4963,8 @@ var experimentsSchema = z10.object({
|
|
|
4835
4963
|
multiFileApplyDiff: z10.boolean().optional(),
|
|
4836
4964
|
preventFocusDisruption: z10.boolean().optional(),
|
|
4837
4965
|
imageGeneration: z10.boolean().optional(),
|
|
4838
|
-
runSlashCommand: z10.boolean().optional()
|
|
4966
|
+
runSlashCommand: z10.boolean().optional(),
|
|
4967
|
+
nativeToolCalling: z10.boolean().optional()
|
|
4839
4968
|
});
|
|
4840
4969
|
|
|
4841
4970
|
// src/telemetry.ts
|
|
@@ -5227,6 +5356,12 @@ var globalSettingsSchema = z14.object({
|
|
|
5227
5356
|
* @default true
|
|
5228
5357
|
*/
|
|
5229
5358
|
includeCurrentCost: z14.boolean().optional(),
|
|
5359
|
+
/**
|
|
5360
|
+
* Maximum number of git status file entries to include in the environment details.
|
|
5361
|
+
* Set to 0 to disable git status. The header (branch, commits) is always included when > 0.
|
|
5362
|
+
* @default 0
|
|
5363
|
+
*/
|
|
5364
|
+
maxGitStatusFiles: z14.number().optional(),
|
|
5230
5365
|
/**
|
|
5231
5366
|
* Whether to include diagnostic messages (errors, warnings) in tool outputs
|
|
5232
5367
|
* @default true
|
|
@@ -5394,6 +5529,7 @@ var EVALS_SETTINGS = {
|
|
|
5394
5529
|
rateLimitSeconds: 0,
|
|
5395
5530
|
maxOpenTabsContext: 20,
|
|
5396
5531
|
maxWorkspaceFiles: 200,
|
|
5532
|
+
maxGitStatusFiles: 20,
|
|
5397
5533
|
showRooIgnoredFiles: true,
|
|
5398
5534
|
maxReadFileLine: -1,
|
|
5399
5535
|
// -1 to enable full file reading.
|
|
@@ -6036,7 +6172,6 @@ export {
|
|
|
6036
6172
|
GLOBAL_SECRET_KEYS,
|
|
6037
6173
|
GLOBAL_SETTINGS_KEYS,
|
|
6038
6174
|
GLOBAL_STATE_KEYS,
|
|
6039
|
-
GPT5_DEFAULT_TEMPERATURE,
|
|
6040
6175
|
HEARTBEAT_INTERVAL_MS,
|
|
6041
6176
|
HUGGINGFACE_API_URL,
|
|
6042
6177
|
HUGGINGFACE_CACHE_DURATION,
|
|
@@ -6071,6 +6206,7 @@ export {
|
|
|
6071
6206
|
RooModelsResponseSchema,
|
|
6072
6207
|
RooPricingSchema,
|
|
6073
6208
|
SECRET_STATE_KEYS,
|
|
6209
|
+
TOOL_PROTOCOL,
|
|
6074
6210
|
TaskBridgeCommandName,
|
|
6075
6211
|
TaskBridgeEventName,
|
|
6076
6212
|
TaskCommandName,
|
|
@@ -6142,6 +6278,7 @@ export {
|
|
|
6142
6278
|
geminiModels,
|
|
6143
6279
|
getApiProtocol,
|
|
6144
6280
|
getClaudeCodeModelId,
|
|
6281
|
+
getEffectiveProtocol,
|
|
6145
6282
|
getModelId,
|
|
6146
6283
|
getProviderDefaultModelId,
|
|
6147
6284
|
gitPropertiesSchema,
|
|
@@ -6173,6 +6310,7 @@ export {
|
|
|
6173
6310
|
isLanguage,
|
|
6174
6311
|
isLocalProvider,
|
|
6175
6312
|
isModelParameter,
|
|
6313
|
+
isNativeProtocol,
|
|
6176
6314
|
isNonBlockingAsk,
|
|
6177
6315
|
isProviderName,
|
|
6178
6316
|
isResumableAsk,
|
|
@@ -6229,8 +6367,12 @@ export {
|
|
|
6229
6367
|
queuedMessageSchema,
|
|
6230
6368
|
qwenCodeDefaultModelId,
|
|
6231
6369
|
qwenCodeModels,
|
|
6370
|
+
reasoningEffortExtendedSchema,
|
|
6371
|
+
reasoningEffortSettingSchema,
|
|
6372
|
+
reasoningEffortSettingValues,
|
|
6232
6373
|
reasoningEffortWithMinimalSchema,
|
|
6233
6374
|
reasoningEfforts,
|
|
6375
|
+
reasoningEffortsExtended,
|
|
6234
6376
|
reasoningEffortsSchema,
|
|
6235
6377
|
requestyDefaultModelId,
|
|
6236
6378
|
requestyDefaultModelInfo,
|
|
@@ -6239,6 +6381,7 @@ export {
|
|
|
6239
6381
|
rooCodeSettingsSchema,
|
|
6240
6382
|
rooCodeTelemetryEventSchema,
|
|
6241
6383
|
rooDefaultModelId,
|
|
6384
|
+
rooModelDefaults,
|
|
6242
6385
|
rooModels,
|
|
6243
6386
|
sambaNovaDefaultModelId,
|
|
6244
6387
|
sambaNovaModels,
|