@roo-code/types 1.79.0 → 1.81.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -23,7 +23,7 @@ __export(index_exports, {
23
23
  ANTHROPIC_DEFAULT_MAX_TOKENS: () => ANTHROPIC_DEFAULT_MAX_TOKENS,
24
24
  ANTHROPIC_STYLE_PROVIDERS: () => ANTHROPIC_STYLE_PROVIDERS,
25
25
  AWS_INFERENCE_PROFILE_MAPPING: () => AWS_INFERENCE_PROFILE_MAPPING,
26
- BEDROCK_CLAUDE_SONNET_4_MODEL_ID: () => BEDROCK_CLAUDE_SONNET_4_MODEL_ID,
26
+ BEDROCK_1M_CONTEXT_MODEL_IDS: () => BEDROCK_1M_CONTEXT_MODEL_IDS,
27
27
  BEDROCK_DEFAULT_CONTEXT: () => BEDROCK_DEFAULT_CONTEXT,
28
28
  BEDROCK_DEFAULT_TEMPERATURE: () => BEDROCK_DEFAULT_TEMPERATURE,
29
29
  BEDROCK_MAX_TOKENS: () => BEDROCK_MAX_TOKENS,
@@ -272,6 +272,7 @@ __export(index_exports, {
272
272
  toolUsageSchema: () => toolUsageSchema,
273
273
  unboundDefaultModelId: () => unboundDefaultModelId,
274
274
  unboundDefaultModelInfo: () => unboundDefaultModelInfo,
275
+ usageStatsSchema: () => usageStatsSchema,
275
276
  userFeaturesSchema: () => userFeaturesSchema,
276
277
  userSettingsConfigSchema: () => userSettingsConfigSchema,
277
278
  userSettingsDataSchema: () => userSettingsDataSchema,
@@ -328,6 +329,7 @@ function isResumableAsk(ask) {
328
329
  return resumableAsks.includes(ask);
329
330
  }
330
331
  var interactiveAsks = [
332
+ "followup",
331
333
  "command",
332
334
  "tool",
333
335
  "browser_action_launch",
@@ -679,6 +681,8 @@ var modelInfoSchema = import_zod5.z.object({
679
681
  minTokensPerCachePoint: import_zod5.z.number().optional(),
680
682
  maxCachePoints: import_zod5.z.number().optional(),
681
683
  cachableFields: import_zod5.z.array(import_zod5.z.string()).optional(),
684
+ // Flag to indicate if the model is deprecated and should not be used
685
+ deprecated: import_zod5.z.boolean().optional(),
682
686
  /**
683
687
  * Service tiers with pricing information.
684
688
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -744,6 +748,39 @@ var codebaseIndexProviderSchema = import_zod6.z.object({
744
748
  // src/providers/anthropic.ts
745
749
  var anthropicDefaultModelId = "claude-sonnet-4-20250514";
746
750
  var anthropicModels = {
751
+ "claude-sonnet-4-5": {
752
+ maxTokens: 64e3,
753
+ // Overridden to 8k if `enableReasoningEffort` is false.
754
+ contextWindow: 2e5,
755
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
756
+ supportsImages: true,
757
+ supportsComputerUse: true,
758
+ supportsPromptCache: true,
759
+ inputPrice: 3,
760
+ // $3 per million input tokens (≤200K context)
761
+ outputPrice: 15,
762
+ // $15 per million output tokens (≤200K context)
763
+ cacheWritesPrice: 3.75,
764
+ // $3.75 per million tokens
765
+ cacheReadsPrice: 0.3,
766
+ // $0.30 per million tokens
767
+ supportsReasoningBudget: true,
768
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
769
+ tiers: [
770
+ {
771
+ contextWindow: 1e6,
772
+ // 1M tokens with beta flag
773
+ inputPrice: 6,
774
+ // $6 per million input tokens (>200K context)
775
+ outputPrice: 22.5,
776
+ // $22.50 per million output tokens (>200K context)
777
+ cacheWritesPrice: 7.5,
778
+ // $7.50 per million tokens (>200K context)
779
+ cacheReadsPrice: 0.6
780
+ // $0.60 per million tokens (>200K context)
781
+ }
782
+ ]
783
+ },
747
784
  "claude-sonnet-4-20250514": {
748
785
  maxTokens: 64e3,
749
786
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -896,6 +933,21 @@ var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
896
933
  var bedrockDefaultModelId = "anthropic.claude-sonnet-4-20250514-v1:0";
897
934
  var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
898
935
  var bedrockModels = {
936
+ "anthropic.claude-sonnet-4-5-20250929-v1:0": {
937
+ maxTokens: 8192,
938
+ contextWindow: 2e5,
939
+ supportsImages: true,
940
+ supportsComputerUse: true,
941
+ supportsPromptCache: true,
942
+ supportsReasoningBudget: true,
943
+ inputPrice: 3,
944
+ outputPrice: 15,
945
+ cacheWritesPrice: 3.75,
946
+ cacheReadsPrice: 0.3,
947
+ minTokensPerCachePoint: 1024,
948
+ maxCachePoints: 4,
949
+ cachableFields: ["system", "messages", "tools"]
950
+ },
899
951
  "amazon.nova-pro-v1:0": {
900
952
  maxTokens: 5e3,
901
953
  contextWindow: 3e5,
@@ -1322,7 +1374,10 @@ var BEDROCK_REGIONS = [
1322
1374
  { value: "us-gov-east-1", label: "us-gov-east-1" },
1323
1375
  { value: "us-gov-west-1", label: "us-gov-west-1" }
1324
1376
  ].sort((a, b) => a.value.localeCompare(b.value));
1325
- var BEDROCK_CLAUDE_SONNET_4_MODEL_ID = "anthropic.claude-sonnet-4-20250514-v1:0";
1377
+ var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1378
+ "anthropic.claude-sonnet-4-20250514-v1:0",
1379
+ "anthropic.claude-sonnet-4-5-20250929-v1:0"
1380
+ ];
1326
1381
 
1327
1382
  // src/providers/cerebras.ts
1328
1383
  var cerebrasDefaultModelId = "qwen-3-coder-480b-free";
@@ -1432,6 +1487,33 @@ var chutesModels = {
1432
1487
  outputPrice: 0,
1433
1488
  description: "DeepSeek V3.1 model."
1434
1489
  },
1490
+ "deepseek-ai/DeepSeek-V3.1-Terminus": {
1491
+ maxTokens: 163840,
1492
+ contextWindow: 163840,
1493
+ supportsImages: false,
1494
+ supportsPromptCache: false,
1495
+ inputPrice: 0.23,
1496
+ outputPrice: 0.9,
1497
+ description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
1498
+ },
1499
+ "deepseek-ai/DeepSeek-V3.1-turbo": {
1500
+ maxTokens: 32768,
1501
+ contextWindow: 163840,
1502
+ supportsImages: false,
1503
+ supportsPromptCache: false,
1504
+ inputPrice: 1,
1505
+ outputPrice: 3,
1506
+ description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
1507
+ },
1508
+ "deepseek-ai/DeepSeek-V3.2-Exp": {
1509
+ maxTokens: 163840,
1510
+ contextWindow: 163840,
1511
+ supportsImages: false,
1512
+ supportsPromptCache: false,
1513
+ inputPrice: 0.25,
1514
+ outputPrice: 0.35,
1515
+ description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
1516
+ },
1435
1517
  "unsloth/Llama-3.3-70B-Instruct": {
1436
1518
  maxTokens: 32768,
1437
1519
  // From Groq
@@ -1632,6 +1714,24 @@ var chutesModels = {
1632
1714
  outputPrice: 0,
1633
1715
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
1634
1716
  },
1717
+ "zai-org/GLM-4.5-turbo": {
1718
+ maxTokens: 32768,
1719
+ contextWindow: 131072,
1720
+ supportsImages: false,
1721
+ supportsPromptCache: false,
1722
+ inputPrice: 1,
1723
+ outputPrice: 3,
1724
+ description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
1725
+ },
1726
+ "zai-org/GLM-4.6-FP8": {
1727
+ maxTokens: 32768,
1728
+ contextWindow: 202752,
1729
+ supportsImages: false,
1730
+ supportsPromptCache: false,
1731
+ inputPrice: 0,
1732
+ outputPrice: 0,
1733
+ description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
1734
+ },
1635
1735
  "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": {
1636
1736
  maxTokens: 32768,
1637
1737
  contextWindow: 262144,
@@ -1685,6 +1785,15 @@ var chutesModels = {
1685
1785
  inputPrice: 0,
1686
1786
  outputPrice: 0,
1687
1787
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
1788
+ },
1789
+ "Qwen/Qwen3-VL-235B-A22B-Thinking": {
1790
+ maxTokens: 262144,
1791
+ contextWindow: 262144,
1792
+ supportsImages: true,
1793
+ supportsPromptCache: false,
1794
+ inputPrice: 0.16,
1795
+ outputPrice: 0.65,
1796
+ description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
1688
1797
  }
1689
1798
  };
1690
1799
 
@@ -1699,6 +1808,15 @@ function getClaudeCodeModelId(baseModelId, useVertex = false) {
1699
1808
  return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
1700
1809
  }
1701
1810
  var claudeCodeModels = {
1811
+ "claude-sonnet-4-5": {
1812
+ ...anthropicModels["claude-sonnet-4-5"],
1813
+ supportsImages: false,
1814
+ supportsPromptCache: true,
1815
+ // Claude Code does report cache tokens
1816
+ supportsReasoningEffort: false,
1817
+ supportsReasoningBudget: false,
1818
+ requiredReasoningBudget: false
1819
+ },
1702
1820
  "claude-sonnet-4-20250514": {
1703
1821
  ...anthropicModels["claude-sonnet-4-20250514"],
1704
1822
  supportsImages: false,
@@ -2506,6 +2624,7 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2506
2624
  "vertex_ai/claude-opus-4-1@20250805",
2507
2625
  "vertex_ai/claude-opus-4@20250514",
2508
2626
  "vertex_ai/claude-sonnet-4@20250514",
2627
+ "vertex_ai/claude-sonnet-4-5@20250929",
2509
2628
  "openrouter/anthropic/claude-3.5-sonnet",
2510
2629
  "openrouter/anthropic/claude-3.5-sonnet:beta",
2511
2630
  "openrouter/anthropic/claude-3.7-sonnet",
@@ -2750,6 +2869,20 @@ var openAiNativeModels = {
2750
2869
  supportsTemperature: false,
2751
2870
  tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }]
2752
2871
  },
2872
+ "gpt-5-codex": {
2873
+ maxTokens: 128e3,
2874
+ contextWindow: 4e5,
2875
+ supportsImages: true,
2876
+ supportsPromptCache: true,
2877
+ supportsReasoningEffort: true,
2878
+ reasoningEffort: "medium",
2879
+ inputPrice: 1.25,
2880
+ outputPrice: 10,
2881
+ cacheReadsPrice: 0.13,
2882
+ description: "GPT-5-Codex: A version of GPT-5 optimized for agentic coding in Codex",
2883
+ supportsVerbosity: true,
2884
+ supportsTemperature: false
2885
+ },
2753
2886
  "gpt-4.1": {
2754
2887
  maxTokens: 32768,
2755
2888
  contextWindow: 1047576,
@@ -3014,6 +3147,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3014
3147
  "anthropic/claude-3.7-sonnet:beta",
3015
3148
  "anthropic/claude-3.7-sonnet:thinking",
3016
3149
  "anthropic/claude-sonnet-4",
3150
+ "anthropic/claude-sonnet-4.5",
3017
3151
  "anthropic/claude-opus-4",
3018
3152
  "anthropic/claude-opus-4.1",
3019
3153
  "google/gemini-2.5-flash-preview",
@@ -3033,6 +3167,7 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
3033
3167
  "anthropic/claude-3.7-sonnet:beta",
3034
3168
  "anthropic/claude-3.7-sonnet:thinking",
3035
3169
  "anthropic/claude-sonnet-4",
3170
+ "anthropic/claude-sonnet-4.5",
3036
3171
  "anthropic/claude-opus-4",
3037
3172
  "anthropic/claude-opus-4.1"
3038
3173
  ]);
@@ -3046,6 +3181,7 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3046
3181
  "anthropic/claude-opus-4",
3047
3182
  "anthropic/claude-opus-4.1",
3048
3183
  "anthropic/claude-sonnet-4",
3184
+ "anthropic/claude-sonnet-4.5",
3049
3185
  "google/gemini-2.5-pro-preview",
3050
3186
  "google/gemini-2.5-pro",
3051
3187
  "google/gemini-2.5-flash-preview-05-20",
@@ -3110,6 +3246,34 @@ var rooModels = {
3110
3246
  inputPrice: 0,
3111
3247
  outputPrice: 0,
3112
3248
  description: "A reasoning model that is blazing fast and excels at agentic coding, accessible for free through Roo Code Cloud for a limited time. (Note: the free prompts and completions are logged by xAI and used to improve the model.)"
3249
+ },
3250
+ "roo/code-supernova-1-million": {
3251
+ maxTokens: 3e4,
3252
+ contextWindow: 1e6,
3253
+ supportsImages: true,
3254
+ supportsPromptCache: true,
3255
+ inputPrice: 0,
3256
+ outputPrice: 0,
3257
+ description: "A versatile agentic coding stealth model with a 1M token context window that supports image inputs, accessible for free through Roo Code Cloud for a limited time. (Note: the free prompts and completions are logged by the model provider and used to improve the model.)"
3258
+ },
3259
+ "xai/grok-4-fast": {
3260
+ maxTokens: 3e4,
3261
+ contextWindow: 2e6,
3262
+ supportsImages: false,
3263
+ supportsPromptCache: false,
3264
+ inputPrice: 0,
3265
+ outputPrice: 0,
3266
+ description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)",
3267
+ deprecated: true
3268
+ },
3269
+ "deepseek/deepseek-chat-v3.1": {
3270
+ maxTokens: 16384,
3271
+ contextWindow: 163840,
3272
+ supportsImages: false,
3273
+ supportsPromptCache: false,
3274
+ inputPrice: 0,
3275
+ outputPrice: 0,
3276
+ description: "DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active). It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference."
3113
3277
  }
3114
3278
  };
3115
3279
 
@@ -3153,6 +3317,15 @@ var sambaNovaModels = {
3153
3317
  outputPrice: 4.5,
3154
3318
  description: "DeepSeek V3 model with 32K context window."
3155
3319
  },
3320
+ "DeepSeek-V3.1": {
3321
+ maxTokens: 8192,
3322
+ contextWindow: 32768,
3323
+ supportsImages: false,
3324
+ supportsPromptCache: false,
3325
+ inputPrice: 3,
3326
+ outputPrice: 4.5,
3327
+ description: "DeepSeek V3.1 model with 32K context window."
3328
+ },
3156
3329
  "DeepSeek-R1-Distill-Llama-70B": {
3157
3330
  maxTokens: 8192,
3158
3331
  contextWindow: 131072,
@@ -3188,6 +3361,15 @@ var sambaNovaModels = {
3188
3361
  inputPrice: 0.4,
3189
3362
  outputPrice: 0.8,
3190
3363
  description: "Alibaba Qwen 3 32B model with 8K context window."
3364
+ },
3365
+ "gpt-oss-120b": {
3366
+ maxTokens: 8192,
3367
+ contextWindow: 131072,
3368
+ supportsImages: false,
3369
+ supportsPromptCache: false,
3370
+ inputPrice: 0.22,
3371
+ outputPrice: 0.59,
3372
+ description: "OpenAI gpt oss 120b model with 128k context window."
3191
3373
  }
3192
3374
  };
3193
3375
 
@@ -3205,7 +3387,7 @@ var unboundDefaultModelInfo = {
3205
3387
  };
3206
3388
 
3207
3389
  // src/providers/vertex.ts
3208
- var vertexDefaultModelId = "claude-sonnet-4@20250514";
3390
+ var vertexDefaultModelId = "claude-sonnet-4-5@20250929";
3209
3391
  var vertexModels = {
3210
3392
  "gemini-2.5-flash-preview-05-20:thinking": {
3211
3393
  maxTokens: 65535,
@@ -3376,6 +3558,18 @@ var vertexModels = {
3376
3558
  cacheReadsPrice: 0.3,
3377
3559
  supportsReasoningBudget: true
3378
3560
  },
3561
+ "claude-sonnet-4-5@20250929": {
3562
+ maxTokens: 8192,
3563
+ contextWindow: 2e5,
3564
+ supportsImages: true,
3565
+ supportsComputerUse: true,
3566
+ supportsPromptCache: true,
3567
+ inputPrice: 3,
3568
+ outputPrice: 15,
3569
+ cacheWritesPrice: 3.75,
3570
+ cacheReadsPrice: 0.3,
3571
+ supportsReasoningBudget: true
3572
+ },
3379
3573
  "claude-opus-4-1@20250805": {
3380
3574
  maxTokens: 8192,
3381
3575
  contextWindow: 2e5,
@@ -3981,6 +4175,17 @@ var internationalZAiModels = {
3981
4175
  cacheWritesPrice: 0,
3982
4176
  cacheReadsPrice: 0.03,
3983
4177
  description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
4178
+ },
4179
+ "glm-4.6": {
4180
+ maxTokens: 98304,
4181
+ contextWindow: 204800,
4182
+ supportsImages: false,
4183
+ supportsPromptCache: true,
4184
+ inputPrice: 0.6,
4185
+ outputPrice: 2.2,
4186
+ cacheWritesPrice: 0,
4187
+ cacheReadsPrice: 0.11,
4188
+ description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
3984
4189
  }
3985
4190
  };
3986
4191
  var mainlandZAiDefaultModelId = "glm-4.5";
@@ -4046,6 +4251,43 @@ var mainlandZAiModels = {
4046
4251
  cacheReadsPrice: 0.02
4047
4252
  }
4048
4253
  ]
4254
+ },
4255
+ "glm-4.6": {
4256
+ maxTokens: 98304,
4257
+ contextWindow: 204800,
4258
+ supportsImages: false,
4259
+ supportsPromptCache: true,
4260
+ inputPrice: 0.29,
4261
+ outputPrice: 1.14,
4262
+ cacheWritesPrice: 0,
4263
+ cacheReadsPrice: 0.057,
4264
+ description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations.",
4265
+ tiers: [
4266
+ {
4267
+ contextWindow: 32e3,
4268
+ inputPrice: 0.21,
4269
+ outputPrice: 1,
4270
+ cacheReadsPrice: 0.043
4271
+ },
4272
+ {
4273
+ contextWindow: 128e3,
4274
+ inputPrice: 0.29,
4275
+ outputPrice: 1.14,
4276
+ cacheReadsPrice: 0.057
4277
+ },
4278
+ {
4279
+ contextWindow: 2e5,
4280
+ inputPrice: 0.29,
4281
+ outputPrice: 1.14,
4282
+ cacheReadsPrice: 0.057
4283
+ },
4284
+ {
4285
+ contextWindow: Infinity,
4286
+ inputPrice: 0.29,
4287
+ outputPrice: 1.14,
4288
+ cacheReadsPrice: 0.057
4289
+ }
4290
+ ]
4049
4291
  }
4050
4292
  };
4051
4293
  var ZAI_DEFAULT_TEMPERATURE = 0;
@@ -4214,7 +4456,8 @@ var openAiSchema = baseProviderSettingsSchema.extend({
4214
4456
  var ollamaSchema = baseProviderSettingsSchema.extend({
4215
4457
  ollamaModelId: import_zod7.z.string().optional(),
4216
4458
  ollamaBaseUrl: import_zod7.z.string().optional(),
4217
- ollamaApiKey: import_zod7.z.string().optional()
4459
+ ollamaApiKey: import_zod7.z.string().optional(),
4460
+ ollamaNumCtx: import_zod7.z.number().int().min(128).optional()
4218
4461
  });
4219
4462
  var vsCodeLmSchema = baseProviderSettingsSchema.extend({
4220
4463
  vsCodeLmModelSelector: import_zod7.z.object({
@@ -4570,7 +4813,7 @@ var MODELS_BY_PROVIDER = {
4570
4813
  },
4571
4814
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
4572
4815
  zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
4573
- // Dynamic providers; models pulled from the respective APIs.
4816
+ // Dynamic providers; models pulled from remote APIs.
4574
4817
  glama: { id: "glama", label: "Glama", models: [] },
4575
4818
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
4576
4819
  litellm: { id: "litellm", label: "LiteLLM", models: [] },
@@ -4578,7 +4821,10 @@ var MODELS_BY_PROVIDER = {
4578
4821
  requesty: { id: "requesty", label: "Requesty", models: [] },
4579
4822
  unbound: { id: "unbound", label: "Unbound", models: [] },
4580
4823
  deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
4581
- "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
4824
+ "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] },
4825
+ // Local providers; models discovered from localhost endpoints.
4826
+ lmstudio: { id: "lmstudio", label: "LM Studio", models: [] },
4827
+ ollama: { id: "ollama", label: "Ollama", models: [] }
4582
4828
  };
4583
4829
 
4584
4830
  // src/history.ts
@@ -4656,11 +4902,15 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
4656
4902
  TelemetryEventName2["ACCOUNT_CONNECT_SUCCESS"] = "Account Connect Success";
4657
4903
  TelemetryEventName2["ACCOUNT_LOGOUT_CLICKED"] = "Account Logout Clicked";
4658
4904
  TelemetryEventName2["ACCOUNT_LOGOUT_SUCCESS"] = "Account Logout Success";
4905
+ TelemetryEventName2["FEATURED_PROVIDER_CLICKED"] = "Featured Provider Clicked";
4906
+ TelemetryEventName2["UPSELL_DISMISSED"] = "Upsell Dismissed";
4907
+ TelemetryEventName2["UPSELL_CLICKED"] = "Upsell Clicked";
4659
4908
  TelemetryEventName2["SCHEMA_VALIDATION_ERROR"] = "Schema Validation Error";
4660
4909
  TelemetryEventName2["DIFF_APPLICATION_ERROR"] = "Diff Application Error";
4661
4910
  TelemetryEventName2["SHELL_INTEGRATION_ERROR"] = "Shell Integration Error";
4662
4911
  TelemetryEventName2["CONSECUTIVE_MISTAKE_ERROR"] = "Consecutive Mistake Error";
4663
4912
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
4913
+ TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
4664
4914
  return TelemetryEventName2;
4665
4915
  })(TelemetryEventName || {});
4666
4916
  var staticAppPropertiesSchema = import_zod10.z.object({
@@ -4735,6 +4985,9 @@ var rooCodeTelemetryEventSchema = import_zod10.z.discriminatedUnion("type", [
4735
4985
  "Account Connect Success" /* ACCOUNT_CONNECT_SUCCESS */,
4736
4986
  "Account Logout Clicked" /* ACCOUNT_LOGOUT_CLICKED */,
4737
4987
  "Account Logout Success" /* ACCOUNT_LOGOUT_SUCCESS */,
4988
+ "Featured Provider Clicked" /* FEATURED_PROVIDER_CLICKED */,
4989
+ "Upsell Dismissed" /* UPSELL_DISMISSED */,
4990
+ "Upsell Clicked" /* UPSELL_CLICKED */,
4738
4991
  "Schema Validation Error" /* SCHEMA_VALIDATION_ERROR */,
4739
4992
  "Diff Application Error" /* DIFF_APPLICATION_ERROR */,
4740
4993
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
@@ -4748,6 +5001,14 @@ var rooCodeTelemetryEventSchema = import_zod10.z.discriminatedUnion("type", [
4748
5001
  ]),
4749
5002
  properties: telemetryPropertiesSchema
4750
5003
  }),
5004
+ import_zod10.z.object({
5005
+ type: import_zod10.z.literal("Telemetry Settings Changed" /* TELEMETRY_SETTINGS_CHANGED */),
5006
+ properties: import_zod10.z.object({
5007
+ ...telemetryPropertiesSchema.shape,
5008
+ previousSetting: telemetrySettingsSchema,
5009
+ newSetting: telemetrySettingsSchema
5010
+ })
5011
+ }),
4751
5012
  import_zod10.z.object({
4752
5013
  type: import_zod10.z.literal("Task Message" /* TASK_MESSAGE */),
4753
5014
  properties: import_zod10.z.object({
@@ -4909,7 +5170,8 @@ var commandIds = [
4909
5170
  "importSettings",
4910
5171
  "focusInput",
4911
5172
  "acceptInput",
4912
- "focusPanel"
5173
+ "focusPanel",
5174
+ "toggleAutoApprove"
4913
5175
  ];
4914
5176
  var languages = [
4915
5177
  "ca",
@@ -5034,6 +5296,7 @@ var globalSettingsSchema = import_zod13.z.object({
5034
5296
  enhancementApiConfigId: import_zod13.z.string().optional(),
5035
5297
  includeTaskHistoryInEnhance: import_zod13.z.boolean().optional(),
5036
5298
  historyPreviewCollapsed: import_zod13.z.boolean().optional(),
5299
+ reasoningBlockCollapsed: import_zod13.z.boolean().optional(),
5037
5300
  profileThresholds: import_zod13.z.record(import_zod13.z.string(), import_zod13.z.number()).optional(),
5038
5301
  hasOpenedModeSelector: import_zod13.z.boolean().optional(),
5039
5302
  lastModeExportPath: import_zod13.z.string().optional(),
@@ -5564,6 +5827,27 @@ var TaskSocketEvents = /* @__PURE__ */ ((TaskSocketEvents2) => {
5564
5827
  TaskSocketEvents2["RELAYED_COMMAND"] = "task:relayed_command";
5565
5828
  return TaskSocketEvents2;
5566
5829
  })(TaskSocketEvents || {});
5830
+ var usageStatsSchema = import_zod15.z.object({
5831
+ success: import_zod15.z.boolean(),
5832
+ data: import_zod15.z.object({
5833
+ dates: import_zod15.z.array(import_zod15.z.string()),
5834
+ // Array of date strings
5835
+ tasks: import_zod15.z.array(import_zod15.z.number()),
5836
+ // Array of task counts
5837
+ tokens: import_zod15.z.array(import_zod15.z.number()),
5838
+ // Array of token counts
5839
+ costs: import_zod15.z.array(import_zod15.z.number()),
5840
+ // Array of costs in USD
5841
+ totals: import_zod15.z.object({
5842
+ tasks: import_zod15.z.number(),
5843
+ tokens: import_zod15.z.number(),
5844
+ cost: import_zod15.z.number()
5845
+ // Total cost in USD
5846
+ })
5847
+ }),
5848
+ period: import_zod15.z.number()
5849
+ // Period in days (e.g., 30)
5850
+ });
5567
5851
 
5568
5852
  // src/cookie-consent.ts
5569
5853
  var CONSENT_COOKIE_NAME = "roo-code-cookie-consent";
@@ -5680,7 +5964,7 @@ var mcpExecutionStatusSchema = import_zod18.z.discriminatedUnion("status", [
5680
5964
 
5681
5965
  // src/single-file-read-models.ts
5682
5966
  function shouldUseSingleFileRead(modelId) {
5683
- return modelId.includes("grok-code-fast-1");
5967
+ return modelId.includes("grok-code-fast-1") || modelId.includes("code-supernova");
5684
5968
  }
5685
5969
 
5686
5970
  // src/todo.ts
@@ -5725,7 +6009,7 @@ var commandExecutionStatusSchema = import_zod20.z.discriminatedUnion("status", [
5725
6009
  ANTHROPIC_DEFAULT_MAX_TOKENS,
5726
6010
  ANTHROPIC_STYLE_PROVIDERS,
5727
6011
  AWS_INFERENCE_PROFILE_MAPPING,
5728
- BEDROCK_CLAUDE_SONNET_4_MODEL_ID,
6012
+ BEDROCK_1M_CONTEXT_MODEL_IDS,
5729
6013
  BEDROCK_DEFAULT_CONTEXT,
5730
6014
  BEDROCK_DEFAULT_TEMPERATURE,
5731
6015
  BEDROCK_MAX_TOKENS,
@@ -5974,6 +6258,7 @@ var commandExecutionStatusSchema = import_zod20.z.discriminatedUnion("status", [
5974
6258
  toolUsageSchema,
5975
6259
  unboundDefaultModelId,
5976
6260
  unboundDefaultModelInfo,
6261
+ usageStatsSchema,
5977
6262
  userFeaturesSchema,
5978
6263
  userSettingsConfigSchema,
5979
6264
  userSettingsDataSchema,