@roo-code/types 1.79.0 → 1.81.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -36,6 +36,7 @@ function isResumableAsk(ask) {
36
36
  return resumableAsks.includes(ask);
37
37
  }
38
38
  var interactiveAsks = [
39
+ "followup",
39
40
  "command",
40
41
  "tool",
41
42
  "browser_action_launch",
@@ -387,6 +388,8 @@ var modelInfoSchema = z5.object({
387
388
  minTokensPerCachePoint: z5.number().optional(),
388
389
  maxCachePoints: z5.number().optional(),
389
390
  cachableFields: z5.array(z5.string()).optional(),
391
+ // Flag to indicate if the model is deprecated and should not be used
392
+ deprecated: z5.boolean().optional(),
390
393
  /**
391
394
  * Service tiers with pricing information.
392
395
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -452,6 +455,39 @@ var codebaseIndexProviderSchema = z6.object({
452
455
  // src/providers/anthropic.ts
453
456
  var anthropicDefaultModelId = "claude-sonnet-4-20250514";
454
457
  var anthropicModels = {
458
+ "claude-sonnet-4-5": {
459
+ maxTokens: 64e3,
460
+ // Overridden to 8k if `enableReasoningEffort` is false.
461
+ contextWindow: 2e5,
462
+ // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
463
+ supportsImages: true,
464
+ supportsComputerUse: true,
465
+ supportsPromptCache: true,
466
+ inputPrice: 3,
467
+ // $3 per million input tokens (≤200K context)
468
+ outputPrice: 15,
469
+ // $15 per million output tokens (≤200K context)
470
+ cacheWritesPrice: 3.75,
471
+ // $3.75 per million tokens
472
+ cacheReadsPrice: 0.3,
473
+ // $0.30 per million tokens
474
+ supportsReasoningBudget: true,
475
+ // Tiered pricing for extended context (requires beta flag 'context-1m-2025-08-07')
476
+ tiers: [
477
+ {
478
+ contextWindow: 1e6,
479
+ // 1M tokens with beta flag
480
+ inputPrice: 6,
481
+ // $6 per million input tokens (>200K context)
482
+ outputPrice: 22.5,
483
+ // $22.50 per million output tokens (>200K context)
484
+ cacheWritesPrice: 7.5,
485
+ // $7.50 per million tokens (>200K context)
486
+ cacheReadsPrice: 0.6
487
+ // $0.60 per million tokens (>200K context)
488
+ }
489
+ ]
490
+ },
455
491
  "claude-sonnet-4-20250514": {
456
492
  maxTokens: 64e3,
457
493
  // Overridden to 8k if `enableReasoningEffort` is false.
@@ -604,6 +640,21 @@ var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
604
640
  var bedrockDefaultModelId = "anthropic.claude-sonnet-4-20250514-v1:0";
605
641
  var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
606
642
  var bedrockModels = {
643
+ "anthropic.claude-sonnet-4-5-20250929-v1:0": {
644
+ maxTokens: 8192,
645
+ contextWindow: 2e5,
646
+ supportsImages: true,
647
+ supportsComputerUse: true,
648
+ supportsPromptCache: true,
649
+ supportsReasoningBudget: true,
650
+ inputPrice: 3,
651
+ outputPrice: 15,
652
+ cacheWritesPrice: 3.75,
653
+ cacheReadsPrice: 0.3,
654
+ minTokensPerCachePoint: 1024,
655
+ maxCachePoints: 4,
656
+ cachableFields: ["system", "messages", "tools"]
657
+ },
607
658
  "amazon.nova-pro-v1:0": {
608
659
  maxTokens: 5e3,
609
660
  contextWindow: 3e5,
@@ -1030,7 +1081,10 @@ var BEDROCK_REGIONS = [
1030
1081
  { value: "us-gov-east-1", label: "us-gov-east-1" },
1031
1082
  { value: "us-gov-west-1", label: "us-gov-west-1" }
1032
1083
  ].sort((a, b) => a.value.localeCompare(b.value));
1033
- var BEDROCK_CLAUDE_SONNET_4_MODEL_ID = "anthropic.claude-sonnet-4-20250514-v1:0";
1084
+ var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1085
+ "anthropic.claude-sonnet-4-20250514-v1:0",
1086
+ "anthropic.claude-sonnet-4-5-20250929-v1:0"
1087
+ ];
1034
1088
 
1035
1089
  // src/providers/cerebras.ts
1036
1090
  var cerebrasDefaultModelId = "qwen-3-coder-480b-free";
@@ -1140,6 +1194,33 @@ var chutesModels = {
1140
1194
  outputPrice: 0,
1141
1195
  description: "DeepSeek V3.1 model."
1142
1196
  },
1197
+ "deepseek-ai/DeepSeek-V3.1-Terminus": {
1198
+ maxTokens: 163840,
1199
+ contextWindow: 163840,
1200
+ supportsImages: false,
1201
+ supportsPromptCache: false,
1202
+ inputPrice: 0.23,
1203
+ outputPrice: 0.9,
1204
+ description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
1205
+ },
1206
+ "deepseek-ai/DeepSeek-V3.1-turbo": {
1207
+ maxTokens: 32768,
1208
+ contextWindow: 163840,
1209
+ supportsImages: false,
1210
+ supportsPromptCache: false,
1211
+ inputPrice: 1,
1212
+ outputPrice: 3,
1213
+ description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
1214
+ },
1215
+ "deepseek-ai/DeepSeek-V3.2-Exp": {
1216
+ maxTokens: 163840,
1217
+ contextWindow: 163840,
1218
+ supportsImages: false,
1219
+ supportsPromptCache: false,
1220
+ inputPrice: 0.25,
1221
+ outputPrice: 0.35,
1222
+ description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
1223
+ },
1143
1224
  "unsloth/Llama-3.3-70B-Instruct": {
1144
1225
  maxTokens: 32768,
1145
1226
  // From Groq
@@ -1340,6 +1421,24 @@ var chutesModels = {
1340
1421
  outputPrice: 0,
1341
1422
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
1342
1423
  },
1424
+ "zai-org/GLM-4.5-turbo": {
1425
+ maxTokens: 32768,
1426
+ contextWindow: 131072,
1427
+ supportsImages: false,
1428
+ supportsPromptCache: false,
1429
+ inputPrice: 1,
1430
+ outputPrice: 3,
1431
+ description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
1432
+ },
1433
+ "zai-org/GLM-4.6-FP8": {
1434
+ maxTokens: 32768,
1435
+ contextWindow: 202752,
1436
+ supportsImages: false,
1437
+ supportsPromptCache: false,
1438
+ inputPrice: 0,
1439
+ outputPrice: 0,
1440
+ description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
1441
+ },
1343
1442
  "Qwen/Qwen3-Coder-480B-A35B-Instruct-FP8": {
1344
1443
  maxTokens: 32768,
1345
1444
  contextWindow: 262144,
@@ -1393,6 +1492,15 @@ var chutesModels = {
1393
1492
  inputPrice: 0,
1394
1493
  outputPrice: 0,
1395
1494
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
1495
+ },
1496
+ "Qwen/Qwen3-VL-235B-A22B-Thinking": {
1497
+ maxTokens: 262144,
1498
+ contextWindow: 262144,
1499
+ supportsImages: true,
1500
+ supportsPromptCache: false,
1501
+ inputPrice: 0.16,
1502
+ outputPrice: 0.65,
1503
+ description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
1396
1504
  }
1397
1505
  };
1398
1506
 
@@ -1407,6 +1515,15 @@ function getClaudeCodeModelId(baseModelId, useVertex = false) {
1407
1515
  return useVertex ? convertModelNameForVertex(baseModelId) : baseModelId;
1408
1516
  }
1409
1517
  var claudeCodeModels = {
1518
+ "claude-sonnet-4-5": {
1519
+ ...anthropicModels["claude-sonnet-4-5"],
1520
+ supportsImages: false,
1521
+ supportsPromptCache: true,
1522
+ // Claude Code does report cache tokens
1523
+ supportsReasoningEffort: false,
1524
+ supportsReasoningBudget: false,
1525
+ requiredReasoningBudget: false
1526
+ },
1410
1527
  "claude-sonnet-4-20250514": {
1411
1528
  ...anthropicModels["claude-sonnet-4-20250514"],
1412
1529
  supportsImages: false,
@@ -2214,6 +2331,7 @@ var LITELLM_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2214
2331
  "vertex_ai/claude-opus-4-1@20250805",
2215
2332
  "vertex_ai/claude-opus-4@20250514",
2216
2333
  "vertex_ai/claude-sonnet-4@20250514",
2334
+ "vertex_ai/claude-sonnet-4-5@20250929",
2217
2335
  "openrouter/anthropic/claude-3.5-sonnet",
2218
2336
  "openrouter/anthropic/claude-3.5-sonnet:beta",
2219
2337
  "openrouter/anthropic/claude-3.7-sonnet",
@@ -2458,6 +2576,20 @@ var openAiNativeModels = {
2458
2576
  supportsTemperature: false,
2459
2577
  tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }]
2460
2578
  },
2579
+ "gpt-5-codex": {
2580
+ maxTokens: 128e3,
2581
+ contextWindow: 4e5,
2582
+ supportsImages: true,
2583
+ supportsPromptCache: true,
2584
+ supportsReasoningEffort: true,
2585
+ reasoningEffort: "medium",
2586
+ inputPrice: 1.25,
2587
+ outputPrice: 10,
2588
+ cacheReadsPrice: 0.13,
2589
+ description: "GPT-5-Codex: A version of GPT-5 optimized for agentic coding in Codex",
2590
+ supportsVerbosity: true,
2591
+ supportsTemperature: false
2592
+ },
2461
2593
  "gpt-4.1": {
2462
2594
  maxTokens: 32768,
2463
2595
  contextWindow: 1047576,
@@ -2722,6 +2854,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
2722
2854
  "anthropic/claude-3.7-sonnet:beta",
2723
2855
  "anthropic/claude-3.7-sonnet:thinking",
2724
2856
  "anthropic/claude-sonnet-4",
2857
+ "anthropic/claude-sonnet-4.5",
2725
2858
  "anthropic/claude-opus-4",
2726
2859
  "anthropic/claude-opus-4.1",
2727
2860
  "google/gemini-2.5-flash-preview",
@@ -2741,6 +2874,7 @@ var OPEN_ROUTER_COMPUTER_USE_MODELS = /* @__PURE__ */ new Set([
2741
2874
  "anthropic/claude-3.7-sonnet:beta",
2742
2875
  "anthropic/claude-3.7-sonnet:thinking",
2743
2876
  "anthropic/claude-sonnet-4",
2877
+ "anthropic/claude-sonnet-4.5",
2744
2878
  "anthropic/claude-opus-4",
2745
2879
  "anthropic/claude-opus-4.1"
2746
2880
  ]);
@@ -2754,6 +2888,7 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
2754
2888
  "anthropic/claude-opus-4",
2755
2889
  "anthropic/claude-opus-4.1",
2756
2890
  "anthropic/claude-sonnet-4",
2891
+ "anthropic/claude-sonnet-4.5",
2757
2892
  "google/gemini-2.5-pro-preview",
2758
2893
  "google/gemini-2.5-pro",
2759
2894
  "google/gemini-2.5-flash-preview-05-20",
@@ -2818,6 +2953,34 @@ var rooModels = {
2818
2953
  inputPrice: 0,
2819
2954
  outputPrice: 0,
2820
2955
  description: "A reasoning model that is blazing fast and excels at agentic coding, accessible for free through Roo Code Cloud for a limited time. (Note: the free prompts and completions are logged by xAI and used to improve the model.)"
2956
+ },
2957
+ "roo/code-supernova-1-million": {
2958
+ maxTokens: 3e4,
2959
+ contextWindow: 1e6,
2960
+ supportsImages: true,
2961
+ supportsPromptCache: true,
2962
+ inputPrice: 0,
2963
+ outputPrice: 0,
2964
+ description: "A versatile agentic coding stealth model with a 1M token context window that supports image inputs, accessible for free through Roo Code Cloud for a limited time. (Note: the free prompts and completions are logged by the model provider and used to improve the model.)"
2965
+ },
2966
+ "xai/grok-4-fast": {
2967
+ maxTokens: 3e4,
2968
+ contextWindow: 2e6,
2969
+ supportsImages: false,
2970
+ supportsPromptCache: false,
2971
+ inputPrice: 0,
2972
+ outputPrice: 0,
2973
+ description: "Grok 4 Fast is xAI's latest multimodal model with SOTA cost-efficiency and a 2M token context window. (Note: prompts and completions are logged by xAI and used to improve the model.)",
2974
+ deprecated: true
2975
+ },
2976
+ "deepseek/deepseek-chat-v3.1": {
2977
+ maxTokens: 16384,
2978
+ contextWindow: 163840,
2979
+ supportsImages: false,
2980
+ supportsPromptCache: false,
2981
+ inputPrice: 0,
2982
+ outputPrice: 0,
2983
+ description: "DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active). It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference."
2821
2984
  }
2822
2985
  };
2823
2986
 
@@ -2861,6 +3024,15 @@ var sambaNovaModels = {
2861
3024
  outputPrice: 4.5,
2862
3025
  description: "DeepSeek V3 model with 32K context window."
2863
3026
  },
3027
+ "DeepSeek-V3.1": {
3028
+ maxTokens: 8192,
3029
+ contextWindow: 32768,
3030
+ supportsImages: false,
3031
+ supportsPromptCache: false,
3032
+ inputPrice: 3,
3033
+ outputPrice: 4.5,
3034
+ description: "DeepSeek V3.1 model with 32K context window."
3035
+ },
2864
3036
  "DeepSeek-R1-Distill-Llama-70B": {
2865
3037
  maxTokens: 8192,
2866
3038
  contextWindow: 131072,
@@ -2896,6 +3068,15 @@ var sambaNovaModels = {
2896
3068
  inputPrice: 0.4,
2897
3069
  outputPrice: 0.8,
2898
3070
  description: "Alibaba Qwen 3 32B model with 8K context window."
3071
+ },
3072
+ "gpt-oss-120b": {
3073
+ maxTokens: 8192,
3074
+ contextWindow: 131072,
3075
+ supportsImages: false,
3076
+ supportsPromptCache: false,
3077
+ inputPrice: 0.22,
3078
+ outputPrice: 0.59,
3079
+ description: "OpenAI gpt oss 120b model with 128k context window."
2899
3080
  }
2900
3081
  };
2901
3082
 
@@ -2913,7 +3094,7 @@ var unboundDefaultModelInfo = {
2913
3094
  };
2914
3095
 
2915
3096
  // src/providers/vertex.ts
2916
- var vertexDefaultModelId = "claude-sonnet-4@20250514";
3097
+ var vertexDefaultModelId = "claude-sonnet-4-5@20250929";
2917
3098
  var vertexModels = {
2918
3099
  "gemini-2.5-flash-preview-05-20:thinking": {
2919
3100
  maxTokens: 65535,
@@ -3084,6 +3265,18 @@ var vertexModels = {
3084
3265
  cacheReadsPrice: 0.3,
3085
3266
  supportsReasoningBudget: true
3086
3267
  },
3268
+ "claude-sonnet-4-5@20250929": {
3269
+ maxTokens: 8192,
3270
+ contextWindow: 2e5,
3271
+ supportsImages: true,
3272
+ supportsComputerUse: true,
3273
+ supportsPromptCache: true,
3274
+ inputPrice: 3,
3275
+ outputPrice: 15,
3276
+ cacheWritesPrice: 3.75,
3277
+ cacheReadsPrice: 0.3,
3278
+ supportsReasoningBudget: true
3279
+ },
3087
3280
  "claude-opus-4-1@20250805": {
3088
3281
  maxTokens: 8192,
3089
3282
  contextWindow: 2e5,
@@ -3689,6 +3882,17 @@ var internationalZAiModels = {
3689
3882
  cacheWritesPrice: 0,
3690
3883
  cacheReadsPrice: 0.03,
3691
3884
  description: "GLM-4.5-Air is the lightweight version of GLM-4.5. It balances performance and cost-effectiveness, and can flexibly switch to hybrid thinking models."
3885
+ },
3886
+ "glm-4.6": {
3887
+ maxTokens: 98304,
3888
+ contextWindow: 204800,
3889
+ supportsImages: false,
3890
+ supportsPromptCache: true,
3891
+ inputPrice: 0.6,
3892
+ outputPrice: 2.2,
3893
+ cacheWritesPrice: 0,
3894
+ cacheReadsPrice: 0.11,
3895
+ description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations."
3692
3896
  }
3693
3897
  };
3694
3898
  var mainlandZAiDefaultModelId = "glm-4.5";
@@ -3754,6 +3958,43 @@ var mainlandZAiModels = {
3754
3958
  cacheReadsPrice: 0.02
3755
3959
  }
3756
3960
  ]
3961
+ },
3962
+ "glm-4.6": {
3963
+ maxTokens: 98304,
3964
+ contextWindow: 204800,
3965
+ supportsImages: false,
3966
+ supportsPromptCache: true,
3967
+ inputPrice: 0.29,
3968
+ outputPrice: 1.14,
3969
+ cacheWritesPrice: 0,
3970
+ cacheReadsPrice: 0.057,
3971
+ description: "GLM-4.6 is Zhipu's newest model with an extended context window of up to 200k tokens, providing enhanced capabilities for processing longer documents and conversations.",
3972
+ tiers: [
3973
+ {
3974
+ contextWindow: 32e3,
3975
+ inputPrice: 0.21,
3976
+ outputPrice: 1,
3977
+ cacheReadsPrice: 0.043
3978
+ },
3979
+ {
3980
+ contextWindow: 128e3,
3981
+ inputPrice: 0.29,
3982
+ outputPrice: 1.14,
3983
+ cacheReadsPrice: 0.057
3984
+ },
3985
+ {
3986
+ contextWindow: 2e5,
3987
+ inputPrice: 0.29,
3988
+ outputPrice: 1.14,
3989
+ cacheReadsPrice: 0.057
3990
+ },
3991
+ {
3992
+ contextWindow: Infinity,
3993
+ inputPrice: 0.29,
3994
+ outputPrice: 1.14,
3995
+ cacheReadsPrice: 0.057
3996
+ }
3997
+ ]
3757
3998
  }
3758
3999
  };
3759
4000
  var ZAI_DEFAULT_TEMPERATURE = 0;
@@ -3922,7 +4163,8 @@ var openAiSchema = baseProviderSettingsSchema.extend({
3922
4163
  var ollamaSchema = baseProviderSettingsSchema.extend({
3923
4164
  ollamaModelId: z7.string().optional(),
3924
4165
  ollamaBaseUrl: z7.string().optional(),
3925
- ollamaApiKey: z7.string().optional()
4166
+ ollamaApiKey: z7.string().optional(),
4167
+ ollamaNumCtx: z7.number().int().min(128).optional()
3926
4168
  });
3927
4169
  var vsCodeLmSchema = baseProviderSettingsSchema.extend({
3928
4170
  vsCodeLmModelSelector: z7.object({
@@ -4278,7 +4520,7 @@ var MODELS_BY_PROVIDER = {
4278
4520
  },
4279
4521
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
4280
4522
  zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
4281
- // Dynamic providers; models pulled from the respective APIs.
4523
+ // Dynamic providers; models pulled from remote APIs.
4282
4524
  glama: { id: "glama", label: "Glama", models: [] },
4283
4525
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
4284
4526
  litellm: { id: "litellm", label: "LiteLLM", models: [] },
@@ -4286,7 +4528,10 @@ var MODELS_BY_PROVIDER = {
4286
4528
  requesty: { id: "requesty", label: "Requesty", models: [] },
4287
4529
  unbound: { id: "unbound", label: "Unbound", models: [] },
4288
4530
  deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
4289
- "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
4531
+ "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] },
4532
+ // Local providers; models discovered from localhost endpoints.
4533
+ lmstudio: { id: "lmstudio", label: "LM Studio", models: [] },
4534
+ ollama: { id: "ollama", label: "Ollama", models: [] }
4290
4535
  };
4291
4536
 
4292
4537
  // src/history.ts
@@ -4364,11 +4609,15 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
4364
4609
  TelemetryEventName2["ACCOUNT_CONNECT_SUCCESS"] = "Account Connect Success";
4365
4610
  TelemetryEventName2["ACCOUNT_LOGOUT_CLICKED"] = "Account Logout Clicked";
4366
4611
  TelemetryEventName2["ACCOUNT_LOGOUT_SUCCESS"] = "Account Logout Success";
4612
+ TelemetryEventName2["FEATURED_PROVIDER_CLICKED"] = "Featured Provider Clicked";
4613
+ TelemetryEventName2["UPSELL_DISMISSED"] = "Upsell Dismissed";
4614
+ TelemetryEventName2["UPSELL_CLICKED"] = "Upsell Clicked";
4367
4615
  TelemetryEventName2["SCHEMA_VALIDATION_ERROR"] = "Schema Validation Error";
4368
4616
  TelemetryEventName2["DIFF_APPLICATION_ERROR"] = "Diff Application Error";
4369
4617
  TelemetryEventName2["SHELL_INTEGRATION_ERROR"] = "Shell Integration Error";
4370
4618
  TelemetryEventName2["CONSECUTIVE_MISTAKE_ERROR"] = "Consecutive Mistake Error";
4371
4619
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
4620
+ TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
4372
4621
  return TelemetryEventName2;
4373
4622
  })(TelemetryEventName || {});
4374
4623
  var staticAppPropertiesSchema = z10.object({
@@ -4443,6 +4692,9 @@ var rooCodeTelemetryEventSchema = z10.discriminatedUnion("type", [
4443
4692
  "Account Connect Success" /* ACCOUNT_CONNECT_SUCCESS */,
4444
4693
  "Account Logout Clicked" /* ACCOUNT_LOGOUT_CLICKED */,
4445
4694
  "Account Logout Success" /* ACCOUNT_LOGOUT_SUCCESS */,
4695
+ "Featured Provider Clicked" /* FEATURED_PROVIDER_CLICKED */,
4696
+ "Upsell Dismissed" /* UPSELL_DISMISSED */,
4697
+ "Upsell Clicked" /* UPSELL_CLICKED */,
4446
4698
  "Schema Validation Error" /* SCHEMA_VALIDATION_ERROR */,
4447
4699
  "Diff Application Error" /* DIFF_APPLICATION_ERROR */,
4448
4700
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
@@ -4456,6 +4708,14 @@ var rooCodeTelemetryEventSchema = z10.discriminatedUnion("type", [
4456
4708
  ]),
4457
4709
  properties: telemetryPropertiesSchema
4458
4710
  }),
4711
+ z10.object({
4712
+ type: z10.literal("Telemetry Settings Changed" /* TELEMETRY_SETTINGS_CHANGED */),
4713
+ properties: z10.object({
4714
+ ...telemetryPropertiesSchema.shape,
4715
+ previousSetting: telemetrySettingsSchema,
4716
+ newSetting: telemetrySettingsSchema
4717
+ })
4718
+ }),
4459
4719
  z10.object({
4460
4720
  type: z10.literal("Task Message" /* TASK_MESSAGE */),
4461
4721
  properties: z10.object({
@@ -4617,7 +4877,8 @@ var commandIds = [
4617
4877
  "importSettings",
4618
4878
  "focusInput",
4619
4879
  "acceptInput",
4620
- "focusPanel"
4880
+ "focusPanel",
4881
+ "toggleAutoApprove"
4621
4882
  ];
4622
4883
  var languages = [
4623
4884
  "ca",
@@ -4742,6 +5003,7 @@ var globalSettingsSchema = z13.object({
4742
5003
  enhancementApiConfigId: z13.string().optional(),
4743
5004
  includeTaskHistoryInEnhance: z13.boolean().optional(),
4744
5005
  historyPreviewCollapsed: z13.boolean().optional(),
5006
+ reasoningBlockCollapsed: z13.boolean().optional(),
4745
5007
  profileThresholds: z13.record(z13.string(), z13.number()).optional(),
4746
5008
  hasOpenedModeSelector: z13.boolean().optional(),
4747
5009
  lastModeExportPath: z13.string().optional(),
@@ -5272,6 +5534,27 @@ var TaskSocketEvents = /* @__PURE__ */ ((TaskSocketEvents2) => {
5272
5534
  TaskSocketEvents2["RELAYED_COMMAND"] = "task:relayed_command";
5273
5535
  return TaskSocketEvents2;
5274
5536
  })(TaskSocketEvents || {});
5537
+ var usageStatsSchema = z15.object({
5538
+ success: z15.boolean(),
5539
+ data: z15.object({
5540
+ dates: z15.array(z15.string()),
5541
+ // Array of date strings
5542
+ tasks: z15.array(z15.number()),
5543
+ // Array of task counts
5544
+ tokens: z15.array(z15.number()),
5545
+ // Array of token counts
5546
+ costs: z15.array(z15.number()),
5547
+ // Array of costs in USD
5548
+ totals: z15.object({
5549
+ tasks: z15.number(),
5550
+ tokens: z15.number(),
5551
+ cost: z15.number()
5552
+ // Total cost in USD
5553
+ })
5554
+ }),
5555
+ period: z15.number()
5556
+ // Period in days (e.g., 30)
5557
+ });
5275
5558
 
5276
5559
  // src/cookie-consent.ts
5277
5560
  var CONSENT_COOKIE_NAME = "roo-code-cookie-consent";
@@ -5388,7 +5671,7 @@ var mcpExecutionStatusSchema = z18.discriminatedUnion("status", [
5388
5671
 
5389
5672
  // src/single-file-read-models.ts
5390
5673
  function shouldUseSingleFileRead(modelId) {
5391
- return modelId.includes("grok-code-fast-1");
5674
+ return modelId.includes("grok-code-fast-1") || modelId.includes("code-supernova");
5392
5675
  }
5393
5676
 
5394
5677
  // src/todo.ts
@@ -5432,7 +5715,7 @@ export {
5432
5715
  ANTHROPIC_DEFAULT_MAX_TOKENS,
5433
5716
  ANTHROPIC_STYLE_PROVIDERS,
5434
5717
  AWS_INFERENCE_PROFILE_MAPPING,
5435
- BEDROCK_CLAUDE_SONNET_4_MODEL_ID,
5718
+ BEDROCK_1M_CONTEXT_MODEL_IDS,
5436
5719
  BEDROCK_DEFAULT_CONTEXT,
5437
5720
  BEDROCK_DEFAULT_TEMPERATURE,
5438
5721
  BEDROCK_MAX_TOKENS,
@@ -5681,6 +5964,7 @@ export {
5681
5964
  toolUsageSchema,
5682
5965
  unboundDefaultModelId,
5683
5966
  unboundDefaultModelInfo,
5967
+ usageStatsSchema,
5684
5968
  userFeaturesSchema,
5685
5969
  userSettingsConfigSchema,
5686
5970
  userSettingsDataSchema,