@roo-code/types 1.74.0 → 1.76.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -30,6 +30,8 @@ __export(index_exports, {
30
30
  BEDROCK_REGIONS: () => BEDROCK_REGIONS,
31
31
  CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS: () => CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
32
32
  CODEBASE_INDEX_DEFAULTS: () => CODEBASE_INDEX_DEFAULTS,
33
+ CONSENT_COOKIE_NAME: () => CONSENT_COOKIE_NAME,
34
+ COOKIE_CONSENT_EVENTS: () => COOKIE_CONSENT_EVENTS,
33
35
  ConnectionState: () => ConnectionState,
34
36
  DEEP_SEEK_DEFAULT_TEMPERATURE: () => DEEP_SEEK_DEFAULT_TEMPERATURE,
35
37
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT: () => DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
@@ -122,6 +124,8 @@ __export(index_exports, {
122
124
  customModePromptsSchema: () => customModePromptsSchema,
123
125
  customModesSettingsSchema: () => customModesSettingsSchema,
124
126
  customSupportPromptsSchema: () => customSupportPromptsSchema,
127
+ deepInfraDefaultModelId: () => deepInfraDefaultModelId,
128
+ deepInfraDefaultModelInfo: () => deepInfraDefaultModelInfo,
125
129
  deepSeekDefaultModelId: () => deepSeekDefaultModelId,
126
130
  deepSeekModels: () => deepSeekModels,
127
131
  discriminatedProviderSettingsWithIdSchema: () => discriminatedProviderSettingsWithIdSchema,
@@ -229,6 +233,8 @@ __export(index_exports, {
229
233
  rooModels: () => rooModels,
230
234
  sambaNovaDefaultModelId: () => sambaNovaDefaultModelId,
231
235
  sambaNovaModels: () => sambaNovaModels,
236
+ serviceTierSchema: () => serviceTierSchema,
237
+ serviceTiers: () => serviceTiers,
232
238
  shareResponseSchema: () => shareResponseSchema,
233
239
  shouldUseSingleFileRead: () => shouldUseSingleFileRead,
234
240
  staticAppPropertiesSchema: () => staticAppPropertiesSchema,
@@ -266,7 +272,9 @@ __export(index_exports, {
266
272
  vscodeLlmDefaultModelId: () => vscodeLlmDefaultModelId,
267
273
  vscodeLlmModels: () => vscodeLlmModels,
268
274
  xaiDefaultModelId: () => xaiDefaultModelId,
269
- xaiModels: () => xaiModels
275
+ xaiModels: () => xaiModels,
276
+ zaiApiLineConfigs: () => zaiApiLineConfigs,
277
+ zaiApiLineSchema: () => zaiApiLineSchema
270
278
  });
271
279
  module.exports = __toCommonJS(index_exports);
272
280
 
@@ -370,6 +378,7 @@ var clineMessageSchema = import_zod.z.object({
370
378
  contextCondense: contextCondenseSchema.optional(),
371
379
  isProtected: import_zod.z.boolean().optional(),
372
380
  apiProtocol: import_zod.z.union([import_zod.z.literal("openai"), import_zod.z.literal("anthropic")]).optional(),
381
+ isAnswered: import_zod.z.boolean().optional(),
373
382
  metadata: import_zod.z.object({
374
383
  gpt5: import_zod.z.object({
375
384
  previous_response_id: import_zod.z.string().optional(),
@@ -629,6 +638,8 @@ var reasoningEffortsSchema = import_zod5.z.enum(reasoningEfforts);
629
638
  var reasoningEffortWithMinimalSchema = import_zod5.z.union([reasoningEffortsSchema, import_zod5.z.literal("minimal")]);
630
639
  var verbosityLevels = ["low", "medium", "high"];
631
640
  var verbosityLevelsSchema = import_zod5.z.enum(verbosityLevels);
641
+ var serviceTiers = ["default", "flex", "priority"];
642
+ var serviceTierSchema = import_zod5.z.enum(serviceTiers);
632
643
  var modelParameters = ["max_tokens", "temperature", "reasoning", "include_reasoning"];
633
644
  var modelParametersSchema = import_zod5.z.enum(modelParameters);
634
645
  var isModelParameter = (value) => modelParameters.includes(value);
@@ -656,8 +667,15 @@ var modelInfoSchema = import_zod5.z.object({
656
667
  minTokensPerCachePoint: import_zod5.z.number().optional(),
657
668
  maxCachePoints: import_zod5.z.number().optional(),
658
669
  cachableFields: import_zod5.z.array(import_zod5.z.string()).optional(),
670
+ /**
671
+ * Service tiers with pricing information.
672
+ * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
673
+ * The top-level input/output/cache* fields represent the default/standard tier.
674
+ */
659
675
  tiers: import_zod5.z.array(
660
676
  import_zod5.z.object({
677
+ name: serviceTierSchema.optional(),
678
+ // Service tier name (flex, priority, etc.)
661
679
  contextWindow: import_zod5.z.number(),
662
680
  inputPrice: import_zod5.z.number().optional(),
663
681
  outputPrice: import_zod5.z.number().optional(),
@@ -1620,6 +1638,15 @@ var chutesModels = {
1620
1638
  outputPrice: 0.5926,
1621
1639
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
1622
1640
  },
1641
+ "moonshotai/Kimi-K2-Instruct-0905": {
1642
+ maxTokens: 32768,
1643
+ contextWindow: 262144,
1644
+ supportsImages: false,
1645
+ supportsPromptCache: false,
1646
+ inputPrice: 0.1999,
1647
+ outputPrice: 0.8001,
1648
+ description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
1649
+ },
1623
1650
  "Qwen/Qwen3-235B-A22B-Thinking-2507": {
1624
1651
  maxTokens: 32768,
1625
1652
  contextWindow: 262144,
@@ -1628,6 +1655,24 @@ var chutesModels = {
1628
1655
  inputPrice: 0.077968332,
1629
1656
  outputPrice: 0.31202496,
1630
1657
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
1658
+ },
1659
+ "Qwen/Qwen3-Next-80B-A3B-Instruct": {
1660
+ maxTokens: 32768,
1661
+ contextWindow: 131072,
1662
+ supportsImages: false,
1663
+ supportsPromptCache: false,
1664
+ inputPrice: 0,
1665
+ outputPrice: 0,
1666
+ description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
1667
+ },
1668
+ "Qwen/Qwen3-Next-80B-A3B-Thinking": {
1669
+ maxTokens: 32768,
1670
+ contextWindow: 131072,
1671
+ supportsImages: false,
1672
+ supportsPromptCache: false,
1673
+ inputPrice: 0,
1674
+ outputPrice: 0,
1675
+ description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
1631
1676
  }
1632
1677
  };
1633
1678
 
@@ -1707,14 +1752,14 @@ var deepSeekModels = {
1707
1752
  contextWindow: 128e3,
1708
1753
  supportsImages: false,
1709
1754
  supportsPromptCache: true,
1710
- inputPrice: 0.27,
1711
- // $0.27 per million tokens (cache miss)
1712
- outputPrice: 1.1,
1713
- // $1.10 per million tokens
1714
- cacheWritesPrice: 0.27,
1715
- // $0.27 per million tokens (cache miss)
1755
+ inputPrice: 0.56,
1756
+ // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1757
+ outputPrice: 1.68,
1758
+ // $1.68 per million tokens - Updated Sept 5, 2025
1759
+ cacheWritesPrice: 0.56,
1760
+ // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1716
1761
  cacheReadsPrice: 0.07,
1717
- // $0.07 per million tokens (cache hit).
1762
+ // $0.07 per million tokens (cache hit) - Updated Sept 5, 2025
1718
1763
  description: `DeepSeek-V3 achieves a significant breakthrough in inference speed over previous models. It tops the leaderboard among open-source models and rivals the most advanced closed-source models globally.`
1719
1764
  },
1720
1765
  "deepseek-reasoner": {
@@ -1723,14 +1768,14 @@ var deepSeekModels = {
1723
1768
  contextWindow: 128e3,
1724
1769
  supportsImages: false,
1725
1770
  supportsPromptCache: true,
1726
- inputPrice: 0.55,
1727
- // $0.55 per million tokens (cache miss)
1728
- outputPrice: 2.19,
1729
- // $2.19 per million tokens
1730
- cacheWritesPrice: 0.55,
1731
- // $0.55 per million tokens (cache miss)
1732
- cacheReadsPrice: 0.14,
1733
- // $0.14 per million tokens (cache hit)
1771
+ inputPrice: 0.56,
1772
+ // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1773
+ outputPrice: 1.68,
1774
+ // $1.68 per million tokens - Updated Sept 5, 2025
1775
+ cacheWritesPrice: 0.56,
1776
+ // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1777
+ cacheReadsPrice: 0.07,
1778
+ // $0.07 per million tokens (cache hit) - Updated Sept 5, 2025
1734
1779
  description: `DeepSeek-R1 achieves performance comparable to OpenAI-o1 across math, code, and reasoning tasks. Supports Chain of Thought reasoning with up to 64K output tokens.`
1735
1780
  }
1736
1781
  };
@@ -1840,8 +1885,18 @@ var featherlessModels = {
1840
1885
  var featherlessDefaultModelId = "deepseek-ai/DeepSeek-R1-0528";
1841
1886
 
1842
1887
  // src/providers/fireworks.ts
1843
- var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct";
1888
+ var fireworksDefaultModelId = "accounts/fireworks/models/kimi-k2-instruct-0905";
1844
1889
  var fireworksModels = {
1890
+ "accounts/fireworks/models/kimi-k2-instruct-0905": {
1891
+ maxTokens: 16384,
1892
+ contextWindow: 262144,
1893
+ supportsImages: false,
1894
+ supportsPromptCache: true,
1895
+ inputPrice: 0.6,
1896
+ outputPrice: 2.5,
1897
+ cacheReadsPrice: 0.15,
1898
+ description: "Kimi K2 model gets a new version update: Agentic coding: more accurate, better generalization across scaffolds. Frontend coding: improved aesthetics and functionalities on web, 3d, and other tasks. Context length: extended from 128k to 256k, providing better long-horizon support."
1899
+ },
1845
1900
  "accounts/fireworks/models/kimi-k2-instruct": {
1846
1901
  maxTokens: 16384,
1847
1902
  contextWindow: 128e3,
@@ -2250,7 +2305,7 @@ var glamaDefaultModelInfo = {
2250
2305
  var GLAMA_DEFAULT_TEMPERATURE = 0;
2251
2306
 
2252
2307
  // src/providers/groq.ts
2253
- var groqDefaultModelId = "llama-3.3-70b-versatile";
2308
+ var groqDefaultModelId = "moonshotai/kimi-k2-instruct-0905";
2254
2309
  var groqModels = {
2255
2310
  // Models based on API response: https://api.groq.com/openai/v1/models
2256
2311
  "llama-3.1-8b-instant": {
@@ -2336,6 +2391,16 @@ var groqModels = {
2336
2391
  // 50% discount for cached input tokens
2337
2392
  description: "Moonshot AI Kimi K2 Instruct 1T model, 128K context."
2338
2393
  },
2394
+ "moonshotai/kimi-k2-instruct-0905": {
2395
+ maxTokens: 16384,
2396
+ contextWindow: 262144,
2397
+ supportsImages: false,
2398
+ supportsPromptCache: true,
2399
+ inputPrice: 0.6,
2400
+ outputPrice: 2.5,
2401
+ cacheReadsPrice: 0.15,
2402
+ description: "Kimi K2 model gets a new version update: Agentic coding: more accurate, better generalization across scaffolds. Frontend coding: improved aesthetics and functionalities on web, 3d, and other tasks. Context length: extended from 128k to 256k, providing better long-horizon support."
2403
+ },
2339
2404
  "openai/gpt-oss-120b": {
2340
2405
  maxTokens: 32766,
2341
2406
  contextWindow: 131072,
@@ -2546,7 +2611,7 @@ var mistralModels = {
2546
2611
  var MISTRAL_DEFAULT_TEMPERATURE = 0;
2547
2612
 
2548
2613
  // src/providers/moonshot.ts
2549
- var moonshotDefaultModelId = "kimi-k2-0711-preview";
2614
+ var moonshotDefaultModelId = "kimi-k2-0905-preview";
2550
2615
  var moonshotModels = {
2551
2616
  "kimi-k2-0711-preview": {
2552
2617
  maxTokens: 32e3,
@@ -2562,6 +2627,31 @@ var moonshotModels = {
2562
2627
  cacheReadsPrice: 0.15,
2563
2628
  // $0.15 per million tokens (cache hit)
2564
2629
  description: `Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters.`
2630
+ },
2631
+ "kimi-k2-0905-preview": {
2632
+ maxTokens: 16384,
2633
+ contextWindow: 262144,
2634
+ supportsImages: false,
2635
+ supportsPromptCache: true,
2636
+ inputPrice: 0.6,
2637
+ outputPrice: 2.5,
2638
+ cacheReadsPrice: 0.15,
2639
+ description: "Kimi K2 model gets a new version update: Agentic coding: more accurate, better generalization across scaffolds. Frontend coding: improved aesthetics and functionalities on web, 3d, and other tasks. Context length: extended from 128k to 256k, providing better long-horizon support."
2640
+ },
2641
+ "kimi-k2-turbo-preview": {
2642
+ maxTokens: 32e3,
2643
+ contextWindow: 262144,
2644
+ supportsImages: false,
2645
+ supportsPromptCache: true,
2646
+ inputPrice: 2.4,
2647
+ // $2.40 per million tokens (cache miss)
2648
+ outputPrice: 10,
2649
+ // $10.00 per million tokens
2650
+ cacheWritesPrice: 0,
2651
+ // $0 per million tokens (cache miss)
2652
+ cacheReadsPrice: 0.6,
2653
+ // $0.60 per million tokens (cache hit)
2654
+ description: `Kimi K2 Turbo is a high-speed version of the state-of-the-art Kimi K2 mixture-of-experts (MoE) language model, with the same 32 billion activated parameters and 1 trillion total parameters, optimized for output speeds of up to 60 tokens per second, peaking at 100 tokens per second.`
2565
2655
  }
2566
2656
  };
2567
2657
  var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
@@ -2609,7 +2699,11 @@ var openAiNativeModels = {
2609
2699
  description: "GPT-5: The best model for coding and agentic tasks across domains",
2610
2700
  // supportsVerbosity is a new capability; ensure ModelInfo includes it
2611
2701
  supportsVerbosity: true,
2612
- supportsTemperature: false
2702
+ supportsTemperature: false,
2703
+ tiers: [
2704
+ { name: "flex", contextWindow: 4e5, inputPrice: 0.625, outputPrice: 5, cacheReadsPrice: 0.0625 },
2705
+ { name: "priority", contextWindow: 4e5, inputPrice: 2.5, outputPrice: 20, cacheReadsPrice: 0.25 }
2706
+ ]
2613
2707
  },
2614
2708
  "gpt-5-mini-2025-08-07": {
2615
2709
  maxTokens: 128e3,
@@ -2623,7 +2717,11 @@ var openAiNativeModels = {
2623
2717
  cacheReadsPrice: 0.03,
2624
2718
  description: "GPT-5 Mini: A faster, more cost-efficient version of GPT-5 for well-defined tasks",
2625
2719
  supportsVerbosity: true,
2626
- supportsTemperature: false
2720
+ supportsTemperature: false,
2721
+ tiers: [
2722
+ { name: "flex", contextWindow: 4e5, inputPrice: 0.125, outputPrice: 1, cacheReadsPrice: 0.0125 },
2723
+ { name: "priority", contextWindow: 4e5, inputPrice: 0.45, outputPrice: 3.6, cacheReadsPrice: 0.045 }
2724
+ ]
2627
2725
  },
2628
2726
  "gpt-5-nano-2025-08-07": {
2629
2727
  maxTokens: 128e3,
@@ -2637,7 +2735,8 @@ var openAiNativeModels = {
2637
2735
  cacheReadsPrice: 0.01,
2638
2736
  description: "GPT-5 Nano: Fastest, most cost-efficient version of GPT-5",
2639
2737
  supportsVerbosity: true,
2640
- supportsTemperature: false
2738
+ supportsTemperature: false,
2739
+ tiers: [{ name: "flex", contextWindow: 4e5, inputPrice: 0.025, outputPrice: 0.2, cacheReadsPrice: 25e-4 }]
2641
2740
  },
2642
2741
  "gpt-4.1": {
2643
2742
  maxTokens: 32768,
@@ -2647,7 +2746,10 @@ var openAiNativeModels = {
2647
2746
  inputPrice: 2,
2648
2747
  outputPrice: 8,
2649
2748
  cacheReadsPrice: 0.5,
2650
- supportsTemperature: true
2749
+ supportsTemperature: true,
2750
+ tiers: [
2751
+ { name: "priority", contextWindow: 1047576, inputPrice: 3.5, outputPrice: 14, cacheReadsPrice: 0.875 }
2752
+ ]
2651
2753
  },
2652
2754
  "gpt-4.1-mini": {
2653
2755
  maxTokens: 32768,
@@ -2657,7 +2759,10 @@ var openAiNativeModels = {
2657
2759
  inputPrice: 0.4,
2658
2760
  outputPrice: 1.6,
2659
2761
  cacheReadsPrice: 0.1,
2660
- supportsTemperature: true
2762
+ supportsTemperature: true,
2763
+ tiers: [
2764
+ { name: "priority", contextWindow: 1047576, inputPrice: 0.7, outputPrice: 2.8, cacheReadsPrice: 0.175 }
2765
+ ]
2661
2766
  },
2662
2767
  "gpt-4.1-nano": {
2663
2768
  maxTokens: 32768,
@@ -2667,7 +2772,10 @@ var openAiNativeModels = {
2667
2772
  inputPrice: 0.1,
2668
2773
  outputPrice: 0.4,
2669
2774
  cacheReadsPrice: 0.025,
2670
- supportsTemperature: true
2775
+ supportsTemperature: true,
2776
+ tiers: [
2777
+ { name: "priority", contextWindow: 1047576, inputPrice: 0.2, outputPrice: 0.8, cacheReadsPrice: 0.05 }
2778
+ ]
2671
2779
  },
2672
2780
  o3: {
2673
2781
  maxTokens: 1e5,
@@ -2679,7 +2787,11 @@ var openAiNativeModels = {
2679
2787
  cacheReadsPrice: 0.5,
2680
2788
  supportsReasoningEffort: true,
2681
2789
  reasoningEffort: "medium",
2682
- supportsTemperature: false
2790
+ supportsTemperature: false,
2791
+ tiers: [
2792
+ { name: "flex", contextWindow: 2e5, inputPrice: 1, outputPrice: 4, cacheReadsPrice: 0.25 },
2793
+ { name: "priority", contextWindow: 2e5, inputPrice: 3.5, outputPrice: 14, cacheReadsPrice: 0.875 }
2794
+ ]
2683
2795
  },
2684
2796
  "o3-high": {
2685
2797
  maxTokens: 1e5,
@@ -2713,7 +2825,11 @@ var openAiNativeModels = {
2713
2825
  cacheReadsPrice: 0.275,
2714
2826
  supportsReasoningEffort: true,
2715
2827
  reasoningEffort: "medium",
2716
- supportsTemperature: false
2828
+ supportsTemperature: false,
2829
+ tiers: [
2830
+ { name: "flex", contextWindow: 2e5, inputPrice: 0.55, outputPrice: 2.2, cacheReadsPrice: 0.138 },
2831
+ { name: "priority", contextWindow: 2e5, inputPrice: 2, outputPrice: 8, cacheReadsPrice: 0.5 }
2832
+ ]
2717
2833
  },
2718
2834
  "o4-mini-high": {
2719
2835
  maxTokens: 1e5,
@@ -2809,7 +2925,10 @@ var openAiNativeModels = {
2809
2925
  inputPrice: 2.5,
2810
2926
  outputPrice: 10,
2811
2927
  cacheReadsPrice: 1.25,
2812
- supportsTemperature: true
2928
+ supportsTemperature: true,
2929
+ tiers: [
2930
+ { name: "priority", contextWindow: 128e3, inputPrice: 4.25, outputPrice: 17, cacheReadsPrice: 2.125 }
2931
+ ]
2813
2932
  },
2814
2933
  "gpt-4o-mini": {
2815
2934
  maxTokens: 16384,
@@ -2819,7 +2938,10 @@ var openAiNativeModels = {
2819
2938
  inputPrice: 0.15,
2820
2939
  outputPrice: 0.6,
2821
2940
  cacheReadsPrice: 0.075,
2822
- supportsTemperature: true
2941
+ supportsTemperature: true,
2942
+ tiers: [
2943
+ { name: "priority", contextWindow: 128e3, inputPrice: 0.25, outputPrice: 1, cacheReadsPrice: 0.125 }
2944
+ ]
2823
2945
  },
2824
2946
  "codex-mini-latest": {
2825
2947
  maxTokens: 16384,
@@ -3360,6 +3482,60 @@ var vertexModels = {
3360
3482
  inputPrice: 0.35,
3361
3483
  outputPrice: 1.15,
3362
3484
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
3485
+ },
3486
+ "deepseek-r1-0528-maas": {
3487
+ maxTokens: 32768,
3488
+ contextWindow: 163840,
3489
+ supportsImages: false,
3490
+ supportsPromptCache: false,
3491
+ inputPrice: 1.35,
3492
+ outputPrice: 5.4,
3493
+ description: "DeepSeek R1 (0528). Available in us-central1"
3494
+ },
3495
+ "deepseek-v3.1-maas": {
3496
+ maxTokens: 32768,
3497
+ contextWindow: 163840,
3498
+ supportsImages: false,
3499
+ supportsPromptCache: false,
3500
+ inputPrice: 0.6,
3501
+ outputPrice: 1.7,
3502
+ description: "DeepSeek V3.1. Available in us-west2"
3503
+ },
3504
+ "gpt-oss-120b-maas": {
3505
+ maxTokens: 32768,
3506
+ contextWindow: 131072,
3507
+ supportsImages: false,
3508
+ supportsPromptCache: false,
3509
+ inputPrice: 0.15,
3510
+ outputPrice: 0.6,
3511
+ description: "OpenAI gpt-oss 120B. Available in us-central1"
3512
+ },
3513
+ "gpt-oss-20b-maas": {
3514
+ maxTokens: 32768,
3515
+ contextWindow: 131072,
3516
+ supportsImages: false,
3517
+ supportsPromptCache: false,
3518
+ inputPrice: 0.075,
3519
+ outputPrice: 0.3,
3520
+ description: "OpenAI gpt-oss 20B. Available in us-central1"
3521
+ },
3522
+ "qwen3-coder-480b-a35b-instruct-maas": {
3523
+ maxTokens: 32768,
3524
+ contextWindow: 262144,
3525
+ supportsImages: false,
3526
+ supportsPromptCache: false,
3527
+ inputPrice: 1,
3528
+ outputPrice: 4,
3529
+ description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
3530
+ },
3531
+ "qwen3-235b-a22b-instruct-2507-maas": {
3532
+ maxTokens: 16384,
3533
+ contextWindow: 262144,
3534
+ supportsImages: false,
3535
+ supportsPromptCache: false,
3536
+ inputPrice: 0.25,
3537
+ outputPrice: 1,
3538
+ description: "Qwen3 235B A22B Instruct. Available in us-south1"
3363
3539
  }
3364
3540
  };
3365
3541
  var VERTEX_REGIONS = [
@@ -3368,6 +3544,7 @@ var VERTEX_REGIONS = [
3368
3544
  { value: "us-east1", label: "us-east1" },
3369
3545
  { value: "us-east4", label: "us-east4" },
3370
3546
  { value: "us-east5", label: "us-east5" },
3547
+ { value: "us-south1", label: "us-south1" },
3371
3548
  { value: "us-west1", label: "us-west1" },
3372
3549
  { value: "us-west2", label: "us-west2" },
3373
3550
  { value: "us-west3", label: "us-west3" },
@@ -3860,6 +4037,28 @@ var mainlandZAiModels = {
3860
4037
  }
3861
4038
  };
3862
4039
  var ZAI_DEFAULT_TEMPERATURE = 0;
4040
+ var zaiApiLineConfigs = {
4041
+ international_coding: {
4042
+ name: "International Coding Plan",
4043
+ baseUrl: "https://api.z.ai/api/coding/paas/v4",
4044
+ isChina: false
4045
+ },
4046
+ international: { name: "International Standard", baseUrl: "https://api.z.ai/api/paas/v4", isChina: false },
4047
+ china_coding: { name: "China Coding Plan", baseUrl: "https://open.bigmodel.cn/api/coding/paas/v4", isChina: true },
4048
+ china: { name: "China Standard", baseUrl: "https://open.bigmodel.cn/api/paas/v4", isChina: true }
4049
+ };
4050
+
4051
+ // src/providers/deepinfra.ts
4052
+ var deepInfraDefaultModelId = "Qwen/Qwen3-Coder-480B-A35B-Instruct-Turbo";
4053
+ var deepInfraDefaultModelInfo = {
4054
+ maxTokens: 16384,
4055
+ contextWindow: 262144,
4056
+ supportsImages: false,
4057
+ supportsPromptCache: false,
4058
+ inputPrice: 0.3,
4059
+ outputPrice: 1.2,
4060
+ description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
4061
+ };
3863
4062
 
3864
4063
  // src/provider-settings.ts
3865
4064
  var providerNames = [
@@ -3879,6 +4078,7 @@ var providerNames = [
3879
4078
  "mistral",
3880
4079
  "moonshot",
3881
4080
  "deepseek",
4081
+ "deepinfra",
3882
4082
  "doubao",
3883
4083
  "qwen-code",
3884
4084
  "unbound",
@@ -4019,7 +4219,10 @@ var geminiCliSchema = apiModelIdProviderModelSchema.extend({
4019
4219
  });
4020
4220
  var openAiNativeSchema = apiModelIdProviderModelSchema.extend({
4021
4221
  openAiNativeApiKey: import_zod7.z.string().optional(),
4022
- openAiNativeBaseUrl: import_zod7.z.string().optional()
4222
+ openAiNativeBaseUrl: import_zod7.z.string().optional(),
4223
+ // OpenAI Responses API service tier for openai-native provider only.
4224
+ // UI should only expose this when the selected model supports flex/priority.
4225
+ openAiNativeServiceTier: serviceTierSchema.optional()
4023
4226
  });
4024
4227
  var mistralSchema = apiModelIdProviderModelSchema.extend({
4025
4228
  mistralApiKey: import_zod7.z.string().optional(),
@@ -4029,6 +4232,11 @@ var deepSeekSchema = apiModelIdProviderModelSchema.extend({
4029
4232
  deepSeekBaseUrl: import_zod7.z.string().optional(),
4030
4233
  deepSeekApiKey: import_zod7.z.string().optional()
4031
4234
  });
4235
+ var deepInfraSchema = apiModelIdProviderModelSchema.extend({
4236
+ deepInfraBaseUrl: import_zod7.z.string().optional(),
4237
+ deepInfraApiKey: import_zod7.z.string().optional(),
4238
+ deepInfraModelId: import_zod7.z.string().optional()
4239
+ });
4032
4240
  var doubaoSchema = apiModelIdProviderModelSchema.extend({
4033
4241
  doubaoBaseUrl: import_zod7.z.string().optional(),
4034
4242
  doubaoApiKey: import_zod7.z.string().optional()
@@ -4076,9 +4284,10 @@ var cerebrasSchema = apiModelIdProviderModelSchema.extend({
4076
4284
  var sambaNovaSchema = apiModelIdProviderModelSchema.extend({
4077
4285
  sambaNovaApiKey: import_zod7.z.string().optional()
4078
4286
  });
4287
+ var zaiApiLineSchema = import_zod7.z.enum(["international_coding", "international", "china_coding", "china"]);
4079
4288
  var zaiSchema = apiModelIdProviderModelSchema.extend({
4080
4289
  zaiApiKey: import_zod7.z.string().optional(),
4081
- zaiApiLine: import_zod7.z.union([import_zod7.z.literal("china"), import_zod7.z.literal("international")]).optional()
4290
+ zaiApiLine: zaiApiLineSchema.optional()
4082
4291
  });
4083
4292
  var fireworksSchema = apiModelIdProviderModelSchema.extend({
4084
4293
  fireworksApiKey: import_zod7.z.string().optional()
@@ -4119,6 +4328,7 @@ var providerSettingsSchemaDiscriminated = import_zod7.z.discriminatedUnion("apiP
4119
4328
  openAiNativeSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("openai-native") })),
4120
4329
  mistralSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("mistral") })),
4121
4330
  deepSeekSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("deepseek") })),
4331
+ deepInfraSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("deepinfra") })),
4122
4332
  doubaoSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("doubao") })),
4123
4333
  moonshotSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("moonshot") })),
4124
4334
  unboundSchema.merge(import_zod7.z.object({ apiProvider: import_zod7.z.literal("unbound") })),
@@ -4158,6 +4368,7 @@ var providerSettingsSchema = import_zod7.z.object({
4158
4368
  ...openAiNativeSchema.shape,
4159
4369
  ...mistralSchema.shape,
4160
4370
  ...deepSeekSchema.shape,
4371
+ ...deepInfraSchema.shape,
4161
4372
  ...doubaoSchema.shape,
4162
4373
  ...moonshotSchema.shape,
4163
4374
  ...unboundSchema.shape,
@@ -4198,7 +4409,8 @@ var MODEL_ID_KEYS = [
4198
4409
  "litellmModelId",
4199
4410
  "huggingFaceModelId",
4200
4411
  "ioIntelligenceModelId",
4201
- "vercelAiGatewayModelId"
4412
+ "vercelAiGatewayModelId",
4413
+ "deepInfraModelId"
4202
4414
  ];
4203
4415
  var getModelId = (settings) => {
4204
4416
  const modelIdKey = MODEL_ID_KEYS.find((key) => settings[key]);
@@ -4307,6 +4519,7 @@ var MODELS_BY_PROVIDER = {
4307
4519
  openrouter: { id: "openrouter", label: "OpenRouter", models: [] },
4308
4520
  requesty: { id: "requesty", label: "Requesty", models: [] },
4309
4521
  unbound: { id: "unbound", label: "Unbound", models: [] },
4522
+ deepinfra: { id: "deepinfra", label: "DeepInfra", models: [] },
4310
4523
  "vercel-ai-gateway": { id: "vercel-ai-gateway", label: "Vercel AI Gateway", models: [] }
4311
4524
  };
4312
4525
  var dynamicProviders = [
@@ -4316,6 +4529,7 @@ var dynamicProviders = [
4316
4529
  "openrouter",
4317
4530
  "requesty",
4318
4531
  "unbound",
4532
+ "deepinfra",
4319
4533
  "vercel-ai-gateway"
4320
4534
  ];
4321
4535
  var isDynamicProvider = (key) => dynamicProviders.includes(key);
@@ -4683,6 +4897,7 @@ var globalSettingsSchema = import_zod13.z.object({
4683
4897
  lastShownAnnouncementId: import_zod13.z.string().optional(),
4684
4898
  customInstructions: import_zod13.z.string().optional(),
4685
4899
  taskHistory: import_zod13.z.array(historyItemSchema).optional(),
4900
+ dismissedUpsells: import_zod13.z.array(import_zod13.z.string()).optional(),
4686
4901
  // Image generation settings (experimental) - flattened for simplicity
4687
4902
  openRouterImageApiKey: import_zod13.z.string().optional(),
4688
4903
  openRouterImageGenerationSelectedModel: import_zod13.z.string().optional(),
@@ -4764,7 +4979,6 @@ var globalSettingsSchema = import_zod13.z.object({
4764
4979
  telemetrySetting: telemetrySettingsSchema.optional(),
4765
4980
  mcpEnabled: import_zod13.z.boolean().optional(),
4766
4981
  enableMcpServerCreation: import_zod13.z.boolean().optional(),
4767
- remoteControlEnabled: import_zod13.z.boolean().optional(),
4768
4982
  mode: import_zod13.z.string().optional(),
4769
4983
  modeApiConfigs: import_zod13.z.record(import_zod13.z.string(), import_zod13.z.string()).optional(),
4770
4984
  customModes: import_zod13.z.array(modeConfigSchema).optional(),
@@ -4803,6 +5017,7 @@ var SECRET_STATE_KEYS = [
4803
5017
  "groqApiKey",
4804
5018
  "chutesApiKey",
4805
5019
  "litellmApiKey",
5020
+ "deepInfraApiKey",
4806
5021
  "codeIndexOpenAiKey",
4807
5022
  "codeIndexQdrantApiKey",
4808
5023
  "codebaseIndexOpenAiCompatibleApiKey",
@@ -4886,7 +5101,6 @@ var EVALS_SETTINGS = {
4886
5101
  language: "en",
4887
5102
  telemetrySetting: "enabled",
4888
5103
  mcpEnabled: false,
4889
- remoteControlEnabled: false,
4890
5104
  mode: "code",
4891
5105
  // "architect",
4892
5106
  customModes: []
@@ -4996,7 +5210,8 @@ var userFeaturesSchema = import_zod15.z.object({
4996
5210
  roomoteControlEnabled: import_zod15.z.boolean().optional()
4997
5211
  });
4998
5212
  var userSettingsConfigSchema = import_zod15.z.object({
4999
- extensionBridgeEnabled: import_zod15.z.boolean().optional()
5213
+ extensionBridgeEnabled: import_zod15.z.boolean().optional(),
5214
+ taskSyncEnabled: import_zod15.z.boolean().optional()
5000
5215
  });
5001
5216
  var userSettingsDataSchema = import_zod15.z.object({
5002
5217
  features: userFeaturesSchema,
@@ -5299,6 +5514,12 @@ var TaskSocketEvents = /* @__PURE__ */ ((TaskSocketEvents2) => {
5299
5514
  return TaskSocketEvents2;
5300
5515
  })(TaskSocketEvents || {});
5301
5516
 
5517
+ // src/cookie-consent.ts
5518
+ var CONSENT_COOKIE_NAME = "roo-code-cookie-consent";
5519
+ var COOKIE_CONSENT_EVENTS = {
5520
+ CHANGED: "cookieConsentChanged"
5521
+ };
5522
+
5302
5523
  // src/followup.ts
5303
5524
  var import_zod16 = require("zod");
5304
5525
  var suggestionItemSchema = import_zod16.z.object({
@@ -5460,6 +5681,8 @@ var commandExecutionStatusSchema = import_zod20.z.discriminatedUnion("status", [
5460
5681
  BEDROCK_REGIONS,
5461
5682
  CLAUDE_CODE_DEFAULT_MAX_OUTPUT_TOKENS,
5462
5683
  CODEBASE_INDEX_DEFAULTS,
5684
+ CONSENT_COOKIE_NAME,
5685
+ COOKIE_CONSENT_EVENTS,
5463
5686
  ConnectionState,
5464
5687
  DEEP_SEEK_DEFAULT_TEMPERATURE,
5465
5688
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
@@ -5552,6 +5775,8 @@ var commandExecutionStatusSchema = import_zod20.z.discriminatedUnion("status", [
5552
5775
  customModePromptsSchema,
5553
5776
  customModesSettingsSchema,
5554
5777
  customSupportPromptsSchema,
5778
+ deepInfraDefaultModelId,
5779
+ deepInfraDefaultModelInfo,
5555
5780
  deepSeekDefaultModelId,
5556
5781
  deepSeekModels,
5557
5782
  discriminatedProviderSettingsWithIdSchema,
@@ -5659,6 +5884,8 @@ var commandExecutionStatusSchema = import_zod20.z.discriminatedUnion("status", [
5659
5884
  rooModels,
5660
5885
  sambaNovaDefaultModelId,
5661
5886
  sambaNovaModels,
5887
+ serviceTierSchema,
5888
+ serviceTiers,
5662
5889
  shareResponseSchema,
5663
5890
  shouldUseSingleFileRead,
5664
5891
  staticAppPropertiesSchema,
@@ -5696,6 +5923,8 @@ var commandExecutionStatusSchema = import_zod20.z.discriminatedUnion("status", [
5696
5923
  vscodeLlmDefaultModelId,
5697
5924
  vscodeLlmModels,
5698
5925
  xaiDefaultModelId,
5699
- xaiModels
5926
+ xaiModels,
5927
+ zaiApiLineConfigs,
5928
+ zaiApiLineSchema
5700
5929
  });
5701
5930
  //# sourceMappingURL=index.cjs.map