@roo-code/types 1.86.0 → 1.88.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -66,6 +66,7 @@ var clineSays = [
66
66
  "shell_integration_warning",
67
67
  "browser_action",
68
68
  "browser_action_result",
69
+ "browser_session_status",
69
70
  "mcp_server_request_started",
70
71
  "mcp_server_response",
71
72
  "subtask_result",
@@ -130,6 +131,8 @@ var toolNames = [
130
131
  "write_to_file",
131
132
  "apply_diff",
132
133
  "insert_content",
134
+ "search_and_replace",
135
+ "apply_patch",
133
136
  "search_files",
134
137
  "list_files",
135
138
  "list_code_definition_names",
@@ -180,6 +183,9 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
180
183
  RooCodeEventName2["TaskPaused"] = "taskPaused";
181
184
  RooCodeEventName2["TaskUnpaused"] = "taskUnpaused";
182
185
  RooCodeEventName2["TaskSpawned"] = "taskSpawned";
186
+ RooCodeEventName2["TaskDelegated"] = "taskDelegated";
187
+ RooCodeEventName2["TaskDelegationCompleted"] = "taskDelegationCompleted";
188
+ RooCodeEventName2["TaskDelegationResumed"] = "taskDelegationResumed";
183
189
  RooCodeEventName2["Message"] = "message";
184
190
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
185
191
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
@@ -213,6 +219,26 @@ var rooCodeEventsSchema = z3.object({
213
219
  ["taskPaused" /* TaskPaused */]: z3.tuple([z3.string()]),
214
220
  ["taskUnpaused" /* TaskUnpaused */]: z3.tuple([z3.string()]),
215
221
  ["taskSpawned" /* TaskSpawned */]: z3.tuple([z3.string(), z3.string()]),
222
+ ["taskDelegated" /* TaskDelegated */]: z3.tuple([
223
+ z3.string(),
224
+ // parentTaskId
225
+ z3.string()
226
+ // childTaskId
227
+ ]),
228
+ ["taskDelegationCompleted" /* TaskDelegationCompleted */]: z3.tuple([
229
+ z3.string(),
230
+ // parentTaskId
231
+ z3.string(),
232
+ // childTaskId
233
+ z3.string()
234
+ // completionResultSummary
235
+ ]),
236
+ ["taskDelegationResumed" /* TaskDelegationResumed */]: z3.tuple([
237
+ z3.string(),
238
+ // parentTaskId
239
+ z3.string()
240
+ // childTaskId
241
+ ]),
216
242
  ["message" /* Message */]: z3.tuple([
217
243
  z3.object({
218
244
  taskId: z3.string(),
@@ -297,6 +323,21 @@ var taskEventSchema = z3.discriminatedUnion("eventName", [
297
323
  payload: rooCodeEventsSchema.shape["taskSpawned" /* TaskSpawned */],
298
324
  taskId: z3.number().optional()
299
325
  }),
326
+ z3.object({
327
+ eventName: z3.literal("taskDelegated" /* TaskDelegated */),
328
+ payload: rooCodeEventsSchema.shape["taskDelegated" /* TaskDelegated */],
329
+ taskId: z3.number().optional()
330
+ }),
331
+ z3.object({
332
+ eventName: z3.literal("taskDelegationCompleted" /* TaskDelegationCompleted */),
333
+ payload: rooCodeEventsSchema.shape["taskDelegationCompleted" /* TaskDelegationCompleted */],
334
+ taskId: z3.number().optional()
335
+ }),
336
+ z3.object({
337
+ eventName: z3.literal("taskDelegationResumed" /* TaskDelegationResumed */),
338
+ payload: rooCodeEventsSchema.shape["taskDelegationResumed" /* TaskDelegationResumed */],
339
+ taskId: z3.number().optional()
340
+ }),
300
341
  // Task Execution
301
342
  z3.object({
302
343
  eventName: z3.literal("message" /* Message */),
@@ -415,6 +456,13 @@ var modelInfoSchema = z5.object({
415
456
  supportsNativeTools: z5.boolean().optional(),
416
457
  // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
417
458
  defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
459
+ // Exclude specific native tools from being available (only applies to native protocol)
460
+ // These tools will be removed from the set of tools available to the model
461
+ excludedTools: z5.array(z5.string()).optional(),
462
+ // Include specific native tools (only applies to native protocol)
463
+ // These tools will be added if they belong to an allowed group in the current mode
464
+ // Cannot force-add tools from groups the mode doesn't allow
465
+ includedTools: z5.array(z5.string()).optional(),
418
466
  /**
419
467
  * Service tiers with pricing information.
420
468
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -448,7 +496,16 @@ var CODEBASE_INDEX_DEFAULTS = {
448
496
  var codebaseIndexConfigSchema = z6.object({
449
497
  codebaseIndexEnabled: z6.boolean().optional(),
450
498
  codebaseIndexQdrantUrl: z6.string().optional(),
451
- codebaseIndexEmbedderProvider: z6.enum(["openai", "ollama", "openai-compatible", "gemini", "mistral", "vercel-ai-gateway", "openrouter"]).optional(),
499
+ codebaseIndexEmbedderProvider: z6.enum([
500
+ "openai",
501
+ "ollama",
502
+ "openai-compatible",
503
+ "gemini",
504
+ "mistral",
505
+ "vercel-ai-gateway",
506
+ "bedrock",
507
+ "openrouter"
508
+ ]).optional(),
452
509
  codebaseIndexEmbedderBaseUrl: z6.string().optional(),
453
510
  codebaseIndexEmbedderModelId: z6.string().optional(),
454
511
  codebaseIndexEmbedderModelDimension: z6.number().optional(),
@@ -456,7 +513,10 @@ var codebaseIndexConfigSchema = z6.object({
456
513
  codebaseIndexSearchMaxResults: z6.number().min(CODEBASE_INDEX_DEFAULTS.MIN_SEARCH_RESULTS).max(CODEBASE_INDEX_DEFAULTS.MAX_SEARCH_RESULTS).optional(),
457
514
  // OpenAI Compatible specific fields
458
515
  codebaseIndexOpenAiCompatibleBaseUrl: z6.string().optional(),
459
- codebaseIndexOpenAiCompatibleModelDimension: z6.number().optional()
516
+ codebaseIndexOpenAiCompatibleModelDimension: z6.number().optional(),
517
+ // Bedrock specific fields
518
+ codebaseIndexBedrockRegion: z6.string().optional(),
519
+ codebaseIndexBedrockProfile: z6.string().optional()
460
520
  });
461
521
  var codebaseIndexModelsSchema = z6.object({
462
522
  openai: z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional(),
@@ -465,7 +525,8 @@ var codebaseIndexModelsSchema = z6.object({
465
525
  gemini: z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional(),
466
526
  mistral: z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional(),
467
527
  "vercel-ai-gateway": z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional(),
468
- openrouter: z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional()
528
+ openrouter: z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional(),
529
+ bedrock: z6.record(z6.string(), z6.object({ dimension: z6.number() })).optional()
469
530
  });
470
531
  var codebaseIndexProviderSchema = z6.object({
471
532
  codeIndexOpenAiKey: z6.string().optional(),
@@ -489,6 +550,7 @@ var anthropicModels = {
489
550
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
490
551
  supportsImages: true,
491
552
  supportsPromptCache: true,
553
+ supportsNativeTools: true,
492
554
  inputPrice: 3,
493
555
  // $3 per million input tokens (≤200K context)
494
556
  outputPrice: 15,
@@ -521,6 +583,7 @@ var anthropicModels = {
521
583
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
522
584
  supportsImages: true,
523
585
  supportsPromptCache: true,
586
+ supportsNativeTools: true,
524
587
  inputPrice: 3,
525
588
  // $3 per million input tokens (≤200K context)
526
589
  outputPrice: 15,
@@ -546,12 +609,30 @@ var anthropicModels = {
546
609
  }
547
610
  ]
548
611
  },
612
+ "claude-opus-4-5-20251101": {
613
+ maxTokens: 32e3,
614
+ // Overridden to 8k if `enableReasoningEffort` is false.
615
+ contextWindow: 2e5,
616
+ supportsImages: true,
617
+ supportsPromptCache: true,
618
+ supportsNativeTools: true,
619
+ inputPrice: 5,
620
+ // $5 per million input tokens
621
+ outputPrice: 25,
622
+ // $25 per million output tokens
623
+ cacheWritesPrice: 6.25,
624
+ // $6.25 per million tokens
625
+ cacheReadsPrice: 0.5,
626
+ // $0.50 per million tokens
627
+ supportsReasoningBudget: true
628
+ },
549
629
  "claude-opus-4-1-20250805": {
550
630
  maxTokens: 32e3,
551
631
  // Overridden to 8k if `enableReasoningEffort` is false.
552
632
  contextWindow: 2e5,
553
633
  supportsImages: true,
554
634
  supportsPromptCache: true,
635
+ supportsNativeTools: true,
555
636
  inputPrice: 15,
556
637
  // $15 per million input tokens
557
638
  outputPrice: 75,
@@ -568,6 +649,7 @@ var anthropicModels = {
568
649
  contextWindow: 2e5,
569
650
  supportsImages: true,
570
651
  supportsPromptCache: true,
652
+ supportsNativeTools: true,
571
653
  inputPrice: 15,
572
654
  // $15 per million input tokens
573
655
  outputPrice: 75,
@@ -584,6 +666,7 @@ var anthropicModels = {
584
666
  contextWindow: 2e5,
585
667
  supportsImages: true,
586
668
  supportsPromptCache: true,
669
+ supportsNativeTools: true,
587
670
  inputPrice: 3,
588
671
  // $3 per million input tokens
589
672
  outputPrice: 15,
@@ -601,6 +684,7 @@ var anthropicModels = {
601
684
  contextWindow: 2e5,
602
685
  supportsImages: true,
603
686
  supportsPromptCache: true,
687
+ supportsNativeTools: true,
604
688
  inputPrice: 3,
605
689
  // $3 per million input tokens
606
690
  outputPrice: 15,
@@ -615,6 +699,7 @@ var anthropicModels = {
615
699
  contextWindow: 2e5,
616
700
  supportsImages: true,
617
701
  supportsPromptCache: true,
702
+ supportsNativeTools: true,
618
703
  inputPrice: 3,
619
704
  // $3 per million input tokens
620
705
  outputPrice: 15,
@@ -629,6 +714,7 @@ var anthropicModels = {
629
714
  contextWindow: 2e5,
630
715
  supportsImages: false,
631
716
  supportsPromptCache: true,
717
+ supportsNativeTools: true,
632
718
  inputPrice: 1,
633
719
  outputPrice: 5,
634
720
  cacheWritesPrice: 1.25,
@@ -639,6 +725,7 @@ var anthropicModels = {
639
725
  contextWindow: 2e5,
640
726
  supportsImages: true,
641
727
  supportsPromptCache: true,
728
+ supportsNativeTools: true,
642
729
  inputPrice: 15,
643
730
  outputPrice: 75,
644
731
  cacheWritesPrice: 18.75,
@@ -649,6 +736,7 @@ var anthropicModels = {
649
736
  contextWindow: 2e5,
650
737
  supportsImages: true,
651
738
  supportsPromptCache: true,
739
+ supportsNativeTools: true,
652
740
  inputPrice: 0.25,
653
741
  outputPrice: 1.25,
654
742
  cacheWritesPrice: 0.3,
@@ -659,6 +747,7 @@ var anthropicModels = {
659
747
  contextWindow: 2e5,
660
748
  supportsImages: true,
661
749
  supportsPromptCache: true,
750
+ supportsNativeTools: true,
662
751
  inputPrice: 1,
663
752
  outputPrice: 5,
664
753
  cacheWritesPrice: 1.25,
@@ -669,6 +758,125 @@ var anthropicModels = {
669
758
  };
670
759
  var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
671
760
 
761
+ // src/providers/baseten.ts
762
+ var basetenModels = {
763
+ "moonshotai/Kimi-K2-Thinking": {
764
+ maxTokens: 163800,
765
+ contextWindow: 262e3,
766
+ supportsImages: false,
767
+ supportsPromptCache: false,
768
+ supportsNativeTools: true,
769
+ inputPrice: 0.6,
770
+ outputPrice: 2.5,
771
+ cacheWritesPrice: 0,
772
+ cacheReadsPrice: 0,
773
+ description: "Kimi K2 Thinking - A model with enhanced reasoning capabilities from Kimi K2"
774
+ },
775
+ "zai-org/GLM-4.6": {
776
+ maxTokens: 2e5,
777
+ contextWindow: 2e5,
778
+ supportsImages: false,
779
+ supportsPromptCache: false,
780
+ supportsNativeTools: true,
781
+ inputPrice: 0.6,
782
+ outputPrice: 2.2,
783
+ cacheWritesPrice: 0,
784
+ cacheReadsPrice: 0,
785
+ description: "Frontier open model with advanced agentic, reasoning and coding capabilities"
786
+ },
787
+ "deepseek-ai/DeepSeek-R1": {
788
+ maxTokens: 131072,
789
+ contextWindow: 163840,
790
+ supportsImages: false,
791
+ supportsPromptCache: false,
792
+ inputPrice: 2.55,
793
+ outputPrice: 5.95,
794
+ cacheWritesPrice: 0,
795
+ cacheReadsPrice: 0,
796
+ description: "DeepSeek's first-generation reasoning model"
797
+ },
798
+ "deepseek-ai/DeepSeek-R1-0528": {
799
+ maxTokens: 131072,
800
+ contextWindow: 163840,
801
+ supportsImages: false,
802
+ supportsPromptCache: false,
803
+ inputPrice: 2.55,
804
+ outputPrice: 5.95,
805
+ cacheWritesPrice: 0,
806
+ cacheReadsPrice: 0,
807
+ description: "The latest revision of DeepSeek's first-generation reasoning model"
808
+ },
809
+ "deepseek-ai/DeepSeek-V3-0324": {
810
+ maxTokens: 131072,
811
+ contextWindow: 163840,
812
+ supportsImages: false,
813
+ supportsPromptCache: false,
814
+ inputPrice: 0.77,
815
+ outputPrice: 0.77,
816
+ cacheWritesPrice: 0,
817
+ cacheReadsPrice: 0,
818
+ description: "Fast general-purpose LLM with enhanced reasoning capabilities"
819
+ },
820
+ "deepseek-ai/DeepSeek-V3.1": {
821
+ maxTokens: 131072,
822
+ contextWindow: 163840,
823
+ supportsImages: false,
824
+ supportsPromptCache: false,
825
+ inputPrice: 0.5,
826
+ outputPrice: 1.5,
827
+ cacheWritesPrice: 0,
828
+ cacheReadsPrice: 0,
829
+ description: "Extremely capable general-purpose LLM with hybrid reasoning capabilities and advanced tool calling"
830
+ },
831
+ "Qwen/Qwen3-235B-A22B-Instruct-2507": {
832
+ maxTokens: 262144,
833
+ contextWindow: 262144,
834
+ supportsImages: false,
835
+ supportsPromptCache: false,
836
+ inputPrice: 0.22,
837
+ outputPrice: 0.8,
838
+ cacheWritesPrice: 0,
839
+ cacheReadsPrice: 0,
840
+ description: "Mixture-of-experts LLM with math and reasoning capabilities"
841
+ },
842
+ "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
843
+ maxTokens: 262144,
844
+ contextWindow: 262144,
845
+ supportsImages: false,
846
+ supportsPromptCache: false,
847
+ inputPrice: 0.38,
848
+ outputPrice: 1.53,
849
+ cacheWritesPrice: 0,
850
+ cacheReadsPrice: 0,
851
+ description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
852
+ },
853
+ "openai/gpt-oss-120b": {
854
+ maxTokens: 128072,
855
+ contextWindow: 128072,
856
+ supportsImages: false,
857
+ supportsPromptCache: false,
858
+ supportsNativeTools: true,
859
+ inputPrice: 0.1,
860
+ outputPrice: 0.5,
861
+ cacheWritesPrice: 0,
862
+ cacheReadsPrice: 0,
863
+ description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
864
+ },
865
+ "moonshotai/Kimi-K2-Instruct-0905": {
866
+ maxTokens: 168e3,
867
+ contextWindow: 262e3,
868
+ supportsImages: false,
869
+ supportsPromptCache: false,
870
+ supportsNativeTools: true,
871
+ inputPrice: 0.6,
872
+ outputPrice: 2.5,
873
+ cacheWritesPrice: 0,
874
+ cacheReadsPrice: 0,
875
+ description: "State of the art language model for agentic and coding tasks. September Update."
876
+ }
877
+ };
878
+ var basetenDefaultModelId = "zai-org/GLM-4.6";
879
+
672
880
  // src/providers/bedrock.ts
673
881
  var bedrockDefaultModelId = "anthropic.claude-sonnet-4-5-20250929-v1:0";
674
882
  var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
@@ -773,6 +981,20 @@ var bedrockModels = {
773
981
  maxCachePoints: 4,
774
982
  cachableFields: ["system", "messages", "tools"]
775
983
  },
984
+ "anthropic.claude-opus-4-5-20251101-v1:0": {
985
+ maxTokens: 8192,
986
+ contextWindow: 2e5,
987
+ supportsImages: true,
988
+ supportsPromptCache: true,
989
+ supportsReasoningBudget: true,
990
+ inputPrice: 5,
991
+ outputPrice: 25,
992
+ cacheWritesPrice: 6.25,
993
+ cacheReadsPrice: 0.5,
994
+ minTokensPerCachePoint: 1024,
995
+ maxCachePoints: 4,
996
+ cachableFields: ["system", "messages", "tools"]
997
+ },
776
998
  "anthropic.claude-opus-4-20250514-v1:0": {
777
999
  maxTokens: 8192,
778
1000
  contextWindow: 2e5,
@@ -1115,7 +1337,8 @@ var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1115
1337
  var BEDROCK_GLOBAL_INFERENCE_MODEL_IDS = [
1116
1338
  "anthropic.claude-sonnet-4-20250514-v1:0",
1117
1339
  "anthropic.claude-sonnet-4-5-20250929-v1:0",
1118
- "anthropic.claude-haiku-4-5-20251001-v1:0"
1340
+ "anthropic.claude-haiku-4-5-20251001-v1:0",
1341
+ "anthropic.claude-opus-4-5-20251101-v1:0"
1119
1342
  ];
1120
1343
 
1121
1344
  // src/providers/cerebras.ts
@@ -1131,24 +1354,6 @@ var cerebrasModels = {
1131
1354
  outputPrice: 0,
1132
1355
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
1133
1356
  },
1134
- "qwen-3-coder-480b-free": {
1135
- maxTokens: 4e4,
1136
- contextWindow: 64e3,
1137
- supportsImages: false,
1138
- supportsPromptCache: false,
1139
- inputPrice: 0,
1140
- outputPrice: 0,
1141
- description: "[SOON TO BE DEPRECATED] SOTA coding model with ~2000 tokens/s ($0 free tier)\n\n\u2022 Use this if you don't have a Cerebras subscription\n\u2022 64K context window\n\u2022 Rate limits: 150K TPM, 1M TPH/TPD, 10 RPM, 100 RPH/RPD\n\nUpgrade for higher limits: [https://cloud.cerebras.ai/?utm=roocode](https://cloud.cerebras.ai/?utm=roocode)"
1142
- },
1143
- "qwen-3-coder-480b": {
1144
- maxTokens: 4e4,
1145
- contextWindow: 128e3,
1146
- supportsImages: false,
1147
- supportsPromptCache: false,
1148
- inputPrice: 0,
1149
- outputPrice: 0,
1150
- description: "[SOON TO BE DEPRECATED] SOTA coding model with ~2000 tokens/s ($50/$250 paid tiers)\n\n\u2022 Use this if you have a Cerebras subscription\n\u2022 131K context window with higher rate limits"
1151
- },
1152
1357
  "qwen-3-235b-a22b-instruct-2507": {
1153
1358
  maxTokens: 64e3,
1154
1359
  contextWindow: 64e3,
@@ -1176,16 +1381,6 @@ var cerebrasModels = {
1176
1381
  outputPrice: 0,
1177
1382
  description: "SOTA coding performance with ~2500 tokens/s"
1178
1383
  },
1179
- "qwen-3-235b-a22b-thinking-2507": {
1180
- maxTokens: 4e4,
1181
- contextWindow: 65e3,
1182
- supportsImages: false,
1183
- supportsPromptCache: false,
1184
- inputPrice: 0,
1185
- outputPrice: 0,
1186
- description: "SOTA performance with ~1500 tokens/s",
1187
- supportsReasoningEffort: true
1188
- },
1189
1384
  "gpt-oss-120b": {
1190
1385
  maxTokens: 8e3,
1191
1386
  contextWindow: 64e3,
@@ -1584,7 +1779,10 @@ var claudeCodeModels = {
1584
1779
  // Claude Code does report cache tokens
1585
1780
  supportsReasoningEffort: false,
1586
1781
  supportsReasoningBudget: false,
1587
- requiredReasoningBudget: false
1782
+ requiredReasoningBudget: false,
1783
+ // Claude Code manages its own tools and temperature via the CLI
1784
+ supportsNativeTools: false,
1785
+ supportsTemperature: false
1588
1786
  },
1589
1787
  "claude-sonnet-4-5-20250929[1m]": {
1590
1788
  ...anthropicModels["claude-sonnet-4-5"],
@@ -1595,7 +1793,10 @@ var claudeCodeModels = {
1595
1793
  // Claude Code does report cache tokens
1596
1794
  supportsReasoningEffort: false,
1597
1795
  supportsReasoningBudget: false,
1598
- requiredReasoningBudget: false
1796
+ requiredReasoningBudget: false,
1797
+ // Claude Code manages its own tools and temperature via the CLI
1798
+ supportsNativeTools: false,
1799
+ supportsTemperature: false
1599
1800
  },
1600
1801
  "claude-sonnet-4-20250514": {
1601
1802
  ...anthropicModels["claude-sonnet-4-20250514"],
@@ -1604,7 +1805,22 @@ var claudeCodeModels = {
1604
1805
  // Claude Code does report cache tokens
1605
1806
  supportsReasoningEffort: false,
1606
1807
  supportsReasoningBudget: false,
1607
- requiredReasoningBudget: false
1808
+ requiredReasoningBudget: false,
1809
+ // Claude Code manages its own tools and temperature via the CLI
1810
+ supportsNativeTools: false,
1811
+ supportsTemperature: false
1812
+ },
1813
+ "claude-opus-4-5-20251101": {
1814
+ ...anthropicModels["claude-opus-4-5-20251101"],
1815
+ supportsImages: false,
1816
+ supportsPromptCache: true,
1817
+ // Claude Code does report cache tokens
1818
+ supportsReasoningEffort: false,
1819
+ supportsReasoningBudget: false,
1820
+ requiredReasoningBudget: false,
1821
+ // Claude Code manages its own tools and temperature via the CLI
1822
+ supportsNativeTools: false,
1823
+ supportsTemperature: false
1608
1824
  },
1609
1825
  "claude-opus-4-1-20250805": {
1610
1826
  ...anthropicModels["claude-opus-4-1-20250805"],
@@ -1613,7 +1829,10 @@ var claudeCodeModels = {
1613
1829
  // Claude Code does report cache tokens
1614
1830
  supportsReasoningEffort: false,
1615
1831
  supportsReasoningBudget: false,
1616
- requiredReasoningBudget: false
1832
+ requiredReasoningBudget: false,
1833
+ // Claude Code manages its own tools and temperature via the CLI
1834
+ supportsNativeTools: false,
1835
+ supportsTemperature: false
1617
1836
  },
1618
1837
  "claude-opus-4-20250514": {
1619
1838
  ...anthropicModels["claude-opus-4-20250514"],
@@ -1622,7 +1841,10 @@ var claudeCodeModels = {
1622
1841
  // Claude Code does report cache tokens
1623
1842
  supportsReasoningEffort: false,
1624
1843
  supportsReasoningBudget: false,
1625
- requiredReasoningBudget: false
1844
+ requiredReasoningBudget: false,
1845
+ // Claude Code manages its own tools and temperature via the CLI
1846
+ supportsNativeTools: false,
1847
+ supportsTemperature: false
1626
1848
  },
1627
1849
  "claude-3-7-sonnet-20250219": {
1628
1850
  ...anthropicModels["claude-3-7-sonnet-20250219"],
@@ -1631,7 +1853,10 @@ var claudeCodeModels = {
1631
1853
  // Claude Code does report cache tokens
1632
1854
  supportsReasoningEffort: false,
1633
1855
  supportsReasoningBudget: false,
1634
- requiredReasoningBudget: false
1856
+ requiredReasoningBudget: false,
1857
+ // Claude Code manages its own tools and temperature via the CLI
1858
+ supportsNativeTools: false,
1859
+ supportsTemperature: false
1635
1860
  },
1636
1861
  "claude-3-5-sonnet-20241022": {
1637
1862
  ...anthropicModels["claude-3-5-sonnet-20241022"],
@@ -1640,7 +1865,10 @@ var claudeCodeModels = {
1640
1865
  // Claude Code does report cache tokens
1641
1866
  supportsReasoningEffort: false,
1642
1867
  supportsReasoningBudget: false,
1643
- requiredReasoningBudget: false
1868
+ requiredReasoningBudget: false,
1869
+ // Claude Code manages its own tools and temperature via the CLI
1870
+ supportsNativeTools: false,
1871
+ supportsTemperature: false
1644
1872
  },
1645
1873
  "claude-3-5-haiku-20241022": {
1646
1874
  ...anthropicModels["claude-3-5-haiku-20241022"],
@@ -1649,7 +1877,10 @@ var claudeCodeModels = {
1649
1877
  // Claude Code does report cache tokens
1650
1878
  supportsReasoningEffort: false,
1651
1879
  supportsReasoningBudget: false,
1652
- requiredReasoningBudget: false
1880
+ requiredReasoningBudget: false,
1881
+ // Claude Code manages its own tools and temperature via the CLI
1882
+ supportsNativeTools: false,
1883
+ supportsTemperature: false
1653
1884
  },
1654
1885
  "claude-haiku-4-5-20251001": {
1655
1886
  ...anthropicModels["claude-haiku-4-5-20251001"],
@@ -1658,7 +1889,10 @@ var claudeCodeModels = {
1658
1889
  // Claude Code does report cache tokens
1659
1890
  supportsReasoningEffort: false,
1660
1891
  supportsReasoningBudget: false,
1661
- requiredReasoningBudget: false
1892
+ requiredReasoningBudget: false,
1893
+ // Claude Code manages its own tools and temperature via the CLI
1894
+ supportsNativeTools: false,
1895
+ supportsTemperature: false
1662
1896
  }
1663
1897
  };
1664
1898
 
@@ -1671,6 +1905,7 @@ var deepSeekModels = {
1671
1905
  contextWindow: 128e3,
1672
1906
  supportsImages: false,
1673
1907
  supportsPromptCache: true,
1908
+ supportsNativeTools: true,
1674
1909
  inputPrice: 0.56,
1675
1910
  // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1676
1911
  outputPrice: 1.68,
@@ -1687,6 +1922,7 @@ var deepSeekModels = {
1687
1922
  contextWindow: 128e3,
1688
1923
  supportsImages: false,
1689
1924
  supportsPromptCache: true,
1925
+ supportsNativeTools: true,
1690
1926
  inputPrice: 0.56,
1691
1927
  // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1692
1928
  outputPrice: 1.68,
@@ -1708,6 +1944,7 @@ var doubaoModels = {
1708
1944
  contextWindow: 128e3,
1709
1945
  supportsImages: true,
1710
1946
  supportsPromptCache: true,
1947
+ supportsNativeTools: true,
1711
1948
  inputPrice: 1e-4,
1712
1949
  // $0.0001 per million tokens (cache miss)
1713
1950
  outputPrice: 4e-4,
@@ -1723,6 +1960,7 @@ var doubaoModels = {
1723
1960
  contextWindow: 128e3,
1724
1961
  supportsImages: true,
1725
1962
  supportsPromptCache: true,
1963
+ supportsNativeTools: true,
1726
1964
  inputPrice: 2e-4,
1727
1965
  // $0.0002 per million tokens
1728
1966
  outputPrice: 8e-4,
@@ -1738,6 +1976,7 @@ var doubaoModels = {
1738
1976
  contextWindow: 128e3,
1739
1977
  supportsImages: true,
1740
1978
  supportsPromptCache: true,
1979
+ supportsNativeTools: true,
1741
1980
  inputPrice: 15e-5,
1742
1981
  // $0.00015 per million tokens
1743
1982
  outputPrice: 6e-4,
@@ -1778,6 +2017,7 @@ var featherlessModels = {
1778
2017
  contextWindow: 32678,
1779
2018
  supportsImages: false,
1780
2019
  supportsPromptCache: false,
2020
+ supportsNativeTools: true,
1781
2021
  inputPrice: 0,
1782
2022
  outputPrice: 0,
1783
2023
  description: "Kimi K2 Instruct model."
@@ -1796,6 +2036,7 @@ var featherlessModels = {
1796
2036
  contextWindow: 32678,
1797
2037
  supportsImages: false,
1798
2038
  supportsPromptCache: false,
2039
+ supportsNativeTools: true,
1799
2040
  inputPrice: 0,
1800
2041
  outputPrice: 0,
1801
2042
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -1811,6 +2052,7 @@ var fireworksModels = {
1811
2052
  contextWindow: 262144,
1812
2053
  supportsImages: false,
1813
2054
  supportsPromptCache: true,
2055
+ supportsNativeTools: true,
1814
2056
  inputPrice: 0.6,
1815
2057
  outputPrice: 2.5,
1816
2058
  cacheReadsPrice: 0.15,
@@ -1821,6 +2063,7 @@ var fireworksModels = {
1821
2063
  contextWindow: 128e3,
1822
2064
  supportsImages: false,
1823
2065
  supportsPromptCache: false,
2066
+ supportsNativeTools: true,
1824
2067
  inputPrice: 0.6,
1825
2068
  outputPrice: 2.5,
1826
2069
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -1830,6 +2073,7 @@ var fireworksModels = {
1830
2073
  contextWindow: 204800,
1831
2074
  supportsImages: false,
1832
2075
  supportsPromptCache: false,
2076
+ supportsNativeTools: true,
1833
2077
  inputPrice: 0.3,
1834
2078
  outputPrice: 1.2,
1835
2079
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -1839,6 +2083,7 @@ var fireworksModels = {
1839
2083
  contextWindow: 256e3,
1840
2084
  supportsImages: false,
1841
2085
  supportsPromptCache: false,
2086
+ supportsNativeTools: true,
1842
2087
  inputPrice: 0.22,
1843
2088
  outputPrice: 0.88,
1844
2089
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -1848,6 +2093,7 @@ var fireworksModels = {
1848
2093
  contextWindow: 256e3,
1849
2094
  supportsImages: false,
1850
2095
  supportsPromptCache: false,
2096
+ supportsNativeTools: true,
1851
2097
  inputPrice: 0.45,
1852
2098
  outputPrice: 1.8,
1853
2099
  description: "Qwen3's most agentic code model to date."
@@ -1857,6 +2103,7 @@ var fireworksModels = {
1857
2103
  contextWindow: 16e4,
1858
2104
  supportsImages: false,
1859
2105
  supportsPromptCache: false,
2106
+ supportsNativeTools: true,
1860
2107
  inputPrice: 3,
1861
2108
  outputPrice: 8,
1862
2109
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -1866,6 +2113,7 @@ var fireworksModels = {
1866
2113
  contextWindow: 128e3,
1867
2114
  supportsImages: false,
1868
2115
  supportsPromptCache: false,
2116
+ supportsNativeTools: true,
1869
2117
  inputPrice: 0.9,
1870
2118
  outputPrice: 0.9,
1871
2119
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -1875,6 +2123,7 @@ var fireworksModels = {
1875
2123
  contextWindow: 163840,
1876
2124
  supportsImages: false,
1877
2125
  supportsPromptCache: false,
2126
+ supportsNativeTools: true,
1878
2127
  inputPrice: 0.56,
1879
2128
  outputPrice: 1.68,
1880
2129
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -1884,6 +2133,7 @@ var fireworksModels = {
1884
2133
  contextWindow: 128e3,
1885
2134
  supportsImages: false,
1886
2135
  supportsPromptCache: false,
2136
+ supportsNativeTools: true,
1887
2137
  inputPrice: 0.55,
1888
2138
  outputPrice: 2.19,
1889
2139
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -1893,6 +2143,7 @@ var fireworksModels = {
1893
2143
  contextWindow: 128e3,
1894
2144
  supportsImages: false,
1895
2145
  supportsPromptCache: false,
2146
+ supportsNativeTools: true,
1896
2147
  inputPrice: 0.55,
1897
2148
  outputPrice: 2.19,
1898
2149
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -1902,6 +2153,7 @@ var fireworksModels = {
1902
2153
  contextWindow: 198e3,
1903
2154
  supportsImages: false,
1904
2155
  supportsPromptCache: false,
2156
+ supportsNativeTools: true,
1905
2157
  inputPrice: 0.55,
1906
2158
  outputPrice: 2.19,
1907
2159
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -1911,6 +2163,7 @@ var fireworksModels = {
1911
2163
  contextWindow: 128e3,
1912
2164
  supportsImages: false,
1913
2165
  supportsPromptCache: false,
2166
+ supportsNativeTools: true,
1914
2167
  inputPrice: 0.07,
1915
2168
  outputPrice: 0.3,
1916
2169
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -1920,6 +2173,7 @@ var fireworksModels = {
1920
2173
  contextWindow: 128e3,
1921
2174
  supportsImages: false,
1922
2175
  supportsPromptCache: false,
2176
+ supportsNativeTools: true,
1923
2177
  inputPrice: 0.15,
1924
2178
  outputPrice: 0.6,
1925
2179
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -1927,133 +2181,39 @@ var fireworksModels = {
1927
2181
  };
1928
2182
 
1929
2183
  // src/providers/gemini.ts
1930
- var geminiDefaultModelId = "gemini-2.0-flash-001";
2184
+ var geminiDefaultModelId = "gemini-2.5-pro";
1931
2185
  var geminiModels = {
1932
- // Latest models (pointing to the most recent stable versions)
1933
- "gemini-flash-latest": {
1934
- maxTokens: 65536,
1935
- contextWindow: 1048576,
1936
- supportsImages: true,
1937
- supportsPromptCache: true,
1938
- inputPrice: 0.3,
1939
- outputPrice: 2.5,
1940
- cacheReadsPrice: 0.075,
1941
- cacheWritesPrice: 1,
1942
- maxThinkingTokens: 24576,
1943
- supportsReasoningBudget: true
1944
- },
1945
- "gemini-flash-lite-latest": {
2186
+ "gemini-3-pro-preview": {
1946
2187
  maxTokens: 65536,
1947
2188
  contextWindow: 1048576,
1948
2189
  supportsImages: true,
2190
+ supportsNativeTools: true,
1949
2191
  supportsPromptCache: true,
1950
- inputPrice: 0.1,
1951
- outputPrice: 0.4,
1952
- cacheReadsPrice: 0.025,
1953
- cacheWritesPrice: 1,
1954
- supportsReasoningBudget: true,
1955
- maxThinkingTokens: 24576
1956
- },
1957
- // 2.5 Flash models (09-2025 versions - most recent)
1958
- "gemini-2.5-flash-preview-09-2025": {
1959
- maxTokens: 65536,
1960
- contextWindow: 1048576,
1961
- supportsImages: true,
1962
- supportsPromptCache: true,
1963
- inputPrice: 0.3,
1964
- outputPrice: 2.5,
1965
- cacheReadsPrice: 0.075,
1966
- cacheWritesPrice: 1,
1967
- maxThinkingTokens: 24576,
1968
- supportsReasoningBudget: true
1969
- },
1970
- "gemini-2.5-flash-lite-preview-09-2025": {
1971
- maxTokens: 65536,
1972
- contextWindow: 1048576,
1973
- supportsImages: true,
1974
- supportsPromptCache: true,
1975
- inputPrice: 0.1,
1976
- outputPrice: 0.4,
1977
- cacheReadsPrice: 0.025,
1978
- cacheWritesPrice: 1,
1979
- supportsReasoningBudget: true,
1980
- maxThinkingTokens: 24576
1981
- },
1982
- // 2.5 Flash models (06-17 version)
1983
- "gemini-2.5-flash-lite-preview-06-17": {
1984
- maxTokens: 64e3,
1985
- contextWindow: 1048576,
1986
- supportsImages: true,
1987
- supportsPromptCache: true,
1988
- inputPrice: 0.1,
1989
- outputPrice: 0.4,
1990
- cacheReadsPrice: 0.025,
1991
- cacheWritesPrice: 1,
1992
- supportsReasoningBudget: true,
1993
- maxThinkingTokens: 24576
1994
- },
1995
- // 2.5 Flash models (05-20 versions)
1996
- "gemini-2.5-flash-preview-05-20:thinking": {
1997
- maxTokens: 65535,
1998
- contextWindow: 1048576,
1999
- supportsImages: true,
2000
- supportsPromptCache: true,
2001
- inputPrice: 0.15,
2002
- outputPrice: 3.5,
2003
- cacheReadsPrice: 0.0375,
2004
- cacheWritesPrice: 1,
2005
- maxThinkingTokens: 24576,
2006
- supportsReasoningBudget: true,
2007
- requiredReasoningBudget: true
2008
- },
2009
- "gemini-2.5-flash-preview-05-20": {
2010
- maxTokens: 65535,
2011
- contextWindow: 1048576,
2012
- supportsImages: true,
2013
- supportsPromptCache: true,
2014
- inputPrice: 0.15,
2015
- outputPrice: 0.6,
2016
- cacheReadsPrice: 0.0375,
2017
- cacheWritesPrice: 1
2018
- },
2019
- // 2.5 Flash models (04-17 versions)
2020
- "gemini-2.5-flash-preview-04-17:thinking": {
2021
- maxTokens: 65535,
2022
- contextWindow: 1048576,
2023
- supportsImages: true,
2024
- supportsPromptCache: false,
2025
- inputPrice: 0.15,
2026
- outputPrice: 3.5,
2027
- maxThinkingTokens: 24576,
2028
- supportsReasoningBudget: true,
2029
- requiredReasoningBudget: true
2030
- },
2031
- "gemini-2.5-flash-preview-04-17": {
2032
- maxTokens: 65535,
2033
- contextWindow: 1048576,
2034
- supportsImages: true,
2035
- supportsPromptCache: false,
2036
- inputPrice: 0.15,
2037
- outputPrice: 0.6
2038
- },
2039
- // 2.5 Flash stable
2040
- "gemini-2.5-flash": {
2041
- maxTokens: 64e3,
2042
- contextWindow: 1048576,
2043
- supportsImages: true,
2044
- supportsPromptCache: true,
2045
- inputPrice: 0.3,
2046
- outputPrice: 2.5,
2047
- cacheReadsPrice: 0.075,
2048
- cacheWritesPrice: 1,
2049
- maxThinkingTokens: 24576,
2050
- supportsReasoningBudget: true
2192
+ supportsReasoningEffort: ["low", "high"],
2193
+ reasoningEffort: "low",
2194
+ supportsTemperature: true,
2195
+ defaultTemperature: 1,
2196
+ inputPrice: 4,
2197
+ outputPrice: 18,
2198
+ tiers: [
2199
+ {
2200
+ contextWindow: 2e5,
2201
+ inputPrice: 2,
2202
+ outputPrice: 12
2203
+ },
2204
+ {
2205
+ contextWindow: Infinity,
2206
+ inputPrice: 4,
2207
+ outputPrice: 18
2208
+ }
2209
+ ]
2051
2210
  },
2052
2211
  // 2.5 Pro models
2053
- "gemini-2.5-pro-preview-06-05": {
2054
- maxTokens: 65535,
2212
+ "gemini-2.5-pro": {
2213
+ maxTokens: 64e3,
2055
2214
  contextWindow: 1048576,
2056
2215
  supportsImages: true,
2216
+ supportsNativeTools: true,
2057
2217
  supportsPromptCache: true,
2058
2218
  inputPrice: 2.5,
2059
2219
  // This is the pricing for prompts above 200k tokens.
@@ -2062,6 +2222,7 @@ var geminiModels = {
2062
2222
  cacheWritesPrice: 4.5,
2063
2223
  maxThinkingTokens: 32768,
2064
2224
  supportsReasoningBudget: true,
2225
+ requiredReasoningBudget: true,
2065
2226
  tiers: [
2066
2227
  {
2067
2228
  contextWindow: 2e5,
@@ -2077,16 +2238,19 @@ var geminiModels = {
2077
2238
  }
2078
2239
  ]
2079
2240
  },
2080
- "gemini-2.5-pro-preview-05-06": {
2241
+ "gemini-2.5-pro-preview-06-05": {
2081
2242
  maxTokens: 65535,
2082
2243
  contextWindow: 1048576,
2083
2244
  supportsImages: true,
2245
+ supportsNativeTools: true,
2084
2246
  supportsPromptCache: true,
2085
2247
  inputPrice: 2.5,
2086
2248
  // This is the pricing for prompts above 200k tokens.
2087
2249
  outputPrice: 15,
2088
2250
  cacheReadsPrice: 0.625,
2089
2251
  cacheWritesPrice: 4.5,
2252
+ maxThinkingTokens: 32768,
2253
+ supportsReasoningBudget: true,
2090
2254
  tiers: [
2091
2255
  {
2092
2256
  contextWindow: 2e5,
@@ -2102,18 +2266,17 @@ var geminiModels = {
2102
2266
  }
2103
2267
  ]
2104
2268
  },
2105
- "gemini-2.5-pro-preview-03-25": {
2269
+ "gemini-2.5-pro-preview-05-06": {
2106
2270
  maxTokens: 65535,
2107
2271
  contextWindow: 1048576,
2108
2272
  supportsImages: true,
2273
+ supportsNativeTools: true,
2109
2274
  supportsPromptCache: true,
2110
2275
  inputPrice: 2.5,
2111
2276
  // This is the pricing for prompts above 200k tokens.
2112
2277
  outputPrice: 15,
2113
2278
  cacheReadsPrice: 0.625,
2114
2279
  cacheWritesPrice: 4.5,
2115
- maxThinkingTokens: 32768,
2116
- supportsReasoningBudget: true,
2117
2280
  tiers: [
2118
2281
  {
2119
2282
  contextWindow: 2e5,
@@ -2129,18 +2292,11 @@ var geminiModels = {
2129
2292
  }
2130
2293
  ]
2131
2294
  },
2132
- "gemini-2.5-pro-exp-03-25": {
2295
+ "gemini-2.5-pro-preview-03-25": {
2133
2296
  maxTokens: 65535,
2134
2297
  contextWindow: 1048576,
2135
2298
  supportsImages: true,
2136
- supportsPromptCache: false,
2137
- inputPrice: 0,
2138
- outputPrice: 0
2139
- },
2140
- "gemini-2.5-pro": {
2141
- maxTokens: 64e3,
2142
- contextWindow: 1048576,
2143
- supportsImages: true,
2299
+ supportsNativeTools: true,
2144
2300
  supportsPromptCache: true,
2145
2301
  inputPrice: 2.5,
2146
2302
  // This is the pricing for prompts above 200k tokens.
@@ -2149,7 +2305,6 @@ var geminiModels = {
2149
2305
  cacheWritesPrice: 4.5,
2150
2306
  maxThinkingTokens: 32768,
2151
2307
  supportsReasoningBudget: true,
2152
- requiredReasoningBudget: true,
2153
2308
  tiers: [
2154
2309
  {
2155
2310
  contextWindow: 2e5,
@@ -2165,125 +2320,72 @@ var geminiModels = {
2165
2320
  }
2166
2321
  ]
2167
2322
  },
2168
- // 2.0 Flash models
2169
- "gemini-2.0-flash-lite-preview-02-05": {
2170
- maxTokens: 8192,
2323
+ // 2.5 Flash models
2324
+ "gemini-flash-latest": {
2325
+ maxTokens: 65536,
2171
2326
  contextWindow: 1048576,
2172
2327
  supportsImages: true,
2173
- supportsPromptCache: false,
2174
- inputPrice: 0,
2175
- outputPrice: 0
2328
+ supportsNativeTools: true,
2329
+ supportsPromptCache: true,
2330
+ inputPrice: 0.3,
2331
+ outputPrice: 2.5,
2332
+ cacheReadsPrice: 0.075,
2333
+ cacheWritesPrice: 1,
2334
+ maxThinkingTokens: 24576,
2335
+ supportsReasoningBudget: true
2176
2336
  },
2177
- "gemini-2.0-flash-thinking-exp-01-21": {
2337
+ "gemini-2.5-flash-preview-09-2025": {
2178
2338
  maxTokens: 65536,
2179
2339
  contextWindow: 1048576,
2180
2340
  supportsImages: true,
2181
- supportsPromptCache: false,
2182
- inputPrice: 0,
2183
- outputPrice: 0
2184
- },
2185
- "gemini-2.0-flash-thinking-exp-1219": {
2186
- maxTokens: 8192,
2187
- contextWindow: 32767,
2188
- supportsImages: true,
2189
- supportsPromptCache: false,
2190
- inputPrice: 0,
2191
- outputPrice: 0
2341
+ supportsNativeTools: true,
2342
+ supportsPromptCache: true,
2343
+ inputPrice: 0.3,
2344
+ outputPrice: 2.5,
2345
+ cacheReadsPrice: 0.075,
2346
+ cacheWritesPrice: 1,
2347
+ maxThinkingTokens: 24576,
2348
+ supportsReasoningBudget: true
2192
2349
  },
2193
- "gemini-2.0-flash-exp": {
2194
- maxTokens: 8192,
2350
+ "gemini-2.5-flash": {
2351
+ maxTokens: 64e3,
2195
2352
  contextWindow: 1048576,
2196
2353
  supportsImages: true,
2197
- supportsPromptCache: false,
2198
- inputPrice: 0,
2199
- outputPrice: 0
2354
+ supportsNativeTools: true,
2355
+ supportsPromptCache: true,
2356
+ inputPrice: 0.3,
2357
+ outputPrice: 2.5,
2358
+ cacheReadsPrice: 0.075,
2359
+ cacheWritesPrice: 1,
2360
+ maxThinkingTokens: 24576,
2361
+ supportsReasoningBudget: true
2200
2362
  },
2201
- "gemini-2.0-flash-001": {
2202
- maxTokens: 8192,
2363
+ // 2.5 Flash Lite models
2364
+ "gemini-flash-lite-latest": {
2365
+ maxTokens: 65536,
2203
2366
  contextWindow: 1048576,
2204
2367
  supportsImages: true,
2368
+ supportsNativeTools: true,
2205
2369
  supportsPromptCache: true,
2206
2370
  inputPrice: 0.1,
2207
2371
  outputPrice: 0.4,
2208
2372
  cacheReadsPrice: 0.025,
2209
- cacheWritesPrice: 1
2210
- },
2211
- // 2.0 Pro models
2212
- "gemini-2.0-pro-exp-02-05": {
2213
- maxTokens: 8192,
2214
- contextWindow: 2097152,
2215
- supportsImages: true,
2216
- supportsPromptCache: false,
2217
- inputPrice: 0,
2218
- outputPrice: 0
2373
+ cacheWritesPrice: 1,
2374
+ supportsReasoningBudget: true,
2375
+ maxThinkingTokens: 24576
2219
2376
  },
2220
- // 1.5 Flash models
2221
- "gemini-1.5-flash-002": {
2222
- maxTokens: 8192,
2377
+ "gemini-2.5-flash-lite-preview-09-2025": {
2378
+ maxTokens: 65536,
2223
2379
  contextWindow: 1048576,
2224
2380
  supportsImages: true,
2381
+ supportsNativeTools: true,
2225
2382
  supportsPromptCache: true,
2226
- inputPrice: 0.15,
2227
- // This is the pricing for prompts above 128k tokens.
2228
- outputPrice: 0.6,
2229
- cacheReadsPrice: 0.0375,
2383
+ inputPrice: 0.1,
2384
+ outputPrice: 0.4,
2385
+ cacheReadsPrice: 0.025,
2230
2386
  cacheWritesPrice: 1,
2231
- tiers: [
2232
- {
2233
- contextWindow: 128e3,
2234
- inputPrice: 0.075,
2235
- outputPrice: 0.3,
2236
- cacheReadsPrice: 0.01875
2237
- },
2238
- {
2239
- contextWindow: Infinity,
2240
- inputPrice: 0.15,
2241
- outputPrice: 0.6,
2242
- cacheReadsPrice: 0.0375
2243
- }
2244
- ]
2245
- },
2246
- "gemini-1.5-flash-exp-0827": {
2247
- maxTokens: 8192,
2248
- contextWindow: 1048576,
2249
- supportsImages: true,
2250
- supportsPromptCache: false,
2251
- inputPrice: 0,
2252
- outputPrice: 0
2253
- },
2254
- "gemini-1.5-flash-8b-exp-0827": {
2255
- maxTokens: 8192,
2256
- contextWindow: 1048576,
2257
- supportsImages: true,
2258
- supportsPromptCache: false,
2259
- inputPrice: 0,
2260
- outputPrice: 0
2261
- },
2262
- // 1.5 Pro models
2263
- "gemini-1.5-pro-002": {
2264
- maxTokens: 8192,
2265
- contextWindow: 2097152,
2266
- supportsImages: true,
2267
- supportsPromptCache: false,
2268
- inputPrice: 0,
2269
- outputPrice: 0
2270
- },
2271
- "gemini-1.5-pro-exp-0827": {
2272
- maxTokens: 8192,
2273
- contextWindow: 2097152,
2274
- supportsImages: true,
2275
- supportsPromptCache: false,
2276
- inputPrice: 0,
2277
- outputPrice: 0
2278
- },
2279
- // Experimental models
2280
- "gemini-exp-1206": {
2281
- maxTokens: 8192,
2282
- contextWindow: 2097152,
2283
- supportsImages: true,
2284
- supportsPromptCache: false,
2285
- inputPrice: 0,
2286
- outputPrice: 0
2387
+ supportsReasoningBudget: true,
2388
+ maxThinkingTokens: 24576
2287
2389
  }
2288
2390
  };
2289
2391
 
@@ -2311,6 +2413,7 @@ var groqModels = {
2311
2413
  contextWindow: 131072,
2312
2414
  supportsImages: false,
2313
2415
  supportsPromptCache: false,
2416
+ supportsNativeTools: true,
2314
2417
  inputPrice: 0.05,
2315
2418
  outputPrice: 0.08,
2316
2419
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2320,6 +2423,7 @@ var groqModels = {
2320
2423
  contextWindow: 131072,
2321
2424
  supportsImages: false,
2322
2425
  supportsPromptCache: false,
2426
+ supportsNativeTools: true,
2323
2427
  inputPrice: 0.59,
2324
2428
  outputPrice: 0.79,
2325
2429
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2329,6 +2433,7 @@ var groqModels = {
2329
2433
  contextWindow: 131072,
2330
2434
  supportsImages: false,
2331
2435
  supportsPromptCache: false,
2436
+ supportsNativeTools: true,
2332
2437
  inputPrice: 0.11,
2333
2438
  outputPrice: 0.34,
2334
2439
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2365,6 +2470,7 @@ var groqModels = {
2365
2470
  contextWindow: 131072,
2366
2471
  supportsImages: false,
2367
2472
  supportsPromptCache: false,
2473
+ supportsNativeTools: true,
2368
2474
  inputPrice: 0.29,
2369
2475
  outputPrice: 0.59,
2370
2476
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2394,6 +2500,7 @@ var groqModels = {
2394
2500
  contextWindow: 262144,
2395
2501
  supportsImages: false,
2396
2502
  supportsPromptCache: true,
2503
+ supportsNativeTools: true,
2397
2504
  inputPrice: 0.6,
2398
2505
  outputPrice: 2.5,
2399
2506
  cacheReadsPrice: 0.15,
@@ -2404,6 +2511,7 @@ var groqModels = {
2404
2511
  contextWindow: 131072,
2405
2512
  supportsImages: false,
2406
2513
  supportsPromptCache: false,
2514
+ supportsNativeTools: true,
2407
2515
  inputPrice: 0.15,
2408
2516
  outputPrice: 0.75,
2409
2517
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2413,6 +2521,7 @@ var groqModels = {
2413
2521
  contextWindow: 131072,
2414
2522
  supportsImages: false,
2415
2523
  supportsPromptCache: false,
2524
+ supportsNativeTools: true,
2416
2525
  inputPrice: 0.1,
2417
2526
  outputPrice: 0.5,
2418
2527
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2439,6 +2548,7 @@ var ioIntelligenceModels = {
2439
2548
  contextWindow: 128e3,
2440
2549
  supportsImages: false,
2441
2550
  supportsPromptCache: false,
2551
+ supportsNativeTools: true,
2442
2552
  description: "DeepSeek R1 reasoning model"
2443
2553
  },
2444
2554
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -2446,6 +2556,7 @@ var ioIntelligenceModels = {
2446
2556
  contextWindow: 43e4,
2447
2557
  supportsImages: true,
2448
2558
  supportsPromptCache: false,
2559
+ supportsNativeTools: true,
2449
2560
  description: "Llama 4 Maverick 17B model"
2450
2561
  },
2451
2562
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -2453,6 +2564,7 @@ var ioIntelligenceModels = {
2453
2564
  contextWindow: 106e3,
2454
2565
  supportsImages: false,
2455
2566
  supportsPromptCache: false,
2567
+ supportsNativeTools: true,
2456
2568
  description: "Qwen3 Coder 480B specialized for coding"
2457
2569
  },
2458
2570
  "openai/gpt-oss-120b": {
@@ -2460,6 +2572,7 @@ var ioIntelligenceModels = {
2460
2572
  contextWindow: 131072,
2461
2573
  supportsImages: false,
2462
2574
  supportsPromptCache: false,
2575
+ supportsNativeTools: true,
2463
2576
  description: "OpenAI GPT-OSS 120B model"
2464
2577
  }
2465
2578
  };
@@ -2500,75 +2613,84 @@ var mistralModels = {
2500
2613
  contextWindow: 128e3,
2501
2614
  supportsImages: true,
2502
2615
  supportsPromptCache: false,
2616
+ supportsNativeTools: true,
2503
2617
  inputPrice: 2,
2504
2618
  outputPrice: 5
2505
2619
  },
2506
2620
  "devstral-medium-latest": {
2507
- maxTokens: 131e3,
2621
+ maxTokens: 8192,
2508
2622
  contextWindow: 131e3,
2509
2623
  supportsImages: true,
2510
2624
  supportsPromptCache: false,
2625
+ supportsNativeTools: true,
2511
2626
  inputPrice: 0.4,
2512
2627
  outputPrice: 2
2513
2628
  },
2514
2629
  "mistral-medium-latest": {
2515
- maxTokens: 131e3,
2630
+ maxTokens: 8192,
2516
2631
  contextWindow: 131e3,
2517
2632
  supportsImages: true,
2518
2633
  supportsPromptCache: false,
2634
+ supportsNativeTools: true,
2519
2635
  inputPrice: 0.4,
2520
2636
  outputPrice: 2
2521
2637
  },
2522
2638
  "codestral-latest": {
2523
- maxTokens: 256e3,
2639
+ maxTokens: 8192,
2524
2640
  contextWindow: 256e3,
2525
2641
  supportsImages: false,
2526
2642
  supportsPromptCache: false,
2643
+ supportsNativeTools: true,
2527
2644
  inputPrice: 0.3,
2528
2645
  outputPrice: 0.9
2529
2646
  },
2530
2647
  "mistral-large-latest": {
2531
- maxTokens: 131e3,
2648
+ maxTokens: 8192,
2532
2649
  contextWindow: 131e3,
2533
2650
  supportsImages: false,
2534
2651
  supportsPromptCache: false,
2652
+ supportsNativeTools: true,
2535
2653
  inputPrice: 2,
2536
2654
  outputPrice: 6
2537
2655
  },
2538
2656
  "ministral-8b-latest": {
2539
- maxTokens: 131e3,
2657
+ maxTokens: 8192,
2540
2658
  contextWindow: 131e3,
2541
2659
  supportsImages: false,
2542
2660
  supportsPromptCache: false,
2661
+ supportsNativeTools: true,
2543
2662
  inputPrice: 0.1,
2544
2663
  outputPrice: 0.1
2545
2664
  },
2546
2665
  "ministral-3b-latest": {
2547
- maxTokens: 131e3,
2666
+ maxTokens: 8192,
2548
2667
  contextWindow: 131e3,
2549
2668
  supportsImages: false,
2550
2669
  supportsPromptCache: false,
2670
+ supportsNativeTools: true,
2551
2671
  inputPrice: 0.04,
2552
2672
  outputPrice: 0.04
2553
2673
  },
2554
2674
  "mistral-small-latest": {
2555
- maxTokens: 32e3,
2675
+ maxTokens: 8192,
2556
2676
  contextWindow: 32e3,
2557
2677
  supportsImages: false,
2558
2678
  supportsPromptCache: false,
2679
+ supportsNativeTools: true,
2559
2680
  inputPrice: 0.2,
2560
2681
  outputPrice: 0.6
2561
2682
  },
2562
2683
  "pixtral-large-latest": {
2563
- maxTokens: 131e3,
2684
+ maxTokens: 8192,
2564
2685
  contextWindow: 131e3,
2565
2686
  supportsImages: true,
2566
2687
  supportsPromptCache: false,
2688
+ supportsNativeTools: true,
2567
2689
  inputPrice: 2,
2568
2690
  outputPrice: 6
2569
2691
  }
2570
2692
  };
2571
- var MISTRAL_DEFAULT_TEMPERATURE = 0;
2693
+ var MISTRAL_DEFAULT_TEMPERATURE = 1;
2572
2694
 
2573
2695
  // src/providers/moonshot.ts
2574
2696
  var moonshotDefaultModelId = "kimi-k2-0905-preview";
@@ -2578,6 +2700,7 @@ var moonshotModels = {
2578
2700
  contextWindow: 131072,
2579
2701
  supportsImages: false,
2580
2702
  supportsPromptCache: true,
2703
+ supportsNativeTools: true,
2581
2704
  inputPrice: 0.6,
2582
2705
  // $0.60 per million tokens (cache miss)
2583
2706
  outputPrice: 2.5,
@@ -2593,6 +2716,7 @@ var moonshotModels = {
2593
2716
  contextWindow: 262144,
2594
2717
  supportsImages: false,
2595
2718
  supportsPromptCache: true,
2719
+ supportsNativeTools: true,
2596
2720
  inputPrice: 0.6,
2597
2721
  outputPrice: 2.5,
2598
2722
  cacheReadsPrice: 0.15,
@@ -2603,6 +2727,7 @@ var moonshotModels = {
2603
2727
  contextWindow: 262144,
2604
2728
  supportsImages: false,
2605
2729
  supportsPromptCache: true,
2730
+ supportsNativeTools: true,
2606
2731
  inputPrice: 2.4,
2607
2732
  // $2.40 per million tokens (cache miss)
2608
2733
  outputPrice: 10,
@@ -2621,6 +2746,7 @@ var moonshotModels = {
2621
2746
  supportsImages: false,
2622
2747
  // Text-only (no image/vision support)
2623
2748
  supportsPromptCache: true,
2749
+ supportsNativeTools: true,
2624
2750
  inputPrice: 0.6,
2625
2751
  // $0.60 per million tokens (cache miss)
2626
2752
  outputPrice: 2.5,
@@ -2658,6 +2784,7 @@ var openAiNativeModels = {
2658
2784
  "gpt-5.1": {
2659
2785
  maxTokens: 128e3,
2660
2786
  contextWindow: 4e5,
2787
+ supportsNativeTools: true,
2661
2788
  supportsImages: true,
2662
2789
  supportsPromptCache: true,
2663
2790
  promptCacheRetention: "24h",
@@ -2677,6 +2804,7 @@ var openAiNativeModels = {
2677
2804
  "gpt-5.1-codex": {
2678
2805
  maxTokens: 128e3,
2679
2806
  contextWindow: 4e5,
2807
+ supportsNativeTools: true,
2680
2808
  supportsImages: true,
2681
2809
  supportsPromptCache: true,
2682
2810
  promptCacheRetention: "24h",
@@ -2692,6 +2820,7 @@ var openAiNativeModels = {
2692
2820
  "gpt-5.1-codex-mini": {
2693
2821
  maxTokens: 128e3,
2694
2822
  contextWindow: 4e5,
2823
+ supportsNativeTools: true,
2695
2824
  supportsImages: true,
2696
2825
  supportsPromptCache: true,
2697
2826
  promptCacheRetention: "24h",
@@ -2706,6 +2835,7 @@ var openAiNativeModels = {
2706
2835
  "gpt-5": {
2707
2836
  maxTokens: 128e3,
2708
2837
  contextWindow: 4e5,
2838
+ supportsNativeTools: true,
2709
2839
  supportsImages: true,
2710
2840
  supportsPromptCache: true,
2711
2841
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -2724,6 +2854,7 @@ var openAiNativeModels = {
2724
2854
  "gpt-5-mini": {
2725
2855
  maxTokens: 128e3,
2726
2856
  contextWindow: 4e5,
2857
+ supportsNativeTools: true,
2727
2858
  supportsImages: true,
2728
2859
  supportsPromptCache: true,
2729
2860
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -2742,6 +2873,7 @@ var openAiNativeModels = {
2742
2873
  "gpt-5-codex": {
2743
2874
  maxTokens: 128e3,
2744
2875
  contextWindow: 4e5,
2876
+ supportsNativeTools: true,
2745
2877
  supportsImages: true,
2746
2878
  supportsPromptCache: true,
2747
2879
  supportsReasoningEffort: ["low", "medium", "high"],
@@ -2756,6 +2888,7 @@ var openAiNativeModels = {
2756
2888
  "gpt-5-nano": {
2757
2889
  maxTokens: 128e3,
2758
2890
  contextWindow: 4e5,
2891
+ supportsNativeTools: true,
2759
2892
  supportsImages: true,
2760
2893
  supportsPromptCache: true,
2761
2894
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -2771,6 +2904,7 @@ var openAiNativeModels = {
2771
2904
  "gpt-5-chat-latest": {
2772
2905
  maxTokens: 128e3,
2773
2906
  contextWindow: 4e5,
2907
+ supportsNativeTools: true,
2774
2908
  supportsImages: true,
2775
2909
  supportsPromptCache: true,
2776
2910
  inputPrice: 1.25,
@@ -2781,6 +2915,7 @@ var openAiNativeModels = {
2781
2915
  "gpt-4.1": {
2782
2916
  maxTokens: 32768,
2783
2917
  contextWindow: 1047576,
2918
+ supportsNativeTools: true,
2784
2919
  supportsImages: true,
2785
2920
  supportsPromptCache: true,
2786
2921
  inputPrice: 2,
@@ -2794,6 +2929,7 @@ var openAiNativeModels = {
2794
2929
  "gpt-4.1-mini": {
2795
2930
  maxTokens: 32768,
2796
2931
  contextWindow: 1047576,
2932
+ supportsNativeTools: true,
2797
2933
  supportsImages: true,
2798
2934
  supportsPromptCache: true,
2799
2935
  inputPrice: 0.4,
@@ -2807,6 +2943,7 @@ var openAiNativeModels = {
2807
2943
  "gpt-4.1-nano": {
2808
2944
  maxTokens: 32768,
2809
2945
  contextWindow: 1047576,
2946
+ supportsNativeTools: true,
2810
2947
  supportsImages: true,
2811
2948
  supportsPromptCache: true,
2812
2949
  inputPrice: 0.1,
@@ -2820,6 +2957,7 @@ var openAiNativeModels = {
2820
2957
  o3: {
2821
2958
  maxTokens: 1e5,
2822
2959
  contextWindow: 2e5,
2960
+ supportsNativeTools: true,
2823
2961
  supportsImages: true,
2824
2962
  supportsPromptCache: true,
2825
2963
  inputPrice: 2,
@@ -2836,6 +2974,7 @@ var openAiNativeModels = {
2836
2974
  "o3-high": {
2837
2975
  maxTokens: 1e5,
2838
2976
  contextWindow: 2e5,
2977
+ supportsNativeTools: true,
2839
2978
  supportsImages: true,
2840
2979
  supportsPromptCache: true,
2841
2980
  inputPrice: 2,
@@ -2847,6 +2986,7 @@ var openAiNativeModels = {
2847
2986
  "o3-low": {
2848
2987
  maxTokens: 1e5,
2849
2988
  contextWindow: 2e5,
2989
+ supportsNativeTools: true,
2850
2990
  supportsImages: true,
2851
2991
  supportsPromptCache: true,
2852
2992
  inputPrice: 2,
@@ -2858,6 +2998,7 @@ var openAiNativeModels = {
2858
2998
  "o4-mini": {
2859
2999
  maxTokens: 1e5,
2860
3000
  contextWindow: 2e5,
3001
+ supportsNativeTools: true,
2861
3002
  supportsImages: true,
2862
3003
  supportsPromptCache: true,
2863
3004
  inputPrice: 1.1,
@@ -2874,6 +3015,7 @@ var openAiNativeModels = {
2874
3015
  "o4-mini-high": {
2875
3016
  maxTokens: 1e5,
2876
3017
  contextWindow: 2e5,
3018
+ supportsNativeTools: true,
2877
3019
  supportsImages: true,
2878
3020
  supportsPromptCache: true,
2879
3021
  inputPrice: 1.1,
@@ -2885,6 +3027,7 @@ var openAiNativeModels = {
2885
3027
  "o4-mini-low": {
2886
3028
  maxTokens: 1e5,
2887
3029
  contextWindow: 2e5,
3030
+ supportsNativeTools: true,
2888
3031
  supportsImages: true,
2889
3032
  supportsPromptCache: true,
2890
3033
  inputPrice: 1.1,
@@ -2896,6 +3039,7 @@ var openAiNativeModels = {
2896
3039
  "o3-mini": {
2897
3040
  maxTokens: 1e5,
2898
3041
  contextWindow: 2e5,
3042
+ supportsNativeTools: true,
2899
3043
  supportsImages: false,
2900
3044
  supportsPromptCache: true,
2901
3045
  inputPrice: 1.1,
@@ -2908,6 +3052,7 @@ var openAiNativeModels = {
2908
3052
  "o3-mini-high": {
2909
3053
  maxTokens: 1e5,
2910
3054
  contextWindow: 2e5,
3055
+ supportsNativeTools: true,
2911
3056
  supportsImages: false,
2912
3057
  supportsPromptCache: true,
2913
3058
  inputPrice: 1.1,
@@ -2919,6 +3064,7 @@ var openAiNativeModels = {
2919
3064
  "o3-mini-low": {
2920
3065
  maxTokens: 1e5,
2921
3066
  contextWindow: 2e5,
3067
+ supportsNativeTools: true,
2922
3068
  supportsImages: false,
2923
3069
  supportsPromptCache: true,
2924
3070
  inputPrice: 1.1,
@@ -2930,6 +3076,7 @@ var openAiNativeModels = {
2930
3076
  o1: {
2931
3077
  maxTokens: 1e5,
2932
3078
  contextWindow: 2e5,
3079
+ supportsNativeTools: true,
2933
3080
  supportsImages: true,
2934
3081
  supportsPromptCache: true,
2935
3082
  inputPrice: 15,
@@ -2940,6 +3087,7 @@ var openAiNativeModels = {
2940
3087
  "o1-preview": {
2941
3088
  maxTokens: 32768,
2942
3089
  contextWindow: 128e3,
3090
+ supportsNativeTools: true,
2943
3091
  supportsImages: true,
2944
3092
  supportsPromptCache: true,
2945
3093
  inputPrice: 15,
@@ -2950,6 +3098,7 @@ var openAiNativeModels = {
2950
3098
  "o1-mini": {
2951
3099
  maxTokens: 65536,
2952
3100
  contextWindow: 128e3,
3101
+ supportsNativeTools: true,
2953
3102
  supportsImages: true,
2954
3103
  supportsPromptCache: true,
2955
3104
  inputPrice: 1.1,
@@ -2960,6 +3109,7 @@ var openAiNativeModels = {
2960
3109
  "gpt-4o": {
2961
3110
  maxTokens: 16384,
2962
3111
  contextWindow: 128e3,
3112
+ supportsNativeTools: true,
2963
3113
  supportsImages: true,
2964
3114
  supportsPromptCache: true,
2965
3115
  inputPrice: 2.5,
@@ -2973,6 +3123,7 @@ var openAiNativeModels = {
2973
3123
  "gpt-4o-mini": {
2974
3124
  maxTokens: 16384,
2975
3125
  contextWindow: 128e3,
3126
+ supportsNativeTools: true,
2976
3127
  supportsImages: true,
2977
3128
  supportsPromptCache: true,
2978
3129
  inputPrice: 0.15,
@@ -2986,6 +3137,7 @@ var openAiNativeModels = {
2986
3137
  "codex-mini-latest": {
2987
3138
  maxTokens: 16384,
2988
3139
  contextWindow: 2e5,
3140
+ supportsNativeTools: true,
2989
3141
  supportsImages: false,
2990
3142
  supportsPromptCache: false,
2991
3143
  inputPrice: 1.5,
@@ -2998,6 +3150,7 @@ var openAiNativeModels = {
2998
3150
  "gpt-5-2025-08-07": {
2999
3151
  maxTokens: 128e3,
3000
3152
  contextWindow: 4e5,
3153
+ supportsNativeTools: true,
3001
3154
  supportsImages: true,
3002
3155
  supportsPromptCache: true,
3003
3156
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3016,6 +3169,7 @@ var openAiNativeModels = {
3016
3169
  "gpt-5-mini-2025-08-07": {
3017
3170
  maxTokens: 128e3,
3018
3171
  contextWindow: 4e5,
3172
+ supportsNativeTools: true,
3019
3173
  supportsImages: true,
3020
3174
  supportsPromptCache: true,
3021
3175
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3034,6 +3188,7 @@ var openAiNativeModels = {
3034
3188
  "gpt-5-nano-2025-08-07": {
3035
3189
  maxTokens: 128e3,
3036
3190
  contextWindow: 4e5,
3191
+ supportsNativeTools: true,
3037
3192
  supportsImages: true,
3038
3193
  supportsPromptCache: true,
3039
3194
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3053,7 +3208,8 @@ var openAiModelInfoSaneDefaults = {
3053
3208
  supportsImages: true,
3054
3209
  supportsPromptCache: false,
3055
3210
  inputPrice: 0,
3056
- outputPrice: 0
3211
+ outputPrice: 0,
3212
+ supportsNativeTools: true
3057
3213
  };
3058
3214
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3059
3215
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3097,6 +3253,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3097
3253
  "anthropic/claude-opus-4",
3098
3254
  "anthropic/claude-opus-4.1",
3099
3255
  "anthropic/claude-haiku-4.5",
3256
+ "anthropic/claude-opus-4.5",
3100
3257
  "google/gemini-2.5-flash-preview",
3101
3258
  "google/gemini-2.5-flash-preview:thinking",
3102
3259
  "google/gemini-2.5-flash-preview-05-20",
@@ -3118,6 +3275,7 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3118
3275
  "anthropic/claude-opus-4.1",
3119
3276
  "anthropic/claude-sonnet-4",
3120
3277
  "anthropic/claude-sonnet-4.5",
3278
+ "anthropic/claude-opus-4.5",
3121
3279
  "anthropic/claude-haiku-4.5",
3122
3280
  "google/gemini-2.5-pro-preview",
3123
3281
  "google/gemini-2.5-pro",
@@ -3175,16 +3333,6 @@ var requestyDefaultModelInfo = {
3175
3333
  import { z as z7 } from "zod";
3176
3334
  var rooDefaultModelId = "xai/grok-code-fast-1";
3177
3335
  var rooModels = {};
3178
- var rooModelDefaults = {
3179
- // Add model-specific defaults below.
3180
- // You can configure defaultToolProtocol and other ModelInfo fields for specific model IDs.
3181
- "anthropic/claude-haiku-4.5": {
3182
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3183
- },
3184
- "minimax/minimax-m2:free": {
3185
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3186
- }
3187
- };
3188
3336
  var RooPricingSchema = z7.object({
3189
3337
  input: z7.string(),
3190
3338
  output: z7.string(),
@@ -3218,6 +3366,7 @@ var sambaNovaModels = {
3218
3366
  contextWindow: 16384,
3219
3367
  supportsImages: false,
3220
3368
  supportsPromptCache: false,
3369
+ supportsNativeTools: true,
3221
3370
  inputPrice: 0.1,
3222
3371
  outputPrice: 0.2,
3223
3372
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3227,6 +3376,7 @@ var sambaNovaModels = {
3227
3376
  contextWindow: 131072,
3228
3377
  supportsImages: false,
3229
3378
  supportsPromptCache: false,
3379
+ supportsNativeTools: true,
3230
3380
  inputPrice: 0.6,
3231
3381
  outputPrice: 1.2,
3232
3382
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3237,6 +3387,7 @@ var sambaNovaModels = {
3237
3387
  supportsImages: false,
3238
3388
  supportsPromptCache: false,
3239
3389
  supportsReasoningBudget: true,
3390
+ supportsNativeTools: true,
3240
3391
  inputPrice: 5,
3241
3392
  outputPrice: 7,
3242
3393
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3246,6 +3397,7 @@ var sambaNovaModels = {
3246
3397
  contextWindow: 32768,
3247
3398
  supportsImages: false,
3248
3399
  supportsPromptCache: false,
3400
+ supportsNativeTools: true,
3249
3401
  inputPrice: 3,
3250
3402
  outputPrice: 4.5,
3251
3403
  description: "DeepSeek V3 model with 32K context window."
@@ -3255,6 +3407,7 @@ var sambaNovaModels = {
3255
3407
  contextWindow: 32768,
3256
3408
  supportsImages: false,
3257
3409
  supportsPromptCache: false,
3410
+ supportsNativeTools: true,
3258
3411
  inputPrice: 3,
3259
3412
  outputPrice: 4.5,
3260
3413
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3273,6 +3426,7 @@ var sambaNovaModels = {
3273
3426
  contextWindow: 131072,
3274
3427
  supportsImages: true,
3275
3428
  supportsPromptCache: false,
3429
+ supportsNativeTools: true,
3276
3430
  inputPrice: 0.63,
3277
3431
  outputPrice: 1.8,
3278
3432
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3291,6 +3445,7 @@ var sambaNovaModels = {
3291
3445
  contextWindow: 8192,
3292
3446
  supportsImages: false,
3293
3447
  supportsPromptCache: false,
3448
+ supportsNativeTools: true,
3294
3449
  inputPrice: 0.4,
3295
3450
  outputPrice: 0.8,
3296
3451
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3300,6 +3455,7 @@ var sambaNovaModels = {
3300
3455
  contextWindow: 131072,
3301
3456
  supportsImages: false,
3302
3457
  supportsPromptCache: false,
3458
+ supportsNativeTools: true,
3303
3459
  inputPrice: 0.22,
3304
3460
  outputPrice: 0.59,
3305
3461
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3322,10 +3478,36 @@ var unboundDefaultModelInfo = {
3322
3478
  // src/providers/vertex.ts
3323
3479
  var vertexDefaultModelId = "claude-sonnet-4-5@20250929";
3324
3480
  var vertexModels = {
3481
+ "gemini-3-pro-preview": {
3482
+ maxTokens: 65536,
3483
+ contextWindow: 1048576,
3484
+ supportsImages: true,
3485
+ supportsNativeTools: true,
3486
+ supportsPromptCache: true,
3487
+ supportsReasoningEffort: ["low", "high"],
3488
+ reasoningEffort: "low",
3489
+ supportsTemperature: true,
3490
+ defaultTemperature: 1,
3491
+ inputPrice: 4,
3492
+ outputPrice: 18,
3493
+ tiers: [
3494
+ {
3495
+ contextWindow: 2e5,
3496
+ inputPrice: 2,
3497
+ outputPrice: 12
3498
+ },
3499
+ {
3500
+ contextWindow: Infinity,
3501
+ inputPrice: 4,
3502
+ outputPrice: 18
3503
+ }
3504
+ ]
3505
+ },
3325
3506
  "gemini-2.5-flash-preview-05-20:thinking": {
3326
3507
  maxTokens: 65535,
3327
3508
  contextWindow: 1048576,
3328
3509
  supportsImages: true,
3510
+ supportsNativeTools: true,
3329
3511
  supportsPromptCache: true,
3330
3512
  inputPrice: 0.15,
3331
3513
  outputPrice: 3.5,
@@ -3337,6 +3519,7 @@ var vertexModels = {
3337
3519
  maxTokens: 65535,
3338
3520
  contextWindow: 1048576,
3339
3521
  supportsImages: true,
3522
+ supportsNativeTools: true,
3340
3523
  supportsPromptCache: true,
3341
3524
  inputPrice: 0.15,
3342
3525
  outputPrice: 0.6
@@ -3345,6 +3528,7 @@ var vertexModels = {
3345
3528
  maxTokens: 64e3,
3346
3529
  contextWindow: 1048576,
3347
3530
  supportsImages: true,
3531
+ supportsNativeTools: true,
3348
3532
  supportsPromptCache: true,
3349
3533
  inputPrice: 0.3,
3350
3534
  outputPrice: 2.5,
@@ -3357,6 +3541,7 @@ var vertexModels = {
3357
3541
  maxTokens: 65535,
3358
3542
  contextWindow: 1048576,
3359
3543
  supportsImages: true,
3544
+ supportsNativeTools: true,
3360
3545
  supportsPromptCache: false,
3361
3546
  inputPrice: 0.15,
3362
3547
  outputPrice: 3.5,
@@ -3368,6 +3553,7 @@ var vertexModels = {
3368
3553
  maxTokens: 65535,
3369
3554
  contextWindow: 1048576,
3370
3555
  supportsImages: true,
3556
+ supportsNativeTools: true,
3371
3557
  supportsPromptCache: false,
3372
3558
  inputPrice: 0.15,
3373
3559
  outputPrice: 0.6
@@ -3376,6 +3562,7 @@ var vertexModels = {
3376
3562
  maxTokens: 65535,
3377
3563
  contextWindow: 1048576,
3378
3564
  supportsImages: true,
3565
+ supportsNativeTools: true,
3379
3566
  supportsPromptCache: true,
3380
3567
  inputPrice: 2.5,
3381
3568
  outputPrice: 15
@@ -3384,6 +3571,7 @@ var vertexModels = {
3384
3571
  maxTokens: 65535,
3385
3572
  contextWindow: 1048576,
3386
3573
  supportsImages: true,
3574
+ supportsNativeTools: true,
3387
3575
  supportsPromptCache: true,
3388
3576
  inputPrice: 2.5,
3389
3577
  outputPrice: 15
@@ -3392,6 +3580,7 @@ var vertexModels = {
3392
3580
  maxTokens: 65535,
3393
3581
  contextWindow: 1048576,
3394
3582
  supportsImages: true,
3583
+ supportsNativeTools: true,
3395
3584
  supportsPromptCache: true,
3396
3585
  inputPrice: 2.5,
3397
3586
  outputPrice: 15,
@@ -3402,6 +3591,7 @@ var vertexModels = {
3402
3591
  maxTokens: 64e3,
3403
3592
  contextWindow: 1048576,
3404
3593
  supportsImages: true,
3594
+ supportsNativeTools: true,
3405
3595
  supportsPromptCache: true,
3406
3596
  inputPrice: 2.5,
3407
3597
  outputPrice: 15,
@@ -3427,6 +3617,7 @@ var vertexModels = {
3427
3617
  maxTokens: 65535,
3428
3618
  contextWindow: 1048576,
3429
3619
  supportsImages: true,
3620
+ supportsNativeTools: true,
3430
3621
  supportsPromptCache: false,
3431
3622
  inputPrice: 0,
3432
3623
  outputPrice: 0
@@ -3435,6 +3626,7 @@ var vertexModels = {
3435
3626
  maxTokens: 8192,
3436
3627
  contextWindow: 2097152,
3437
3628
  supportsImages: true,
3629
+ supportsNativeTools: true,
3438
3630
  supportsPromptCache: false,
3439
3631
  inputPrice: 0,
3440
3632
  outputPrice: 0
@@ -3443,6 +3635,7 @@ var vertexModels = {
3443
3635
  maxTokens: 8192,
3444
3636
  contextWindow: 1048576,
3445
3637
  supportsImages: true,
3638
+ supportsNativeTools: true,
3446
3639
  supportsPromptCache: true,
3447
3640
  inputPrice: 0.15,
3448
3641
  outputPrice: 0.6
@@ -3451,6 +3644,7 @@ var vertexModels = {
3451
3644
  maxTokens: 8192,
3452
3645
  contextWindow: 1048576,
3453
3646
  supportsImages: true,
3647
+ supportsNativeTools: true,
3454
3648
  supportsPromptCache: false,
3455
3649
  inputPrice: 0.075,
3456
3650
  outputPrice: 0.3
@@ -3459,6 +3653,7 @@ var vertexModels = {
3459
3653
  maxTokens: 8192,
3460
3654
  contextWindow: 32768,
3461
3655
  supportsImages: true,
3656
+ supportsNativeTools: true,
3462
3657
  supportsPromptCache: false,
3463
3658
  inputPrice: 0,
3464
3659
  outputPrice: 0
@@ -3467,6 +3662,7 @@ var vertexModels = {
3467
3662
  maxTokens: 8192,
3468
3663
  contextWindow: 1048576,
3469
3664
  supportsImages: true,
3665
+ supportsNativeTools: true,
3470
3666
  supportsPromptCache: true,
3471
3667
  inputPrice: 0.075,
3472
3668
  outputPrice: 0.3
@@ -3475,6 +3671,7 @@ var vertexModels = {
3475
3671
  maxTokens: 8192,
3476
3672
  contextWindow: 2097152,
3477
3673
  supportsImages: true,
3674
+ supportsNativeTools: true,
3478
3675
  supportsPromptCache: false,
3479
3676
  inputPrice: 1.25,
3480
3677
  outputPrice: 5
@@ -3512,6 +3709,17 @@ var vertexModels = {
3512
3709
  cacheReadsPrice: 0.1,
3513
3710
  supportsReasoningBudget: true
3514
3711
  },
3712
+ "claude-opus-4-5@20251101": {
3713
+ maxTokens: 8192,
3714
+ contextWindow: 2e5,
3715
+ supportsImages: true,
3716
+ supportsPromptCache: true,
3717
+ inputPrice: 5,
3718
+ outputPrice: 25,
3719
+ cacheWritesPrice: 6.25,
3720
+ cacheReadsPrice: 0.5,
3721
+ supportsReasoningBudget: true
3722
+ },
3515
3723
  "claude-opus-4-1@20250805": {
3516
3724
  maxTokens: 8192,
3517
3725
  contextWindow: 2e5,
@@ -3609,6 +3817,7 @@ var vertexModels = {
3609
3817
  maxTokens: 64e3,
3610
3818
  contextWindow: 1048576,
3611
3819
  supportsImages: true,
3820
+ supportsNativeTools: true,
3612
3821
  supportsPromptCache: true,
3613
3822
  inputPrice: 0.1,
3614
3823
  outputPrice: 0.4,
@@ -3910,17 +4119,67 @@ var xaiModels = {
3910
4119
  contextWindow: 262144,
3911
4120
  supportsImages: false,
3912
4121
  supportsPromptCache: true,
4122
+ supportsNativeTools: true,
3913
4123
  inputPrice: 0.2,
3914
4124
  outputPrice: 1.5,
3915
4125
  cacheWritesPrice: 0.02,
3916
4126
  cacheReadsPrice: 0.02,
3917
4127
  description: "xAI's Grok Code Fast model with 256K context window"
3918
4128
  },
4129
+ "grok-4-1-fast-reasoning": {
4130
+ maxTokens: 65536,
4131
+ contextWindow: 2e6,
4132
+ supportsImages: true,
4133
+ supportsPromptCache: true,
4134
+ supportsNativeTools: true,
4135
+ inputPrice: 0.2,
4136
+ outputPrice: 0.5,
4137
+ cacheWritesPrice: 0.05,
4138
+ cacheReadsPrice: 0.05,
4139
+ description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning"
4140
+ },
4141
+ "grok-4-1-fast-non-reasoning": {
4142
+ maxTokens: 65536,
4143
+ contextWindow: 2e6,
4144
+ supportsImages: true,
4145
+ supportsPromptCache: true,
4146
+ supportsNativeTools: true,
4147
+ inputPrice: 0.2,
4148
+ outputPrice: 0.5,
4149
+ cacheWritesPrice: 0.05,
4150
+ cacheReadsPrice: 0.05,
4151
+ description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling"
4152
+ },
4153
+ "grok-4-fast-reasoning": {
4154
+ maxTokens: 65536,
4155
+ contextWindow: 2e6,
4156
+ supportsImages: true,
4157
+ supportsPromptCache: true,
4158
+ supportsNativeTools: true,
4159
+ inputPrice: 0.2,
4160
+ outputPrice: 0.5,
4161
+ cacheWritesPrice: 0.05,
4162
+ cacheReadsPrice: 0.05,
4163
+ description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning"
4164
+ },
4165
+ "grok-4-fast-non-reasoning": {
4166
+ maxTokens: 65536,
4167
+ contextWindow: 2e6,
4168
+ supportsImages: true,
4169
+ supportsPromptCache: true,
4170
+ supportsNativeTools: true,
4171
+ inputPrice: 0.2,
4172
+ outputPrice: 0.5,
4173
+ cacheWritesPrice: 0.05,
4174
+ cacheReadsPrice: 0.05,
4175
+ description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling"
4176
+ },
3919
4177
  "grok-4": {
3920
4178
  maxTokens: 8192,
3921
4179
  contextWindow: 256e3,
3922
4180
  supportsImages: true,
3923
4181
  supportsPromptCache: true,
4182
+ supportsNativeTools: true,
3924
4183
  inputPrice: 3,
3925
4184
  outputPrice: 15,
3926
4185
  cacheWritesPrice: 0.75,
@@ -3932,6 +4191,7 @@ var xaiModels = {
3932
4191
  contextWindow: 131072,
3933
4192
  supportsImages: false,
3934
4193
  supportsPromptCache: true,
4194
+ supportsNativeTools: true,
3935
4195
  inputPrice: 3,
3936
4196
  outputPrice: 15,
3937
4197
  cacheWritesPrice: 0.75,
@@ -3943,6 +4203,7 @@ var xaiModels = {
3943
4203
  contextWindow: 131072,
3944
4204
  supportsImages: false,
3945
4205
  supportsPromptCache: true,
4206
+ supportsNativeTools: true,
3946
4207
  inputPrice: 5,
3947
4208
  outputPrice: 25,
3948
4209
  cacheWritesPrice: 1.25,
@@ -3954,6 +4215,7 @@ var xaiModels = {
3954
4215
  contextWindow: 131072,
3955
4216
  supportsImages: false,
3956
4217
  supportsPromptCache: true,
4218
+ supportsNativeTools: true,
3957
4219
  inputPrice: 0.3,
3958
4220
  outputPrice: 0.5,
3959
4221
  cacheWritesPrice: 0.07,
@@ -3966,6 +4228,7 @@ var xaiModels = {
3966
4228
  contextWindow: 131072,
3967
4229
  supportsImages: false,
3968
4230
  supportsPromptCache: true,
4231
+ supportsNativeTools: true,
3969
4232
  inputPrice: 0.6,
3970
4233
  outputPrice: 4,
3971
4234
  cacheWritesPrice: 0.15,
@@ -3978,6 +4241,7 @@ var xaiModels = {
3978
4241
  contextWindow: 131072,
3979
4242
  supportsImages: false,
3980
4243
  supportsPromptCache: false,
4244
+ supportsNativeTools: true,
3981
4245
  inputPrice: 2,
3982
4246
  outputPrice: 10,
3983
4247
  description: "xAI's Grok-2 model (version 1212) with 128K context window"
@@ -3987,6 +4251,7 @@ var xaiModels = {
3987
4251
  contextWindow: 32768,
3988
4252
  supportsImages: true,
3989
4253
  supportsPromptCache: false,
4254
+ supportsNativeTools: true,
3990
4255
  inputPrice: 2,
3991
4256
  outputPrice: 10,
3992
4257
  description: "xAI's Grok-2 Vision model (version 1212) with image support and 32K context window"
@@ -4095,6 +4360,7 @@ var internationalZAiModels = {
4095
4360
  contextWindow: 131072,
4096
4361
  supportsImages: false,
4097
4362
  supportsPromptCache: true,
4363
+ supportsNativeTools: true,
4098
4364
  supportsReasoningBinary: true,
4099
4365
  inputPrice: 0.6,
4100
4366
  outputPrice: 2.2,
@@ -4107,6 +4373,7 @@ var internationalZAiModels = {
4107
4373
  contextWindow: 131072,
4108
4374
  supportsImages: false,
4109
4375
  supportsPromptCache: true,
4376
+ supportsNativeTools: true,
4110
4377
  inputPrice: 0.2,
4111
4378
  outputPrice: 1.1,
4112
4379
  cacheWritesPrice: 0,
@@ -4118,6 +4385,7 @@ var internationalZAiModels = {
4118
4385
  contextWindow: 131072,
4119
4386
  supportsImages: false,
4120
4387
  supportsPromptCache: true,
4388
+ supportsNativeTools: true,
4121
4389
  inputPrice: 2.2,
4122
4390
  outputPrice: 8.9,
4123
4391
  cacheWritesPrice: 0,
@@ -4129,6 +4397,7 @@ var internationalZAiModels = {
4129
4397
  contextWindow: 131072,
4130
4398
  supportsImages: false,
4131
4399
  supportsPromptCache: true,
4400
+ supportsNativeTools: true,
4132
4401
  inputPrice: 1.1,
4133
4402
  outputPrice: 4.5,
4134
4403
  cacheWritesPrice: 0,
@@ -4140,6 +4409,7 @@ var internationalZAiModels = {
4140
4409
  contextWindow: 131072,
4141
4410
  supportsImages: false,
4142
4411
  supportsPromptCache: true,
4412
+ supportsNativeTools: true,
4143
4413
  inputPrice: 0,
4144
4414
  outputPrice: 0,
4145
4415
  cacheWritesPrice: 0,
@@ -4151,6 +4421,7 @@ var internationalZAiModels = {
4151
4421
  contextWindow: 131072,
4152
4422
  supportsImages: true,
4153
4423
  supportsPromptCache: true,
4424
+ supportsNativeTools: true,
4154
4425
  inputPrice: 0.6,
4155
4426
  outputPrice: 1.8,
4156
4427
  cacheWritesPrice: 0,
@@ -4162,6 +4433,7 @@ var internationalZAiModels = {
4162
4433
  contextWindow: 2e5,
4163
4434
  supportsImages: false,
4164
4435
  supportsPromptCache: true,
4436
+ supportsNativeTools: true,
4165
4437
  supportsReasoningBinary: true,
4166
4438
  inputPrice: 0.6,
4167
4439
  outputPrice: 2.2,
@@ -4174,6 +4446,7 @@ var internationalZAiModels = {
4174
4446
  contextWindow: 131072,
4175
4447
  supportsImages: false,
4176
4448
  supportsPromptCache: false,
4449
+ supportsNativeTools: true,
4177
4450
  inputPrice: 0.1,
4178
4451
  outputPrice: 0.1,
4179
4452
  cacheWritesPrice: 0,
@@ -4188,6 +4461,7 @@ var mainlandZAiModels = {
4188
4461
  contextWindow: 131072,
4189
4462
  supportsImages: false,
4190
4463
  supportsPromptCache: true,
4464
+ supportsNativeTools: true,
4191
4465
  supportsReasoningBinary: true,
4192
4466
  inputPrice: 0.29,
4193
4467
  outputPrice: 1.14,
@@ -4200,6 +4474,7 @@ var mainlandZAiModels = {
4200
4474
  contextWindow: 131072,
4201
4475
  supportsImages: false,
4202
4476
  supportsPromptCache: true,
4477
+ supportsNativeTools: true,
4203
4478
  inputPrice: 0.1,
4204
4479
  outputPrice: 0.6,
4205
4480
  cacheWritesPrice: 0,
@@ -4211,6 +4486,7 @@ var mainlandZAiModels = {
4211
4486
  contextWindow: 131072,
4212
4487
  supportsImages: false,
4213
4488
  supportsPromptCache: true,
4489
+ supportsNativeTools: true,
4214
4490
  inputPrice: 0.29,
4215
4491
  outputPrice: 1.14,
4216
4492
  cacheWritesPrice: 0,
@@ -4222,6 +4498,7 @@ var mainlandZAiModels = {
4222
4498
  contextWindow: 131072,
4223
4499
  supportsImages: false,
4224
4500
  supportsPromptCache: true,
4501
+ supportsNativeTools: true,
4225
4502
  inputPrice: 0.1,
4226
4503
  outputPrice: 0.6,
4227
4504
  cacheWritesPrice: 0,
@@ -4233,6 +4510,7 @@ var mainlandZAiModels = {
4233
4510
  contextWindow: 131072,
4234
4511
  supportsImages: false,
4235
4512
  supportsPromptCache: true,
4513
+ supportsNativeTools: true,
4236
4514
  inputPrice: 0,
4237
4515
  outputPrice: 0,
4238
4516
  cacheWritesPrice: 0,
@@ -4244,6 +4522,7 @@ var mainlandZAiModels = {
4244
4522
  contextWindow: 131072,
4245
4523
  supportsImages: true,
4246
4524
  supportsPromptCache: true,
4525
+ supportsNativeTools: true,
4247
4526
  inputPrice: 0.29,
4248
4527
  outputPrice: 0.93,
4249
4528
  cacheWritesPrice: 0,
@@ -4255,6 +4534,7 @@ var mainlandZAiModels = {
4255
4534
  contextWindow: 204800,
4256
4535
  supportsImages: false,
4257
4536
  supportsPromptCache: true,
4537
+ supportsNativeTools: true,
4258
4538
  supportsReasoningBinary: true,
4259
4539
  inputPrice: 0.29,
4260
4540
  outputPrice: 1.14,
@@ -4284,6 +4564,7 @@ var deepInfraDefaultModelInfo = {
4284
4564
  contextWindow: 262144,
4285
4565
  supportsImages: false,
4286
4566
  supportsPromptCache: false,
4567
+ supportsNativeTools: true,
4287
4568
  inputPrice: 0.3,
4288
4569
  outputPrice: 1.2,
4289
4570
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -4297,11 +4578,12 @@ var minimaxModels = {
4297
4578
  contextWindow: 192e3,
4298
4579
  supportsImages: false,
4299
4580
  supportsPromptCache: true,
4581
+ supportsNativeTools: true,
4582
+ preserveReasoning: true,
4300
4583
  inputPrice: 0.3,
4301
4584
  outputPrice: 1.2,
4302
4585
  cacheWritesPrice: 0.375,
4303
4586
  cacheReadsPrice: 0.03,
4304
- preserveReasoning: true,
4305
4587
  description: "MiniMax M2, a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4306
4588
  },
4307
4589
  "MiniMax-M2-Stable": {
@@ -4309,14 +4591,17 @@ var minimaxModels = {
4309
4591
  contextWindow: 192e3,
4310
4592
  supportsImages: false,
4311
4593
  supportsPromptCache: true,
4594
+ supportsNativeTools: true,
4595
+ preserveReasoning: true,
4312
4596
  inputPrice: 0.3,
4313
4597
  outputPrice: 1.2,
4314
4598
  cacheWritesPrice: 0.375,
4315
4599
  cacheReadsPrice: 0.03,
4316
- preserveReasoning: true,
4317
4600
  description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4318
4601
  }
4319
4602
  };
4603
+ var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
4604
+ var MINIMAX_DEFAULT_MAX_TOKENS = 16384;
4320
4605
  var MINIMAX_DEFAULT_TEMPERATURE = 1;
4321
4606
 
4322
4607
  // src/providers/index.ts
@@ -4340,6 +4625,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
4340
4625
  return "meta-llama/Llama-3.3-70B-Instruct";
4341
4626
  case "chutes":
4342
4627
  return chutesDefaultModelId;
4628
+ case "baseten":
4629
+ return basetenDefaultModelId;
4343
4630
  case "bedrock":
4344
4631
  return bedrockDefaultModelId;
4345
4632
  case "vertex":
@@ -4433,6 +4720,7 @@ var providerNames = [
4433
4720
  ...fauxProviders,
4434
4721
  "anthropic",
4435
4722
  "bedrock",
4723
+ "baseten",
4436
4724
  "cerebras",
4437
4725
  "claude-code",
4438
4726
  "doubao",
@@ -4672,6 +4960,9 @@ var vercelAiGatewaySchema = baseProviderSettingsSchema.extend({
4672
4960
  vercelAiGatewayApiKey: z8.string().optional(),
4673
4961
  vercelAiGatewayModelId: z8.string().optional()
4674
4962
  });
4963
+ var basetenSchema = apiModelIdProviderModelSchema.extend({
4964
+ basetenApiKey: z8.string().optional()
4965
+ });
4675
4966
  var defaultSchema = z8.object({
4676
4967
  apiProvider: z8.undefined()
4677
4968
  });
@@ -4701,6 +4992,7 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
4701
4992
  fakeAiSchema.merge(z8.object({ apiProvider: z8.literal("fake-ai") })),
4702
4993
  xaiSchema.merge(z8.object({ apiProvider: z8.literal("xai") })),
4703
4994
  groqSchema.merge(z8.object({ apiProvider: z8.literal("groq") })),
4995
+ basetenSchema.merge(z8.object({ apiProvider: z8.literal("baseten") })),
4704
4996
  huggingFaceSchema.merge(z8.object({ apiProvider: z8.literal("huggingface") })),
4705
4997
  chutesSchema.merge(z8.object({ apiProvider: z8.literal("chutes") })),
4706
4998
  litellmSchema.merge(z8.object({ apiProvider: z8.literal("litellm") })),
@@ -4742,6 +5034,7 @@ var providerSettingsSchema = z8.object({
4742
5034
  ...fakeAiSchema.shape,
4743
5035
  ...xaiSchema.shape,
4744
5036
  ...groqSchema.shape,
5037
+ ...basetenSchema.shape,
4745
5038
  ...huggingFaceSchema.shape,
4746
5039
  ...chutesSchema.shape,
4747
5040
  ...litellmSchema.shape,
@@ -4805,6 +5098,7 @@ var modelIdKeysByProvider = {
4805
5098
  requesty: "requestyModelId",
4806
5099
  xai: "apiModelId",
4807
5100
  groq: "apiModelId",
5101
+ baseten: "apiModelId",
4808
5102
  chutes: "apiModelId",
4809
5103
  litellm: "litellmModelId",
4810
5104
  huggingface: "huggingFaceModelId",
@@ -4817,7 +5111,7 @@ var modelIdKeysByProvider = {
4817
5111
  roo: "apiModelId",
4818
5112
  "vercel-ai-gateway": "vercelAiGatewayModelId"
4819
5113
  };
4820
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock"];
5114
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
4821
5115
  var getApiProtocol = (provider, modelId) => {
4822
5116
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
4823
5117
  return "anthropic";
@@ -4912,7 +5206,8 @@ var MODELS_BY_PROVIDER = {
4912
5206
  models: Object.keys(vscodeLlmModels)
4913
5207
  },
4914
5208
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
4915
- zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
5209
+ zai: { id: "zai", label: "Z.ai", models: Object.keys(internationalZAiModels) },
5210
+ baseten: { id: "baseten", label: "Baseten", models: Object.keys(basetenModels) },
4916
5211
  // Dynamic providers; models pulled from remote APIs.
4917
5212
  glama: { id: "glama", label: "Glama", models: [] },
4918
5213
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
@@ -4944,7 +5239,18 @@ var historyItemSchema = z9.object({
4944
5239
  totalCost: z9.number(),
4945
5240
  size: z9.number().optional(),
4946
5241
  workspace: z9.string().optional(),
4947
- mode: z9.string().optional()
5242
+ mode: z9.string().optional(),
5243
+ status: z9.enum(["active", "completed", "delegated"]).optional(),
5244
+ delegatedToId: z9.string().optional(),
5245
+ // Last child this parent delegated to
5246
+ childIds: z9.array(z9.string()).optional(),
5247
+ // All children spawned by this task
5248
+ awaitingChildId: z9.string().optional(),
5249
+ // Child currently awaited (set when delegated)
5250
+ completedByChildId: z9.string().optional(),
5251
+ // Child that completed and resumed this parent
5252
+ completionResultSummary: z9.string().optional()
5253
+ // Summary from completed child
4948
5254
  });
4949
5255
 
4950
5256
  // src/experiment.ts
@@ -4955,7 +5261,7 @@ var experimentIds = [
4955
5261
  "preventFocusDisruption",
4956
5262
  "imageGeneration",
4957
5263
  "runSlashCommand",
4958
- "nativeToolCalling"
5264
+ "multipleNativeToolCalls"
4959
5265
  ];
4960
5266
  var experimentIdsSchema = z10.enum(experimentIds);
4961
5267
  var experimentsSchema = z10.object({
@@ -4964,7 +5270,7 @@ var experimentsSchema = z10.object({
4964
5270
  preventFocusDisruption: z10.boolean().optional(),
4965
5271
  imageGeneration: z10.boolean().optional(),
4966
5272
  runSlashCommand: z10.boolean().optional(),
4967
- nativeToolCalling: z10.boolean().optional()
5273
+ multipleNativeToolCalls: z10.boolean().optional()
4968
5274
  });
4969
5275
 
4970
5276
  // src/telemetry.ts
@@ -5014,6 +5320,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
5014
5320
  TelemetryEventName2["CONSECUTIVE_MISTAKE_ERROR"] = "Consecutive Mistake Error";
5015
5321
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
5016
5322
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
5323
+ TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5017
5324
  return TelemetryEventName2;
5018
5325
  })(TelemetryEventName || {});
5019
5326
  var staticAppPropertiesSchema = z11.object({
@@ -5097,6 +5404,7 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
5097
5404
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
5098
5405
  "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */,
5099
5406
  "Code Index Error" /* CODE_INDEX_ERROR */,
5407
+ "Model Cache Empty Response" /* MODEL_CACHE_EMPTY_RESPONSE */,
5100
5408
  "Context Condensed" /* CONTEXT_CONDENSED */,
5101
5409
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
5102
5410
  "Tab Shown" /* TAB_SHOWN */,
@@ -5257,8 +5565,6 @@ var terminalActionIds = ["terminalAddToContext", "terminalFixCommand", "terminal
5257
5565
  var commandIds = [
5258
5566
  "activationCompleted",
5259
5567
  "plusButtonClicked",
5260
- "promptsButtonClicked",
5261
- "mcpButtonClicked",
5262
5568
  "historyButtonClicked",
5263
5569
  "marketplaceButtonClicked",
5264
5570
  "popoutButtonClicked",
@@ -5315,6 +5621,7 @@ var globalSettingsSchema = z14.object({
5315
5621
  taskHistory: z14.array(historyItemSchema).optional(),
5316
5622
  dismissedUpsells: z14.array(z14.string()).optional(),
5317
5623
  // Image generation settings (experimental) - flattened for simplicity
5624
+ imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
5318
5625
  openRouterImageApiKey: z14.string().optional(),
5319
5626
  openRouterImageGenerationSelectedModel: z14.string().optional(),
5320
5627
  condensingApiConfigId: z14.string().optional(),
@@ -5466,7 +5773,8 @@ var SECRET_STATE_KEYS = [
5466
5773
  "fireworksApiKey",
5467
5774
  "featherlessApiKey",
5468
5775
  "ioIntelligenceApiKey",
5469
- "vercelAiGatewayApiKey"
5776
+ "vercelAiGatewayApiKey",
5777
+ "basetenApiKey"
5470
5778
  ];
5471
5779
  var GLOBAL_SECRET_KEYS = [
5472
5780
  "openRouterImageApiKey"
@@ -5731,6 +6039,9 @@ var ExtensionBridgeEventName = ((ExtensionBridgeEventName2) => {
5731
6039
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskPaused"] = "taskPaused" /* TaskPaused */] = "TaskPaused";
5732
6040
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskUnpaused"] = "taskUnpaused" /* TaskUnpaused */] = "TaskUnpaused";
5733
6041
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskSpawned"] = "taskSpawned" /* TaskSpawned */] = "TaskSpawned";
6042
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegated"] = "taskDelegated" /* TaskDelegated */] = "TaskDelegated";
6043
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegationCompleted"] = "taskDelegationCompleted" /* TaskDelegationCompleted */] = "TaskDelegationCompleted";
6044
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegationResumed"] = "taskDelegationResumed" /* TaskDelegationResumed */] = "TaskDelegationResumed";
5734
6045
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskUserMessage"] = "taskUserMessage" /* TaskUserMessage */] = "TaskUserMessage";
5735
6046
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated" /* TaskTokenUsageUpdated */] = "TaskTokenUsageUpdated";
5736
6047
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["ModeChanged"] = "modeChanged" /* ModeChanged */] = "ModeChanged";
@@ -5806,6 +6117,21 @@ var extensionBridgeEventSchema = z16.discriminatedUnion("type", [
5806
6117
  instance: extensionInstanceSchema,
5807
6118
  timestamp: z16.number()
5808
6119
  }),
6120
+ z16.object({
6121
+ type: z16.literal(ExtensionBridgeEventName.TaskDelegated),
6122
+ instance: extensionInstanceSchema,
6123
+ timestamp: z16.number()
6124
+ }),
6125
+ z16.object({
6126
+ type: z16.literal(ExtensionBridgeEventName.TaskDelegationCompleted),
6127
+ instance: extensionInstanceSchema,
6128
+ timestamp: z16.number()
6129
+ }),
6130
+ z16.object({
6131
+ type: z16.literal(ExtensionBridgeEventName.TaskDelegationResumed),
6132
+ instance: extensionInstanceSchema,
6133
+ timestamp: z16.number()
6134
+ }),
5809
6135
  z16.object({
5810
6136
  type: z16.literal(ExtensionBridgeEventName.TaskUserMessage),
5811
6137
  instance: extensionInstanceSchema,
@@ -5994,6 +6320,30 @@ var followUpDataSchema = z17.object({
5994
6320
  suggest: z17.array(suggestionItemSchema).optional()
5995
6321
  });
5996
6322
 
6323
+ // src/image-generation.ts
6324
+ var IMAGE_GENERATION_MODELS = [
6325
+ // OpenRouter models
6326
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image", provider: "openrouter" },
6327
+ { value: "google/gemini-3-pro-image-preview", label: "Gemini 3 Pro Image Preview", provider: "openrouter" },
6328
+ { value: "openai/gpt-5-image", label: "GPT-5 Image", provider: "openrouter" },
6329
+ { value: "openai/gpt-5-image-mini", label: "GPT-5 Image Mini", provider: "openrouter" },
6330
+ { value: "black-forest-labs/flux.2-flex", label: "Black Forest Labs FLUX.2 Flex", provider: "openrouter" },
6331
+ { value: "black-forest-labs/flux.2-pro", label: "Black Forest Labs FLUX.2 Pro", provider: "openrouter" },
6332
+ // Roo Code Cloud models
6333
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image", provider: "roo" },
6334
+ { value: "google/gemini-3-pro-image", label: "Gemini 3 Pro Image", provider: "roo" },
6335
+ {
6336
+ value: "bfl/flux-2-pro:free",
6337
+ label: "Black Forest Labs FLUX.2 Pro (Free)",
6338
+ provider: "roo",
6339
+ apiMethod: "images_api"
6340
+ }
6341
+ ];
6342
+ var IMAGE_GENERATION_MODEL_IDS = IMAGE_GENERATION_MODELS.map((m) => m.value);
6343
+ function getImageGenerationProvider(explicitProvider, hasExistingModel) {
6344
+ return explicitProvider !== void 0 ? explicitProvider : hasExistingModel ? "openrouter" : "roo";
6345
+ }
6346
+
5997
6347
  // src/ipc.ts
5998
6348
  import { z as z18 } from "zod";
5999
6349
  var IpcMessageType = /* @__PURE__ */ ((IpcMessageType2) => {
@@ -6181,12 +6531,15 @@ export {
6181
6531
  HUGGINGFACE_SLIDER_MIN,
6182
6532
  HUGGINGFACE_SLIDER_STEP,
6183
6533
  HUGGINGFACE_TEMPERATURE_MAX_VALUE,
6534
+ IMAGE_GENERATION_MODELS,
6535
+ IMAGE_GENERATION_MODEL_IDS,
6184
6536
  INSTANCE_TTL_SECONDS,
6185
6537
  IO_INTELLIGENCE_CACHE_DURATION,
6186
6538
  IpcMessageType,
6187
6539
  IpcOrigin,
6188
6540
  LMSTUDIO_DEFAULT_TEMPERATURE,
6189
6541
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6542
+ MINIMAX_DEFAULT_MAX_TOKENS,
6190
6543
  MINIMAX_DEFAULT_TEMPERATURE,
6191
6544
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
6192
6545
  MISTRAL_DEFAULT_TEMPERATURE,
@@ -6224,6 +6577,8 @@ export {
6224
6577
  anthropicModels,
6225
6578
  appPropertiesSchema,
6226
6579
  azureOpenAiDefaultApiVersion,
6580
+ basetenDefaultModelId,
6581
+ basetenModels,
6227
6582
  bedrockDefaultModelId,
6228
6583
  bedrockDefaultPromptRouterModelId,
6229
6584
  bedrockModels,
@@ -6279,6 +6634,7 @@ export {
6279
6634
  getApiProtocol,
6280
6635
  getClaudeCodeModelId,
6281
6636
  getEffectiveProtocol,
6637
+ getImageGenerationProvider,
6282
6638
  getModelId,
6283
6639
  getProviderDefaultModelId,
6284
6640
  gitPropertiesSchema,
@@ -6332,6 +6688,7 @@ export {
6332
6688
  mcpMarketplaceItemSchema,
6333
6689
  mcpParameterSchema,
6334
6690
  minimaxDefaultModelId,
6691
+ minimaxDefaultModelInfo,
6335
6692
  minimaxModels,
6336
6693
  mistralDefaultModelId,
6337
6694
  mistralModels,
@@ -6381,7 +6738,6 @@ export {
6381
6738
  rooCodeSettingsSchema,
6382
6739
  rooCodeTelemetryEventSchema,
6383
6740
  rooDefaultModelId,
6384
- rooModelDefaults,
6385
6741
  rooModels,
6386
6742
  sambaNovaDefaultModelId,
6387
6743
  sambaNovaModels,