@roo-code/types 1.86.0 → 1.88.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -60,12 +60,15 @@ __export(index_exports, {
60
60
  HUGGINGFACE_SLIDER_MIN: () => HUGGINGFACE_SLIDER_MIN,
61
61
  HUGGINGFACE_SLIDER_STEP: () => HUGGINGFACE_SLIDER_STEP,
62
62
  HUGGINGFACE_TEMPERATURE_MAX_VALUE: () => HUGGINGFACE_TEMPERATURE_MAX_VALUE,
63
+ IMAGE_GENERATION_MODELS: () => IMAGE_GENERATION_MODELS,
64
+ IMAGE_GENERATION_MODEL_IDS: () => IMAGE_GENERATION_MODEL_IDS,
63
65
  INSTANCE_TTL_SECONDS: () => INSTANCE_TTL_SECONDS,
64
66
  IO_INTELLIGENCE_CACHE_DURATION: () => IO_INTELLIGENCE_CACHE_DURATION,
65
67
  IpcMessageType: () => IpcMessageType,
66
68
  IpcOrigin: () => IpcOrigin,
67
69
  LMSTUDIO_DEFAULT_TEMPERATURE: () => LMSTUDIO_DEFAULT_TEMPERATURE,
68
70
  MAX_CHECKPOINT_TIMEOUT_SECONDS: () => MAX_CHECKPOINT_TIMEOUT_SECONDS,
71
+ MINIMAX_DEFAULT_MAX_TOKENS: () => MINIMAX_DEFAULT_MAX_TOKENS,
69
72
  MINIMAX_DEFAULT_TEMPERATURE: () => MINIMAX_DEFAULT_TEMPERATURE,
70
73
  MIN_CHECKPOINT_TIMEOUT_SECONDS: () => MIN_CHECKPOINT_TIMEOUT_SECONDS,
71
74
  MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
@@ -103,6 +106,8 @@ __export(index_exports, {
103
106
  anthropicModels: () => anthropicModels,
104
107
  appPropertiesSchema: () => appPropertiesSchema,
105
108
  azureOpenAiDefaultApiVersion: () => azureOpenAiDefaultApiVersion,
109
+ basetenDefaultModelId: () => basetenDefaultModelId,
110
+ basetenModels: () => basetenModels,
106
111
  bedrockDefaultModelId: () => bedrockDefaultModelId,
107
112
  bedrockDefaultPromptRouterModelId: () => bedrockDefaultPromptRouterModelId,
108
113
  bedrockModels: () => bedrockModels,
@@ -158,6 +163,7 @@ __export(index_exports, {
158
163
  getApiProtocol: () => getApiProtocol,
159
164
  getClaudeCodeModelId: () => getClaudeCodeModelId,
160
165
  getEffectiveProtocol: () => getEffectiveProtocol,
166
+ getImageGenerationProvider: () => getImageGenerationProvider,
161
167
  getModelId: () => getModelId,
162
168
  getProviderDefaultModelId: () => getProviderDefaultModelId,
163
169
  gitPropertiesSchema: () => gitPropertiesSchema,
@@ -211,6 +217,7 @@ __export(index_exports, {
211
217
  mcpMarketplaceItemSchema: () => mcpMarketplaceItemSchema,
212
218
  mcpParameterSchema: () => mcpParameterSchema,
213
219
  minimaxDefaultModelId: () => minimaxDefaultModelId,
220
+ minimaxDefaultModelInfo: () => minimaxDefaultModelInfo,
214
221
  minimaxModels: () => minimaxModels,
215
222
  mistralDefaultModelId: () => mistralDefaultModelId,
216
223
  mistralModels: () => mistralModels,
@@ -260,7 +267,6 @@ __export(index_exports, {
260
267
  rooCodeSettingsSchema: () => rooCodeSettingsSchema,
261
268
  rooCodeTelemetryEventSchema: () => rooCodeTelemetryEventSchema,
262
269
  rooDefaultModelId: () => rooDefaultModelId,
263
- rooModelDefaults: () => rooModelDefaults,
264
270
  rooModels: () => rooModels,
265
271
  sambaNovaDefaultModelId: () => sambaNovaDefaultModelId,
266
272
  sambaNovaModels: () => sambaNovaModels,
@@ -378,6 +384,7 @@ var clineSays = [
378
384
  "shell_integration_warning",
379
385
  "browser_action",
380
386
  "browser_action_result",
387
+ "browser_session_status",
381
388
  "mcp_server_request_started",
382
389
  "mcp_server_response",
383
390
  "subtask_result",
@@ -442,6 +449,8 @@ var toolNames = [
442
449
  "write_to_file",
443
450
  "apply_diff",
444
451
  "insert_content",
452
+ "search_and_replace",
453
+ "apply_patch",
445
454
  "search_files",
446
455
  "list_files",
447
456
  "list_code_definition_names",
@@ -492,6 +501,9 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
492
501
  RooCodeEventName2["TaskPaused"] = "taskPaused";
493
502
  RooCodeEventName2["TaskUnpaused"] = "taskUnpaused";
494
503
  RooCodeEventName2["TaskSpawned"] = "taskSpawned";
504
+ RooCodeEventName2["TaskDelegated"] = "taskDelegated";
505
+ RooCodeEventName2["TaskDelegationCompleted"] = "taskDelegationCompleted";
506
+ RooCodeEventName2["TaskDelegationResumed"] = "taskDelegationResumed";
495
507
  RooCodeEventName2["Message"] = "message";
496
508
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
497
509
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
@@ -525,6 +537,26 @@ var rooCodeEventsSchema = import_zod3.z.object({
525
537
  ["taskPaused" /* TaskPaused */]: import_zod3.z.tuple([import_zod3.z.string()]),
526
538
  ["taskUnpaused" /* TaskUnpaused */]: import_zod3.z.tuple([import_zod3.z.string()]),
527
539
  ["taskSpawned" /* TaskSpawned */]: import_zod3.z.tuple([import_zod3.z.string(), import_zod3.z.string()]),
540
+ ["taskDelegated" /* TaskDelegated */]: import_zod3.z.tuple([
541
+ import_zod3.z.string(),
542
+ // parentTaskId
543
+ import_zod3.z.string()
544
+ // childTaskId
545
+ ]),
546
+ ["taskDelegationCompleted" /* TaskDelegationCompleted */]: import_zod3.z.tuple([
547
+ import_zod3.z.string(),
548
+ // parentTaskId
549
+ import_zod3.z.string(),
550
+ // childTaskId
551
+ import_zod3.z.string()
552
+ // completionResultSummary
553
+ ]),
554
+ ["taskDelegationResumed" /* TaskDelegationResumed */]: import_zod3.z.tuple([
555
+ import_zod3.z.string(),
556
+ // parentTaskId
557
+ import_zod3.z.string()
558
+ // childTaskId
559
+ ]),
528
560
  ["message" /* Message */]: import_zod3.z.tuple([
529
561
  import_zod3.z.object({
530
562
  taskId: import_zod3.z.string(),
@@ -609,6 +641,21 @@ var taskEventSchema = import_zod3.z.discriminatedUnion("eventName", [
609
641
  payload: rooCodeEventsSchema.shape["taskSpawned" /* TaskSpawned */],
610
642
  taskId: import_zod3.z.number().optional()
611
643
  }),
644
+ import_zod3.z.object({
645
+ eventName: import_zod3.z.literal("taskDelegated" /* TaskDelegated */),
646
+ payload: rooCodeEventsSchema.shape["taskDelegated" /* TaskDelegated */],
647
+ taskId: import_zod3.z.number().optional()
648
+ }),
649
+ import_zod3.z.object({
650
+ eventName: import_zod3.z.literal("taskDelegationCompleted" /* TaskDelegationCompleted */),
651
+ payload: rooCodeEventsSchema.shape["taskDelegationCompleted" /* TaskDelegationCompleted */],
652
+ taskId: import_zod3.z.number().optional()
653
+ }),
654
+ import_zod3.z.object({
655
+ eventName: import_zod3.z.literal("taskDelegationResumed" /* TaskDelegationResumed */),
656
+ payload: rooCodeEventsSchema.shape["taskDelegationResumed" /* TaskDelegationResumed */],
657
+ taskId: import_zod3.z.number().optional()
658
+ }),
612
659
  // Task Execution
613
660
  import_zod3.z.object({
614
661
  eventName: import_zod3.z.literal("message" /* Message */),
@@ -727,6 +774,13 @@ var modelInfoSchema = import_zod5.z.object({
727
774
  supportsNativeTools: import_zod5.z.boolean().optional(),
728
775
  // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
729
776
  defaultToolProtocol: import_zod5.z.enum(["xml", "native"]).optional(),
777
+ // Exclude specific native tools from being available (only applies to native protocol)
778
+ // These tools will be removed from the set of tools available to the model
779
+ excludedTools: import_zod5.z.array(import_zod5.z.string()).optional(),
780
+ // Include specific native tools (only applies to native protocol)
781
+ // These tools will be added if they belong to an allowed group in the current mode
782
+ // Cannot force-add tools from groups the mode doesn't allow
783
+ includedTools: import_zod5.z.array(import_zod5.z.string()).optional(),
730
784
  /**
731
785
  * Service tiers with pricing information.
732
786
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -760,7 +814,16 @@ var CODEBASE_INDEX_DEFAULTS = {
760
814
  var codebaseIndexConfigSchema = import_zod6.z.object({
761
815
  codebaseIndexEnabled: import_zod6.z.boolean().optional(),
762
816
  codebaseIndexQdrantUrl: import_zod6.z.string().optional(),
763
- codebaseIndexEmbedderProvider: import_zod6.z.enum(["openai", "ollama", "openai-compatible", "gemini", "mistral", "vercel-ai-gateway", "openrouter"]).optional(),
817
+ codebaseIndexEmbedderProvider: import_zod6.z.enum([
818
+ "openai",
819
+ "ollama",
820
+ "openai-compatible",
821
+ "gemini",
822
+ "mistral",
823
+ "vercel-ai-gateway",
824
+ "bedrock",
825
+ "openrouter"
826
+ ]).optional(),
764
827
  codebaseIndexEmbedderBaseUrl: import_zod6.z.string().optional(),
765
828
  codebaseIndexEmbedderModelId: import_zod6.z.string().optional(),
766
829
  codebaseIndexEmbedderModelDimension: import_zod6.z.number().optional(),
@@ -768,7 +831,10 @@ var codebaseIndexConfigSchema = import_zod6.z.object({
768
831
  codebaseIndexSearchMaxResults: import_zod6.z.number().min(CODEBASE_INDEX_DEFAULTS.MIN_SEARCH_RESULTS).max(CODEBASE_INDEX_DEFAULTS.MAX_SEARCH_RESULTS).optional(),
769
832
  // OpenAI Compatible specific fields
770
833
  codebaseIndexOpenAiCompatibleBaseUrl: import_zod6.z.string().optional(),
771
- codebaseIndexOpenAiCompatibleModelDimension: import_zod6.z.number().optional()
834
+ codebaseIndexOpenAiCompatibleModelDimension: import_zod6.z.number().optional(),
835
+ // Bedrock specific fields
836
+ codebaseIndexBedrockRegion: import_zod6.z.string().optional(),
837
+ codebaseIndexBedrockProfile: import_zod6.z.string().optional()
772
838
  });
773
839
  var codebaseIndexModelsSchema = import_zod6.z.object({
774
840
  openai: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
@@ -777,7 +843,8 @@ var codebaseIndexModelsSchema = import_zod6.z.object({
777
843
  gemini: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
778
844
  mistral: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
779
845
  "vercel-ai-gateway": import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
780
- openrouter: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional()
846
+ openrouter: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
847
+ bedrock: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional()
781
848
  });
782
849
  var codebaseIndexProviderSchema = import_zod6.z.object({
783
850
  codeIndexOpenAiKey: import_zod6.z.string().optional(),
@@ -801,6 +868,7 @@ var anthropicModels = {
801
868
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
802
869
  supportsImages: true,
803
870
  supportsPromptCache: true,
871
+ supportsNativeTools: true,
804
872
  inputPrice: 3,
805
873
  // $3 per million input tokens (≤200K context)
806
874
  outputPrice: 15,
@@ -833,6 +901,7 @@ var anthropicModels = {
833
901
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
834
902
  supportsImages: true,
835
903
  supportsPromptCache: true,
904
+ supportsNativeTools: true,
836
905
  inputPrice: 3,
837
906
  // $3 per million input tokens (≤200K context)
838
907
  outputPrice: 15,
@@ -858,12 +927,30 @@ var anthropicModels = {
858
927
  }
859
928
  ]
860
929
  },
930
+ "claude-opus-4-5-20251101": {
931
+ maxTokens: 32e3,
932
+ // Overridden to 8k if `enableReasoningEffort` is false.
933
+ contextWindow: 2e5,
934
+ supportsImages: true,
935
+ supportsPromptCache: true,
936
+ supportsNativeTools: true,
937
+ inputPrice: 5,
938
+ // $5 per million input tokens
939
+ outputPrice: 25,
940
+ // $25 per million output tokens
941
+ cacheWritesPrice: 6.25,
942
+ // $6.25 per million tokens
943
+ cacheReadsPrice: 0.5,
944
+ // $0.50 per million tokens
945
+ supportsReasoningBudget: true
946
+ },
861
947
  "claude-opus-4-1-20250805": {
862
948
  maxTokens: 32e3,
863
949
  // Overridden to 8k if `enableReasoningEffort` is false.
864
950
  contextWindow: 2e5,
865
951
  supportsImages: true,
866
952
  supportsPromptCache: true,
953
+ supportsNativeTools: true,
867
954
  inputPrice: 15,
868
955
  // $15 per million input tokens
869
956
  outputPrice: 75,
@@ -880,6 +967,7 @@ var anthropicModels = {
880
967
  contextWindow: 2e5,
881
968
  supportsImages: true,
882
969
  supportsPromptCache: true,
970
+ supportsNativeTools: true,
883
971
  inputPrice: 15,
884
972
  // $15 per million input tokens
885
973
  outputPrice: 75,
@@ -896,6 +984,7 @@ var anthropicModels = {
896
984
  contextWindow: 2e5,
897
985
  supportsImages: true,
898
986
  supportsPromptCache: true,
987
+ supportsNativeTools: true,
899
988
  inputPrice: 3,
900
989
  // $3 per million input tokens
901
990
  outputPrice: 15,
@@ -913,6 +1002,7 @@ var anthropicModels = {
913
1002
  contextWindow: 2e5,
914
1003
  supportsImages: true,
915
1004
  supportsPromptCache: true,
1005
+ supportsNativeTools: true,
916
1006
  inputPrice: 3,
917
1007
  // $3 per million input tokens
918
1008
  outputPrice: 15,
@@ -927,6 +1017,7 @@ var anthropicModels = {
927
1017
  contextWindow: 2e5,
928
1018
  supportsImages: true,
929
1019
  supportsPromptCache: true,
1020
+ supportsNativeTools: true,
930
1021
  inputPrice: 3,
931
1022
  // $3 per million input tokens
932
1023
  outputPrice: 15,
@@ -941,6 +1032,7 @@ var anthropicModels = {
941
1032
  contextWindow: 2e5,
942
1033
  supportsImages: false,
943
1034
  supportsPromptCache: true,
1035
+ supportsNativeTools: true,
944
1036
  inputPrice: 1,
945
1037
  outputPrice: 5,
946
1038
  cacheWritesPrice: 1.25,
@@ -951,6 +1043,7 @@ var anthropicModels = {
951
1043
  contextWindow: 2e5,
952
1044
  supportsImages: true,
953
1045
  supportsPromptCache: true,
1046
+ supportsNativeTools: true,
954
1047
  inputPrice: 15,
955
1048
  outputPrice: 75,
956
1049
  cacheWritesPrice: 18.75,
@@ -961,6 +1054,7 @@ var anthropicModels = {
961
1054
  contextWindow: 2e5,
962
1055
  supportsImages: true,
963
1056
  supportsPromptCache: true,
1057
+ supportsNativeTools: true,
964
1058
  inputPrice: 0.25,
965
1059
  outputPrice: 1.25,
966
1060
  cacheWritesPrice: 0.3,
@@ -971,6 +1065,7 @@ var anthropicModels = {
971
1065
  contextWindow: 2e5,
972
1066
  supportsImages: true,
973
1067
  supportsPromptCache: true,
1068
+ supportsNativeTools: true,
974
1069
  inputPrice: 1,
975
1070
  outputPrice: 5,
976
1071
  cacheWritesPrice: 1.25,
@@ -981,6 +1076,125 @@ var anthropicModels = {
981
1076
  };
982
1077
  var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
983
1078
 
1079
+ // src/providers/baseten.ts
1080
+ var basetenModels = {
1081
+ "moonshotai/Kimi-K2-Thinking": {
1082
+ maxTokens: 163800,
1083
+ contextWindow: 262e3,
1084
+ supportsImages: false,
1085
+ supportsPromptCache: false,
1086
+ supportsNativeTools: true,
1087
+ inputPrice: 0.6,
1088
+ outputPrice: 2.5,
1089
+ cacheWritesPrice: 0,
1090
+ cacheReadsPrice: 0,
1091
+ description: "Kimi K2 Thinking - A model with enhanced reasoning capabilities from Kimi K2"
1092
+ },
1093
+ "zai-org/GLM-4.6": {
1094
+ maxTokens: 2e5,
1095
+ contextWindow: 2e5,
1096
+ supportsImages: false,
1097
+ supportsPromptCache: false,
1098
+ supportsNativeTools: true,
1099
+ inputPrice: 0.6,
1100
+ outputPrice: 2.2,
1101
+ cacheWritesPrice: 0,
1102
+ cacheReadsPrice: 0,
1103
+ description: "Frontier open model with advanced agentic, reasoning and coding capabilities"
1104
+ },
1105
+ "deepseek-ai/DeepSeek-R1": {
1106
+ maxTokens: 131072,
1107
+ contextWindow: 163840,
1108
+ supportsImages: false,
1109
+ supportsPromptCache: false,
1110
+ inputPrice: 2.55,
1111
+ outputPrice: 5.95,
1112
+ cacheWritesPrice: 0,
1113
+ cacheReadsPrice: 0,
1114
+ description: "DeepSeek's first-generation reasoning model"
1115
+ },
1116
+ "deepseek-ai/DeepSeek-R1-0528": {
1117
+ maxTokens: 131072,
1118
+ contextWindow: 163840,
1119
+ supportsImages: false,
1120
+ supportsPromptCache: false,
1121
+ inputPrice: 2.55,
1122
+ outputPrice: 5.95,
1123
+ cacheWritesPrice: 0,
1124
+ cacheReadsPrice: 0,
1125
+ description: "The latest revision of DeepSeek's first-generation reasoning model"
1126
+ },
1127
+ "deepseek-ai/DeepSeek-V3-0324": {
1128
+ maxTokens: 131072,
1129
+ contextWindow: 163840,
1130
+ supportsImages: false,
1131
+ supportsPromptCache: false,
1132
+ inputPrice: 0.77,
1133
+ outputPrice: 0.77,
1134
+ cacheWritesPrice: 0,
1135
+ cacheReadsPrice: 0,
1136
+ description: "Fast general-purpose LLM with enhanced reasoning capabilities"
1137
+ },
1138
+ "deepseek-ai/DeepSeek-V3.1": {
1139
+ maxTokens: 131072,
1140
+ contextWindow: 163840,
1141
+ supportsImages: false,
1142
+ supportsPromptCache: false,
1143
+ inputPrice: 0.5,
1144
+ outputPrice: 1.5,
1145
+ cacheWritesPrice: 0,
1146
+ cacheReadsPrice: 0,
1147
+ description: "Extremely capable general-purpose LLM with hybrid reasoning capabilities and advanced tool calling"
1148
+ },
1149
+ "Qwen/Qwen3-235B-A22B-Instruct-2507": {
1150
+ maxTokens: 262144,
1151
+ contextWindow: 262144,
1152
+ supportsImages: false,
1153
+ supportsPromptCache: false,
1154
+ inputPrice: 0.22,
1155
+ outputPrice: 0.8,
1156
+ cacheWritesPrice: 0,
1157
+ cacheReadsPrice: 0,
1158
+ description: "Mixture-of-experts LLM with math and reasoning capabilities"
1159
+ },
1160
+ "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
1161
+ maxTokens: 262144,
1162
+ contextWindow: 262144,
1163
+ supportsImages: false,
1164
+ supportsPromptCache: false,
1165
+ inputPrice: 0.38,
1166
+ outputPrice: 1.53,
1167
+ cacheWritesPrice: 0,
1168
+ cacheReadsPrice: 0,
1169
+ description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
1170
+ },
1171
+ "openai/gpt-oss-120b": {
1172
+ maxTokens: 128072,
1173
+ contextWindow: 128072,
1174
+ supportsImages: false,
1175
+ supportsPromptCache: false,
1176
+ supportsNativeTools: true,
1177
+ inputPrice: 0.1,
1178
+ outputPrice: 0.5,
1179
+ cacheWritesPrice: 0,
1180
+ cacheReadsPrice: 0,
1181
+ description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
1182
+ },
1183
+ "moonshotai/Kimi-K2-Instruct-0905": {
1184
+ maxTokens: 168e3,
1185
+ contextWindow: 262e3,
1186
+ supportsImages: false,
1187
+ supportsPromptCache: false,
1188
+ supportsNativeTools: true,
1189
+ inputPrice: 0.6,
1190
+ outputPrice: 2.5,
1191
+ cacheWritesPrice: 0,
1192
+ cacheReadsPrice: 0,
1193
+ description: "State of the art language model for agentic and coding tasks. September Update."
1194
+ }
1195
+ };
1196
+ var basetenDefaultModelId = "zai-org/GLM-4.6";
1197
+
984
1198
  // src/providers/bedrock.ts
985
1199
  var bedrockDefaultModelId = "anthropic.claude-sonnet-4-5-20250929-v1:0";
986
1200
  var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
@@ -1085,6 +1299,20 @@ var bedrockModels = {
1085
1299
  maxCachePoints: 4,
1086
1300
  cachableFields: ["system", "messages", "tools"]
1087
1301
  },
1302
+ "anthropic.claude-opus-4-5-20251101-v1:0": {
1303
+ maxTokens: 8192,
1304
+ contextWindow: 2e5,
1305
+ supportsImages: true,
1306
+ supportsPromptCache: true,
1307
+ supportsReasoningBudget: true,
1308
+ inputPrice: 5,
1309
+ outputPrice: 25,
1310
+ cacheWritesPrice: 6.25,
1311
+ cacheReadsPrice: 0.5,
1312
+ minTokensPerCachePoint: 1024,
1313
+ maxCachePoints: 4,
1314
+ cachableFields: ["system", "messages", "tools"]
1315
+ },
1088
1316
  "anthropic.claude-opus-4-20250514-v1:0": {
1089
1317
  maxTokens: 8192,
1090
1318
  contextWindow: 2e5,
@@ -1427,7 +1655,8 @@ var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1427
1655
  var BEDROCK_GLOBAL_INFERENCE_MODEL_IDS = [
1428
1656
  "anthropic.claude-sonnet-4-20250514-v1:0",
1429
1657
  "anthropic.claude-sonnet-4-5-20250929-v1:0",
1430
- "anthropic.claude-haiku-4-5-20251001-v1:0"
1658
+ "anthropic.claude-haiku-4-5-20251001-v1:0",
1659
+ "anthropic.claude-opus-4-5-20251101-v1:0"
1431
1660
  ];
1432
1661
 
1433
1662
  // src/providers/cerebras.ts
@@ -1443,24 +1672,6 @@ var cerebrasModels = {
1443
1672
  outputPrice: 0,
1444
1673
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
1445
1674
  },
1446
- "qwen-3-coder-480b-free": {
1447
- maxTokens: 4e4,
1448
- contextWindow: 64e3,
1449
- supportsImages: false,
1450
- supportsPromptCache: false,
1451
- inputPrice: 0,
1452
- outputPrice: 0,
1453
- description: "[SOON TO BE DEPRECATED] SOTA coding model with ~2000 tokens/s ($0 free tier)\n\n\u2022 Use this if you don't have a Cerebras subscription\n\u2022 64K context window\n\u2022 Rate limits: 150K TPM, 1M TPH/TPD, 10 RPM, 100 RPH/RPD\n\nUpgrade for higher limits: [https://cloud.cerebras.ai/?utm=roocode](https://cloud.cerebras.ai/?utm=roocode)"
1454
- },
1455
- "qwen-3-coder-480b": {
1456
- maxTokens: 4e4,
1457
- contextWindow: 128e3,
1458
- supportsImages: false,
1459
- supportsPromptCache: false,
1460
- inputPrice: 0,
1461
- outputPrice: 0,
1462
- description: "[SOON TO BE DEPRECATED] SOTA coding model with ~2000 tokens/s ($50/$250 paid tiers)\n\n\u2022 Use this if you have a Cerebras subscription\n\u2022 131K context window with higher rate limits"
1463
- },
1464
1675
  "qwen-3-235b-a22b-instruct-2507": {
1465
1676
  maxTokens: 64e3,
1466
1677
  contextWindow: 64e3,
@@ -1488,16 +1699,6 @@ var cerebrasModels = {
1488
1699
  outputPrice: 0,
1489
1700
  description: "SOTA coding performance with ~2500 tokens/s"
1490
1701
  },
1491
- "qwen-3-235b-a22b-thinking-2507": {
1492
- maxTokens: 4e4,
1493
- contextWindow: 65e3,
1494
- supportsImages: false,
1495
- supportsPromptCache: false,
1496
- inputPrice: 0,
1497
- outputPrice: 0,
1498
- description: "SOTA performance with ~1500 tokens/s",
1499
- supportsReasoningEffort: true
1500
- },
1501
1702
  "gpt-oss-120b": {
1502
1703
  maxTokens: 8e3,
1503
1704
  contextWindow: 64e3,
@@ -1896,7 +2097,10 @@ var claudeCodeModels = {
1896
2097
  // Claude Code does report cache tokens
1897
2098
  supportsReasoningEffort: false,
1898
2099
  supportsReasoningBudget: false,
1899
- requiredReasoningBudget: false
2100
+ requiredReasoningBudget: false,
2101
+ // Claude Code manages its own tools and temperature via the CLI
2102
+ supportsNativeTools: false,
2103
+ supportsTemperature: false
1900
2104
  },
1901
2105
  "claude-sonnet-4-5-20250929[1m]": {
1902
2106
  ...anthropicModels["claude-sonnet-4-5"],
@@ -1907,7 +2111,10 @@ var claudeCodeModels = {
1907
2111
  // Claude Code does report cache tokens
1908
2112
  supportsReasoningEffort: false,
1909
2113
  supportsReasoningBudget: false,
1910
- requiredReasoningBudget: false
2114
+ requiredReasoningBudget: false,
2115
+ // Claude Code manages its own tools and temperature via the CLI
2116
+ supportsNativeTools: false,
2117
+ supportsTemperature: false
1911
2118
  },
1912
2119
  "claude-sonnet-4-20250514": {
1913
2120
  ...anthropicModels["claude-sonnet-4-20250514"],
@@ -1916,7 +2123,22 @@ var claudeCodeModels = {
1916
2123
  // Claude Code does report cache tokens
1917
2124
  supportsReasoningEffort: false,
1918
2125
  supportsReasoningBudget: false,
1919
- requiredReasoningBudget: false
2126
+ requiredReasoningBudget: false,
2127
+ // Claude Code manages its own tools and temperature via the CLI
2128
+ supportsNativeTools: false,
2129
+ supportsTemperature: false
2130
+ },
2131
+ "claude-opus-4-5-20251101": {
2132
+ ...anthropicModels["claude-opus-4-5-20251101"],
2133
+ supportsImages: false,
2134
+ supportsPromptCache: true,
2135
+ // Claude Code does report cache tokens
2136
+ supportsReasoningEffort: false,
2137
+ supportsReasoningBudget: false,
2138
+ requiredReasoningBudget: false,
2139
+ // Claude Code manages its own tools and temperature via the CLI
2140
+ supportsNativeTools: false,
2141
+ supportsTemperature: false
1920
2142
  },
1921
2143
  "claude-opus-4-1-20250805": {
1922
2144
  ...anthropicModels["claude-opus-4-1-20250805"],
@@ -1925,7 +2147,10 @@ var claudeCodeModels = {
1925
2147
  // Claude Code does report cache tokens
1926
2148
  supportsReasoningEffort: false,
1927
2149
  supportsReasoningBudget: false,
1928
- requiredReasoningBudget: false
2150
+ requiredReasoningBudget: false,
2151
+ // Claude Code manages its own tools and temperature via the CLI
2152
+ supportsNativeTools: false,
2153
+ supportsTemperature: false
1929
2154
  },
1930
2155
  "claude-opus-4-20250514": {
1931
2156
  ...anthropicModels["claude-opus-4-20250514"],
@@ -1934,7 +2159,10 @@ var claudeCodeModels = {
1934
2159
  // Claude Code does report cache tokens
1935
2160
  supportsReasoningEffort: false,
1936
2161
  supportsReasoningBudget: false,
1937
- requiredReasoningBudget: false
2162
+ requiredReasoningBudget: false,
2163
+ // Claude Code manages its own tools and temperature via the CLI
2164
+ supportsNativeTools: false,
2165
+ supportsTemperature: false
1938
2166
  },
1939
2167
  "claude-3-7-sonnet-20250219": {
1940
2168
  ...anthropicModels["claude-3-7-sonnet-20250219"],
@@ -1943,7 +2171,10 @@ var claudeCodeModels = {
1943
2171
  // Claude Code does report cache tokens
1944
2172
  supportsReasoningEffort: false,
1945
2173
  supportsReasoningBudget: false,
1946
- requiredReasoningBudget: false
2174
+ requiredReasoningBudget: false,
2175
+ // Claude Code manages its own tools and temperature via the CLI
2176
+ supportsNativeTools: false,
2177
+ supportsTemperature: false
1947
2178
  },
1948
2179
  "claude-3-5-sonnet-20241022": {
1949
2180
  ...anthropicModels["claude-3-5-sonnet-20241022"],
@@ -1952,7 +2183,10 @@ var claudeCodeModels = {
1952
2183
  // Claude Code does report cache tokens
1953
2184
  supportsReasoningEffort: false,
1954
2185
  supportsReasoningBudget: false,
1955
- requiredReasoningBudget: false
2186
+ requiredReasoningBudget: false,
2187
+ // Claude Code manages its own tools and temperature via the CLI
2188
+ supportsNativeTools: false,
2189
+ supportsTemperature: false
1956
2190
  },
1957
2191
  "claude-3-5-haiku-20241022": {
1958
2192
  ...anthropicModels["claude-3-5-haiku-20241022"],
@@ -1961,7 +2195,10 @@ var claudeCodeModels = {
1961
2195
  // Claude Code does report cache tokens
1962
2196
  supportsReasoningEffort: false,
1963
2197
  supportsReasoningBudget: false,
1964
- requiredReasoningBudget: false
2198
+ requiredReasoningBudget: false,
2199
+ // Claude Code manages its own tools and temperature via the CLI
2200
+ supportsNativeTools: false,
2201
+ supportsTemperature: false
1965
2202
  },
1966
2203
  "claude-haiku-4-5-20251001": {
1967
2204
  ...anthropicModels["claude-haiku-4-5-20251001"],
@@ -1970,7 +2207,10 @@ var claudeCodeModels = {
1970
2207
  // Claude Code does report cache tokens
1971
2208
  supportsReasoningEffort: false,
1972
2209
  supportsReasoningBudget: false,
1973
- requiredReasoningBudget: false
2210
+ requiredReasoningBudget: false,
2211
+ // Claude Code manages its own tools and temperature via the CLI
2212
+ supportsNativeTools: false,
2213
+ supportsTemperature: false
1974
2214
  }
1975
2215
  };
1976
2216
 
@@ -1983,6 +2223,7 @@ var deepSeekModels = {
1983
2223
  contextWindow: 128e3,
1984
2224
  supportsImages: false,
1985
2225
  supportsPromptCache: true,
2226
+ supportsNativeTools: true,
1986
2227
  inputPrice: 0.56,
1987
2228
  // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1988
2229
  outputPrice: 1.68,
@@ -1999,6 +2240,7 @@ var deepSeekModels = {
1999
2240
  contextWindow: 128e3,
2000
2241
  supportsImages: false,
2001
2242
  supportsPromptCache: true,
2243
+ supportsNativeTools: true,
2002
2244
  inputPrice: 0.56,
2003
2245
  // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
2004
2246
  outputPrice: 1.68,
@@ -2020,6 +2262,7 @@ var doubaoModels = {
2020
2262
  contextWindow: 128e3,
2021
2263
  supportsImages: true,
2022
2264
  supportsPromptCache: true,
2265
+ supportsNativeTools: true,
2023
2266
  inputPrice: 1e-4,
2024
2267
  // $0.0001 per million tokens (cache miss)
2025
2268
  outputPrice: 4e-4,
@@ -2035,6 +2278,7 @@ var doubaoModels = {
2035
2278
  contextWindow: 128e3,
2036
2279
  supportsImages: true,
2037
2280
  supportsPromptCache: true,
2281
+ supportsNativeTools: true,
2038
2282
  inputPrice: 2e-4,
2039
2283
  // $0.0002 per million tokens
2040
2284
  outputPrice: 8e-4,
@@ -2050,6 +2294,7 @@ var doubaoModels = {
2050
2294
  contextWindow: 128e3,
2051
2295
  supportsImages: true,
2052
2296
  supportsPromptCache: true,
2297
+ supportsNativeTools: true,
2053
2298
  inputPrice: 15e-5,
2054
2299
  // $0.00015 per million tokens
2055
2300
  outputPrice: 6e-4,
@@ -2090,6 +2335,7 @@ var featherlessModels = {
2090
2335
  contextWindow: 32678,
2091
2336
  supportsImages: false,
2092
2337
  supportsPromptCache: false,
2338
+ supportsNativeTools: true,
2093
2339
  inputPrice: 0,
2094
2340
  outputPrice: 0,
2095
2341
  description: "Kimi K2 Instruct model."
@@ -2108,6 +2354,7 @@ var featherlessModels = {
2108
2354
  contextWindow: 32678,
2109
2355
  supportsImages: false,
2110
2356
  supportsPromptCache: false,
2357
+ supportsNativeTools: true,
2111
2358
  inputPrice: 0,
2112
2359
  outputPrice: 0,
2113
2360
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -2123,6 +2370,7 @@ var fireworksModels = {
2123
2370
  contextWindow: 262144,
2124
2371
  supportsImages: false,
2125
2372
  supportsPromptCache: true,
2373
+ supportsNativeTools: true,
2126
2374
  inputPrice: 0.6,
2127
2375
  outputPrice: 2.5,
2128
2376
  cacheReadsPrice: 0.15,
@@ -2133,6 +2381,7 @@ var fireworksModels = {
2133
2381
  contextWindow: 128e3,
2134
2382
  supportsImages: false,
2135
2383
  supportsPromptCache: false,
2384
+ supportsNativeTools: true,
2136
2385
  inputPrice: 0.6,
2137
2386
  outputPrice: 2.5,
2138
2387
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2142,6 +2391,7 @@ var fireworksModels = {
2142
2391
  contextWindow: 204800,
2143
2392
  supportsImages: false,
2144
2393
  supportsPromptCache: false,
2394
+ supportsNativeTools: true,
2145
2395
  inputPrice: 0.3,
2146
2396
  outputPrice: 1.2,
2147
2397
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2151,6 +2401,7 @@ var fireworksModels = {
2151
2401
  contextWindow: 256e3,
2152
2402
  supportsImages: false,
2153
2403
  supportsPromptCache: false,
2404
+ supportsNativeTools: true,
2154
2405
  inputPrice: 0.22,
2155
2406
  outputPrice: 0.88,
2156
2407
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2160,6 +2411,7 @@ var fireworksModels = {
2160
2411
  contextWindow: 256e3,
2161
2412
  supportsImages: false,
2162
2413
  supportsPromptCache: false,
2414
+ supportsNativeTools: true,
2163
2415
  inputPrice: 0.45,
2164
2416
  outputPrice: 1.8,
2165
2417
  description: "Qwen3's most agentic code model to date."
@@ -2169,6 +2421,7 @@ var fireworksModels = {
2169
2421
  contextWindow: 16e4,
2170
2422
  supportsImages: false,
2171
2423
  supportsPromptCache: false,
2424
+ supportsNativeTools: true,
2172
2425
  inputPrice: 3,
2173
2426
  outputPrice: 8,
2174
2427
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2178,6 +2431,7 @@ var fireworksModels = {
2178
2431
  contextWindow: 128e3,
2179
2432
  supportsImages: false,
2180
2433
  supportsPromptCache: false,
2434
+ supportsNativeTools: true,
2181
2435
  inputPrice: 0.9,
2182
2436
  outputPrice: 0.9,
2183
2437
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2187,6 +2441,7 @@ var fireworksModels = {
2187
2441
  contextWindow: 163840,
2188
2442
  supportsImages: false,
2189
2443
  supportsPromptCache: false,
2444
+ supportsNativeTools: true,
2190
2445
  inputPrice: 0.56,
2191
2446
  outputPrice: 1.68,
2192
2447
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2196,6 +2451,7 @@ var fireworksModels = {
2196
2451
  contextWindow: 128e3,
2197
2452
  supportsImages: false,
2198
2453
  supportsPromptCache: false,
2454
+ supportsNativeTools: true,
2199
2455
  inputPrice: 0.55,
2200
2456
  outputPrice: 2.19,
2201
2457
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2205,6 +2461,7 @@ var fireworksModels = {
2205
2461
  contextWindow: 128e3,
2206
2462
  supportsImages: false,
2207
2463
  supportsPromptCache: false,
2464
+ supportsNativeTools: true,
2208
2465
  inputPrice: 0.55,
2209
2466
  outputPrice: 2.19,
2210
2467
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2214,6 +2471,7 @@ var fireworksModels = {
2214
2471
  contextWindow: 198e3,
2215
2472
  supportsImages: false,
2216
2473
  supportsPromptCache: false,
2474
+ supportsNativeTools: true,
2217
2475
  inputPrice: 0.55,
2218
2476
  outputPrice: 2.19,
2219
2477
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2223,6 +2481,7 @@ var fireworksModels = {
2223
2481
  contextWindow: 128e3,
2224
2482
  supportsImages: false,
2225
2483
  supportsPromptCache: false,
2484
+ supportsNativeTools: true,
2226
2485
  inputPrice: 0.07,
2227
2486
  outputPrice: 0.3,
2228
2487
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2232,6 +2491,7 @@ var fireworksModels = {
2232
2491
  contextWindow: 128e3,
2233
2492
  supportsImages: false,
2234
2493
  supportsPromptCache: false,
2494
+ supportsNativeTools: true,
2235
2495
  inputPrice: 0.15,
2236
2496
  outputPrice: 0.6,
2237
2497
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2239,133 +2499,39 @@ var fireworksModels = {
2239
2499
  };
2240
2500
 
2241
2501
  // src/providers/gemini.ts
2242
- var geminiDefaultModelId = "gemini-2.0-flash-001";
2502
+ var geminiDefaultModelId = "gemini-2.5-pro";
2243
2503
  var geminiModels = {
2244
- // Latest models (pointing to the most recent stable versions)
2245
- "gemini-flash-latest": {
2246
- maxTokens: 65536,
2247
- contextWindow: 1048576,
2248
- supportsImages: true,
2249
- supportsPromptCache: true,
2250
- inputPrice: 0.3,
2251
- outputPrice: 2.5,
2252
- cacheReadsPrice: 0.075,
2253
- cacheWritesPrice: 1,
2254
- maxThinkingTokens: 24576,
2255
- supportsReasoningBudget: true
2256
- },
2257
- "gemini-flash-lite-latest": {
2504
+ "gemini-3-pro-preview": {
2258
2505
  maxTokens: 65536,
2259
2506
  contextWindow: 1048576,
2260
2507
  supportsImages: true,
2508
+ supportsNativeTools: true,
2261
2509
  supportsPromptCache: true,
2262
- inputPrice: 0.1,
2263
- outputPrice: 0.4,
2264
- cacheReadsPrice: 0.025,
2265
- cacheWritesPrice: 1,
2266
- supportsReasoningBudget: true,
2267
- maxThinkingTokens: 24576
2268
- },
2269
- // 2.5 Flash models (09-2025 versions - most recent)
2270
- "gemini-2.5-flash-preview-09-2025": {
2271
- maxTokens: 65536,
2272
- contextWindow: 1048576,
2273
- supportsImages: true,
2274
- supportsPromptCache: true,
2275
- inputPrice: 0.3,
2276
- outputPrice: 2.5,
2277
- cacheReadsPrice: 0.075,
2278
- cacheWritesPrice: 1,
2279
- maxThinkingTokens: 24576,
2280
- supportsReasoningBudget: true
2281
- },
2282
- "gemini-2.5-flash-lite-preview-09-2025": {
2283
- maxTokens: 65536,
2284
- contextWindow: 1048576,
2285
- supportsImages: true,
2286
- supportsPromptCache: true,
2287
- inputPrice: 0.1,
2288
- outputPrice: 0.4,
2289
- cacheReadsPrice: 0.025,
2290
- cacheWritesPrice: 1,
2291
- supportsReasoningBudget: true,
2292
- maxThinkingTokens: 24576
2293
- },
2294
- // 2.5 Flash models (06-17 version)
2295
- "gemini-2.5-flash-lite-preview-06-17": {
2296
- maxTokens: 64e3,
2297
- contextWindow: 1048576,
2298
- supportsImages: true,
2299
- supportsPromptCache: true,
2300
- inputPrice: 0.1,
2301
- outputPrice: 0.4,
2302
- cacheReadsPrice: 0.025,
2303
- cacheWritesPrice: 1,
2304
- supportsReasoningBudget: true,
2305
- maxThinkingTokens: 24576
2306
- },
2307
- // 2.5 Flash models (05-20 versions)
2308
- "gemini-2.5-flash-preview-05-20:thinking": {
2309
- maxTokens: 65535,
2310
- contextWindow: 1048576,
2311
- supportsImages: true,
2312
- supportsPromptCache: true,
2313
- inputPrice: 0.15,
2314
- outputPrice: 3.5,
2315
- cacheReadsPrice: 0.0375,
2316
- cacheWritesPrice: 1,
2317
- maxThinkingTokens: 24576,
2318
- supportsReasoningBudget: true,
2319
- requiredReasoningBudget: true
2320
- },
2321
- "gemini-2.5-flash-preview-05-20": {
2322
- maxTokens: 65535,
2323
- contextWindow: 1048576,
2324
- supportsImages: true,
2325
- supportsPromptCache: true,
2326
- inputPrice: 0.15,
2327
- outputPrice: 0.6,
2328
- cacheReadsPrice: 0.0375,
2329
- cacheWritesPrice: 1
2330
- },
2331
- // 2.5 Flash models (04-17 versions)
2332
- "gemini-2.5-flash-preview-04-17:thinking": {
2333
- maxTokens: 65535,
2334
- contextWindow: 1048576,
2335
- supportsImages: true,
2336
- supportsPromptCache: false,
2337
- inputPrice: 0.15,
2338
- outputPrice: 3.5,
2339
- maxThinkingTokens: 24576,
2340
- supportsReasoningBudget: true,
2341
- requiredReasoningBudget: true
2342
- },
2343
- "gemini-2.5-flash-preview-04-17": {
2344
- maxTokens: 65535,
2345
- contextWindow: 1048576,
2346
- supportsImages: true,
2347
- supportsPromptCache: false,
2348
- inputPrice: 0.15,
2349
- outputPrice: 0.6
2350
- },
2351
- // 2.5 Flash stable
2352
- "gemini-2.5-flash": {
2353
- maxTokens: 64e3,
2354
- contextWindow: 1048576,
2355
- supportsImages: true,
2356
- supportsPromptCache: true,
2357
- inputPrice: 0.3,
2358
- outputPrice: 2.5,
2359
- cacheReadsPrice: 0.075,
2360
- cacheWritesPrice: 1,
2361
- maxThinkingTokens: 24576,
2362
- supportsReasoningBudget: true
2510
+ supportsReasoningEffort: ["low", "high"],
2511
+ reasoningEffort: "low",
2512
+ supportsTemperature: true,
2513
+ defaultTemperature: 1,
2514
+ inputPrice: 4,
2515
+ outputPrice: 18,
2516
+ tiers: [
2517
+ {
2518
+ contextWindow: 2e5,
2519
+ inputPrice: 2,
2520
+ outputPrice: 12
2521
+ },
2522
+ {
2523
+ contextWindow: Infinity,
2524
+ inputPrice: 4,
2525
+ outputPrice: 18
2526
+ }
2527
+ ]
2363
2528
  },
2364
2529
  // 2.5 Pro models
2365
- "gemini-2.5-pro-preview-06-05": {
2366
- maxTokens: 65535,
2530
+ "gemini-2.5-pro": {
2531
+ maxTokens: 64e3,
2367
2532
  contextWindow: 1048576,
2368
2533
  supportsImages: true,
2534
+ supportsNativeTools: true,
2369
2535
  supportsPromptCache: true,
2370
2536
  inputPrice: 2.5,
2371
2537
  // This is the pricing for prompts above 200k tokens.
@@ -2374,6 +2540,7 @@ var geminiModels = {
2374
2540
  cacheWritesPrice: 4.5,
2375
2541
  maxThinkingTokens: 32768,
2376
2542
  supportsReasoningBudget: true,
2543
+ requiredReasoningBudget: true,
2377
2544
  tiers: [
2378
2545
  {
2379
2546
  contextWindow: 2e5,
@@ -2389,16 +2556,19 @@ var geminiModels = {
2389
2556
  }
2390
2557
  ]
2391
2558
  },
2392
- "gemini-2.5-pro-preview-05-06": {
2559
+ "gemini-2.5-pro-preview-06-05": {
2393
2560
  maxTokens: 65535,
2394
2561
  contextWindow: 1048576,
2395
2562
  supportsImages: true,
2563
+ supportsNativeTools: true,
2396
2564
  supportsPromptCache: true,
2397
2565
  inputPrice: 2.5,
2398
2566
  // This is the pricing for prompts above 200k tokens.
2399
2567
  outputPrice: 15,
2400
2568
  cacheReadsPrice: 0.625,
2401
2569
  cacheWritesPrice: 4.5,
2570
+ maxThinkingTokens: 32768,
2571
+ supportsReasoningBudget: true,
2402
2572
  tiers: [
2403
2573
  {
2404
2574
  contextWindow: 2e5,
@@ -2414,18 +2584,17 @@ var geminiModels = {
2414
2584
  }
2415
2585
  ]
2416
2586
  },
2417
- "gemini-2.5-pro-preview-03-25": {
2587
+ "gemini-2.5-pro-preview-05-06": {
2418
2588
  maxTokens: 65535,
2419
2589
  contextWindow: 1048576,
2420
2590
  supportsImages: true,
2591
+ supportsNativeTools: true,
2421
2592
  supportsPromptCache: true,
2422
2593
  inputPrice: 2.5,
2423
2594
  // This is the pricing for prompts above 200k tokens.
2424
2595
  outputPrice: 15,
2425
2596
  cacheReadsPrice: 0.625,
2426
2597
  cacheWritesPrice: 4.5,
2427
- maxThinkingTokens: 32768,
2428
- supportsReasoningBudget: true,
2429
2598
  tiers: [
2430
2599
  {
2431
2600
  contextWindow: 2e5,
@@ -2441,18 +2610,11 @@ var geminiModels = {
2441
2610
  }
2442
2611
  ]
2443
2612
  },
2444
- "gemini-2.5-pro-exp-03-25": {
2613
+ "gemini-2.5-pro-preview-03-25": {
2445
2614
  maxTokens: 65535,
2446
2615
  contextWindow: 1048576,
2447
2616
  supportsImages: true,
2448
- supportsPromptCache: false,
2449
- inputPrice: 0,
2450
- outputPrice: 0
2451
- },
2452
- "gemini-2.5-pro": {
2453
- maxTokens: 64e3,
2454
- contextWindow: 1048576,
2455
- supportsImages: true,
2617
+ supportsNativeTools: true,
2456
2618
  supportsPromptCache: true,
2457
2619
  inputPrice: 2.5,
2458
2620
  // This is the pricing for prompts above 200k tokens.
@@ -2461,7 +2623,6 @@ var geminiModels = {
2461
2623
  cacheWritesPrice: 4.5,
2462
2624
  maxThinkingTokens: 32768,
2463
2625
  supportsReasoningBudget: true,
2464
- requiredReasoningBudget: true,
2465
2626
  tiers: [
2466
2627
  {
2467
2628
  contextWindow: 2e5,
@@ -2477,125 +2638,72 @@ var geminiModels = {
2477
2638
  }
2478
2639
  ]
2479
2640
  },
2480
- // 2.0 Flash models
2481
- "gemini-2.0-flash-lite-preview-02-05": {
2482
- maxTokens: 8192,
2641
+ // 2.5 Flash models
2642
+ "gemini-flash-latest": {
2643
+ maxTokens: 65536,
2483
2644
  contextWindow: 1048576,
2484
2645
  supportsImages: true,
2485
- supportsPromptCache: false,
2486
- inputPrice: 0,
2487
- outputPrice: 0
2646
+ supportsNativeTools: true,
2647
+ supportsPromptCache: true,
2648
+ inputPrice: 0.3,
2649
+ outputPrice: 2.5,
2650
+ cacheReadsPrice: 0.075,
2651
+ cacheWritesPrice: 1,
2652
+ maxThinkingTokens: 24576,
2653
+ supportsReasoningBudget: true
2488
2654
  },
2489
- "gemini-2.0-flash-thinking-exp-01-21": {
2655
+ "gemini-2.5-flash-preview-09-2025": {
2490
2656
  maxTokens: 65536,
2491
2657
  contextWindow: 1048576,
2492
2658
  supportsImages: true,
2493
- supportsPromptCache: false,
2494
- inputPrice: 0,
2495
- outputPrice: 0
2496
- },
2497
- "gemini-2.0-flash-thinking-exp-1219": {
2498
- maxTokens: 8192,
2499
- contextWindow: 32767,
2500
- supportsImages: true,
2501
- supportsPromptCache: false,
2502
- inputPrice: 0,
2503
- outputPrice: 0
2659
+ supportsNativeTools: true,
2660
+ supportsPromptCache: true,
2661
+ inputPrice: 0.3,
2662
+ outputPrice: 2.5,
2663
+ cacheReadsPrice: 0.075,
2664
+ cacheWritesPrice: 1,
2665
+ maxThinkingTokens: 24576,
2666
+ supportsReasoningBudget: true
2504
2667
  },
2505
- "gemini-2.0-flash-exp": {
2506
- maxTokens: 8192,
2668
+ "gemini-2.5-flash": {
2669
+ maxTokens: 64e3,
2507
2670
  contextWindow: 1048576,
2508
2671
  supportsImages: true,
2509
- supportsPromptCache: false,
2510
- inputPrice: 0,
2511
- outputPrice: 0
2672
+ supportsNativeTools: true,
2673
+ supportsPromptCache: true,
2674
+ inputPrice: 0.3,
2675
+ outputPrice: 2.5,
2676
+ cacheReadsPrice: 0.075,
2677
+ cacheWritesPrice: 1,
2678
+ maxThinkingTokens: 24576,
2679
+ supportsReasoningBudget: true
2512
2680
  },
2513
- "gemini-2.0-flash-001": {
2514
- maxTokens: 8192,
2681
+ // 2.5 Flash Lite models
2682
+ "gemini-flash-lite-latest": {
2683
+ maxTokens: 65536,
2515
2684
  contextWindow: 1048576,
2516
2685
  supportsImages: true,
2686
+ supportsNativeTools: true,
2517
2687
  supportsPromptCache: true,
2518
2688
  inputPrice: 0.1,
2519
2689
  outputPrice: 0.4,
2520
2690
  cacheReadsPrice: 0.025,
2521
- cacheWritesPrice: 1
2522
- },
2523
- // 2.0 Pro models
2524
- "gemini-2.0-pro-exp-02-05": {
2525
- maxTokens: 8192,
2526
- contextWindow: 2097152,
2527
- supportsImages: true,
2528
- supportsPromptCache: false,
2529
- inputPrice: 0,
2530
- outputPrice: 0
2691
+ cacheWritesPrice: 1,
2692
+ supportsReasoningBudget: true,
2693
+ maxThinkingTokens: 24576
2531
2694
  },
2532
- // 1.5 Flash models
2533
- "gemini-1.5-flash-002": {
2534
- maxTokens: 8192,
2695
+ "gemini-2.5-flash-lite-preview-09-2025": {
2696
+ maxTokens: 65536,
2535
2697
  contextWindow: 1048576,
2536
2698
  supportsImages: true,
2699
+ supportsNativeTools: true,
2537
2700
  supportsPromptCache: true,
2538
- inputPrice: 0.15,
2539
- // This is the pricing for prompts above 128k tokens.
2540
- outputPrice: 0.6,
2541
- cacheReadsPrice: 0.0375,
2701
+ inputPrice: 0.1,
2702
+ outputPrice: 0.4,
2703
+ cacheReadsPrice: 0.025,
2542
2704
  cacheWritesPrice: 1,
2543
- tiers: [
2544
- {
2545
- contextWindow: 128e3,
2546
- inputPrice: 0.075,
2547
- outputPrice: 0.3,
2548
- cacheReadsPrice: 0.01875
2549
- },
2550
- {
2551
- contextWindow: Infinity,
2552
- inputPrice: 0.15,
2553
- outputPrice: 0.6,
2554
- cacheReadsPrice: 0.0375
2555
- }
2556
- ]
2557
- },
2558
- "gemini-1.5-flash-exp-0827": {
2559
- maxTokens: 8192,
2560
- contextWindow: 1048576,
2561
- supportsImages: true,
2562
- supportsPromptCache: false,
2563
- inputPrice: 0,
2564
- outputPrice: 0
2565
- },
2566
- "gemini-1.5-flash-8b-exp-0827": {
2567
- maxTokens: 8192,
2568
- contextWindow: 1048576,
2569
- supportsImages: true,
2570
- supportsPromptCache: false,
2571
- inputPrice: 0,
2572
- outputPrice: 0
2573
- },
2574
- // 1.5 Pro models
2575
- "gemini-1.5-pro-002": {
2576
- maxTokens: 8192,
2577
- contextWindow: 2097152,
2578
- supportsImages: true,
2579
- supportsPromptCache: false,
2580
- inputPrice: 0,
2581
- outputPrice: 0
2582
- },
2583
- "gemini-1.5-pro-exp-0827": {
2584
- maxTokens: 8192,
2585
- contextWindow: 2097152,
2586
- supportsImages: true,
2587
- supportsPromptCache: false,
2588
- inputPrice: 0,
2589
- outputPrice: 0
2590
- },
2591
- // Experimental models
2592
- "gemini-exp-1206": {
2593
- maxTokens: 8192,
2594
- contextWindow: 2097152,
2595
- supportsImages: true,
2596
- supportsPromptCache: false,
2597
- inputPrice: 0,
2598
- outputPrice: 0
2705
+ supportsReasoningBudget: true,
2706
+ maxThinkingTokens: 24576
2599
2707
  }
2600
2708
  };
2601
2709
 
@@ -2623,6 +2731,7 @@ var groqModels = {
2623
2731
  contextWindow: 131072,
2624
2732
  supportsImages: false,
2625
2733
  supportsPromptCache: false,
2734
+ supportsNativeTools: true,
2626
2735
  inputPrice: 0.05,
2627
2736
  outputPrice: 0.08,
2628
2737
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2632,6 +2741,7 @@ var groqModels = {
2632
2741
  contextWindow: 131072,
2633
2742
  supportsImages: false,
2634
2743
  supportsPromptCache: false,
2744
+ supportsNativeTools: true,
2635
2745
  inputPrice: 0.59,
2636
2746
  outputPrice: 0.79,
2637
2747
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2641,6 +2751,7 @@ var groqModels = {
2641
2751
  contextWindow: 131072,
2642
2752
  supportsImages: false,
2643
2753
  supportsPromptCache: false,
2754
+ supportsNativeTools: true,
2644
2755
  inputPrice: 0.11,
2645
2756
  outputPrice: 0.34,
2646
2757
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2677,6 +2788,7 @@ var groqModels = {
2677
2788
  contextWindow: 131072,
2678
2789
  supportsImages: false,
2679
2790
  supportsPromptCache: false,
2791
+ supportsNativeTools: true,
2680
2792
  inputPrice: 0.29,
2681
2793
  outputPrice: 0.59,
2682
2794
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2706,6 +2818,7 @@ var groqModels = {
2706
2818
  contextWindow: 262144,
2707
2819
  supportsImages: false,
2708
2820
  supportsPromptCache: true,
2821
+ supportsNativeTools: true,
2709
2822
  inputPrice: 0.6,
2710
2823
  outputPrice: 2.5,
2711
2824
  cacheReadsPrice: 0.15,
@@ -2716,6 +2829,7 @@ var groqModels = {
2716
2829
  contextWindow: 131072,
2717
2830
  supportsImages: false,
2718
2831
  supportsPromptCache: false,
2832
+ supportsNativeTools: true,
2719
2833
  inputPrice: 0.15,
2720
2834
  outputPrice: 0.75,
2721
2835
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2725,6 +2839,7 @@ var groqModels = {
2725
2839
  contextWindow: 131072,
2726
2840
  supportsImages: false,
2727
2841
  supportsPromptCache: false,
2842
+ supportsNativeTools: true,
2728
2843
  inputPrice: 0.1,
2729
2844
  outputPrice: 0.5,
2730
2845
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2751,6 +2866,7 @@ var ioIntelligenceModels = {
2751
2866
  contextWindow: 128e3,
2752
2867
  supportsImages: false,
2753
2868
  supportsPromptCache: false,
2869
+ supportsNativeTools: true,
2754
2870
  description: "DeepSeek R1 reasoning model"
2755
2871
  },
2756
2872
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -2758,6 +2874,7 @@ var ioIntelligenceModels = {
2758
2874
  contextWindow: 43e4,
2759
2875
  supportsImages: true,
2760
2876
  supportsPromptCache: false,
2877
+ supportsNativeTools: true,
2761
2878
  description: "Llama 4 Maverick 17B model"
2762
2879
  },
2763
2880
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -2765,6 +2882,7 @@ var ioIntelligenceModels = {
2765
2882
  contextWindow: 106e3,
2766
2883
  supportsImages: false,
2767
2884
  supportsPromptCache: false,
2885
+ supportsNativeTools: true,
2768
2886
  description: "Qwen3 Coder 480B specialized for coding"
2769
2887
  },
2770
2888
  "openai/gpt-oss-120b": {
@@ -2772,6 +2890,7 @@ var ioIntelligenceModels = {
2772
2890
  contextWindow: 131072,
2773
2891
  supportsImages: false,
2774
2892
  supportsPromptCache: false,
2893
+ supportsNativeTools: true,
2775
2894
  description: "OpenAI GPT-OSS 120B model"
2776
2895
  }
2777
2896
  };
@@ -2812,75 +2931,84 @@ var mistralModels = {
2812
2931
  contextWindow: 128e3,
2813
2932
  supportsImages: true,
2814
2933
  supportsPromptCache: false,
2934
+ supportsNativeTools: true,
2815
2935
  inputPrice: 2,
2816
2936
  outputPrice: 5
2817
2937
  },
2818
2938
  "devstral-medium-latest": {
2819
- maxTokens: 131e3,
2939
+ maxTokens: 8192,
2820
2940
  contextWindow: 131e3,
2821
2941
  supportsImages: true,
2822
2942
  supportsPromptCache: false,
2943
+ supportsNativeTools: true,
2823
2944
  inputPrice: 0.4,
2824
2945
  outputPrice: 2
2825
2946
  },
2826
2947
  "mistral-medium-latest": {
2827
- maxTokens: 131e3,
2948
+ maxTokens: 8192,
2828
2949
  contextWindow: 131e3,
2829
2950
  supportsImages: true,
2830
2951
  supportsPromptCache: false,
2952
+ supportsNativeTools: true,
2831
2953
  inputPrice: 0.4,
2832
2954
  outputPrice: 2
2833
2955
  },
2834
2956
  "codestral-latest": {
2835
- maxTokens: 256e3,
2957
+ maxTokens: 8192,
2836
2958
  contextWindow: 256e3,
2837
2959
  supportsImages: false,
2838
2960
  supportsPromptCache: false,
2961
+ supportsNativeTools: true,
2839
2962
  inputPrice: 0.3,
2840
2963
  outputPrice: 0.9
2841
2964
  },
2842
2965
  "mistral-large-latest": {
2843
- maxTokens: 131e3,
2966
+ maxTokens: 8192,
2844
2967
  contextWindow: 131e3,
2845
2968
  supportsImages: false,
2846
2969
  supportsPromptCache: false,
2970
+ supportsNativeTools: true,
2847
2971
  inputPrice: 2,
2848
2972
  outputPrice: 6
2849
2973
  },
2850
2974
  "ministral-8b-latest": {
2851
- maxTokens: 131e3,
2975
+ maxTokens: 8192,
2852
2976
  contextWindow: 131e3,
2853
2977
  supportsImages: false,
2854
2978
  supportsPromptCache: false,
2979
+ supportsNativeTools: true,
2855
2980
  inputPrice: 0.1,
2856
2981
  outputPrice: 0.1
2857
2982
  },
2858
2983
  "ministral-3b-latest": {
2859
- maxTokens: 131e3,
2984
+ maxTokens: 8192,
2860
2985
  contextWindow: 131e3,
2861
2986
  supportsImages: false,
2862
2987
  supportsPromptCache: false,
2988
+ supportsNativeTools: true,
2863
2989
  inputPrice: 0.04,
2864
2990
  outputPrice: 0.04
2865
2991
  },
2866
2992
  "mistral-small-latest": {
2867
- maxTokens: 32e3,
2993
+ maxTokens: 8192,
2868
2994
  contextWindow: 32e3,
2869
2995
  supportsImages: false,
2870
2996
  supportsPromptCache: false,
2997
+ supportsNativeTools: true,
2871
2998
  inputPrice: 0.2,
2872
2999
  outputPrice: 0.6
2873
3000
  },
2874
3001
  "pixtral-large-latest": {
2875
- maxTokens: 131e3,
3002
+ maxTokens: 8192,
2876
3003
  contextWindow: 131e3,
2877
3004
  supportsImages: true,
2878
3005
  supportsPromptCache: false,
3006
+ supportsNativeTools: true,
2879
3007
  inputPrice: 2,
2880
3008
  outputPrice: 6
2881
3009
  }
2882
3010
  };
2883
- var MISTRAL_DEFAULT_TEMPERATURE = 0;
3011
+ var MISTRAL_DEFAULT_TEMPERATURE = 1;
2884
3012
 
2885
3013
  // src/providers/moonshot.ts
2886
3014
  var moonshotDefaultModelId = "kimi-k2-0905-preview";
@@ -2890,6 +3018,7 @@ var moonshotModels = {
2890
3018
  contextWindow: 131072,
2891
3019
  supportsImages: false,
2892
3020
  supportsPromptCache: true,
3021
+ supportsNativeTools: true,
2893
3022
  inputPrice: 0.6,
2894
3023
  // $0.60 per million tokens (cache miss)
2895
3024
  outputPrice: 2.5,
@@ -2905,6 +3034,7 @@ var moonshotModels = {
2905
3034
  contextWindow: 262144,
2906
3035
  supportsImages: false,
2907
3036
  supportsPromptCache: true,
3037
+ supportsNativeTools: true,
2908
3038
  inputPrice: 0.6,
2909
3039
  outputPrice: 2.5,
2910
3040
  cacheReadsPrice: 0.15,
@@ -2915,6 +3045,7 @@ var moonshotModels = {
2915
3045
  contextWindow: 262144,
2916
3046
  supportsImages: false,
2917
3047
  supportsPromptCache: true,
3048
+ supportsNativeTools: true,
2918
3049
  inputPrice: 2.4,
2919
3050
  // $2.40 per million tokens (cache miss)
2920
3051
  outputPrice: 10,
@@ -2933,6 +3064,7 @@ var moonshotModels = {
2933
3064
  supportsImages: false,
2934
3065
  // Text-only (no image/vision support)
2935
3066
  supportsPromptCache: true,
3067
+ supportsNativeTools: true,
2936
3068
  inputPrice: 0.6,
2937
3069
  // $0.60 per million tokens (cache miss)
2938
3070
  outputPrice: 2.5,
@@ -2970,6 +3102,7 @@ var openAiNativeModels = {
2970
3102
  "gpt-5.1": {
2971
3103
  maxTokens: 128e3,
2972
3104
  contextWindow: 4e5,
3105
+ supportsNativeTools: true,
2973
3106
  supportsImages: true,
2974
3107
  supportsPromptCache: true,
2975
3108
  promptCacheRetention: "24h",
@@ -2989,6 +3122,7 @@ var openAiNativeModels = {
2989
3122
  "gpt-5.1-codex": {
2990
3123
  maxTokens: 128e3,
2991
3124
  contextWindow: 4e5,
3125
+ supportsNativeTools: true,
2992
3126
  supportsImages: true,
2993
3127
  supportsPromptCache: true,
2994
3128
  promptCacheRetention: "24h",
@@ -3004,6 +3138,7 @@ var openAiNativeModels = {
3004
3138
  "gpt-5.1-codex-mini": {
3005
3139
  maxTokens: 128e3,
3006
3140
  contextWindow: 4e5,
3141
+ supportsNativeTools: true,
3007
3142
  supportsImages: true,
3008
3143
  supportsPromptCache: true,
3009
3144
  promptCacheRetention: "24h",
@@ -3018,6 +3153,7 @@ var openAiNativeModels = {
3018
3153
  "gpt-5": {
3019
3154
  maxTokens: 128e3,
3020
3155
  contextWindow: 4e5,
3156
+ supportsNativeTools: true,
3021
3157
  supportsImages: true,
3022
3158
  supportsPromptCache: true,
3023
3159
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3036,6 +3172,7 @@ var openAiNativeModels = {
3036
3172
  "gpt-5-mini": {
3037
3173
  maxTokens: 128e3,
3038
3174
  contextWindow: 4e5,
3175
+ supportsNativeTools: true,
3039
3176
  supportsImages: true,
3040
3177
  supportsPromptCache: true,
3041
3178
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3054,6 +3191,7 @@ var openAiNativeModels = {
3054
3191
  "gpt-5-codex": {
3055
3192
  maxTokens: 128e3,
3056
3193
  contextWindow: 4e5,
3194
+ supportsNativeTools: true,
3057
3195
  supportsImages: true,
3058
3196
  supportsPromptCache: true,
3059
3197
  supportsReasoningEffort: ["low", "medium", "high"],
@@ -3068,6 +3206,7 @@ var openAiNativeModels = {
3068
3206
  "gpt-5-nano": {
3069
3207
  maxTokens: 128e3,
3070
3208
  contextWindow: 4e5,
3209
+ supportsNativeTools: true,
3071
3210
  supportsImages: true,
3072
3211
  supportsPromptCache: true,
3073
3212
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3083,6 +3222,7 @@ var openAiNativeModels = {
3083
3222
  "gpt-5-chat-latest": {
3084
3223
  maxTokens: 128e3,
3085
3224
  contextWindow: 4e5,
3225
+ supportsNativeTools: true,
3086
3226
  supportsImages: true,
3087
3227
  supportsPromptCache: true,
3088
3228
  inputPrice: 1.25,
@@ -3093,6 +3233,7 @@ var openAiNativeModels = {
3093
3233
  "gpt-4.1": {
3094
3234
  maxTokens: 32768,
3095
3235
  contextWindow: 1047576,
3236
+ supportsNativeTools: true,
3096
3237
  supportsImages: true,
3097
3238
  supportsPromptCache: true,
3098
3239
  inputPrice: 2,
@@ -3106,6 +3247,7 @@ var openAiNativeModels = {
3106
3247
  "gpt-4.1-mini": {
3107
3248
  maxTokens: 32768,
3108
3249
  contextWindow: 1047576,
3250
+ supportsNativeTools: true,
3109
3251
  supportsImages: true,
3110
3252
  supportsPromptCache: true,
3111
3253
  inputPrice: 0.4,
@@ -3119,6 +3261,7 @@ var openAiNativeModels = {
3119
3261
  "gpt-4.1-nano": {
3120
3262
  maxTokens: 32768,
3121
3263
  contextWindow: 1047576,
3264
+ supportsNativeTools: true,
3122
3265
  supportsImages: true,
3123
3266
  supportsPromptCache: true,
3124
3267
  inputPrice: 0.1,
@@ -3132,6 +3275,7 @@ var openAiNativeModels = {
3132
3275
  o3: {
3133
3276
  maxTokens: 1e5,
3134
3277
  contextWindow: 2e5,
3278
+ supportsNativeTools: true,
3135
3279
  supportsImages: true,
3136
3280
  supportsPromptCache: true,
3137
3281
  inputPrice: 2,
@@ -3148,6 +3292,7 @@ var openAiNativeModels = {
3148
3292
  "o3-high": {
3149
3293
  maxTokens: 1e5,
3150
3294
  contextWindow: 2e5,
3295
+ supportsNativeTools: true,
3151
3296
  supportsImages: true,
3152
3297
  supportsPromptCache: true,
3153
3298
  inputPrice: 2,
@@ -3159,6 +3304,7 @@ var openAiNativeModels = {
3159
3304
  "o3-low": {
3160
3305
  maxTokens: 1e5,
3161
3306
  contextWindow: 2e5,
3307
+ supportsNativeTools: true,
3162
3308
  supportsImages: true,
3163
3309
  supportsPromptCache: true,
3164
3310
  inputPrice: 2,
@@ -3170,6 +3316,7 @@ var openAiNativeModels = {
3170
3316
  "o4-mini": {
3171
3317
  maxTokens: 1e5,
3172
3318
  contextWindow: 2e5,
3319
+ supportsNativeTools: true,
3173
3320
  supportsImages: true,
3174
3321
  supportsPromptCache: true,
3175
3322
  inputPrice: 1.1,
@@ -3186,6 +3333,7 @@ var openAiNativeModels = {
3186
3333
  "o4-mini-high": {
3187
3334
  maxTokens: 1e5,
3188
3335
  contextWindow: 2e5,
3336
+ supportsNativeTools: true,
3189
3337
  supportsImages: true,
3190
3338
  supportsPromptCache: true,
3191
3339
  inputPrice: 1.1,
@@ -3197,6 +3345,7 @@ var openAiNativeModels = {
3197
3345
  "o4-mini-low": {
3198
3346
  maxTokens: 1e5,
3199
3347
  contextWindow: 2e5,
3348
+ supportsNativeTools: true,
3200
3349
  supportsImages: true,
3201
3350
  supportsPromptCache: true,
3202
3351
  inputPrice: 1.1,
@@ -3208,6 +3357,7 @@ var openAiNativeModels = {
3208
3357
  "o3-mini": {
3209
3358
  maxTokens: 1e5,
3210
3359
  contextWindow: 2e5,
3360
+ supportsNativeTools: true,
3211
3361
  supportsImages: false,
3212
3362
  supportsPromptCache: true,
3213
3363
  inputPrice: 1.1,
@@ -3220,6 +3370,7 @@ var openAiNativeModels = {
3220
3370
  "o3-mini-high": {
3221
3371
  maxTokens: 1e5,
3222
3372
  contextWindow: 2e5,
3373
+ supportsNativeTools: true,
3223
3374
  supportsImages: false,
3224
3375
  supportsPromptCache: true,
3225
3376
  inputPrice: 1.1,
@@ -3231,6 +3382,7 @@ var openAiNativeModels = {
3231
3382
  "o3-mini-low": {
3232
3383
  maxTokens: 1e5,
3233
3384
  contextWindow: 2e5,
3385
+ supportsNativeTools: true,
3234
3386
  supportsImages: false,
3235
3387
  supportsPromptCache: true,
3236
3388
  inputPrice: 1.1,
@@ -3242,6 +3394,7 @@ var openAiNativeModels = {
3242
3394
  o1: {
3243
3395
  maxTokens: 1e5,
3244
3396
  contextWindow: 2e5,
3397
+ supportsNativeTools: true,
3245
3398
  supportsImages: true,
3246
3399
  supportsPromptCache: true,
3247
3400
  inputPrice: 15,
@@ -3252,6 +3405,7 @@ var openAiNativeModels = {
3252
3405
  "o1-preview": {
3253
3406
  maxTokens: 32768,
3254
3407
  contextWindow: 128e3,
3408
+ supportsNativeTools: true,
3255
3409
  supportsImages: true,
3256
3410
  supportsPromptCache: true,
3257
3411
  inputPrice: 15,
@@ -3262,6 +3416,7 @@ var openAiNativeModels = {
3262
3416
  "o1-mini": {
3263
3417
  maxTokens: 65536,
3264
3418
  contextWindow: 128e3,
3419
+ supportsNativeTools: true,
3265
3420
  supportsImages: true,
3266
3421
  supportsPromptCache: true,
3267
3422
  inputPrice: 1.1,
@@ -3272,6 +3427,7 @@ var openAiNativeModels = {
3272
3427
  "gpt-4o": {
3273
3428
  maxTokens: 16384,
3274
3429
  contextWindow: 128e3,
3430
+ supportsNativeTools: true,
3275
3431
  supportsImages: true,
3276
3432
  supportsPromptCache: true,
3277
3433
  inputPrice: 2.5,
@@ -3285,6 +3441,7 @@ var openAiNativeModels = {
3285
3441
  "gpt-4o-mini": {
3286
3442
  maxTokens: 16384,
3287
3443
  contextWindow: 128e3,
3444
+ supportsNativeTools: true,
3288
3445
  supportsImages: true,
3289
3446
  supportsPromptCache: true,
3290
3447
  inputPrice: 0.15,
@@ -3298,6 +3455,7 @@ var openAiNativeModels = {
3298
3455
  "codex-mini-latest": {
3299
3456
  maxTokens: 16384,
3300
3457
  contextWindow: 2e5,
3458
+ supportsNativeTools: true,
3301
3459
  supportsImages: false,
3302
3460
  supportsPromptCache: false,
3303
3461
  inputPrice: 1.5,
@@ -3310,6 +3468,7 @@ var openAiNativeModels = {
3310
3468
  "gpt-5-2025-08-07": {
3311
3469
  maxTokens: 128e3,
3312
3470
  contextWindow: 4e5,
3471
+ supportsNativeTools: true,
3313
3472
  supportsImages: true,
3314
3473
  supportsPromptCache: true,
3315
3474
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3328,6 +3487,7 @@ var openAiNativeModels = {
3328
3487
  "gpt-5-mini-2025-08-07": {
3329
3488
  maxTokens: 128e3,
3330
3489
  contextWindow: 4e5,
3490
+ supportsNativeTools: true,
3331
3491
  supportsImages: true,
3332
3492
  supportsPromptCache: true,
3333
3493
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3346,6 +3506,7 @@ var openAiNativeModels = {
3346
3506
  "gpt-5-nano-2025-08-07": {
3347
3507
  maxTokens: 128e3,
3348
3508
  contextWindow: 4e5,
3509
+ supportsNativeTools: true,
3349
3510
  supportsImages: true,
3350
3511
  supportsPromptCache: true,
3351
3512
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
@@ -3365,7 +3526,8 @@ var openAiModelInfoSaneDefaults = {
3365
3526
  supportsImages: true,
3366
3527
  supportsPromptCache: false,
3367
3528
  inputPrice: 0,
3368
- outputPrice: 0
3529
+ outputPrice: 0,
3530
+ supportsNativeTools: true
3369
3531
  };
3370
3532
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3371
3533
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3409,6 +3571,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3409
3571
  "anthropic/claude-opus-4",
3410
3572
  "anthropic/claude-opus-4.1",
3411
3573
  "anthropic/claude-haiku-4.5",
3574
+ "anthropic/claude-opus-4.5",
3412
3575
  "google/gemini-2.5-flash-preview",
3413
3576
  "google/gemini-2.5-flash-preview:thinking",
3414
3577
  "google/gemini-2.5-flash-preview-05-20",
@@ -3430,6 +3593,7 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3430
3593
  "anthropic/claude-opus-4.1",
3431
3594
  "anthropic/claude-sonnet-4",
3432
3595
  "anthropic/claude-sonnet-4.5",
3596
+ "anthropic/claude-opus-4.5",
3433
3597
  "anthropic/claude-haiku-4.5",
3434
3598
  "google/gemini-2.5-pro-preview",
3435
3599
  "google/gemini-2.5-pro",
@@ -3487,16 +3651,6 @@ var requestyDefaultModelInfo = {
3487
3651
  var import_zod7 = require("zod");
3488
3652
  var rooDefaultModelId = "xai/grok-code-fast-1";
3489
3653
  var rooModels = {};
3490
- var rooModelDefaults = {
3491
- // Add model-specific defaults below.
3492
- // You can configure defaultToolProtocol and other ModelInfo fields for specific model IDs.
3493
- "anthropic/claude-haiku-4.5": {
3494
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3495
- },
3496
- "minimax/minimax-m2:free": {
3497
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
3498
- }
3499
- };
3500
3654
  var RooPricingSchema = import_zod7.z.object({
3501
3655
  input: import_zod7.z.string(),
3502
3656
  output: import_zod7.z.string(),
@@ -3530,6 +3684,7 @@ var sambaNovaModels = {
3530
3684
  contextWindow: 16384,
3531
3685
  supportsImages: false,
3532
3686
  supportsPromptCache: false,
3687
+ supportsNativeTools: true,
3533
3688
  inputPrice: 0.1,
3534
3689
  outputPrice: 0.2,
3535
3690
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3539,6 +3694,7 @@ var sambaNovaModels = {
3539
3694
  contextWindow: 131072,
3540
3695
  supportsImages: false,
3541
3696
  supportsPromptCache: false,
3697
+ supportsNativeTools: true,
3542
3698
  inputPrice: 0.6,
3543
3699
  outputPrice: 1.2,
3544
3700
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3549,6 +3705,7 @@ var sambaNovaModels = {
3549
3705
  supportsImages: false,
3550
3706
  supportsPromptCache: false,
3551
3707
  supportsReasoningBudget: true,
3708
+ supportsNativeTools: true,
3552
3709
  inputPrice: 5,
3553
3710
  outputPrice: 7,
3554
3711
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3558,6 +3715,7 @@ var sambaNovaModels = {
3558
3715
  contextWindow: 32768,
3559
3716
  supportsImages: false,
3560
3717
  supportsPromptCache: false,
3718
+ supportsNativeTools: true,
3561
3719
  inputPrice: 3,
3562
3720
  outputPrice: 4.5,
3563
3721
  description: "DeepSeek V3 model with 32K context window."
@@ -3567,6 +3725,7 @@ var sambaNovaModels = {
3567
3725
  contextWindow: 32768,
3568
3726
  supportsImages: false,
3569
3727
  supportsPromptCache: false,
3728
+ supportsNativeTools: true,
3570
3729
  inputPrice: 3,
3571
3730
  outputPrice: 4.5,
3572
3731
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3585,6 +3744,7 @@ var sambaNovaModels = {
3585
3744
  contextWindow: 131072,
3586
3745
  supportsImages: true,
3587
3746
  supportsPromptCache: false,
3747
+ supportsNativeTools: true,
3588
3748
  inputPrice: 0.63,
3589
3749
  outputPrice: 1.8,
3590
3750
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3603,6 +3763,7 @@ var sambaNovaModels = {
3603
3763
  contextWindow: 8192,
3604
3764
  supportsImages: false,
3605
3765
  supportsPromptCache: false,
3766
+ supportsNativeTools: true,
3606
3767
  inputPrice: 0.4,
3607
3768
  outputPrice: 0.8,
3608
3769
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3612,6 +3773,7 @@ var sambaNovaModels = {
3612
3773
  contextWindow: 131072,
3613
3774
  supportsImages: false,
3614
3775
  supportsPromptCache: false,
3776
+ supportsNativeTools: true,
3615
3777
  inputPrice: 0.22,
3616
3778
  outputPrice: 0.59,
3617
3779
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3634,10 +3796,36 @@ var unboundDefaultModelInfo = {
3634
3796
  // src/providers/vertex.ts
3635
3797
  var vertexDefaultModelId = "claude-sonnet-4-5@20250929";
3636
3798
  var vertexModels = {
3799
+ "gemini-3-pro-preview": {
3800
+ maxTokens: 65536,
3801
+ contextWindow: 1048576,
3802
+ supportsImages: true,
3803
+ supportsNativeTools: true,
3804
+ supportsPromptCache: true,
3805
+ supportsReasoningEffort: ["low", "high"],
3806
+ reasoningEffort: "low",
3807
+ supportsTemperature: true,
3808
+ defaultTemperature: 1,
3809
+ inputPrice: 4,
3810
+ outputPrice: 18,
3811
+ tiers: [
3812
+ {
3813
+ contextWindow: 2e5,
3814
+ inputPrice: 2,
3815
+ outputPrice: 12
3816
+ },
3817
+ {
3818
+ contextWindow: Infinity,
3819
+ inputPrice: 4,
3820
+ outputPrice: 18
3821
+ }
3822
+ ]
3823
+ },
3637
3824
  "gemini-2.5-flash-preview-05-20:thinking": {
3638
3825
  maxTokens: 65535,
3639
3826
  contextWindow: 1048576,
3640
3827
  supportsImages: true,
3828
+ supportsNativeTools: true,
3641
3829
  supportsPromptCache: true,
3642
3830
  inputPrice: 0.15,
3643
3831
  outputPrice: 3.5,
@@ -3649,6 +3837,7 @@ var vertexModels = {
3649
3837
  maxTokens: 65535,
3650
3838
  contextWindow: 1048576,
3651
3839
  supportsImages: true,
3840
+ supportsNativeTools: true,
3652
3841
  supportsPromptCache: true,
3653
3842
  inputPrice: 0.15,
3654
3843
  outputPrice: 0.6
@@ -3657,6 +3846,7 @@ var vertexModels = {
3657
3846
  maxTokens: 64e3,
3658
3847
  contextWindow: 1048576,
3659
3848
  supportsImages: true,
3849
+ supportsNativeTools: true,
3660
3850
  supportsPromptCache: true,
3661
3851
  inputPrice: 0.3,
3662
3852
  outputPrice: 2.5,
@@ -3669,6 +3859,7 @@ var vertexModels = {
3669
3859
  maxTokens: 65535,
3670
3860
  contextWindow: 1048576,
3671
3861
  supportsImages: true,
3862
+ supportsNativeTools: true,
3672
3863
  supportsPromptCache: false,
3673
3864
  inputPrice: 0.15,
3674
3865
  outputPrice: 3.5,
@@ -3680,6 +3871,7 @@ var vertexModels = {
3680
3871
  maxTokens: 65535,
3681
3872
  contextWindow: 1048576,
3682
3873
  supportsImages: true,
3874
+ supportsNativeTools: true,
3683
3875
  supportsPromptCache: false,
3684
3876
  inputPrice: 0.15,
3685
3877
  outputPrice: 0.6
@@ -3688,6 +3880,7 @@ var vertexModels = {
3688
3880
  maxTokens: 65535,
3689
3881
  contextWindow: 1048576,
3690
3882
  supportsImages: true,
3883
+ supportsNativeTools: true,
3691
3884
  supportsPromptCache: true,
3692
3885
  inputPrice: 2.5,
3693
3886
  outputPrice: 15
@@ -3696,6 +3889,7 @@ var vertexModels = {
3696
3889
  maxTokens: 65535,
3697
3890
  contextWindow: 1048576,
3698
3891
  supportsImages: true,
3892
+ supportsNativeTools: true,
3699
3893
  supportsPromptCache: true,
3700
3894
  inputPrice: 2.5,
3701
3895
  outputPrice: 15
@@ -3704,6 +3898,7 @@ var vertexModels = {
3704
3898
  maxTokens: 65535,
3705
3899
  contextWindow: 1048576,
3706
3900
  supportsImages: true,
3901
+ supportsNativeTools: true,
3707
3902
  supportsPromptCache: true,
3708
3903
  inputPrice: 2.5,
3709
3904
  outputPrice: 15,
@@ -3714,6 +3909,7 @@ var vertexModels = {
3714
3909
  maxTokens: 64e3,
3715
3910
  contextWindow: 1048576,
3716
3911
  supportsImages: true,
3912
+ supportsNativeTools: true,
3717
3913
  supportsPromptCache: true,
3718
3914
  inputPrice: 2.5,
3719
3915
  outputPrice: 15,
@@ -3739,6 +3935,7 @@ var vertexModels = {
3739
3935
  maxTokens: 65535,
3740
3936
  contextWindow: 1048576,
3741
3937
  supportsImages: true,
3938
+ supportsNativeTools: true,
3742
3939
  supportsPromptCache: false,
3743
3940
  inputPrice: 0,
3744
3941
  outputPrice: 0
@@ -3747,6 +3944,7 @@ var vertexModels = {
3747
3944
  maxTokens: 8192,
3748
3945
  contextWindow: 2097152,
3749
3946
  supportsImages: true,
3947
+ supportsNativeTools: true,
3750
3948
  supportsPromptCache: false,
3751
3949
  inputPrice: 0,
3752
3950
  outputPrice: 0
@@ -3755,6 +3953,7 @@ var vertexModels = {
3755
3953
  maxTokens: 8192,
3756
3954
  contextWindow: 1048576,
3757
3955
  supportsImages: true,
3956
+ supportsNativeTools: true,
3758
3957
  supportsPromptCache: true,
3759
3958
  inputPrice: 0.15,
3760
3959
  outputPrice: 0.6
@@ -3763,6 +3962,7 @@ var vertexModels = {
3763
3962
  maxTokens: 8192,
3764
3963
  contextWindow: 1048576,
3765
3964
  supportsImages: true,
3965
+ supportsNativeTools: true,
3766
3966
  supportsPromptCache: false,
3767
3967
  inputPrice: 0.075,
3768
3968
  outputPrice: 0.3
@@ -3771,6 +3971,7 @@ var vertexModels = {
3771
3971
  maxTokens: 8192,
3772
3972
  contextWindow: 32768,
3773
3973
  supportsImages: true,
3974
+ supportsNativeTools: true,
3774
3975
  supportsPromptCache: false,
3775
3976
  inputPrice: 0,
3776
3977
  outputPrice: 0
@@ -3779,6 +3980,7 @@ var vertexModels = {
3779
3980
  maxTokens: 8192,
3780
3981
  contextWindow: 1048576,
3781
3982
  supportsImages: true,
3983
+ supportsNativeTools: true,
3782
3984
  supportsPromptCache: true,
3783
3985
  inputPrice: 0.075,
3784
3986
  outputPrice: 0.3
@@ -3787,6 +3989,7 @@ var vertexModels = {
3787
3989
  maxTokens: 8192,
3788
3990
  contextWindow: 2097152,
3789
3991
  supportsImages: true,
3992
+ supportsNativeTools: true,
3790
3993
  supportsPromptCache: false,
3791
3994
  inputPrice: 1.25,
3792
3995
  outputPrice: 5
@@ -3824,6 +4027,17 @@ var vertexModels = {
3824
4027
  cacheReadsPrice: 0.1,
3825
4028
  supportsReasoningBudget: true
3826
4029
  },
4030
+ "claude-opus-4-5@20251101": {
4031
+ maxTokens: 8192,
4032
+ contextWindow: 2e5,
4033
+ supportsImages: true,
4034
+ supportsPromptCache: true,
4035
+ inputPrice: 5,
4036
+ outputPrice: 25,
4037
+ cacheWritesPrice: 6.25,
4038
+ cacheReadsPrice: 0.5,
4039
+ supportsReasoningBudget: true
4040
+ },
3827
4041
  "claude-opus-4-1@20250805": {
3828
4042
  maxTokens: 8192,
3829
4043
  contextWindow: 2e5,
@@ -3921,6 +4135,7 @@ var vertexModels = {
3921
4135
  maxTokens: 64e3,
3922
4136
  contextWindow: 1048576,
3923
4137
  supportsImages: true,
4138
+ supportsNativeTools: true,
3924
4139
  supportsPromptCache: true,
3925
4140
  inputPrice: 0.1,
3926
4141
  outputPrice: 0.4,
@@ -4222,17 +4437,67 @@ var xaiModels = {
4222
4437
  contextWindow: 262144,
4223
4438
  supportsImages: false,
4224
4439
  supportsPromptCache: true,
4440
+ supportsNativeTools: true,
4225
4441
  inputPrice: 0.2,
4226
4442
  outputPrice: 1.5,
4227
4443
  cacheWritesPrice: 0.02,
4228
4444
  cacheReadsPrice: 0.02,
4229
4445
  description: "xAI's Grok Code Fast model with 256K context window"
4230
4446
  },
4447
+ "grok-4-1-fast-reasoning": {
4448
+ maxTokens: 65536,
4449
+ contextWindow: 2e6,
4450
+ supportsImages: true,
4451
+ supportsPromptCache: true,
4452
+ supportsNativeTools: true,
4453
+ inputPrice: 0.2,
4454
+ outputPrice: 0.5,
4455
+ cacheWritesPrice: 0.05,
4456
+ cacheReadsPrice: 0.05,
4457
+ description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning"
4458
+ },
4459
+ "grok-4-1-fast-non-reasoning": {
4460
+ maxTokens: 65536,
4461
+ contextWindow: 2e6,
4462
+ supportsImages: true,
4463
+ supportsPromptCache: true,
4464
+ supportsNativeTools: true,
4465
+ inputPrice: 0.2,
4466
+ outputPrice: 0.5,
4467
+ cacheWritesPrice: 0.05,
4468
+ cacheReadsPrice: 0.05,
4469
+ description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling"
4470
+ },
4471
+ "grok-4-fast-reasoning": {
4472
+ maxTokens: 65536,
4473
+ contextWindow: 2e6,
4474
+ supportsImages: true,
4475
+ supportsPromptCache: true,
4476
+ supportsNativeTools: true,
4477
+ inputPrice: 0.2,
4478
+ outputPrice: 0.5,
4479
+ cacheWritesPrice: 0.05,
4480
+ cacheReadsPrice: 0.05,
4481
+ description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning"
4482
+ },
4483
+ "grok-4-fast-non-reasoning": {
4484
+ maxTokens: 65536,
4485
+ contextWindow: 2e6,
4486
+ supportsImages: true,
4487
+ supportsPromptCache: true,
4488
+ supportsNativeTools: true,
4489
+ inputPrice: 0.2,
4490
+ outputPrice: 0.5,
4491
+ cacheWritesPrice: 0.05,
4492
+ cacheReadsPrice: 0.05,
4493
+ description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling"
4494
+ },
4231
4495
  "grok-4": {
4232
4496
  maxTokens: 8192,
4233
4497
  contextWindow: 256e3,
4234
4498
  supportsImages: true,
4235
4499
  supportsPromptCache: true,
4500
+ supportsNativeTools: true,
4236
4501
  inputPrice: 3,
4237
4502
  outputPrice: 15,
4238
4503
  cacheWritesPrice: 0.75,
@@ -4244,6 +4509,7 @@ var xaiModels = {
4244
4509
  contextWindow: 131072,
4245
4510
  supportsImages: false,
4246
4511
  supportsPromptCache: true,
4512
+ supportsNativeTools: true,
4247
4513
  inputPrice: 3,
4248
4514
  outputPrice: 15,
4249
4515
  cacheWritesPrice: 0.75,
@@ -4255,6 +4521,7 @@ var xaiModels = {
4255
4521
  contextWindow: 131072,
4256
4522
  supportsImages: false,
4257
4523
  supportsPromptCache: true,
4524
+ supportsNativeTools: true,
4258
4525
  inputPrice: 5,
4259
4526
  outputPrice: 25,
4260
4527
  cacheWritesPrice: 1.25,
@@ -4266,6 +4533,7 @@ var xaiModels = {
4266
4533
  contextWindow: 131072,
4267
4534
  supportsImages: false,
4268
4535
  supportsPromptCache: true,
4536
+ supportsNativeTools: true,
4269
4537
  inputPrice: 0.3,
4270
4538
  outputPrice: 0.5,
4271
4539
  cacheWritesPrice: 0.07,
@@ -4278,6 +4546,7 @@ var xaiModels = {
4278
4546
  contextWindow: 131072,
4279
4547
  supportsImages: false,
4280
4548
  supportsPromptCache: true,
4549
+ supportsNativeTools: true,
4281
4550
  inputPrice: 0.6,
4282
4551
  outputPrice: 4,
4283
4552
  cacheWritesPrice: 0.15,
@@ -4290,6 +4559,7 @@ var xaiModels = {
4290
4559
  contextWindow: 131072,
4291
4560
  supportsImages: false,
4292
4561
  supportsPromptCache: false,
4562
+ supportsNativeTools: true,
4293
4563
  inputPrice: 2,
4294
4564
  outputPrice: 10,
4295
4565
  description: "xAI's Grok-2 model (version 1212) with 128K context window"
@@ -4299,6 +4569,7 @@ var xaiModels = {
4299
4569
  contextWindow: 32768,
4300
4570
  supportsImages: true,
4301
4571
  supportsPromptCache: false,
4572
+ supportsNativeTools: true,
4302
4573
  inputPrice: 2,
4303
4574
  outputPrice: 10,
4304
4575
  description: "xAI's Grok-2 Vision model (version 1212) with image support and 32K context window"
@@ -4407,6 +4678,7 @@ var internationalZAiModels = {
4407
4678
  contextWindow: 131072,
4408
4679
  supportsImages: false,
4409
4680
  supportsPromptCache: true,
4681
+ supportsNativeTools: true,
4410
4682
  supportsReasoningBinary: true,
4411
4683
  inputPrice: 0.6,
4412
4684
  outputPrice: 2.2,
@@ -4419,6 +4691,7 @@ var internationalZAiModels = {
4419
4691
  contextWindow: 131072,
4420
4692
  supportsImages: false,
4421
4693
  supportsPromptCache: true,
4694
+ supportsNativeTools: true,
4422
4695
  inputPrice: 0.2,
4423
4696
  outputPrice: 1.1,
4424
4697
  cacheWritesPrice: 0,
@@ -4430,6 +4703,7 @@ var internationalZAiModels = {
4430
4703
  contextWindow: 131072,
4431
4704
  supportsImages: false,
4432
4705
  supportsPromptCache: true,
4706
+ supportsNativeTools: true,
4433
4707
  inputPrice: 2.2,
4434
4708
  outputPrice: 8.9,
4435
4709
  cacheWritesPrice: 0,
@@ -4441,6 +4715,7 @@ var internationalZAiModels = {
4441
4715
  contextWindow: 131072,
4442
4716
  supportsImages: false,
4443
4717
  supportsPromptCache: true,
4718
+ supportsNativeTools: true,
4444
4719
  inputPrice: 1.1,
4445
4720
  outputPrice: 4.5,
4446
4721
  cacheWritesPrice: 0,
@@ -4452,6 +4727,7 @@ var internationalZAiModels = {
4452
4727
  contextWindow: 131072,
4453
4728
  supportsImages: false,
4454
4729
  supportsPromptCache: true,
4730
+ supportsNativeTools: true,
4455
4731
  inputPrice: 0,
4456
4732
  outputPrice: 0,
4457
4733
  cacheWritesPrice: 0,
@@ -4463,6 +4739,7 @@ var internationalZAiModels = {
4463
4739
  contextWindow: 131072,
4464
4740
  supportsImages: true,
4465
4741
  supportsPromptCache: true,
4742
+ supportsNativeTools: true,
4466
4743
  inputPrice: 0.6,
4467
4744
  outputPrice: 1.8,
4468
4745
  cacheWritesPrice: 0,
@@ -4474,6 +4751,7 @@ var internationalZAiModels = {
4474
4751
  contextWindow: 2e5,
4475
4752
  supportsImages: false,
4476
4753
  supportsPromptCache: true,
4754
+ supportsNativeTools: true,
4477
4755
  supportsReasoningBinary: true,
4478
4756
  inputPrice: 0.6,
4479
4757
  outputPrice: 2.2,
@@ -4486,6 +4764,7 @@ var internationalZAiModels = {
4486
4764
  contextWindow: 131072,
4487
4765
  supportsImages: false,
4488
4766
  supportsPromptCache: false,
4767
+ supportsNativeTools: true,
4489
4768
  inputPrice: 0.1,
4490
4769
  outputPrice: 0.1,
4491
4770
  cacheWritesPrice: 0,
@@ -4500,6 +4779,7 @@ var mainlandZAiModels = {
4500
4779
  contextWindow: 131072,
4501
4780
  supportsImages: false,
4502
4781
  supportsPromptCache: true,
4782
+ supportsNativeTools: true,
4503
4783
  supportsReasoningBinary: true,
4504
4784
  inputPrice: 0.29,
4505
4785
  outputPrice: 1.14,
@@ -4512,6 +4792,7 @@ var mainlandZAiModels = {
4512
4792
  contextWindow: 131072,
4513
4793
  supportsImages: false,
4514
4794
  supportsPromptCache: true,
4795
+ supportsNativeTools: true,
4515
4796
  inputPrice: 0.1,
4516
4797
  outputPrice: 0.6,
4517
4798
  cacheWritesPrice: 0,
@@ -4523,6 +4804,7 @@ var mainlandZAiModels = {
4523
4804
  contextWindow: 131072,
4524
4805
  supportsImages: false,
4525
4806
  supportsPromptCache: true,
4807
+ supportsNativeTools: true,
4526
4808
  inputPrice: 0.29,
4527
4809
  outputPrice: 1.14,
4528
4810
  cacheWritesPrice: 0,
@@ -4534,6 +4816,7 @@ var mainlandZAiModels = {
4534
4816
  contextWindow: 131072,
4535
4817
  supportsImages: false,
4536
4818
  supportsPromptCache: true,
4819
+ supportsNativeTools: true,
4537
4820
  inputPrice: 0.1,
4538
4821
  outputPrice: 0.6,
4539
4822
  cacheWritesPrice: 0,
@@ -4545,6 +4828,7 @@ var mainlandZAiModels = {
4545
4828
  contextWindow: 131072,
4546
4829
  supportsImages: false,
4547
4830
  supportsPromptCache: true,
4831
+ supportsNativeTools: true,
4548
4832
  inputPrice: 0,
4549
4833
  outputPrice: 0,
4550
4834
  cacheWritesPrice: 0,
@@ -4556,6 +4840,7 @@ var mainlandZAiModels = {
4556
4840
  contextWindow: 131072,
4557
4841
  supportsImages: true,
4558
4842
  supportsPromptCache: true,
4843
+ supportsNativeTools: true,
4559
4844
  inputPrice: 0.29,
4560
4845
  outputPrice: 0.93,
4561
4846
  cacheWritesPrice: 0,
@@ -4567,6 +4852,7 @@ var mainlandZAiModels = {
4567
4852
  contextWindow: 204800,
4568
4853
  supportsImages: false,
4569
4854
  supportsPromptCache: true,
4855
+ supportsNativeTools: true,
4570
4856
  supportsReasoningBinary: true,
4571
4857
  inputPrice: 0.29,
4572
4858
  outputPrice: 1.14,
@@ -4596,6 +4882,7 @@ var deepInfraDefaultModelInfo = {
4596
4882
  contextWindow: 262144,
4597
4883
  supportsImages: false,
4598
4884
  supportsPromptCache: false,
4885
+ supportsNativeTools: true,
4599
4886
  inputPrice: 0.3,
4600
4887
  outputPrice: 1.2,
4601
4888
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -4609,11 +4896,12 @@ var minimaxModels = {
4609
4896
  contextWindow: 192e3,
4610
4897
  supportsImages: false,
4611
4898
  supportsPromptCache: true,
4899
+ supportsNativeTools: true,
4900
+ preserveReasoning: true,
4612
4901
  inputPrice: 0.3,
4613
4902
  outputPrice: 1.2,
4614
4903
  cacheWritesPrice: 0.375,
4615
4904
  cacheReadsPrice: 0.03,
4616
- preserveReasoning: true,
4617
4905
  description: "MiniMax M2, a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4618
4906
  },
4619
4907
  "MiniMax-M2-Stable": {
@@ -4621,14 +4909,17 @@ var minimaxModels = {
4621
4909
  contextWindow: 192e3,
4622
4910
  supportsImages: false,
4623
4911
  supportsPromptCache: true,
4912
+ supportsNativeTools: true,
4913
+ preserveReasoning: true,
4624
4914
  inputPrice: 0.3,
4625
4915
  outputPrice: 1.2,
4626
4916
  cacheWritesPrice: 0.375,
4627
4917
  cacheReadsPrice: 0.03,
4628
- preserveReasoning: true,
4629
4918
  description: "MiniMax M2 Stable (High Concurrency, Commercial Use), a model born for Agents and code, featuring Top-tier Coding Capabilities, Powerful Agentic Performance, and Ultimate Cost-Effectiveness & Speed."
4630
4919
  }
4631
4920
  };
4921
+ var minimaxDefaultModelInfo = minimaxModels[minimaxDefaultModelId];
4922
+ var MINIMAX_DEFAULT_MAX_TOKENS = 16384;
4632
4923
  var MINIMAX_DEFAULT_TEMPERATURE = 1;
4633
4924
 
4634
4925
  // src/providers/index.ts
@@ -4652,6 +4943,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
4652
4943
  return "meta-llama/Llama-3.3-70B-Instruct";
4653
4944
  case "chutes":
4654
4945
  return chutesDefaultModelId;
4946
+ case "baseten":
4947
+ return basetenDefaultModelId;
4655
4948
  case "bedrock":
4656
4949
  return bedrockDefaultModelId;
4657
4950
  case "vertex":
@@ -4745,6 +5038,7 @@ var providerNames = [
4745
5038
  ...fauxProviders,
4746
5039
  "anthropic",
4747
5040
  "bedrock",
5041
+ "baseten",
4748
5042
  "cerebras",
4749
5043
  "claude-code",
4750
5044
  "doubao",
@@ -4984,6 +5278,9 @@ var vercelAiGatewaySchema = baseProviderSettingsSchema.extend({
4984
5278
  vercelAiGatewayApiKey: import_zod8.z.string().optional(),
4985
5279
  vercelAiGatewayModelId: import_zod8.z.string().optional()
4986
5280
  });
5281
+ var basetenSchema = apiModelIdProviderModelSchema.extend({
5282
+ basetenApiKey: import_zod8.z.string().optional()
5283
+ });
4987
5284
  var defaultSchema = import_zod8.z.object({
4988
5285
  apiProvider: import_zod8.z.undefined()
4989
5286
  });
@@ -5013,6 +5310,7 @@ var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiP
5013
5310
  fakeAiSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("fake-ai") })),
5014
5311
  xaiSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("xai") })),
5015
5312
  groqSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("groq") })),
5313
+ basetenSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("baseten") })),
5016
5314
  huggingFaceSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("huggingface") })),
5017
5315
  chutesSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("chutes") })),
5018
5316
  litellmSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("litellm") })),
@@ -5054,6 +5352,7 @@ var providerSettingsSchema = import_zod8.z.object({
5054
5352
  ...fakeAiSchema.shape,
5055
5353
  ...xaiSchema.shape,
5056
5354
  ...groqSchema.shape,
5355
+ ...basetenSchema.shape,
5057
5356
  ...huggingFaceSchema.shape,
5058
5357
  ...chutesSchema.shape,
5059
5358
  ...litellmSchema.shape,
@@ -5117,6 +5416,7 @@ var modelIdKeysByProvider = {
5117
5416
  requesty: "requestyModelId",
5118
5417
  xai: "apiModelId",
5119
5418
  groq: "apiModelId",
5419
+ baseten: "apiModelId",
5120
5420
  chutes: "apiModelId",
5121
5421
  litellm: "litellmModelId",
5122
5422
  huggingface: "huggingFaceModelId",
@@ -5129,7 +5429,7 @@ var modelIdKeysByProvider = {
5129
5429
  roo: "apiModelId",
5130
5430
  "vercel-ai-gateway": "vercelAiGatewayModelId"
5131
5431
  };
5132
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock"];
5432
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
5133
5433
  var getApiProtocol = (provider, modelId) => {
5134
5434
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
5135
5435
  return "anthropic";
@@ -5224,7 +5524,8 @@ var MODELS_BY_PROVIDER = {
5224
5524
  models: Object.keys(vscodeLlmModels)
5225
5525
  },
5226
5526
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
5227
- zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
5527
+ zai: { id: "zai", label: "Z.ai", models: Object.keys(internationalZAiModels) },
5528
+ baseten: { id: "baseten", label: "Baseten", models: Object.keys(basetenModels) },
5228
5529
  // Dynamic providers; models pulled from remote APIs.
5229
5530
  glama: { id: "glama", label: "Glama", models: [] },
5230
5531
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
@@ -5256,7 +5557,18 @@ var historyItemSchema = import_zod9.z.object({
5256
5557
  totalCost: import_zod9.z.number(),
5257
5558
  size: import_zod9.z.number().optional(),
5258
5559
  workspace: import_zod9.z.string().optional(),
5259
- mode: import_zod9.z.string().optional()
5560
+ mode: import_zod9.z.string().optional(),
5561
+ status: import_zod9.z.enum(["active", "completed", "delegated"]).optional(),
5562
+ delegatedToId: import_zod9.z.string().optional(),
5563
+ // Last child this parent delegated to
5564
+ childIds: import_zod9.z.array(import_zod9.z.string()).optional(),
5565
+ // All children spawned by this task
5566
+ awaitingChildId: import_zod9.z.string().optional(),
5567
+ // Child currently awaited (set when delegated)
5568
+ completedByChildId: import_zod9.z.string().optional(),
5569
+ // Child that completed and resumed this parent
5570
+ completionResultSummary: import_zod9.z.string().optional()
5571
+ // Summary from completed child
5260
5572
  });
5261
5573
 
5262
5574
  // src/experiment.ts
@@ -5267,7 +5579,7 @@ var experimentIds = [
5267
5579
  "preventFocusDisruption",
5268
5580
  "imageGeneration",
5269
5581
  "runSlashCommand",
5270
- "nativeToolCalling"
5582
+ "multipleNativeToolCalls"
5271
5583
  ];
5272
5584
  var experimentIdsSchema = import_zod10.z.enum(experimentIds);
5273
5585
  var experimentsSchema = import_zod10.z.object({
@@ -5276,7 +5588,7 @@ var experimentsSchema = import_zod10.z.object({
5276
5588
  preventFocusDisruption: import_zod10.z.boolean().optional(),
5277
5589
  imageGeneration: import_zod10.z.boolean().optional(),
5278
5590
  runSlashCommand: import_zod10.z.boolean().optional(),
5279
- nativeToolCalling: import_zod10.z.boolean().optional()
5591
+ multipleNativeToolCalls: import_zod10.z.boolean().optional()
5280
5592
  });
5281
5593
 
5282
5594
  // src/telemetry.ts
@@ -5326,6 +5638,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
5326
5638
  TelemetryEventName2["CONSECUTIVE_MISTAKE_ERROR"] = "Consecutive Mistake Error";
5327
5639
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
5328
5640
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
5641
+ TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5329
5642
  return TelemetryEventName2;
5330
5643
  })(TelemetryEventName || {});
5331
5644
  var staticAppPropertiesSchema = import_zod11.z.object({
@@ -5409,6 +5722,7 @@ var rooCodeTelemetryEventSchema = import_zod11.z.discriminatedUnion("type", [
5409
5722
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
5410
5723
  "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */,
5411
5724
  "Code Index Error" /* CODE_INDEX_ERROR */,
5725
+ "Model Cache Empty Response" /* MODEL_CACHE_EMPTY_RESPONSE */,
5412
5726
  "Context Condensed" /* CONTEXT_CONDENSED */,
5413
5727
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
5414
5728
  "Tab Shown" /* TAB_SHOWN */,
@@ -5569,8 +5883,6 @@ var terminalActionIds = ["terminalAddToContext", "terminalFixCommand", "terminal
5569
5883
  var commandIds = [
5570
5884
  "activationCompleted",
5571
5885
  "plusButtonClicked",
5572
- "promptsButtonClicked",
5573
- "mcpButtonClicked",
5574
5886
  "historyButtonClicked",
5575
5887
  "marketplaceButtonClicked",
5576
5888
  "popoutButtonClicked",
@@ -5627,6 +5939,7 @@ var globalSettingsSchema = import_zod14.z.object({
5627
5939
  taskHistory: import_zod14.z.array(historyItemSchema).optional(),
5628
5940
  dismissedUpsells: import_zod14.z.array(import_zod14.z.string()).optional(),
5629
5941
  // Image generation settings (experimental) - flattened for simplicity
5942
+ imageGenerationProvider: import_zod14.z.enum(["openrouter", "roo"]).optional(),
5630
5943
  openRouterImageApiKey: import_zod14.z.string().optional(),
5631
5944
  openRouterImageGenerationSelectedModel: import_zod14.z.string().optional(),
5632
5945
  condensingApiConfigId: import_zod14.z.string().optional(),
@@ -5778,7 +6091,8 @@ var SECRET_STATE_KEYS = [
5778
6091
  "fireworksApiKey",
5779
6092
  "featherlessApiKey",
5780
6093
  "ioIntelligenceApiKey",
5781
- "vercelAiGatewayApiKey"
6094
+ "vercelAiGatewayApiKey",
6095
+ "basetenApiKey"
5782
6096
  ];
5783
6097
  var GLOBAL_SECRET_KEYS = [
5784
6098
  "openRouterImageApiKey"
@@ -6043,6 +6357,9 @@ var ExtensionBridgeEventName = ((ExtensionBridgeEventName2) => {
6043
6357
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskPaused"] = "taskPaused" /* TaskPaused */] = "TaskPaused";
6044
6358
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskUnpaused"] = "taskUnpaused" /* TaskUnpaused */] = "TaskUnpaused";
6045
6359
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskSpawned"] = "taskSpawned" /* TaskSpawned */] = "TaskSpawned";
6360
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegated"] = "taskDelegated" /* TaskDelegated */] = "TaskDelegated";
6361
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegationCompleted"] = "taskDelegationCompleted" /* TaskDelegationCompleted */] = "TaskDelegationCompleted";
6362
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegationResumed"] = "taskDelegationResumed" /* TaskDelegationResumed */] = "TaskDelegationResumed";
6046
6363
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskUserMessage"] = "taskUserMessage" /* TaskUserMessage */] = "TaskUserMessage";
6047
6364
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated" /* TaskTokenUsageUpdated */] = "TaskTokenUsageUpdated";
6048
6365
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["ModeChanged"] = "modeChanged" /* ModeChanged */] = "ModeChanged";
@@ -6118,6 +6435,21 @@ var extensionBridgeEventSchema = import_zod16.z.discriminatedUnion("type", [
6118
6435
  instance: extensionInstanceSchema,
6119
6436
  timestamp: import_zod16.z.number()
6120
6437
  }),
6438
+ import_zod16.z.object({
6439
+ type: import_zod16.z.literal(ExtensionBridgeEventName.TaskDelegated),
6440
+ instance: extensionInstanceSchema,
6441
+ timestamp: import_zod16.z.number()
6442
+ }),
6443
+ import_zod16.z.object({
6444
+ type: import_zod16.z.literal(ExtensionBridgeEventName.TaskDelegationCompleted),
6445
+ instance: extensionInstanceSchema,
6446
+ timestamp: import_zod16.z.number()
6447
+ }),
6448
+ import_zod16.z.object({
6449
+ type: import_zod16.z.literal(ExtensionBridgeEventName.TaskDelegationResumed),
6450
+ instance: extensionInstanceSchema,
6451
+ timestamp: import_zod16.z.number()
6452
+ }),
6121
6453
  import_zod16.z.object({
6122
6454
  type: import_zod16.z.literal(ExtensionBridgeEventName.TaskUserMessage),
6123
6455
  instance: extensionInstanceSchema,
@@ -6306,6 +6638,30 @@ var followUpDataSchema = import_zod17.z.object({
6306
6638
  suggest: import_zod17.z.array(suggestionItemSchema).optional()
6307
6639
  });
6308
6640
 
6641
+ // src/image-generation.ts
6642
+ var IMAGE_GENERATION_MODELS = [
6643
+ // OpenRouter models
6644
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image", provider: "openrouter" },
6645
+ { value: "google/gemini-3-pro-image-preview", label: "Gemini 3 Pro Image Preview", provider: "openrouter" },
6646
+ { value: "openai/gpt-5-image", label: "GPT-5 Image", provider: "openrouter" },
6647
+ { value: "openai/gpt-5-image-mini", label: "GPT-5 Image Mini", provider: "openrouter" },
6648
+ { value: "black-forest-labs/flux.2-flex", label: "Black Forest Labs FLUX.2 Flex", provider: "openrouter" },
6649
+ { value: "black-forest-labs/flux.2-pro", label: "Black Forest Labs FLUX.2 Pro", provider: "openrouter" },
6650
+ // Roo Code Cloud models
6651
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image", provider: "roo" },
6652
+ { value: "google/gemini-3-pro-image", label: "Gemini 3 Pro Image", provider: "roo" },
6653
+ {
6654
+ value: "bfl/flux-2-pro:free",
6655
+ label: "Black Forest Labs FLUX.2 Pro (Free)",
6656
+ provider: "roo",
6657
+ apiMethod: "images_api"
6658
+ }
6659
+ ];
6660
+ var IMAGE_GENERATION_MODEL_IDS = IMAGE_GENERATION_MODELS.map((m) => m.value);
6661
+ function getImageGenerationProvider(explicitProvider, hasExistingModel) {
6662
+ return explicitProvider !== void 0 ? explicitProvider : hasExistingModel ? "openrouter" : "roo";
6663
+ }
6664
+
6309
6665
  // src/ipc.ts
6310
6666
  var import_zod18 = require("zod");
6311
6667
  var IpcMessageType = /* @__PURE__ */ ((IpcMessageType2) => {
@@ -6494,12 +6850,15 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6494
6850
  HUGGINGFACE_SLIDER_MIN,
6495
6851
  HUGGINGFACE_SLIDER_STEP,
6496
6852
  HUGGINGFACE_TEMPERATURE_MAX_VALUE,
6853
+ IMAGE_GENERATION_MODELS,
6854
+ IMAGE_GENERATION_MODEL_IDS,
6497
6855
  INSTANCE_TTL_SECONDS,
6498
6856
  IO_INTELLIGENCE_CACHE_DURATION,
6499
6857
  IpcMessageType,
6500
6858
  IpcOrigin,
6501
6859
  LMSTUDIO_DEFAULT_TEMPERATURE,
6502
6860
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6861
+ MINIMAX_DEFAULT_MAX_TOKENS,
6503
6862
  MINIMAX_DEFAULT_TEMPERATURE,
6504
6863
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
6505
6864
  MISTRAL_DEFAULT_TEMPERATURE,
@@ -6537,6 +6896,8 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6537
6896
  anthropicModels,
6538
6897
  appPropertiesSchema,
6539
6898
  azureOpenAiDefaultApiVersion,
6899
+ basetenDefaultModelId,
6900
+ basetenModels,
6540
6901
  bedrockDefaultModelId,
6541
6902
  bedrockDefaultPromptRouterModelId,
6542
6903
  bedrockModels,
@@ -6592,6 +6953,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6592
6953
  getApiProtocol,
6593
6954
  getClaudeCodeModelId,
6594
6955
  getEffectiveProtocol,
6956
+ getImageGenerationProvider,
6595
6957
  getModelId,
6596
6958
  getProviderDefaultModelId,
6597
6959
  gitPropertiesSchema,
@@ -6645,6 +7007,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6645
7007
  mcpMarketplaceItemSchema,
6646
7008
  mcpParameterSchema,
6647
7009
  minimaxDefaultModelId,
7010
+ minimaxDefaultModelInfo,
6648
7011
  minimaxModels,
6649
7012
  mistralDefaultModelId,
6650
7013
  mistralModels,
@@ -6694,7 +7057,6 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6694
7057
  rooCodeSettingsSchema,
6695
7058
  rooCodeTelemetryEventSchema,
6696
7059
  rooDefaultModelId,
6697
- rooModelDefaults,
6698
7060
  rooModels,
6699
7061
  sambaNovaDefaultModelId,
6700
7062
  sambaNovaModels,