@roo-code/types 1.87.0 → 1.88.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -106,6 +106,8 @@ __export(index_exports, {
106
106
  anthropicModels: () => anthropicModels,
107
107
  appPropertiesSchema: () => appPropertiesSchema,
108
108
  azureOpenAiDefaultApiVersion: () => azureOpenAiDefaultApiVersion,
109
+ basetenDefaultModelId: () => basetenDefaultModelId,
110
+ basetenModels: () => basetenModels,
109
111
  bedrockDefaultModelId: () => bedrockDefaultModelId,
110
112
  bedrockDefaultPromptRouterModelId: () => bedrockDefaultPromptRouterModelId,
111
113
  bedrockModels: () => bedrockModels,
@@ -161,6 +163,7 @@ __export(index_exports, {
161
163
  getApiProtocol: () => getApiProtocol,
162
164
  getClaudeCodeModelId: () => getClaudeCodeModelId,
163
165
  getEffectiveProtocol: () => getEffectiveProtocol,
166
+ getImageGenerationProvider: () => getImageGenerationProvider,
164
167
  getModelId: () => getModelId,
165
168
  getProviderDefaultModelId: () => getProviderDefaultModelId,
166
169
  gitPropertiesSchema: () => gitPropertiesSchema,
@@ -381,6 +384,7 @@ var clineSays = [
381
384
  "shell_integration_warning",
382
385
  "browser_action",
383
386
  "browser_action_result",
387
+ "browser_session_status",
384
388
  "mcp_server_request_started",
385
389
  "mcp_server_response",
386
390
  "subtask_result",
@@ -445,6 +449,8 @@ var toolNames = [
445
449
  "write_to_file",
446
450
  "apply_diff",
447
451
  "insert_content",
452
+ "search_and_replace",
453
+ "apply_patch",
448
454
  "search_files",
449
455
  "list_files",
450
456
  "list_code_definition_names",
@@ -495,6 +501,9 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
495
501
  RooCodeEventName2["TaskPaused"] = "taskPaused";
496
502
  RooCodeEventName2["TaskUnpaused"] = "taskUnpaused";
497
503
  RooCodeEventName2["TaskSpawned"] = "taskSpawned";
504
+ RooCodeEventName2["TaskDelegated"] = "taskDelegated";
505
+ RooCodeEventName2["TaskDelegationCompleted"] = "taskDelegationCompleted";
506
+ RooCodeEventName2["TaskDelegationResumed"] = "taskDelegationResumed";
498
507
  RooCodeEventName2["Message"] = "message";
499
508
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
500
509
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
@@ -528,6 +537,26 @@ var rooCodeEventsSchema = import_zod3.z.object({
528
537
  ["taskPaused" /* TaskPaused */]: import_zod3.z.tuple([import_zod3.z.string()]),
529
538
  ["taskUnpaused" /* TaskUnpaused */]: import_zod3.z.tuple([import_zod3.z.string()]),
530
539
  ["taskSpawned" /* TaskSpawned */]: import_zod3.z.tuple([import_zod3.z.string(), import_zod3.z.string()]),
540
+ ["taskDelegated" /* TaskDelegated */]: import_zod3.z.tuple([
541
+ import_zod3.z.string(),
542
+ // parentTaskId
543
+ import_zod3.z.string()
544
+ // childTaskId
545
+ ]),
546
+ ["taskDelegationCompleted" /* TaskDelegationCompleted */]: import_zod3.z.tuple([
547
+ import_zod3.z.string(),
548
+ // parentTaskId
549
+ import_zod3.z.string(),
550
+ // childTaskId
551
+ import_zod3.z.string()
552
+ // completionResultSummary
553
+ ]),
554
+ ["taskDelegationResumed" /* TaskDelegationResumed */]: import_zod3.z.tuple([
555
+ import_zod3.z.string(),
556
+ // parentTaskId
557
+ import_zod3.z.string()
558
+ // childTaskId
559
+ ]),
531
560
  ["message" /* Message */]: import_zod3.z.tuple([
532
561
  import_zod3.z.object({
533
562
  taskId: import_zod3.z.string(),
@@ -612,6 +641,21 @@ var taskEventSchema = import_zod3.z.discriminatedUnion("eventName", [
612
641
  payload: rooCodeEventsSchema.shape["taskSpawned" /* TaskSpawned */],
613
642
  taskId: import_zod3.z.number().optional()
614
643
  }),
644
+ import_zod3.z.object({
645
+ eventName: import_zod3.z.literal("taskDelegated" /* TaskDelegated */),
646
+ payload: rooCodeEventsSchema.shape["taskDelegated" /* TaskDelegated */],
647
+ taskId: import_zod3.z.number().optional()
648
+ }),
649
+ import_zod3.z.object({
650
+ eventName: import_zod3.z.literal("taskDelegationCompleted" /* TaskDelegationCompleted */),
651
+ payload: rooCodeEventsSchema.shape["taskDelegationCompleted" /* TaskDelegationCompleted */],
652
+ taskId: import_zod3.z.number().optional()
653
+ }),
654
+ import_zod3.z.object({
655
+ eventName: import_zod3.z.literal("taskDelegationResumed" /* TaskDelegationResumed */),
656
+ payload: rooCodeEventsSchema.shape["taskDelegationResumed" /* TaskDelegationResumed */],
657
+ taskId: import_zod3.z.number().optional()
658
+ }),
615
659
  // Task Execution
616
660
  import_zod3.z.object({
617
661
  eventName: import_zod3.z.literal("message" /* Message */),
@@ -730,6 +774,13 @@ var modelInfoSchema = import_zod5.z.object({
730
774
  supportsNativeTools: import_zod5.z.boolean().optional(),
731
775
  // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
732
776
  defaultToolProtocol: import_zod5.z.enum(["xml", "native"]).optional(),
777
+ // Exclude specific native tools from being available (only applies to native protocol)
778
+ // These tools will be removed from the set of tools available to the model
779
+ excludedTools: import_zod5.z.array(import_zod5.z.string()).optional(),
780
+ // Include specific native tools (only applies to native protocol)
781
+ // These tools will be added if they belong to an allowed group in the current mode
782
+ // Cannot force-add tools from groups the mode doesn't allow
783
+ includedTools: import_zod5.z.array(import_zod5.z.string()).optional(),
733
784
  /**
734
785
  * Service tiers with pricing information.
735
786
  * Each tier can have a name (for OpenAI service tiers) and pricing overrides.
@@ -763,7 +814,16 @@ var CODEBASE_INDEX_DEFAULTS = {
763
814
  var codebaseIndexConfigSchema = import_zod6.z.object({
764
815
  codebaseIndexEnabled: import_zod6.z.boolean().optional(),
765
816
  codebaseIndexQdrantUrl: import_zod6.z.string().optional(),
766
- codebaseIndexEmbedderProvider: import_zod6.z.enum(["openai", "ollama", "openai-compatible", "gemini", "mistral", "vercel-ai-gateway", "openrouter"]).optional(),
817
+ codebaseIndexEmbedderProvider: import_zod6.z.enum([
818
+ "openai",
819
+ "ollama",
820
+ "openai-compatible",
821
+ "gemini",
822
+ "mistral",
823
+ "vercel-ai-gateway",
824
+ "bedrock",
825
+ "openrouter"
826
+ ]).optional(),
767
827
  codebaseIndexEmbedderBaseUrl: import_zod6.z.string().optional(),
768
828
  codebaseIndexEmbedderModelId: import_zod6.z.string().optional(),
769
829
  codebaseIndexEmbedderModelDimension: import_zod6.z.number().optional(),
@@ -771,7 +831,10 @@ var codebaseIndexConfigSchema = import_zod6.z.object({
771
831
  codebaseIndexSearchMaxResults: import_zod6.z.number().min(CODEBASE_INDEX_DEFAULTS.MIN_SEARCH_RESULTS).max(CODEBASE_INDEX_DEFAULTS.MAX_SEARCH_RESULTS).optional(),
772
832
  // OpenAI Compatible specific fields
773
833
  codebaseIndexOpenAiCompatibleBaseUrl: import_zod6.z.string().optional(),
774
- codebaseIndexOpenAiCompatibleModelDimension: import_zod6.z.number().optional()
834
+ codebaseIndexOpenAiCompatibleModelDimension: import_zod6.z.number().optional(),
835
+ // Bedrock specific fields
836
+ codebaseIndexBedrockRegion: import_zod6.z.string().optional(),
837
+ codebaseIndexBedrockProfile: import_zod6.z.string().optional()
775
838
  });
776
839
  var codebaseIndexModelsSchema = import_zod6.z.object({
777
840
  openai: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
@@ -780,7 +843,8 @@ var codebaseIndexModelsSchema = import_zod6.z.object({
780
843
  gemini: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
781
844
  mistral: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
782
845
  "vercel-ai-gateway": import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
783
- openrouter: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional()
846
+ openrouter: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional(),
847
+ bedrock: import_zod6.z.record(import_zod6.z.string(), import_zod6.z.object({ dimension: import_zod6.z.number() })).optional()
784
848
  });
785
849
  var codebaseIndexProviderSchema = import_zod6.z.object({
786
850
  codeIndexOpenAiKey: import_zod6.z.string().optional(),
@@ -804,6 +868,7 @@ var anthropicModels = {
804
868
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
805
869
  supportsImages: true,
806
870
  supportsPromptCache: true,
871
+ supportsNativeTools: true,
807
872
  inputPrice: 3,
808
873
  // $3 per million input tokens (≤200K context)
809
874
  outputPrice: 15,
@@ -836,6 +901,7 @@ var anthropicModels = {
836
901
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
837
902
  supportsImages: true,
838
903
  supportsPromptCache: true,
904
+ supportsNativeTools: true,
839
905
  inputPrice: 3,
840
906
  // $3 per million input tokens (≤200K context)
841
907
  outputPrice: 15,
@@ -861,12 +927,30 @@ var anthropicModels = {
861
927
  }
862
928
  ]
863
929
  },
930
+ "claude-opus-4-5-20251101": {
931
+ maxTokens: 32e3,
932
+ // Overridden to 8k if `enableReasoningEffort` is false.
933
+ contextWindow: 2e5,
934
+ supportsImages: true,
935
+ supportsPromptCache: true,
936
+ supportsNativeTools: true,
937
+ inputPrice: 5,
938
+ // $5 per million input tokens
939
+ outputPrice: 25,
940
+ // $25 per million output tokens
941
+ cacheWritesPrice: 6.25,
942
+ // $6.25 per million tokens
943
+ cacheReadsPrice: 0.5,
944
+ // $0.50 per million tokens
945
+ supportsReasoningBudget: true
946
+ },
864
947
  "claude-opus-4-1-20250805": {
865
948
  maxTokens: 32e3,
866
949
  // Overridden to 8k if `enableReasoningEffort` is false.
867
950
  contextWindow: 2e5,
868
951
  supportsImages: true,
869
952
  supportsPromptCache: true,
953
+ supportsNativeTools: true,
870
954
  inputPrice: 15,
871
955
  // $15 per million input tokens
872
956
  outputPrice: 75,
@@ -883,6 +967,7 @@ var anthropicModels = {
883
967
  contextWindow: 2e5,
884
968
  supportsImages: true,
885
969
  supportsPromptCache: true,
970
+ supportsNativeTools: true,
886
971
  inputPrice: 15,
887
972
  // $15 per million input tokens
888
973
  outputPrice: 75,
@@ -899,6 +984,7 @@ var anthropicModels = {
899
984
  contextWindow: 2e5,
900
985
  supportsImages: true,
901
986
  supportsPromptCache: true,
987
+ supportsNativeTools: true,
902
988
  inputPrice: 3,
903
989
  // $3 per million input tokens
904
990
  outputPrice: 15,
@@ -916,6 +1002,7 @@ var anthropicModels = {
916
1002
  contextWindow: 2e5,
917
1003
  supportsImages: true,
918
1004
  supportsPromptCache: true,
1005
+ supportsNativeTools: true,
919
1006
  inputPrice: 3,
920
1007
  // $3 per million input tokens
921
1008
  outputPrice: 15,
@@ -930,6 +1017,7 @@ var anthropicModels = {
930
1017
  contextWindow: 2e5,
931
1018
  supportsImages: true,
932
1019
  supportsPromptCache: true,
1020
+ supportsNativeTools: true,
933
1021
  inputPrice: 3,
934
1022
  // $3 per million input tokens
935
1023
  outputPrice: 15,
@@ -944,6 +1032,7 @@ var anthropicModels = {
944
1032
  contextWindow: 2e5,
945
1033
  supportsImages: false,
946
1034
  supportsPromptCache: true,
1035
+ supportsNativeTools: true,
947
1036
  inputPrice: 1,
948
1037
  outputPrice: 5,
949
1038
  cacheWritesPrice: 1.25,
@@ -954,6 +1043,7 @@ var anthropicModels = {
954
1043
  contextWindow: 2e5,
955
1044
  supportsImages: true,
956
1045
  supportsPromptCache: true,
1046
+ supportsNativeTools: true,
957
1047
  inputPrice: 15,
958
1048
  outputPrice: 75,
959
1049
  cacheWritesPrice: 18.75,
@@ -964,6 +1054,7 @@ var anthropicModels = {
964
1054
  contextWindow: 2e5,
965
1055
  supportsImages: true,
966
1056
  supportsPromptCache: true,
1057
+ supportsNativeTools: true,
967
1058
  inputPrice: 0.25,
968
1059
  outputPrice: 1.25,
969
1060
  cacheWritesPrice: 0.3,
@@ -974,6 +1065,7 @@ var anthropicModels = {
974
1065
  contextWindow: 2e5,
975
1066
  supportsImages: true,
976
1067
  supportsPromptCache: true,
1068
+ supportsNativeTools: true,
977
1069
  inputPrice: 1,
978
1070
  outputPrice: 5,
979
1071
  cacheWritesPrice: 1.25,
@@ -984,6 +1076,125 @@ var anthropicModels = {
984
1076
  };
985
1077
  var ANTHROPIC_DEFAULT_MAX_TOKENS = 8192;
986
1078
 
1079
+ // src/providers/baseten.ts
1080
+ var basetenModels = {
1081
+ "moonshotai/Kimi-K2-Thinking": {
1082
+ maxTokens: 163800,
1083
+ contextWindow: 262e3,
1084
+ supportsImages: false,
1085
+ supportsPromptCache: false,
1086
+ supportsNativeTools: true,
1087
+ inputPrice: 0.6,
1088
+ outputPrice: 2.5,
1089
+ cacheWritesPrice: 0,
1090
+ cacheReadsPrice: 0,
1091
+ description: "Kimi K2 Thinking - A model with enhanced reasoning capabilities from Kimi K2"
1092
+ },
1093
+ "zai-org/GLM-4.6": {
1094
+ maxTokens: 2e5,
1095
+ contextWindow: 2e5,
1096
+ supportsImages: false,
1097
+ supportsPromptCache: false,
1098
+ supportsNativeTools: true,
1099
+ inputPrice: 0.6,
1100
+ outputPrice: 2.2,
1101
+ cacheWritesPrice: 0,
1102
+ cacheReadsPrice: 0,
1103
+ description: "Frontier open model with advanced agentic, reasoning and coding capabilities"
1104
+ },
1105
+ "deepseek-ai/DeepSeek-R1": {
1106
+ maxTokens: 131072,
1107
+ contextWindow: 163840,
1108
+ supportsImages: false,
1109
+ supportsPromptCache: false,
1110
+ inputPrice: 2.55,
1111
+ outputPrice: 5.95,
1112
+ cacheWritesPrice: 0,
1113
+ cacheReadsPrice: 0,
1114
+ description: "DeepSeek's first-generation reasoning model"
1115
+ },
1116
+ "deepseek-ai/DeepSeek-R1-0528": {
1117
+ maxTokens: 131072,
1118
+ contextWindow: 163840,
1119
+ supportsImages: false,
1120
+ supportsPromptCache: false,
1121
+ inputPrice: 2.55,
1122
+ outputPrice: 5.95,
1123
+ cacheWritesPrice: 0,
1124
+ cacheReadsPrice: 0,
1125
+ description: "The latest revision of DeepSeek's first-generation reasoning model"
1126
+ },
1127
+ "deepseek-ai/DeepSeek-V3-0324": {
1128
+ maxTokens: 131072,
1129
+ contextWindow: 163840,
1130
+ supportsImages: false,
1131
+ supportsPromptCache: false,
1132
+ inputPrice: 0.77,
1133
+ outputPrice: 0.77,
1134
+ cacheWritesPrice: 0,
1135
+ cacheReadsPrice: 0,
1136
+ description: "Fast general-purpose LLM with enhanced reasoning capabilities"
1137
+ },
1138
+ "deepseek-ai/DeepSeek-V3.1": {
1139
+ maxTokens: 131072,
1140
+ contextWindow: 163840,
1141
+ supportsImages: false,
1142
+ supportsPromptCache: false,
1143
+ inputPrice: 0.5,
1144
+ outputPrice: 1.5,
1145
+ cacheWritesPrice: 0,
1146
+ cacheReadsPrice: 0,
1147
+ description: "Extremely capable general-purpose LLM with hybrid reasoning capabilities and advanced tool calling"
1148
+ },
1149
+ "Qwen/Qwen3-235B-A22B-Instruct-2507": {
1150
+ maxTokens: 262144,
1151
+ contextWindow: 262144,
1152
+ supportsImages: false,
1153
+ supportsPromptCache: false,
1154
+ inputPrice: 0.22,
1155
+ outputPrice: 0.8,
1156
+ cacheWritesPrice: 0,
1157
+ cacheReadsPrice: 0,
1158
+ description: "Mixture-of-experts LLM with math and reasoning capabilities"
1159
+ },
1160
+ "Qwen/Qwen3-Coder-480B-A35B-Instruct": {
1161
+ maxTokens: 262144,
1162
+ contextWindow: 262144,
1163
+ supportsImages: false,
1164
+ supportsPromptCache: false,
1165
+ inputPrice: 0.38,
1166
+ outputPrice: 1.53,
1167
+ cacheWritesPrice: 0,
1168
+ cacheReadsPrice: 0,
1169
+ description: "Mixture-of-experts LLM with advanced coding and reasoning capabilities"
1170
+ },
1171
+ "openai/gpt-oss-120b": {
1172
+ maxTokens: 128072,
1173
+ contextWindow: 128072,
1174
+ supportsImages: false,
1175
+ supportsPromptCache: false,
1176
+ supportsNativeTools: true,
1177
+ inputPrice: 0.1,
1178
+ outputPrice: 0.5,
1179
+ cacheWritesPrice: 0,
1180
+ cacheReadsPrice: 0,
1181
+ description: "Extremely capable general-purpose LLM with strong, controllable reasoning capabilities"
1182
+ },
1183
+ "moonshotai/Kimi-K2-Instruct-0905": {
1184
+ maxTokens: 168e3,
1185
+ contextWindow: 262e3,
1186
+ supportsImages: false,
1187
+ supportsPromptCache: false,
1188
+ supportsNativeTools: true,
1189
+ inputPrice: 0.6,
1190
+ outputPrice: 2.5,
1191
+ cacheWritesPrice: 0,
1192
+ cacheReadsPrice: 0,
1193
+ description: "State of the art language model for agentic and coding tasks. September Update."
1194
+ }
1195
+ };
1196
+ var basetenDefaultModelId = "zai-org/GLM-4.6";
1197
+
987
1198
  // src/providers/bedrock.ts
988
1199
  var bedrockDefaultModelId = "anthropic.claude-sonnet-4-5-20250929-v1:0";
989
1200
  var bedrockDefaultPromptRouterModelId = "anthropic.claude-3-sonnet-20240229-v1:0";
@@ -1088,6 +1299,20 @@ var bedrockModels = {
1088
1299
  maxCachePoints: 4,
1089
1300
  cachableFields: ["system", "messages", "tools"]
1090
1301
  },
1302
+ "anthropic.claude-opus-4-5-20251101-v1:0": {
1303
+ maxTokens: 8192,
1304
+ contextWindow: 2e5,
1305
+ supportsImages: true,
1306
+ supportsPromptCache: true,
1307
+ supportsReasoningBudget: true,
1308
+ inputPrice: 5,
1309
+ outputPrice: 25,
1310
+ cacheWritesPrice: 6.25,
1311
+ cacheReadsPrice: 0.5,
1312
+ minTokensPerCachePoint: 1024,
1313
+ maxCachePoints: 4,
1314
+ cachableFields: ["system", "messages", "tools"]
1315
+ },
1091
1316
  "anthropic.claude-opus-4-20250514-v1:0": {
1092
1317
  maxTokens: 8192,
1093
1318
  contextWindow: 2e5,
@@ -1430,7 +1655,8 @@ var BEDROCK_1M_CONTEXT_MODEL_IDS = [
1430
1655
  var BEDROCK_GLOBAL_INFERENCE_MODEL_IDS = [
1431
1656
  "anthropic.claude-sonnet-4-20250514-v1:0",
1432
1657
  "anthropic.claude-sonnet-4-5-20250929-v1:0",
1433
- "anthropic.claude-haiku-4-5-20251001-v1:0"
1658
+ "anthropic.claude-haiku-4-5-20251001-v1:0",
1659
+ "anthropic.claude-opus-4-5-20251101-v1:0"
1434
1660
  ];
1435
1661
 
1436
1662
  // src/providers/cerebras.ts
@@ -1446,24 +1672,6 @@ var cerebrasModels = {
1446
1672
  outputPrice: 0,
1447
1673
  description: "Highly intelligent general purpose model with up to 1,000 tokens/s"
1448
1674
  },
1449
- "qwen-3-coder-480b-free": {
1450
- maxTokens: 4e4,
1451
- contextWindow: 64e3,
1452
- supportsImages: false,
1453
- supportsPromptCache: false,
1454
- inputPrice: 0,
1455
- outputPrice: 0,
1456
- description: "[SOON TO BE DEPRECATED] SOTA coding model with ~2000 tokens/s ($0 free tier)\n\n\u2022 Use this if you don't have a Cerebras subscription\n\u2022 64K context window\n\u2022 Rate limits: 150K TPM, 1M TPH/TPD, 10 RPM, 100 RPH/RPD\n\nUpgrade for higher limits: [https://cloud.cerebras.ai/?utm=roocode](https://cloud.cerebras.ai/?utm=roocode)"
1457
- },
1458
- "qwen-3-coder-480b": {
1459
- maxTokens: 4e4,
1460
- contextWindow: 128e3,
1461
- supportsImages: false,
1462
- supportsPromptCache: false,
1463
- inputPrice: 0,
1464
- outputPrice: 0,
1465
- description: "[SOON TO BE DEPRECATED] SOTA coding model with ~2000 tokens/s ($50/$250 paid tiers)\n\n\u2022 Use this if you have a Cerebras subscription\n\u2022 131K context window with higher rate limits"
1466
- },
1467
1675
  "qwen-3-235b-a22b-instruct-2507": {
1468
1676
  maxTokens: 64e3,
1469
1677
  contextWindow: 64e3,
@@ -1491,16 +1699,6 @@ var cerebrasModels = {
1491
1699
  outputPrice: 0,
1492
1700
  description: "SOTA coding performance with ~2500 tokens/s"
1493
1701
  },
1494
- "qwen-3-235b-a22b-thinking-2507": {
1495
- maxTokens: 4e4,
1496
- contextWindow: 65e3,
1497
- supportsImages: false,
1498
- supportsPromptCache: false,
1499
- inputPrice: 0,
1500
- outputPrice: 0,
1501
- description: "SOTA performance with ~1500 tokens/s",
1502
- supportsReasoningEffort: true
1503
- },
1504
1702
  "gpt-oss-120b": {
1505
1703
  maxTokens: 8e3,
1506
1704
  contextWindow: 64e3,
@@ -1899,7 +2097,10 @@ var claudeCodeModels = {
1899
2097
  // Claude Code does report cache tokens
1900
2098
  supportsReasoningEffort: false,
1901
2099
  supportsReasoningBudget: false,
1902
- requiredReasoningBudget: false
2100
+ requiredReasoningBudget: false,
2101
+ // Claude Code manages its own tools and temperature via the CLI
2102
+ supportsNativeTools: false,
2103
+ supportsTemperature: false
1903
2104
  },
1904
2105
  "claude-sonnet-4-5-20250929[1m]": {
1905
2106
  ...anthropicModels["claude-sonnet-4-5"],
@@ -1910,7 +2111,10 @@ var claudeCodeModels = {
1910
2111
  // Claude Code does report cache tokens
1911
2112
  supportsReasoningEffort: false,
1912
2113
  supportsReasoningBudget: false,
1913
- requiredReasoningBudget: false
2114
+ requiredReasoningBudget: false,
2115
+ // Claude Code manages its own tools and temperature via the CLI
2116
+ supportsNativeTools: false,
2117
+ supportsTemperature: false
1914
2118
  },
1915
2119
  "claude-sonnet-4-20250514": {
1916
2120
  ...anthropicModels["claude-sonnet-4-20250514"],
@@ -1919,7 +2123,22 @@ var claudeCodeModels = {
1919
2123
  // Claude Code does report cache tokens
1920
2124
  supportsReasoningEffort: false,
1921
2125
  supportsReasoningBudget: false,
1922
- requiredReasoningBudget: false
2126
+ requiredReasoningBudget: false,
2127
+ // Claude Code manages its own tools and temperature via the CLI
2128
+ supportsNativeTools: false,
2129
+ supportsTemperature: false
2130
+ },
2131
+ "claude-opus-4-5-20251101": {
2132
+ ...anthropicModels["claude-opus-4-5-20251101"],
2133
+ supportsImages: false,
2134
+ supportsPromptCache: true,
2135
+ // Claude Code does report cache tokens
2136
+ supportsReasoningEffort: false,
2137
+ supportsReasoningBudget: false,
2138
+ requiredReasoningBudget: false,
2139
+ // Claude Code manages its own tools and temperature via the CLI
2140
+ supportsNativeTools: false,
2141
+ supportsTemperature: false
1923
2142
  },
1924
2143
  "claude-opus-4-1-20250805": {
1925
2144
  ...anthropicModels["claude-opus-4-1-20250805"],
@@ -1928,7 +2147,10 @@ var claudeCodeModels = {
1928
2147
  // Claude Code does report cache tokens
1929
2148
  supportsReasoningEffort: false,
1930
2149
  supportsReasoningBudget: false,
1931
- requiredReasoningBudget: false
2150
+ requiredReasoningBudget: false,
2151
+ // Claude Code manages its own tools and temperature via the CLI
2152
+ supportsNativeTools: false,
2153
+ supportsTemperature: false
1932
2154
  },
1933
2155
  "claude-opus-4-20250514": {
1934
2156
  ...anthropicModels["claude-opus-4-20250514"],
@@ -1937,7 +2159,10 @@ var claudeCodeModels = {
1937
2159
  // Claude Code does report cache tokens
1938
2160
  supportsReasoningEffort: false,
1939
2161
  supportsReasoningBudget: false,
1940
- requiredReasoningBudget: false
2162
+ requiredReasoningBudget: false,
2163
+ // Claude Code manages its own tools and temperature via the CLI
2164
+ supportsNativeTools: false,
2165
+ supportsTemperature: false
1941
2166
  },
1942
2167
  "claude-3-7-sonnet-20250219": {
1943
2168
  ...anthropicModels["claude-3-7-sonnet-20250219"],
@@ -1946,7 +2171,10 @@ var claudeCodeModels = {
1946
2171
  // Claude Code does report cache tokens
1947
2172
  supportsReasoningEffort: false,
1948
2173
  supportsReasoningBudget: false,
1949
- requiredReasoningBudget: false
2174
+ requiredReasoningBudget: false,
2175
+ // Claude Code manages its own tools and temperature via the CLI
2176
+ supportsNativeTools: false,
2177
+ supportsTemperature: false
1950
2178
  },
1951
2179
  "claude-3-5-sonnet-20241022": {
1952
2180
  ...anthropicModels["claude-3-5-sonnet-20241022"],
@@ -1955,7 +2183,10 @@ var claudeCodeModels = {
1955
2183
  // Claude Code does report cache tokens
1956
2184
  supportsReasoningEffort: false,
1957
2185
  supportsReasoningBudget: false,
1958
- requiredReasoningBudget: false
2186
+ requiredReasoningBudget: false,
2187
+ // Claude Code manages its own tools and temperature via the CLI
2188
+ supportsNativeTools: false,
2189
+ supportsTemperature: false
1959
2190
  },
1960
2191
  "claude-3-5-haiku-20241022": {
1961
2192
  ...anthropicModels["claude-3-5-haiku-20241022"],
@@ -1964,7 +2195,10 @@ var claudeCodeModels = {
1964
2195
  // Claude Code does report cache tokens
1965
2196
  supportsReasoningEffort: false,
1966
2197
  supportsReasoningBudget: false,
1967
- requiredReasoningBudget: false
2198
+ requiredReasoningBudget: false,
2199
+ // Claude Code manages its own tools and temperature via the CLI
2200
+ supportsNativeTools: false,
2201
+ supportsTemperature: false
1968
2202
  },
1969
2203
  "claude-haiku-4-5-20251001": {
1970
2204
  ...anthropicModels["claude-haiku-4-5-20251001"],
@@ -1973,7 +2207,10 @@ var claudeCodeModels = {
1973
2207
  // Claude Code does report cache tokens
1974
2208
  supportsReasoningEffort: false,
1975
2209
  supportsReasoningBudget: false,
1976
- requiredReasoningBudget: false
2210
+ requiredReasoningBudget: false,
2211
+ // Claude Code manages its own tools and temperature via the CLI
2212
+ supportsNativeTools: false,
2213
+ supportsTemperature: false
1977
2214
  }
1978
2215
  };
1979
2216
 
@@ -1986,6 +2223,7 @@ var deepSeekModels = {
1986
2223
  contextWindow: 128e3,
1987
2224
  supportsImages: false,
1988
2225
  supportsPromptCache: true,
2226
+ supportsNativeTools: true,
1989
2227
  inputPrice: 0.56,
1990
2228
  // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
1991
2229
  outputPrice: 1.68,
@@ -2002,6 +2240,7 @@ var deepSeekModels = {
2002
2240
  contextWindow: 128e3,
2003
2241
  supportsImages: false,
2004
2242
  supportsPromptCache: true,
2243
+ supportsNativeTools: true,
2005
2244
  inputPrice: 0.56,
2006
2245
  // $0.56 per million tokens (cache miss) - Updated Sept 5, 2025
2007
2246
  outputPrice: 1.68,
@@ -2023,6 +2262,7 @@ var doubaoModels = {
2023
2262
  contextWindow: 128e3,
2024
2263
  supportsImages: true,
2025
2264
  supportsPromptCache: true,
2265
+ supportsNativeTools: true,
2026
2266
  inputPrice: 1e-4,
2027
2267
  // $0.0001 per million tokens (cache miss)
2028
2268
  outputPrice: 4e-4,
@@ -2038,6 +2278,7 @@ var doubaoModels = {
2038
2278
  contextWindow: 128e3,
2039
2279
  supportsImages: true,
2040
2280
  supportsPromptCache: true,
2281
+ supportsNativeTools: true,
2041
2282
  inputPrice: 2e-4,
2042
2283
  // $0.0002 per million tokens
2043
2284
  outputPrice: 8e-4,
@@ -2053,6 +2294,7 @@ var doubaoModels = {
2053
2294
  contextWindow: 128e3,
2054
2295
  supportsImages: true,
2055
2296
  supportsPromptCache: true,
2297
+ supportsNativeTools: true,
2056
2298
  inputPrice: 15e-5,
2057
2299
  // $0.00015 per million tokens
2058
2300
  outputPrice: 6e-4,
@@ -2093,6 +2335,7 @@ var featherlessModels = {
2093
2335
  contextWindow: 32678,
2094
2336
  supportsImages: false,
2095
2337
  supportsPromptCache: false,
2338
+ supportsNativeTools: true,
2096
2339
  inputPrice: 0,
2097
2340
  outputPrice: 0,
2098
2341
  description: "Kimi K2 Instruct model."
@@ -2111,6 +2354,7 @@ var featherlessModels = {
2111
2354
  contextWindow: 32678,
2112
2355
  supportsImages: false,
2113
2356
  supportsPromptCache: false,
2357
+ supportsNativeTools: true,
2114
2358
  inputPrice: 0,
2115
2359
  outputPrice: 0,
2116
2360
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -2126,6 +2370,7 @@ var fireworksModels = {
2126
2370
  contextWindow: 262144,
2127
2371
  supportsImages: false,
2128
2372
  supportsPromptCache: true,
2373
+ supportsNativeTools: true,
2129
2374
  inputPrice: 0.6,
2130
2375
  outputPrice: 2.5,
2131
2376
  cacheReadsPrice: 0.15,
@@ -2136,6 +2381,7 @@ var fireworksModels = {
2136
2381
  contextWindow: 128e3,
2137
2382
  supportsImages: false,
2138
2383
  supportsPromptCache: false,
2384
+ supportsNativeTools: true,
2139
2385
  inputPrice: 0.6,
2140
2386
  outputPrice: 2.5,
2141
2387
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2145,6 +2391,7 @@ var fireworksModels = {
2145
2391
  contextWindow: 204800,
2146
2392
  supportsImages: false,
2147
2393
  supportsPromptCache: false,
2394
+ supportsNativeTools: true,
2148
2395
  inputPrice: 0.3,
2149
2396
  outputPrice: 1.2,
2150
2397
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2154,6 +2401,7 @@ var fireworksModels = {
2154
2401
  contextWindow: 256e3,
2155
2402
  supportsImages: false,
2156
2403
  supportsPromptCache: false,
2404
+ supportsNativeTools: true,
2157
2405
  inputPrice: 0.22,
2158
2406
  outputPrice: 0.88,
2159
2407
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2163,6 +2411,7 @@ var fireworksModels = {
2163
2411
  contextWindow: 256e3,
2164
2412
  supportsImages: false,
2165
2413
  supportsPromptCache: false,
2414
+ supportsNativeTools: true,
2166
2415
  inputPrice: 0.45,
2167
2416
  outputPrice: 1.8,
2168
2417
  description: "Qwen3's most agentic code model to date."
@@ -2172,6 +2421,7 @@ var fireworksModels = {
2172
2421
  contextWindow: 16e4,
2173
2422
  supportsImages: false,
2174
2423
  supportsPromptCache: false,
2424
+ supportsNativeTools: true,
2175
2425
  inputPrice: 3,
2176
2426
  outputPrice: 8,
2177
2427
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2181,6 +2431,7 @@ var fireworksModels = {
2181
2431
  contextWindow: 128e3,
2182
2432
  supportsImages: false,
2183
2433
  supportsPromptCache: false,
2434
+ supportsNativeTools: true,
2184
2435
  inputPrice: 0.9,
2185
2436
  outputPrice: 0.9,
2186
2437
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2190,6 +2441,7 @@ var fireworksModels = {
2190
2441
  contextWindow: 163840,
2191
2442
  supportsImages: false,
2192
2443
  supportsPromptCache: false,
2444
+ supportsNativeTools: true,
2193
2445
  inputPrice: 0.56,
2194
2446
  outputPrice: 1.68,
2195
2447
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2199,6 +2451,7 @@ var fireworksModels = {
2199
2451
  contextWindow: 128e3,
2200
2452
  supportsImages: false,
2201
2453
  supportsPromptCache: false,
2454
+ supportsNativeTools: true,
2202
2455
  inputPrice: 0.55,
2203
2456
  outputPrice: 2.19,
2204
2457
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2208,6 +2461,7 @@ var fireworksModels = {
2208
2461
  contextWindow: 128e3,
2209
2462
  supportsImages: false,
2210
2463
  supportsPromptCache: false,
2464
+ supportsNativeTools: true,
2211
2465
  inputPrice: 0.55,
2212
2466
  outputPrice: 2.19,
2213
2467
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2217,6 +2471,7 @@ var fireworksModels = {
2217
2471
  contextWindow: 198e3,
2218
2472
  supportsImages: false,
2219
2473
  supportsPromptCache: false,
2474
+ supportsNativeTools: true,
2220
2475
  inputPrice: 0.55,
2221
2476
  outputPrice: 2.19,
2222
2477
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2226,6 +2481,7 @@ var fireworksModels = {
2226
2481
  contextWindow: 128e3,
2227
2482
  supportsImages: false,
2228
2483
  supportsPromptCache: false,
2484
+ supportsNativeTools: true,
2229
2485
  inputPrice: 0.07,
2230
2486
  outputPrice: 0.3,
2231
2487
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2235,6 +2491,7 @@ var fireworksModels = {
2235
2491
  contextWindow: 128e3,
2236
2492
  supportsImages: false,
2237
2493
  supportsPromptCache: false,
2494
+ supportsNativeTools: true,
2238
2495
  inputPrice: 0.15,
2239
2496
  outputPrice: 0.6,
2240
2497
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
@@ -2474,6 +2731,7 @@ var groqModels = {
2474
2731
  contextWindow: 131072,
2475
2732
  supportsImages: false,
2476
2733
  supportsPromptCache: false,
2734
+ supportsNativeTools: true,
2477
2735
  inputPrice: 0.05,
2478
2736
  outputPrice: 0.08,
2479
2737
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2483,6 +2741,7 @@ var groqModels = {
2483
2741
  contextWindow: 131072,
2484
2742
  supportsImages: false,
2485
2743
  supportsPromptCache: false,
2744
+ supportsNativeTools: true,
2486
2745
  inputPrice: 0.59,
2487
2746
  outputPrice: 0.79,
2488
2747
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2492,6 +2751,7 @@ var groqModels = {
2492
2751
  contextWindow: 131072,
2493
2752
  supportsImages: false,
2494
2753
  supportsPromptCache: false,
2754
+ supportsNativeTools: true,
2495
2755
  inputPrice: 0.11,
2496
2756
  outputPrice: 0.34,
2497
2757
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2528,6 +2788,7 @@ var groqModels = {
2528
2788
  contextWindow: 131072,
2529
2789
  supportsImages: false,
2530
2790
  supportsPromptCache: false,
2791
+ supportsNativeTools: true,
2531
2792
  inputPrice: 0.29,
2532
2793
  outputPrice: 0.59,
2533
2794
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2557,6 +2818,7 @@ var groqModels = {
2557
2818
  contextWindow: 262144,
2558
2819
  supportsImages: false,
2559
2820
  supportsPromptCache: true,
2821
+ supportsNativeTools: true,
2560
2822
  inputPrice: 0.6,
2561
2823
  outputPrice: 2.5,
2562
2824
  cacheReadsPrice: 0.15,
@@ -2567,6 +2829,7 @@ var groqModels = {
2567
2829
  contextWindow: 131072,
2568
2830
  supportsImages: false,
2569
2831
  supportsPromptCache: false,
2832
+ supportsNativeTools: true,
2570
2833
  inputPrice: 0.15,
2571
2834
  outputPrice: 0.75,
2572
2835
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2576,6 +2839,7 @@ var groqModels = {
2576
2839
  contextWindow: 131072,
2577
2840
  supportsImages: false,
2578
2841
  supportsPromptCache: false,
2842
+ supportsNativeTools: true,
2579
2843
  inputPrice: 0.1,
2580
2844
  outputPrice: 0.5,
2581
2845
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2602,6 +2866,7 @@ var ioIntelligenceModels = {
2602
2866
  contextWindow: 128e3,
2603
2867
  supportsImages: false,
2604
2868
  supportsPromptCache: false,
2869
+ supportsNativeTools: true,
2605
2870
  description: "DeepSeek R1 reasoning model"
2606
2871
  },
2607
2872
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -2609,6 +2874,7 @@ var ioIntelligenceModels = {
2609
2874
  contextWindow: 43e4,
2610
2875
  supportsImages: true,
2611
2876
  supportsPromptCache: false,
2877
+ supportsNativeTools: true,
2612
2878
  description: "Llama 4 Maverick 17B model"
2613
2879
  },
2614
2880
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -2616,6 +2882,7 @@ var ioIntelligenceModels = {
2616
2882
  contextWindow: 106e3,
2617
2883
  supportsImages: false,
2618
2884
  supportsPromptCache: false,
2885
+ supportsNativeTools: true,
2619
2886
  description: "Qwen3 Coder 480B specialized for coding"
2620
2887
  },
2621
2888
  "openai/gpt-oss-120b": {
@@ -2623,6 +2890,7 @@ var ioIntelligenceModels = {
2623
2890
  contextWindow: 131072,
2624
2891
  supportsImages: false,
2625
2892
  supportsPromptCache: false,
2893
+ supportsNativeTools: true,
2626
2894
  description: "OpenAI GPT-OSS 120B model"
2627
2895
  }
2628
2896
  };
@@ -2663,75 +2931,84 @@ var mistralModels = {
2663
2931
  contextWindow: 128e3,
2664
2932
  supportsImages: true,
2665
2933
  supportsPromptCache: false,
2934
+ supportsNativeTools: true,
2666
2935
  inputPrice: 2,
2667
2936
  outputPrice: 5
2668
2937
  },
2669
2938
  "devstral-medium-latest": {
2670
- maxTokens: 131e3,
2939
+ maxTokens: 8192,
2671
2940
  contextWindow: 131e3,
2672
2941
  supportsImages: true,
2673
2942
  supportsPromptCache: false,
2943
+ supportsNativeTools: true,
2674
2944
  inputPrice: 0.4,
2675
2945
  outputPrice: 2
2676
2946
  },
2677
2947
  "mistral-medium-latest": {
2678
- maxTokens: 131e3,
2948
+ maxTokens: 8192,
2679
2949
  contextWindow: 131e3,
2680
2950
  supportsImages: true,
2681
2951
  supportsPromptCache: false,
2952
+ supportsNativeTools: true,
2682
2953
  inputPrice: 0.4,
2683
2954
  outputPrice: 2
2684
2955
  },
2685
2956
  "codestral-latest": {
2686
- maxTokens: 256e3,
2957
+ maxTokens: 8192,
2687
2958
  contextWindow: 256e3,
2688
2959
  supportsImages: false,
2689
2960
  supportsPromptCache: false,
2961
+ supportsNativeTools: true,
2690
2962
  inputPrice: 0.3,
2691
2963
  outputPrice: 0.9
2692
2964
  },
2693
2965
  "mistral-large-latest": {
2694
- maxTokens: 131e3,
2966
+ maxTokens: 8192,
2695
2967
  contextWindow: 131e3,
2696
2968
  supportsImages: false,
2697
2969
  supportsPromptCache: false,
2970
+ supportsNativeTools: true,
2698
2971
  inputPrice: 2,
2699
2972
  outputPrice: 6
2700
2973
  },
2701
2974
  "ministral-8b-latest": {
2702
- maxTokens: 131e3,
2975
+ maxTokens: 8192,
2703
2976
  contextWindow: 131e3,
2704
2977
  supportsImages: false,
2705
2978
  supportsPromptCache: false,
2979
+ supportsNativeTools: true,
2706
2980
  inputPrice: 0.1,
2707
2981
  outputPrice: 0.1
2708
2982
  },
2709
2983
  "ministral-3b-latest": {
2710
- maxTokens: 131e3,
2984
+ maxTokens: 8192,
2711
2985
  contextWindow: 131e3,
2712
2986
  supportsImages: false,
2713
2987
  supportsPromptCache: false,
2988
+ supportsNativeTools: true,
2714
2989
  inputPrice: 0.04,
2715
2990
  outputPrice: 0.04
2716
2991
  },
2717
2992
  "mistral-small-latest": {
2718
- maxTokens: 32e3,
2993
+ maxTokens: 8192,
2719
2994
  contextWindow: 32e3,
2720
2995
  supportsImages: false,
2721
2996
  supportsPromptCache: false,
2997
+ supportsNativeTools: true,
2722
2998
  inputPrice: 0.2,
2723
2999
  outputPrice: 0.6
2724
3000
  },
2725
3001
  "pixtral-large-latest": {
2726
- maxTokens: 131e3,
3002
+ maxTokens: 8192,
2727
3003
  contextWindow: 131e3,
2728
3004
  supportsImages: true,
2729
3005
  supportsPromptCache: false,
3006
+ supportsNativeTools: true,
2730
3007
  inputPrice: 2,
2731
3008
  outputPrice: 6
2732
3009
  }
2733
3010
  };
2734
- var MISTRAL_DEFAULT_TEMPERATURE = 0;
3011
+ var MISTRAL_DEFAULT_TEMPERATURE = 1;
2735
3012
 
2736
3013
  // src/providers/moonshot.ts
2737
3014
  var moonshotDefaultModelId = "kimi-k2-0905-preview";
@@ -2741,6 +3018,7 @@ var moonshotModels = {
2741
3018
  contextWindow: 131072,
2742
3019
  supportsImages: false,
2743
3020
  supportsPromptCache: true,
3021
+ supportsNativeTools: true,
2744
3022
  inputPrice: 0.6,
2745
3023
  // $0.60 per million tokens (cache miss)
2746
3024
  outputPrice: 2.5,
@@ -2756,6 +3034,7 @@ var moonshotModels = {
2756
3034
  contextWindow: 262144,
2757
3035
  supportsImages: false,
2758
3036
  supportsPromptCache: true,
3037
+ supportsNativeTools: true,
2759
3038
  inputPrice: 0.6,
2760
3039
  outputPrice: 2.5,
2761
3040
  cacheReadsPrice: 0.15,
@@ -2766,6 +3045,7 @@ var moonshotModels = {
2766
3045
  contextWindow: 262144,
2767
3046
  supportsImages: false,
2768
3047
  supportsPromptCache: true,
3048
+ supportsNativeTools: true,
2769
3049
  inputPrice: 2.4,
2770
3050
  // $2.40 per million tokens (cache miss)
2771
3051
  outputPrice: 10,
@@ -2784,6 +3064,7 @@ var moonshotModels = {
2784
3064
  supportsImages: false,
2785
3065
  // Text-only (no image/vision support)
2786
3066
  supportsPromptCache: true,
3067
+ supportsNativeTools: true,
2787
3068
  inputPrice: 0.6,
2788
3069
  // $0.60 per million tokens (cache miss)
2789
3070
  outputPrice: 2.5,
@@ -3290,6 +3571,7 @@ var OPEN_ROUTER_PROMPT_CACHING_MODELS = /* @__PURE__ */ new Set([
3290
3571
  "anthropic/claude-opus-4",
3291
3572
  "anthropic/claude-opus-4.1",
3292
3573
  "anthropic/claude-haiku-4.5",
3574
+ "anthropic/claude-opus-4.5",
3293
3575
  "google/gemini-2.5-flash-preview",
3294
3576
  "google/gemini-2.5-flash-preview:thinking",
3295
3577
  "google/gemini-2.5-flash-preview-05-20",
@@ -3311,6 +3593,7 @@ var OPEN_ROUTER_REASONING_BUDGET_MODELS = /* @__PURE__ */ new Set([
3311
3593
  "anthropic/claude-opus-4.1",
3312
3594
  "anthropic/claude-sonnet-4",
3313
3595
  "anthropic/claude-sonnet-4.5",
3596
+ "anthropic/claude-opus-4.5",
3314
3597
  "anthropic/claude-haiku-4.5",
3315
3598
  "google/gemini-2.5-pro-preview",
3316
3599
  "google/gemini-2.5-pro",
@@ -3401,6 +3684,7 @@ var sambaNovaModels = {
3401
3684
  contextWindow: 16384,
3402
3685
  supportsImages: false,
3403
3686
  supportsPromptCache: false,
3687
+ supportsNativeTools: true,
3404
3688
  inputPrice: 0.1,
3405
3689
  outputPrice: 0.2,
3406
3690
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3410,6 +3694,7 @@ var sambaNovaModels = {
3410
3694
  contextWindow: 131072,
3411
3695
  supportsImages: false,
3412
3696
  supportsPromptCache: false,
3697
+ supportsNativeTools: true,
3413
3698
  inputPrice: 0.6,
3414
3699
  outputPrice: 1.2,
3415
3700
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3420,6 +3705,7 @@ var sambaNovaModels = {
3420
3705
  supportsImages: false,
3421
3706
  supportsPromptCache: false,
3422
3707
  supportsReasoningBudget: true,
3708
+ supportsNativeTools: true,
3423
3709
  inputPrice: 5,
3424
3710
  outputPrice: 7,
3425
3711
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3429,6 +3715,7 @@ var sambaNovaModels = {
3429
3715
  contextWindow: 32768,
3430
3716
  supportsImages: false,
3431
3717
  supportsPromptCache: false,
3718
+ supportsNativeTools: true,
3432
3719
  inputPrice: 3,
3433
3720
  outputPrice: 4.5,
3434
3721
  description: "DeepSeek V3 model with 32K context window."
@@ -3438,6 +3725,7 @@ var sambaNovaModels = {
3438
3725
  contextWindow: 32768,
3439
3726
  supportsImages: false,
3440
3727
  supportsPromptCache: false,
3728
+ supportsNativeTools: true,
3441
3729
  inputPrice: 3,
3442
3730
  outputPrice: 4.5,
3443
3731
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3456,6 +3744,7 @@ var sambaNovaModels = {
3456
3744
  contextWindow: 131072,
3457
3745
  supportsImages: true,
3458
3746
  supportsPromptCache: false,
3747
+ supportsNativeTools: true,
3459
3748
  inputPrice: 0.63,
3460
3749
  outputPrice: 1.8,
3461
3750
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3474,6 +3763,7 @@ var sambaNovaModels = {
3474
3763
  contextWindow: 8192,
3475
3764
  supportsImages: false,
3476
3765
  supportsPromptCache: false,
3766
+ supportsNativeTools: true,
3477
3767
  inputPrice: 0.4,
3478
3768
  outputPrice: 0.8,
3479
3769
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3483,6 +3773,7 @@ var sambaNovaModels = {
3483
3773
  contextWindow: 131072,
3484
3774
  supportsImages: false,
3485
3775
  supportsPromptCache: false,
3776
+ supportsNativeTools: true,
3486
3777
  inputPrice: 0.22,
3487
3778
  outputPrice: 0.59,
3488
3779
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3509,6 +3800,7 @@ var vertexModels = {
3509
3800
  maxTokens: 65536,
3510
3801
  contextWindow: 1048576,
3511
3802
  supportsImages: true,
3803
+ supportsNativeTools: true,
3512
3804
  supportsPromptCache: true,
3513
3805
  supportsReasoningEffort: ["low", "high"],
3514
3806
  reasoningEffort: "low",
@@ -3533,6 +3825,7 @@ var vertexModels = {
3533
3825
  maxTokens: 65535,
3534
3826
  contextWindow: 1048576,
3535
3827
  supportsImages: true,
3828
+ supportsNativeTools: true,
3536
3829
  supportsPromptCache: true,
3537
3830
  inputPrice: 0.15,
3538
3831
  outputPrice: 3.5,
@@ -3544,6 +3837,7 @@ var vertexModels = {
3544
3837
  maxTokens: 65535,
3545
3838
  contextWindow: 1048576,
3546
3839
  supportsImages: true,
3840
+ supportsNativeTools: true,
3547
3841
  supportsPromptCache: true,
3548
3842
  inputPrice: 0.15,
3549
3843
  outputPrice: 0.6
@@ -3552,6 +3846,7 @@ var vertexModels = {
3552
3846
  maxTokens: 64e3,
3553
3847
  contextWindow: 1048576,
3554
3848
  supportsImages: true,
3849
+ supportsNativeTools: true,
3555
3850
  supportsPromptCache: true,
3556
3851
  inputPrice: 0.3,
3557
3852
  outputPrice: 2.5,
@@ -3564,6 +3859,7 @@ var vertexModels = {
3564
3859
  maxTokens: 65535,
3565
3860
  contextWindow: 1048576,
3566
3861
  supportsImages: true,
3862
+ supportsNativeTools: true,
3567
3863
  supportsPromptCache: false,
3568
3864
  inputPrice: 0.15,
3569
3865
  outputPrice: 3.5,
@@ -3575,6 +3871,7 @@ var vertexModels = {
3575
3871
  maxTokens: 65535,
3576
3872
  contextWindow: 1048576,
3577
3873
  supportsImages: true,
3874
+ supportsNativeTools: true,
3578
3875
  supportsPromptCache: false,
3579
3876
  inputPrice: 0.15,
3580
3877
  outputPrice: 0.6
@@ -3583,6 +3880,7 @@ var vertexModels = {
3583
3880
  maxTokens: 65535,
3584
3881
  contextWindow: 1048576,
3585
3882
  supportsImages: true,
3883
+ supportsNativeTools: true,
3586
3884
  supportsPromptCache: true,
3587
3885
  inputPrice: 2.5,
3588
3886
  outputPrice: 15
@@ -3591,6 +3889,7 @@ var vertexModels = {
3591
3889
  maxTokens: 65535,
3592
3890
  contextWindow: 1048576,
3593
3891
  supportsImages: true,
3892
+ supportsNativeTools: true,
3594
3893
  supportsPromptCache: true,
3595
3894
  inputPrice: 2.5,
3596
3895
  outputPrice: 15
@@ -3599,6 +3898,7 @@ var vertexModels = {
3599
3898
  maxTokens: 65535,
3600
3899
  contextWindow: 1048576,
3601
3900
  supportsImages: true,
3901
+ supportsNativeTools: true,
3602
3902
  supportsPromptCache: true,
3603
3903
  inputPrice: 2.5,
3604
3904
  outputPrice: 15,
@@ -3609,6 +3909,7 @@ var vertexModels = {
3609
3909
  maxTokens: 64e3,
3610
3910
  contextWindow: 1048576,
3611
3911
  supportsImages: true,
3912
+ supportsNativeTools: true,
3612
3913
  supportsPromptCache: true,
3613
3914
  inputPrice: 2.5,
3614
3915
  outputPrice: 15,
@@ -3634,6 +3935,7 @@ var vertexModels = {
3634
3935
  maxTokens: 65535,
3635
3936
  contextWindow: 1048576,
3636
3937
  supportsImages: true,
3938
+ supportsNativeTools: true,
3637
3939
  supportsPromptCache: false,
3638
3940
  inputPrice: 0,
3639
3941
  outputPrice: 0
@@ -3642,6 +3944,7 @@ var vertexModels = {
3642
3944
  maxTokens: 8192,
3643
3945
  contextWindow: 2097152,
3644
3946
  supportsImages: true,
3947
+ supportsNativeTools: true,
3645
3948
  supportsPromptCache: false,
3646
3949
  inputPrice: 0,
3647
3950
  outputPrice: 0
@@ -3650,6 +3953,7 @@ var vertexModels = {
3650
3953
  maxTokens: 8192,
3651
3954
  contextWindow: 1048576,
3652
3955
  supportsImages: true,
3956
+ supportsNativeTools: true,
3653
3957
  supportsPromptCache: true,
3654
3958
  inputPrice: 0.15,
3655
3959
  outputPrice: 0.6
@@ -3658,6 +3962,7 @@ var vertexModels = {
3658
3962
  maxTokens: 8192,
3659
3963
  contextWindow: 1048576,
3660
3964
  supportsImages: true,
3965
+ supportsNativeTools: true,
3661
3966
  supportsPromptCache: false,
3662
3967
  inputPrice: 0.075,
3663
3968
  outputPrice: 0.3
@@ -3666,6 +3971,7 @@ var vertexModels = {
3666
3971
  maxTokens: 8192,
3667
3972
  contextWindow: 32768,
3668
3973
  supportsImages: true,
3974
+ supportsNativeTools: true,
3669
3975
  supportsPromptCache: false,
3670
3976
  inputPrice: 0,
3671
3977
  outputPrice: 0
@@ -3674,6 +3980,7 @@ var vertexModels = {
3674
3980
  maxTokens: 8192,
3675
3981
  contextWindow: 1048576,
3676
3982
  supportsImages: true,
3983
+ supportsNativeTools: true,
3677
3984
  supportsPromptCache: true,
3678
3985
  inputPrice: 0.075,
3679
3986
  outputPrice: 0.3
@@ -3682,6 +3989,7 @@ var vertexModels = {
3682
3989
  maxTokens: 8192,
3683
3990
  contextWindow: 2097152,
3684
3991
  supportsImages: true,
3992
+ supportsNativeTools: true,
3685
3993
  supportsPromptCache: false,
3686
3994
  inputPrice: 1.25,
3687
3995
  outputPrice: 5
@@ -3719,6 +4027,17 @@ var vertexModels = {
3719
4027
  cacheReadsPrice: 0.1,
3720
4028
  supportsReasoningBudget: true
3721
4029
  },
4030
+ "claude-opus-4-5@20251101": {
4031
+ maxTokens: 8192,
4032
+ contextWindow: 2e5,
4033
+ supportsImages: true,
4034
+ supportsPromptCache: true,
4035
+ inputPrice: 5,
4036
+ outputPrice: 25,
4037
+ cacheWritesPrice: 6.25,
4038
+ cacheReadsPrice: 0.5,
4039
+ supportsReasoningBudget: true
4040
+ },
3722
4041
  "claude-opus-4-1@20250805": {
3723
4042
  maxTokens: 8192,
3724
4043
  contextWindow: 2e5,
@@ -3816,6 +4135,7 @@ var vertexModels = {
3816
4135
  maxTokens: 64e3,
3817
4136
  contextWindow: 1048576,
3818
4137
  supportsImages: true,
4138
+ supportsNativeTools: true,
3819
4139
  supportsPromptCache: true,
3820
4140
  inputPrice: 0.1,
3821
4141
  outputPrice: 0.4,
@@ -4117,17 +4437,67 @@ var xaiModels = {
4117
4437
  contextWindow: 262144,
4118
4438
  supportsImages: false,
4119
4439
  supportsPromptCache: true,
4440
+ supportsNativeTools: true,
4120
4441
  inputPrice: 0.2,
4121
4442
  outputPrice: 1.5,
4122
4443
  cacheWritesPrice: 0.02,
4123
4444
  cacheReadsPrice: 0.02,
4124
4445
  description: "xAI's Grok Code Fast model with 256K context window"
4125
4446
  },
4447
+ "grok-4-1-fast-reasoning": {
4448
+ maxTokens: 65536,
4449
+ contextWindow: 2e6,
4450
+ supportsImages: true,
4451
+ supportsPromptCache: true,
4452
+ supportsNativeTools: true,
4453
+ inputPrice: 0.2,
4454
+ outputPrice: 0.5,
4455
+ cacheWritesPrice: 0.05,
4456
+ cacheReadsPrice: 0.05,
4457
+ description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning"
4458
+ },
4459
+ "grok-4-1-fast-non-reasoning": {
4460
+ maxTokens: 65536,
4461
+ contextWindow: 2e6,
4462
+ supportsImages: true,
4463
+ supportsPromptCache: true,
4464
+ supportsNativeTools: true,
4465
+ inputPrice: 0.2,
4466
+ outputPrice: 0.5,
4467
+ cacheWritesPrice: 0.05,
4468
+ cacheReadsPrice: 0.05,
4469
+ description: "xAI's Grok 4.1 Fast model with 2M context window, optimized for high-performance agentic tool calling"
4470
+ },
4471
+ "grok-4-fast-reasoning": {
4472
+ maxTokens: 65536,
4473
+ contextWindow: 2e6,
4474
+ supportsImages: true,
4475
+ supportsPromptCache: true,
4476
+ supportsNativeTools: true,
4477
+ inputPrice: 0.2,
4478
+ outputPrice: 0.5,
4479
+ cacheWritesPrice: 0.05,
4480
+ cacheReadsPrice: 0.05,
4481
+ description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling with reasoning"
4482
+ },
4483
+ "grok-4-fast-non-reasoning": {
4484
+ maxTokens: 65536,
4485
+ contextWindow: 2e6,
4486
+ supportsImages: true,
4487
+ supportsPromptCache: true,
4488
+ supportsNativeTools: true,
4489
+ inputPrice: 0.2,
4490
+ outputPrice: 0.5,
4491
+ cacheWritesPrice: 0.05,
4492
+ cacheReadsPrice: 0.05,
4493
+ description: "xAI's Grok 4 Fast model with 2M context window, optimized for high-performance agentic tool calling"
4494
+ },
4126
4495
  "grok-4": {
4127
4496
  maxTokens: 8192,
4128
4497
  contextWindow: 256e3,
4129
4498
  supportsImages: true,
4130
4499
  supportsPromptCache: true,
4500
+ supportsNativeTools: true,
4131
4501
  inputPrice: 3,
4132
4502
  outputPrice: 15,
4133
4503
  cacheWritesPrice: 0.75,
@@ -4139,6 +4509,7 @@ var xaiModels = {
4139
4509
  contextWindow: 131072,
4140
4510
  supportsImages: false,
4141
4511
  supportsPromptCache: true,
4512
+ supportsNativeTools: true,
4142
4513
  inputPrice: 3,
4143
4514
  outputPrice: 15,
4144
4515
  cacheWritesPrice: 0.75,
@@ -4150,6 +4521,7 @@ var xaiModels = {
4150
4521
  contextWindow: 131072,
4151
4522
  supportsImages: false,
4152
4523
  supportsPromptCache: true,
4524
+ supportsNativeTools: true,
4153
4525
  inputPrice: 5,
4154
4526
  outputPrice: 25,
4155
4527
  cacheWritesPrice: 1.25,
@@ -4161,6 +4533,7 @@ var xaiModels = {
4161
4533
  contextWindow: 131072,
4162
4534
  supportsImages: false,
4163
4535
  supportsPromptCache: true,
4536
+ supportsNativeTools: true,
4164
4537
  inputPrice: 0.3,
4165
4538
  outputPrice: 0.5,
4166
4539
  cacheWritesPrice: 0.07,
@@ -4173,6 +4546,7 @@ var xaiModels = {
4173
4546
  contextWindow: 131072,
4174
4547
  supportsImages: false,
4175
4548
  supportsPromptCache: true,
4549
+ supportsNativeTools: true,
4176
4550
  inputPrice: 0.6,
4177
4551
  outputPrice: 4,
4178
4552
  cacheWritesPrice: 0.15,
@@ -4185,6 +4559,7 @@ var xaiModels = {
4185
4559
  contextWindow: 131072,
4186
4560
  supportsImages: false,
4187
4561
  supportsPromptCache: false,
4562
+ supportsNativeTools: true,
4188
4563
  inputPrice: 2,
4189
4564
  outputPrice: 10,
4190
4565
  description: "xAI's Grok-2 model (version 1212) with 128K context window"
@@ -4194,6 +4569,7 @@ var xaiModels = {
4194
4569
  contextWindow: 32768,
4195
4570
  supportsImages: true,
4196
4571
  supportsPromptCache: false,
4572
+ supportsNativeTools: true,
4197
4573
  inputPrice: 2,
4198
4574
  outputPrice: 10,
4199
4575
  description: "xAI's Grok-2 Vision model (version 1212) with image support and 32K context window"
@@ -4302,6 +4678,7 @@ var internationalZAiModels = {
4302
4678
  contextWindow: 131072,
4303
4679
  supportsImages: false,
4304
4680
  supportsPromptCache: true,
4681
+ supportsNativeTools: true,
4305
4682
  supportsReasoningBinary: true,
4306
4683
  inputPrice: 0.6,
4307
4684
  outputPrice: 2.2,
@@ -4314,6 +4691,7 @@ var internationalZAiModels = {
4314
4691
  contextWindow: 131072,
4315
4692
  supportsImages: false,
4316
4693
  supportsPromptCache: true,
4694
+ supportsNativeTools: true,
4317
4695
  inputPrice: 0.2,
4318
4696
  outputPrice: 1.1,
4319
4697
  cacheWritesPrice: 0,
@@ -4325,6 +4703,7 @@ var internationalZAiModels = {
4325
4703
  contextWindow: 131072,
4326
4704
  supportsImages: false,
4327
4705
  supportsPromptCache: true,
4706
+ supportsNativeTools: true,
4328
4707
  inputPrice: 2.2,
4329
4708
  outputPrice: 8.9,
4330
4709
  cacheWritesPrice: 0,
@@ -4336,6 +4715,7 @@ var internationalZAiModels = {
4336
4715
  contextWindow: 131072,
4337
4716
  supportsImages: false,
4338
4717
  supportsPromptCache: true,
4718
+ supportsNativeTools: true,
4339
4719
  inputPrice: 1.1,
4340
4720
  outputPrice: 4.5,
4341
4721
  cacheWritesPrice: 0,
@@ -4347,6 +4727,7 @@ var internationalZAiModels = {
4347
4727
  contextWindow: 131072,
4348
4728
  supportsImages: false,
4349
4729
  supportsPromptCache: true,
4730
+ supportsNativeTools: true,
4350
4731
  inputPrice: 0,
4351
4732
  outputPrice: 0,
4352
4733
  cacheWritesPrice: 0,
@@ -4358,6 +4739,7 @@ var internationalZAiModels = {
4358
4739
  contextWindow: 131072,
4359
4740
  supportsImages: true,
4360
4741
  supportsPromptCache: true,
4742
+ supportsNativeTools: true,
4361
4743
  inputPrice: 0.6,
4362
4744
  outputPrice: 1.8,
4363
4745
  cacheWritesPrice: 0,
@@ -4369,6 +4751,7 @@ var internationalZAiModels = {
4369
4751
  contextWindow: 2e5,
4370
4752
  supportsImages: false,
4371
4753
  supportsPromptCache: true,
4754
+ supportsNativeTools: true,
4372
4755
  supportsReasoningBinary: true,
4373
4756
  inputPrice: 0.6,
4374
4757
  outputPrice: 2.2,
@@ -4381,6 +4764,7 @@ var internationalZAiModels = {
4381
4764
  contextWindow: 131072,
4382
4765
  supportsImages: false,
4383
4766
  supportsPromptCache: false,
4767
+ supportsNativeTools: true,
4384
4768
  inputPrice: 0.1,
4385
4769
  outputPrice: 0.1,
4386
4770
  cacheWritesPrice: 0,
@@ -4395,6 +4779,7 @@ var mainlandZAiModels = {
4395
4779
  contextWindow: 131072,
4396
4780
  supportsImages: false,
4397
4781
  supportsPromptCache: true,
4782
+ supportsNativeTools: true,
4398
4783
  supportsReasoningBinary: true,
4399
4784
  inputPrice: 0.29,
4400
4785
  outputPrice: 1.14,
@@ -4407,6 +4792,7 @@ var mainlandZAiModels = {
4407
4792
  contextWindow: 131072,
4408
4793
  supportsImages: false,
4409
4794
  supportsPromptCache: true,
4795
+ supportsNativeTools: true,
4410
4796
  inputPrice: 0.1,
4411
4797
  outputPrice: 0.6,
4412
4798
  cacheWritesPrice: 0,
@@ -4418,6 +4804,7 @@ var mainlandZAiModels = {
4418
4804
  contextWindow: 131072,
4419
4805
  supportsImages: false,
4420
4806
  supportsPromptCache: true,
4807
+ supportsNativeTools: true,
4421
4808
  inputPrice: 0.29,
4422
4809
  outputPrice: 1.14,
4423
4810
  cacheWritesPrice: 0,
@@ -4429,6 +4816,7 @@ var mainlandZAiModels = {
4429
4816
  contextWindow: 131072,
4430
4817
  supportsImages: false,
4431
4818
  supportsPromptCache: true,
4819
+ supportsNativeTools: true,
4432
4820
  inputPrice: 0.1,
4433
4821
  outputPrice: 0.6,
4434
4822
  cacheWritesPrice: 0,
@@ -4440,6 +4828,7 @@ var mainlandZAiModels = {
4440
4828
  contextWindow: 131072,
4441
4829
  supportsImages: false,
4442
4830
  supportsPromptCache: true,
4831
+ supportsNativeTools: true,
4443
4832
  inputPrice: 0,
4444
4833
  outputPrice: 0,
4445
4834
  cacheWritesPrice: 0,
@@ -4451,6 +4840,7 @@ var mainlandZAiModels = {
4451
4840
  contextWindow: 131072,
4452
4841
  supportsImages: true,
4453
4842
  supportsPromptCache: true,
4843
+ supportsNativeTools: true,
4454
4844
  inputPrice: 0.29,
4455
4845
  outputPrice: 0.93,
4456
4846
  cacheWritesPrice: 0,
@@ -4462,6 +4852,7 @@ var mainlandZAiModels = {
4462
4852
  contextWindow: 204800,
4463
4853
  supportsImages: false,
4464
4854
  supportsPromptCache: true,
4855
+ supportsNativeTools: true,
4465
4856
  supportsReasoningBinary: true,
4466
4857
  inputPrice: 0.29,
4467
4858
  outputPrice: 1.14,
@@ -4491,6 +4882,7 @@ var deepInfraDefaultModelInfo = {
4491
4882
  contextWindow: 262144,
4492
4883
  supportsImages: false,
4493
4884
  supportsPromptCache: false,
4885
+ supportsNativeTools: true,
4494
4886
  inputPrice: 0.3,
4495
4887
  outputPrice: 1.2,
4496
4888
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -4551,6 +4943,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
4551
4943
  return "meta-llama/Llama-3.3-70B-Instruct";
4552
4944
  case "chutes":
4553
4945
  return chutesDefaultModelId;
4946
+ case "baseten":
4947
+ return basetenDefaultModelId;
4554
4948
  case "bedrock":
4555
4949
  return bedrockDefaultModelId;
4556
4950
  case "vertex":
@@ -4644,6 +5038,7 @@ var providerNames = [
4644
5038
  ...fauxProviders,
4645
5039
  "anthropic",
4646
5040
  "bedrock",
5041
+ "baseten",
4647
5042
  "cerebras",
4648
5043
  "claude-code",
4649
5044
  "doubao",
@@ -4883,6 +5278,9 @@ var vercelAiGatewaySchema = baseProviderSettingsSchema.extend({
4883
5278
  vercelAiGatewayApiKey: import_zod8.z.string().optional(),
4884
5279
  vercelAiGatewayModelId: import_zod8.z.string().optional()
4885
5280
  });
5281
+ var basetenSchema = apiModelIdProviderModelSchema.extend({
5282
+ basetenApiKey: import_zod8.z.string().optional()
5283
+ });
4886
5284
  var defaultSchema = import_zod8.z.object({
4887
5285
  apiProvider: import_zod8.z.undefined()
4888
5286
  });
@@ -4912,6 +5310,7 @@ var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiP
4912
5310
  fakeAiSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("fake-ai") })),
4913
5311
  xaiSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("xai") })),
4914
5312
  groqSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("groq") })),
5313
+ basetenSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("baseten") })),
4915
5314
  huggingFaceSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("huggingface") })),
4916
5315
  chutesSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("chutes") })),
4917
5316
  litellmSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("litellm") })),
@@ -4953,6 +5352,7 @@ var providerSettingsSchema = import_zod8.z.object({
4953
5352
  ...fakeAiSchema.shape,
4954
5353
  ...xaiSchema.shape,
4955
5354
  ...groqSchema.shape,
5355
+ ...basetenSchema.shape,
4956
5356
  ...huggingFaceSchema.shape,
4957
5357
  ...chutesSchema.shape,
4958
5358
  ...litellmSchema.shape,
@@ -5016,6 +5416,7 @@ var modelIdKeysByProvider = {
5016
5416
  requesty: "requestyModelId",
5017
5417
  xai: "apiModelId",
5018
5418
  groq: "apiModelId",
5419
+ baseten: "apiModelId",
5019
5420
  chutes: "apiModelId",
5020
5421
  litellm: "litellmModelId",
5021
5422
  huggingface: "huggingFaceModelId",
@@ -5123,7 +5524,8 @@ var MODELS_BY_PROVIDER = {
5123
5524
  models: Object.keys(vscodeLlmModels)
5124
5525
  },
5125
5526
  xai: { id: "xai", label: "xAI (Grok)", models: Object.keys(xaiModels) },
5126
- zai: { id: "zai", label: "Zai", models: Object.keys(internationalZAiModels) },
5527
+ zai: { id: "zai", label: "Z.ai", models: Object.keys(internationalZAiModels) },
5528
+ baseten: { id: "baseten", label: "Baseten", models: Object.keys(basetenModels) },
5127
5529
  // Dynamic providers; models pulled from remote APIs.
5128
5530
  glama: { id: "glama", label: "Glama", models: [] },
5129
5531
  huggingface: { id: "huggingface", label: "Hugging Face", models: [] },
@@ -5155,7 +5557,18 @@ var historyItemSchema = import_zod9.z.object({
5155
5557
  totalCost: import_zod9.z.number(),
5156
5558
  size: import_zod9.z.number().optional(),
5157
5559
  workspace: import_zod9.z.string().optional(),
5158
- mode: import_zod9.z.string().optional()
5560
+ mode: import_zod9.z.string().optional(),
5561
+ status: import_zod9.z.enum(["active", "completed", "delegated"]).optional(),
5562
+ delegatedToId: import_zod9.z.string().optional(),
5563
+ // Last child this parent delegated to
5564
+ childIds: import_zod9.z.array(import_zod9.z.string()).optional(),
5565
+ // All children spawned by this task
5566
+ awaitingChildId: import_zod9.z.string().optional(),
5567
+ // Child currently awaited (set when delegated)
5568
+ completedByChildId: import_zod9.z.string().optional(),
5569
+ // Child that completed and resumed this parent
5570
+ completionResultSummary: import_zod9.z.string().optional()
5571
+ // Summary from completed child
5159
5572
  });
5160
5573
 
5161
5574
  // src/experiment.ts
@@ -5165,7 +5578,8 @@ var experimentIds = [
5165
5578
  "multiFileApplyDiff",
5166
5579
  "preventFocusDisruption",
5167
5580
  "imageGeneration",
5168
- "runSlashCommand"
5581
+ "runSlashCommand",
5582
+ "multipleNativeToolCalls"
5169
5583
  ];
5170
5584
  var experimentIdsSchema = import_zod10.z.enum(experimentIds);
5171
5585
  var experimentsSchema = import_zod10.z.object({
@@ -5173,7 +5587,8 @@ var experimentsSchema = import_zod10.z.object({
5173
5587
  multiFileApplyDiff: import_zod10.z.boolean().optional(),
5174
5588
  preventFocusDisruption: import_zod10.z.boolean().optional(),
5175
5589
  imageGeneration: import_zod10.z.boolean().optional(),
5176
- runSlashCommand: import_zod10.z.boolean().optional()
5590
+ runSlashCommand: import_zod10.z.boolean().optional(),
5591
+ multipleNativeToolCalls: import_zod10.z.boolean().optional()
5177
5592
  });
5178
5593
 
5179
5594
  // src/telemetry.ts
@@ -5223,6 +5638,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
5223
5638
  TelemetryEventName2["CONSECUTIVE_MISTAKE_ERROR"] = "Consecutive Mistake Error";
5224
5639
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
5225
5640
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
5641
+ TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5226
5642
  return TelemetryEventName2;
5227
5643
  })(TelemetryEventName || {});
5228
5644
  var staticAppPropertiesSchema = import_zod11.z.object({
@@ -5306,6 +5722,7 @@ var rooCodeTelemetryEventSchema = import_zod11.z.discriminatedUnion("type", [
5306
5722
  "Shell Integration Error" /* SHELL_INTEGRATION_ERROR */,
5307
5723
  "Consecutive Mistake Error" /* CONSECUTIVE_MISTAKE_ERROR */,
5308
5724
  "Code Index Error" /* CODE_INDEX_ERROR */,
5725
+ "Model Cache Empty Response" /* MODEL_CACHE_EMPTY_RESPONSE */,
5309
5726
  "Context Condensed" /* CONTEXT_CONDENSED */,
5310
5727
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
5311
5728
  "Tab Shown" /* TAB_SHOWN */,
@@ -5466,8 +5883,6 @@ var terminalActionIds = ["terminalAddToContext", "terminalFixCommand", "terminal
5466
5883
  var commandIds = [
5467
5884
  "activationCompleted",
5468
5885
  "plusButtonClicked",
5469
- "promptsButtonClicked",
5470
- "mcpButtonClicked",
5471
5886
  "historyButtonClicked",
5472
5887
  "marketplaceButtonClicked",
5473
5888
  "popoutButtonClicked",
@@ -5524,6 +5939,7 @@ var globalSettingsSchema = import_zod14.z.object({
5524
5939
  taskHistory: import_zod14.z.array(historyItemSchema).optional(),
5525
5940
  dismissedUpsells: import_zod14.z.array(import_zod14.z.string()).optional(),
5526
5941
  // Image generation settings (experimental) - flattened for simplicity
5942
+ imageGenerationProvider: import_zod14.z.enum(["openrouter", "roo"]).optional(),
5527
5943
  openRouterImageApiKey: import_zod14.z.string().optional(),
5528
5944
  openRouterImageGenerationSelectedModel: import_zod14.z.string().optional(),
5529
5945
  condensingApiConfigId: import_zod14.z.string().optional(),
@@ -5675,7 +6091,8 @@ var SECRET_STATE_KEYS = [
5675
6091
  "fireworksApiKey",
5676
6092
  "featherlessApiKey",
5677
6093
  "ioIntelligenceApiKey",
5678
- "vercelAiGatewayApiKey"
6094
+ "vercelAiGatewayApiKey",
6095
+ "basetenApiKey"
5679
6096
  ];
5680
6097
  var GLOBAL_SECRET_KEYS = [
5681
6098
  "openRouterImageApiKey"
@@ -5940,6 +6357,9 @@ var ExtensionBridgeEventName = ((ExtensionBridgeEventName2) => {
5940
6357
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskPaused"] = "taskPaused" /* TaskPaused */] = "TaskPaused";
5941
6358
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskUnpaused"] = "taskUnpaused" /* TaskUnpaused */] = "TaskUnpaused";
5942
6359
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskSpawned"] = "taskSpawned" /* TaskSpawned */] = "TaskSpawned";
6360
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegated"] = "taskDelegated" /* TaskDelegated */] = "TaskDelegated";
6361
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegationCompleted"] = "taskDelegationCompleted" /* TaskDelegationCompleted */] = "TaskDelegationCompleted";
6362
+ ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskDelegationResumed"] = "taskDelegationResumed" /* TaskDelegationResumed */] = "TaskDelegationResumed";
5943
6363
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskUserMessage"] = "taskUserMessage" /* TaskUserMessage */] = "TaskUserMessage";
5944
6364
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated" /* TaskTokenUsageUpdated */] = "TaskTokenUsageUpdated";
5945
6365
  ExtensionBridgeEventName2[ExtensionBridgeEventName2["ModeChanged"] = "modeChanged" /* ModeChanged */] = "ModeChanged";
@@ -6015,6 +6435,21 @@ var extensionBridgeEventSchema = import_zod16.z.discriminatedUnion("type", [
6015
6435
  instance: extensionInstanceSchema,
6016
6436
  timestamp: import_zod16.z.number()
6017
6437
  }),
6438
+ import_zod16.z.object({
6439
+ type: import_zod16.z.literal(ExtensionBridgeEventName.TaskDelegated),
6440
+ instance: extensionInstanceSchema,
6441
+ timestamp: import_zod16.z.number()
6442
+ }),
6443
+ import_zod16.z.object({
6444
+ type: import_zod16.z.literal(ExtensionBridgeEventName.TaskDelegationCompleted),
6445
+ instance: extensionInstanceSchema,
6446
+ timestamp: import_zod16.z.number()
6447
+ }),
6448
+ import_zod16.z.object({
6449
+ type: import_zod16.z.literal(ExtensionBridgeEventName.TaskDelegationResumed),
6450
+ instance: extensionInstanceSchema,
6451
+ timestamp: import_zod16.z.number()
6452
+ }),
6018
6453
  import_zod16.z.object({
6019
6454
  type: import_zod16.z.literal(ExtensionBridgeEventName.TaskUserMessage),
6020
6455
  instance: extensionInstanceSchema,
@@ -6205,12 +6640,27 @@ var followUpDataSchema = import_zod17.z.object({
6205
6640
 
6206
6641
  // src/image-generation.ts
6207
6642
  var IMAGE_GENERATION_MODELS = [
6208
- { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image" },
6209
- { value: "google/gemini-3-pro-image-preview", label: "Gemini 3 Pro Image Preview" },
6210
- { value: "openai/gpt-5-image", label: "GPT-5 Image" },
6211
- { value: "openai/gpt-5-image-mini", label: "GPT-5 Image Mini" }
6643
+ // OpenRouter models
6644
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image", provider: "openrouter" },
6645
+ { value: "google/gemini-3-pro-image-preview", label: "Gemini 3 Pro Image Preview", provider: "openrouter" },
6646
+ { value: "openai/gpt-5-image", label: "GPT-5 Image", provider: "openrouter" },
6647
+ { value: "openai/gpt-5-image-mini", label: "GPT-5 Image Mini", provider: "openrouter" },
6648
+ { value: "black-forest-labs/flux.2-flex", label: "Black Forest Labs FLUX.2 Flex", provider: "openrouter" },
6649
+ { value: "black-forest-labs/flux.2-pro", label: "Black Forest Labs FLUX.2 Pro", provider: "openrouter" },
6650
+ // Roo Code Cloud models
6651
+ { value: "google/gemini-2.5-flash-image", label: "Gemini 2.5 Flash Image", provider: "roo" },
6652
+ { value: "google/gemini-3-pro-image", label: "Gemini 3 Pro Image", provider: "roo" },
6653
+ {
6654
+ value: "bfl/flux-2-pro:free",
6655
+ label: "Black Forest Labs FLUX.2 Pro (Free)",
6656
+ provider: "roo",
6657
+ apiMethod: "images_api"
6658
+ }
6212
6659
  ];
6213
6660
  var IMAGE_GENERATION_MODEL_IDS = IMAGE_GENERATION_MODELS.map((m) => m.value);
6661
+ function getImageGenerationProvider(explicitProvider, hasExistingModel) {
6662
+ return explicitProvider !== void 0 ? explicitProvider : hasExistingModel ? "openrouter" : "roo";
6663
+ }
6214
6664
 
6215
6665
  // src/ipc.ts
6216
6666
  var import_zod18 = require("zod");
@@ -6446,6 +6896,8 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6446
6896
  anthropicModels,
6447
6897
  appPropertiesSchema,
6448
6898
  azureOpenAiDefaultApiVersion,
6899
+ basetenDefaultModelId,
6900
+ basetenModels,
6449
6901
  bedrockDefaultModelId,
6450
6902
  bedrockDefaultPromptRouterModelId,
6451
6903
  bedrockModels,
@@ -6501,6 +6953,7 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
6501
6953
  getApiProtocol,
6502
6954
  getClaudeCodeModelId,
6503
6955
  getEffectiveProtocol,
6956
+ getImageGenerationProvider,
6504
6957
  getModelId,
6505
6958
  getProviderDefaultModelId,
6506
6959
  gitPropertiesSchema,