@roo-code/types 1.105.0 → 1.107.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.cjs CHANGED
@@ -42,7 +42,7 @@ __export(index_exports, {
42
42
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS: () => DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
43
43
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT: () => DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
44
44
  DEFAULT_MODES: () => DEFAULT_MODES,
45
- DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT: () => DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
45
+ DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE: () => DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
46
46
  DEFAULT_WRITE_DELAY_MS: () => DEFAULT_WRITE_DELAY_MS,
47
47
  DOUBAO_API_BASE_URL: () => DOUBAO_API_BASE_URL,
48
48
  DOUBAO_API_CHAT_PATH: () => DOUBAO_API_CHAT_PATH,
@@ -72,13 +72,13 @@ __export(index_exports, {
72
72
  IpcOrigin: () => IpcOrigin,
73
73
  LMSTUDIO_DEFAULT_TEMPERATURE: () => LMSTUDIO_DEFAULT_TEMPERATURE,
74
74
  MAX_CHECKPOINT_TIMEOUT_SECONDS: () => MAX_CHECKPOINT_TIMEOUT_SECONDS,
75
+ MAX_MCP_TOOLS_THRESHOLD: () => MAX_MCP_TOOLS_THRESHOLD,
75
76
  MINIMAX_DEFAULT_MAX_TOKENS: () => MINIMAX_DEFAULT_MAX_TOKENS,
76
77
  MINIMAX_DEFAULT_TEMPERATURE: () => MINIMAX_DEFAULT_TEMPERATURE,
77
78
  MIN_CHECKPOINT_TIMEOUT_SECONDS: () => MIN_CHECKPOINT_TIMEOUT_SECONDS,
78
79
  MISTRAL_DEFAULT_TEMPERATURE: () => MISTRAL_DEFAULT_TEMPERATURE,
79
80
  MODELS_BY_PROVIDER: () => MODELS_BY_PROVIDER,
80
81
  MOONSHOT_DEFAULT_TEMPERATURE: () => MOONSHOT_DEFAULT_TEMPERATURE,
81
- NATIVE_TOOL_DEFAULTS: () => NATIVE_TOOL_DEFAULTS,
82
82
  OPENAI_AZURE_AI_INFERENCE_PATH: () => OPENAI_AZURE_AI_INFERENCE_PATH,
83
83
  OPENAI_NATIVE_DEFAULT_TEMPERATURE: () => OPENAI_NATIVE_DEFAULT_TEMPERATURE,
84
84
  OPENROUTER_DEFAULT_PROVIDER_NAME: () => OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -93,7 +93,11 @@ __export(index_exports, {
93
93
  RooModelsResponseSchema: () => RooModelsResponseSchema,
94
94
  RooPricingSchema: () => RooPricingSchema,
95
95
  SECRET_STATE_KEYS: () => SECRET_STATE_KEYS,
96
- TOOL_PROTOCOL: () => TOOL_PROTOCOL,
96
+ SKILL_NAME_MAX_LENGTH: () => SKILL_NAME_MAX_LENGTH,
97
+ SKILL_NAME_MIN_LENGTH: () => SKILL_NAME_MIN_LENGTH,
98
+ SKILL_NAME_REGEX: () => SKILL_NAME_REGEX,
99
+ SkillNameValidationError: () => SkillNameValidationError,
100
+ TERMINAL_PREVIEW_BYTES: () => TERMINAL_PREVIEW_BYTES,
97
101
  TaskBridgeCommandName: () => TaskBridgeCommandName,
98
102
  TaskBridgeEventName: () => TaskBridgeEventName,
99
103
  TaskCommandName: () => TaskCommandName,
@@ -125,9 +129,6 @@ __export(index_exports, {
125
129
  chutesDefaultModelId: () => chutesDefaultModelId,
126
130
  chutesDefaultModelInfo: () => chutesDefaultModelInfo,
127
131
  chutesModels: () => chutesModels,
128
- claudeCodeDefaultModelId: () => claudeCodeDefaultModelId,
129
- claudeCodeModels: () => claudeCodeModels,
130
- claudeCodeReasoningConfig: () => claudeCodeReasoningConfig,
131
132
  clineAskSchema: () => clineAskSchema,
132
133
  clineAsks: () => clineAsks,
133
134
  clineMessageSchema: () => clineMessageSchema,
@@ -142,6 +143,7 @@ __export(index_exports, {
142
143
  commandIds: () => commandIds,
143
144
  contextCondenseSchema: () => contextCondenseSchema,
144
145
  contextTruncationSchema: () => contextTruncationSchema,
146
+ countEnabledMcpTools: () => countEnabledMcpTools,
145
147
  customModePromptsSchema: () => customModePromptsSchema,
146
148
  customModesSettingsSchema: () => customModesSettingsSchema,
147
149
  customProviders: () => customProviders,
@@ -175,7 +177,6 @@ __export(index_exports, {
175
177
  geminiDefaultModelId: () => geminiDefaultModelId,
176
178
  geminiModels: () => geminiModels,
177
179
  getApiProtocol: () => getApiProtocol,
178
- getEffectiveProtocol: () => getEffectiveProtocol,
179
180
  getErrorMessage: () => getErrorMessage,
180
181
  getErrorStatusCode: () => getErrorStatusCode,
181
182
  getImageGenerationProvider: () => getImageGenerationProvider,
@@ -210,9 +211,9 @@ __export(index_exports, {
210
211
  isInteractiveAsk: () => isInteractiveAsk,
211
212
  isInternalProvider: () => isInternalProvider,
212
213
  isLanguage: () => isLanguage,
214
+ isLegacyReadFileParams: () => isLegacyReadFileParams,
213
215
  isLocalProvider: () => isLocalProvider,
214
216
  isModelParameter: () => isModelParameter,
215
- isNativeProtocol: () => isNativeProtocol,
216
217
  isNonBlockingAsk: () => isNonBlockingAsk,
217
218
  isProviderName: () => isProviderName,
218
219
  isResumableAsk: () => isResumableAsk,
@@ -248,9 +249,10 @@ __export(index_exports, {
248
249
  moonshotDefaultModelId: () => moonshotDefaultModelId,
249
250
  moonshotModels: () => moonshotModels,
250
251
  nonBlockingAsks: () => nonBlockingAsks,
251
- normalizeClaudeCodeModelId: () => normalizeClaudeCodeModelId,
252
252
  ollamaDefaultModelId: () => ollamaDefaultModelId,
253
253
  ollamaDefaultModelInfo: () => ollamaDefaultModelInfo,
254
+ openAiCodexDefaultModelId: () => openAiCodexDefaultModelId,
255
+ openAiCodexModels: () => openAiCodexModels,
254
256
  openAiModelInfoSaneDefaults: () => openAiModelInfoSaneDefaults,
255
257
  openAiNativeDefaultModelId: () => openAiNativeDefaultModelId,
256
258
  openAiNativeModels: () => openAiNativeModels,
@@ -320,6 +322,7 @@ __export(index_exports, {
320
322
  userFeaturesSchema: () => userFeaturesSchema,
321
323
  userSettingsConfigSchema: () => userSettingsConfigSchema,
322
324
  userSettingsDataSchema: () => userSettingsDataSchema,
325
+ validateSkillName: () => validateSkillName,
323
326
  verbosityLevels: () => verbosityLevels,
324
327
  verbosityLevelsSchema: () => verbosityLevelsSchema,
325
328
  vercelAiGatewayDefaultModelId: () => vercelAiGatewayDefaultModelId,
@@ -415,7 +418,9 @@ var clineSays = [
415
418
  "condense_context_error",
416
419
  "sliding_window_truncation",
417
420
  "codebase_search_result",
418
- "user_edit_todos"
421
+ "user_edit_todos",
422
+ "too_many_tools_warning",
423
+ "tool"
419
424
  ];
420
425
  var clineSaySchema = import_zod.z.enum(clineSays);
421
426
  var toolProgressStatusSchema = import_zod.z.object({
@@ -483,6 +488,7 @@ var toolGroupsSchema = import_zod2.z.enum(toolGroups);
483
488
  var toolNames = [
484
489
  "execute_command",
485
490
  "read_file",
491
+ "read_command_output",
486
492
  "write_to_file",
487
493
  "apply_diff",
488
494
  "search_and_replace",
@@ -498,10 +504,10 @@ var toolNames = [
498
504
  "attempt_completion",
499
505
  "switch_mode",
500
506
  "new_task",
501
- "fetch_instructions",
502
507
  "codebase_search",
503
508
  "update_todo_list",
504
509
  "run_slash_command",
510
+ "skill",
505
511
  "generate_image",
506
512
  "custom_tool"
507
513
  ];
@@ -513,20 +519,6 @@ var toolUsageSchema = import_zod2.z.record(
513
519
  failures: import_zod2.z.number()
514
520
  })
515
521
  );
516
- var TOOL_PROTOCOL = {
517
- XML: "xml",
518
- NATIVE: "native"
519
- };
520
- var NATIVE_TOOL_DEFAULTS = {
521
- supportsNativeTools: true,
522
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
523
- };
524
- function isNativeProtocol(protocol) {
525
- return protocol === TOOL_PROTOCOL.NATIVE;
526
- }
527
- function getEffectiveProtocol(toolProtocol) {
528
- return toolProtocol || TOOL_PROTOCOL.XML;
529
- }
530
522
 
531
523
  // src/events.ts
532
524
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
@@ -550,6 +542,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
550
542
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
551
543
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
552
544
  RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
545
+ RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
553
546
  RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
554
547
  RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
555
548
  RooCodeEventName2["ModeChanged"] = "modeChanged";
@@ -609,6 +602,7 @@ var rooCodeEventsSchema = import_zod3.z.object({
609
602
  ["taskModeSwitched" /* TaskModeSwitched */]: import_zod3.z.tuple([import_zod3.z.string(), import_zod3.z.string()]),
610
603
  ["taskAskResponded" /* TaskAskResponded */]: import_zod3.z.tuple([import_zod3.z.string()]),
611
604
  ["taskUserMessage" /* TaskUserMessage */]: import_zod3.z.tuple([import_zod3.z.string()]),
605
+ ["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: import_zod3.z.tuple([import_zod3.z.string(), import_zod3.z.array(queuedMessageSchema)]),
612
606
  ["taskToolFailed" /* TaskToolFailed */]: import_zod3.z.tuple([import_zod3.z.string(), toolNamesSchema, import_zod3.z.string()]),
613
607
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: import_zod3.z.tuple([import_zod3.z.string(), tokenUsageSchema, toolUsageSchema]),
614
608
  ["modeChanged" /* ModeChanged */]: import_zod3.z.tuple([import_zod3.z.string()]),
@@ -714,6 +708,11 @@ var taskEventSchema = import_zod3.z.discriminatedUnion("eventName", [
714
708
  payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
715
709
  taskId: import_zod3.z.number().optional()
716
710
  }),
711
+ import_zod3.z.object({
712
+ eventName: import_zod3.z.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
713
+ payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
714
+ taskId: import_zod3.z.number().optional()
715
+ }),
717
716
  // Task Analytics
718
717
  import_zod3.z.object({
719
718
  eventName: import_zod3.z.literal("taskToolFailed" /* TaskToolFailed */),
@@ -814,10 +813,6 @@ var modelInfoSchema = import_zod5.z.object({
814
813
  isStealthModel: import_zod5.z.boolean().optional(),
815
814
  // Flag to indicate if the model is free (no cost)
816
815
  isFree: import_zod5.z.boolean().optional(),
817
- // Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
818
- supportsNativeTools: import_zod5.z.boolean().optional(),
819
- // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
820
- defaultToolProtocol: import_zod5.z.enum(["xml", "native"]).optional(),
821
816
  // Exclude specific native tools from being available (only applies to native protocol)
822
817
  // These tools will be removed from the set of tools available to the model
823
818
  excludedTools: import_zod5.z.array(import_zod5.z.string()).optional(),
@@ -914,8 +909,6 @@ var anthropicModels = {
914
909
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
915
910
  supportsImages: true,
916
911
  supportsPromptCache: true,
917
- supportsNativeTools: true,
918
- defaultToolProtocol: "native",
919
912
  inputPrice: 3,
920
913
  // $3 per million input tokens (≤200K context)
921
914
  outputPrice: 15,
@@ -948,8 +941,6 @@ var anthropicModels = {
948
941
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
949
942
  supportsImages: true,
950
943
  supportsPromptCache: true,
951
- supportsNativeTools: true,
952
- defaultToolProtocol: "native",
953
944
  inputPrice: 3,
954
945
  // $3 per million input tokens (≤200K context)
955
946
  outputPrice: 15,
@@ -981,8 +972,6 @@ var anthropicModels = {
981
972
  contextWindow: 2e5,
982
973
  supportsImages: true,
983
974
  supportsPromptCache: true,
984
- supportsNativeTools: true,
985
- defaultToolProtocol: "native",
986
975
  inputPrice: 5,
987
976
  // $5 per million input tokens
988
977
  outputPrice: 25,
@@ -999,8 +988,6 @@ var anthropicModels = {
999
988
  contextWindow: 2e5,
1000
989
  supportsImages: true,
1001
990
  supportsPromptCache: true,
1002
- supportsNativeTools: true,
1003
- defaultToolProtocol: "native",
1004
991
  inputPrice: 15,
1005
992
  // $15 per million input tokens
1006
993
  outputPrice: 75,
@@ -1017,8 +1004,6 @@ var anthropicModels = {
1017
1004
  contextWindow: 2e5,
1018
1005
  supportsImages: true,
1019
1006
  supportsPromptCache: true,
1020
- supportsNativeTools: true,
1021
- defaultToolProtocol: "native",
1022
1007
  inputPrice: 15,
1023
1008
  // $15 per million input tokens
1024
1009
  outputPrice: 75,
@@ -1035,8 +1020,6 @@ var anthropicModels = {
1035
1020
  contextWindow: 2e5,
1036
1021
  supportsImages: true,
1037
1022
  supportsPromptCache: true,
1038
- supportsNativeTools: true,
1039
- defaultToolProtocol: "native",
1040
1023
  inputPrice: 3,
1041
1024
  // $3 per million input tokens
1042
1025
  outputPrice: 15,
@@ -1054,8 +1037,6 @@ var anthropicModels = {
1054
1037
  contextWindow: 2e5,
1055
1038
  supportsImages: true,
1056
1039
  supportsPromptCache: true,
1057
- supportsNativeTools: true,
1058
- defaultToolProtocol: "native",
1059
1040
  inputPrice: 3,
1060
1041
  // $3 per million input tokens
1061
1042
  outputPrice: 15,
@@ -1070,8 +1051,6 @@ var anthropicModels = {
1070
1051
  contextWindow: 2e5,
1071
1052
  supportsImages: true,
1072
1053
  supportsPromptCache: true,
1073
- supportsNativeTools: true,
1074
- defaultToolProtocol: "native",
1075
1054
  inputPrice: 3,
1076
1055
  // $3 per million input tokens
1077
1056
  outputPrice: 15,
@@ -1086,8 +1065,6 @@ var anthropicModels = {
1086
1065
  contextWindow: 2e5,
1087
1066
  supportsImages: false,
1088
1067
  supportsPromptCache: true,
1089
- supportsNativeTools: true,
1090
- defaultToolProtocol: "native",
1091
1068
  inputPrice: 1,
1092
1069
  outputPrice: 5,
1093
1070
  cacheWritesPrice: 1.25,
@@ -1098,8 +1075,6 @@ var anthropicModels = {
1098
1075
  contextWindow: 2e5,
1099
1076
  supportsImages: true,
1100
1077
  supportsPromptCache: true,
1101
- supportsNativeTools: true,
1102
- defaultToolProtocol: "native",
1103
1078
  inputPrice: 15,
1104
1079
  outputPrice: 75,
1105
1080
  cacheWritesPrice: 18.75,
@@ -1110,8 +1085,6 @@ var anthropicModels = {
1110
1085
  contextWindow: 2e5,
1111
1086
  supportsImages: true,
1112
1087
  supportsPromptCache: true,
1113
- supportsNativeTools: true,
1114
- defaultToolProtocol: "native",
1115
1088
  inputPrice: 0.25,
1116
1089
  outputPrice: 1.25,
1117
1090
  cacheWritesPrice: 0.3,
@@ -1122,8 +1095,6 @@ var anthropicModels = {
1122
1095
  contextWindow: 2e5,
1123
1096
  supportsImages: true,
1124
1097
  supportsPromptCache: true,
1125
- supportsNativeTools: true,
1126
- defaultToolProtocol: "native",
1127
1098
  inputPrice: 1,
1128
1099
  outputPrice: 5,
1129
1100
  cacheWritesPrice: 1.25,
@@ -1141,7 +1112,6 @@ var basetenModels = {
1141
1112
  contextWindow: 262e3,
1142
1113
  supportsImages: false,
1143
1114
  supportsPromptCache: false,
1144
- supportsNativeTools: true,
1145
1115
  inputPrice: 0.6,
1146
1116
  outputPrice: 2.5,
1147
1117
  cacheWritesPrice: 0,
@@ -1153,7 +1123,6 @@ var basetenModels = {
1153
1123
  contextWindow: 2e5,
1154
1124
  supportsImages: false,
1155
1125
  supportsPromptCache: false,
1156
- supportsNativeTools: true,
1157
1126
  inputPrice: 0.6,
1158
1127
  outputPrice: 2.2,
1159
1128
  cacheWritesPrice: 0,
@@ -1165,7 +1134,6 @@ var basetenModels = {
1165
1134
  contextWindow: 163840,
1166
1135
  supportsImages: false,
1167
1136
  supportsPromptCache: false,
1168
- supportsNativeTools: true,
1169
1137
  inputPrice: 2.55,
1170
1138
  outputPrice: 5.95,
1171
1139
  cacheWritesPrice: 0,
@@ -1177,7 +1145,6 @@ var basetenModels = {
1177
1145
  contextWindow: 163840,
1178
1146
  supportsImages: false,
1179
1147
  supportsPromptCache: false,
1180
- supportsNativeTools: true,
1181
1148
  inputPrice: 2.55,
1182
1149
  outputPrice: 5.95,
1183
1150
  cacheWritesPrice: 0,
@@ -1189,7 +1156,6 @@ var basetenModels = {
1189
1156
  contextWindow: 163840,
1190
1157
  supportsImages: false,
1191
1158
  supportsPromptCache: false,
1192
- supportsNativeTools: true,
1193
1159
  inputPrice: 0.77,
1194
1160
  outputPrice: 0.77,
1195
1161
  cacheWritesPrice: 0,
@@ -1201,7 +1167,6 @@ var basetenModels = {
1201
1167
  contextWindow: 163840,
1202
1168
  supportsImages: false,
1203
1169
  supportsPromptCache: false,
1204
- supportsNativeTools: true,
1205
1170
  inputPrice: 0.5,
1206
1171
  outputPrice: 1.5,
1207
1172
  cacheWritesPrice: 0,
@@ -1213,7 +1178,6 @@ var basetenModels = {
1213
1178
  contextWindow: 163840,
1214
1179
  supportsImages: false,
1215
1180
  supportsPromptCache: false,
1216
- supportsNativeTools: true,
1217
1181
  inputPrice: 0.3,
1218
1182
  outputPrice: 0.45,
1219
1183
  cacheWritesPrice: 0,
@@ -1225,7 +1189,6 @@ var basetenModels = {
1225
1189
  contextWindow: 128072,
1226
1190
  supportsImages: false,
1227
1191
  supportsPromptCache: false,
1228
- supportsNativeTools: true,
1229
1192
  inputPrice: 0.1,
1230
1193
  outputPrice: 0.5,
1231
1194
  cacheWritesPrice: 0,
@@ -1237,7 +1200,6 @@ var basetenModels = {
1237
1200
  contextWindow: 262144,
1238
1201
  supportsImages: false,
1239
1202
  supportsPromptCache: false,
1240
- supportsNativeTools: true,
1241
1203
  inputPrice: 0.22,
1242
1204
  outputPrice: 0.8,
1243
1205
  cacheWritesPrice: 0,
@@ -1249,7 +1211,6 @@ var basetenModels = {
1249
1211
  contextWindow: 262144,
1250
1212
  supportsImages: false,
1251
1213
  supportsPromptCache: false,
1252
- supportsNativeTools: true,
1253
1214
  inputPrice: 0.38,
1254
1215
  outputPrice: 1.53,
1255
1216
  cacheWritesPrice: 0,
@@ -1261,7 +1222,6 @@ var basetenModels = {
1261
1222
  contextWindow: 262e3,
1262
1223
  supportsImages: false,
1263
1224
  supportsPromptCache: false,
1264
- supportsNativeTools: true,
1265
1225
  inputPrice: 0.6,
1266
1226
  outputPrice: 2.5,
1267
1227
  cacheWritesPrice: 0,
@@ -1281,8 +1241,6 @@ var bedrockModels = {
1281
1241
  supportsImages: true,
1282
1242
  supportsPromptCache: true,
1283
1243
  supportsReasoningBudget: true,
1284
- supportsNativeTools: true,
1285
- defaultToolProtocol: "native",
1286
1244
  inputPrice: 3,
1287
1245
  outputPrice: 15,
1288
1246
  cacheWritesPrice: 3.75,
@@ -1296,7 +1254,6 @@ var bedrockModels = {
1296
1254
  contextWindow: 3e5,
1297
1255
  supportsImages: true,
1298
1256
  supportsPromptCache: true,
1299
- supportsNativeTools: true,
1300
1257
  inputPrice: 0.8,
1301
1258
  outputPrice: 3.2,
1302
1259
  cacheWritesPrice: 0.8,
@@ -1312,7 +1269,6 @@ var bedrockModels = {
1312
1269
  contextWindow: 3e5,
1313
1270
  supportsImages: true,
1314
1271
  supportsPromptCache: false,
1315
- supportsNativeTools: true,
1316
1272
  inputPrice: 1,
1317
1273
  outputPrice: 4,
1318
1274
  cacheWritesPrice: 1,
@@ -1326,7 +1282,6 @@ var bedrockModels = {
1326
1282
  contextWindow: 3e5,
1327
1283
  supportsImages: true,
1328
1284
  supportsPromptCache: true,
1329
- supportsNativeTools: true,
1330
1285
  inputPrice: 0.06,
1331
1286
  outputPrice: 0.24,
1332
1287
  cacheWritesPrice: 0.06,
@@ -1342,7 +1297,6 @@ var bedrockModels = {
1342
1297
  contextWindow: 1e6,
1343
1298
  supportsImages: true,
1344
1299
  supportsPromptCache: true,
1345
- supportsNativeTools: true,
1346
1300
  inputPrice: 0.33,
1347
1301
  outputPrice: 2.75,
1348
1302
  cacheWritesPrice: 0,
@@ -1358,7 +1312,6 @@ var bedrockModels = {
1358
1312
  contextWindow: 128e3,
1359
1313
  supportsImages: false,
1360
1314
  supportsPromptCache: true,
1361
- supportsNativeTools: true,
1362
1315
  inputPrice: 0.035,
1363
1316
  outputPrice: 0.14,
1364
1317
  cacheWritesPrice: 0.035,
@@ -1375,8 +1328,6 @@ var bedrockModels = {
1375
1328
  supportsImages: true,
1376
1329
  supportsPromptCache: true,
1377
1330
  supportsReasoningBudget: true,
1378
- supportsNativeTools: true,
1379
- defaultToolProtocol: "native",
1380
1331
  inputPrice: 3,
1381
1332
  outputPrice: 15,
1382
1333
  cacheWritesPrice: 3.75,
@@ -1391,8 +1342,6 @@ var bedrockModels = {
1391
1342
  supportsImages: true,
1392
1343
  supportsPromptCache: true,
1393
1344
  supportsReasoningBudget: true,
1394
- supportsNativeTools: true,
1395
- defaultToolProtocol: "native",
1396
1345
  inputPrice: 15,
1397
1346
  outputPrice: 75,
1398
1347
  cacheWritesPrice: 18.75,
@@ -1407,8 +1356,6 @@ var bedrockModels = {
1407
1356
  supportsImages: true,
1408
1357
  supportsPromptCache: true,
1409
1358
  supportsReasoningBudget: true,
1410
- supportsNativeTools: true,
1411
- defaultToolProtocol: "native",
1412
1359
  inputPrice: 5,
1413
1360
  outputPrice: 25,
1414
1361
  cacheWritesPrice: 6.25,
@@ -1423,8 +1370,6 @@ var bedrockModels = {
1423
1370
  supportsImages: true,
1424
1371
  supportsPromptCache: true,
1425
1372
  supportsReasoningBudget: true,
1426
- supportsNativeTools: true,
1427
- defaultToolProtocol: "native",
1428
1373
  inputPrice: 15,
1429
1374
  outputPrice: 75,
1430
1375
  cacheWritesPrice: 18.75,
@@ -1439,8 +1384,6 @@ var bedrockModels = {
1439
1384
  supportsImages: true,
1440
1385
  supportsPromptCache: true,
1441
1386
  supportsReasoningBudget: true,
1442
- supportsNativeTools: true,
1443
- defaultToolProtocol: "native",
1444
1387
  inputPrice: 3,
1445
1388
  outputPrice: 15,
1446
1389
  cacheWritesPrice: 3.75,
@@ -1454,8 +1397,6 @@ var bedrockModels = {
1454
1397
  contextWindow: 2e5,
1455
1398
  supportsImages: true,
1456
1399
  supportsPromptCache: true,
1457
- supportsNativeTools: true,
1458
- defaultToolProtocol: "native",
1459
1400
  inputPrice: 3,
1460
1401
  outputPrice: 15,
1461
1402
  cacheWritesPrice: 3.75,
@@ -1469,8 +1410,6 @@ var bedrockModels = {
1469
1410
  contextWindow: 2e5,
1470
1411
  supportsImages: false,
1471
1412
  supportsPromptCache: true,
1472
- supportsNativeTools: true,
1473
- defaultToolProtocol: "native",
1474
1413
  inputPrice: 0.8,
1475
1414
  outputPrice: 4,
1476
1415
  cacheWritesPrice: 1,
@@ -1485,8 +1424,6 @@ var bedrockModels = {
1485
1424
  supportsImages: true,
1486
1425
  supportsPromptCache: true,
1487
1426
  supportsReasoningBudget: true,
1488
- supportsNativeTools: true,
1489
- defaultToolProtocol: "native",
1490
1427
  inputPrice: 1,
1491
1428
  outputPrice: 5,
1492
1429
  cacheWritesPrice: 1.25,
@@ -1502,8 +1439,6 @@ var bedrockModels = {
1502
1439
  contextWindow: 2e5,
1503
1440
  supportsImages: true,
1504
1441
  supportsPromptCache: false,
1505
- supportsNativeTools: true,
1506
- defaultToolProtocol: "native",
1507
1442
  inputPrice: 3,
1508
1443
  outputPrice: 15
1509
1444
  },
@@ -1512,8 +1447,6 @@ var bedrockModels = {
1512
1447
  contextWindow: 2e5,
1513
1448
  supportsImages: true,
1514
1449
  supportsPromptCache: false,
1515
- supportsNativeTools: true,
1516
- defaultToolProtocol: "native",
1517
1450
  inputPrice: 15,
1518
1451
  outputPrice: 75
1519
1452
  },
@@ -1522,8 +1455,6 @@ var bedrockModels = {
1522
1455
  contextWindow: 2e5,
1523
1456
  supportsImages: true,
1524
1457
  supportsPromptCache: false,
1525
- supportsNativeTools: true,
1526
- defaultToolProtocol: "native",
1527
1458
  inputPrice: 3,
1528
1459
  outputPrice: 15
1529
1460
  },
@@ -1532,8 +1463,6 @@ var bedrockModels = {
1532
1463
  contextWindow: 2e5,
1533
1464
  supportsImages: true,
1534
1465
  supportsPromptCache: false,
1535
- supportsNativeTools: true,
1536
- defaultToolProtocol: "native",
1537
1466
  inputPrice: 0.25,
1538
1467
  outputPrice: 1.25
1539
1468
  },
@@ -1542,7 +1471,6 @@ var bedrockModels = {
1542
1471
  contextWindow: 128e3,
1543
1472
  supportsImages: false,
1544
1473
  supportsPromptCache: false,
1545
- supportsNativeTools: true,
1546
1474
  inputPrice: 1.35,
1547
1475
  outputPrice: 5.4
1548
1476
  },
@@ -1551,7 +1479,6 @@ var bedrockModels = {
1551
1479
  contextWindow: 128e3,
1552
1480
  supportsImages: false,
1553
1481
  supportsPromptCache: false,
1554
- supportsNativeTools: true,
1555
1482
  inputPrice: 0.5,
1556
1483
  outputPrice: 1.5,
1557
1484
  description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
@@ -1561,7 +1488,6 @@ var bedrockModels = {
1561
1488
  contextWindow: 128e3,
1562
1489
  supportsImages: false,
1563
1490
  supportsPromptCache: false,
1564
- supportsNativeTools: true,
1565
1491
  inputPrice: 2,
1566
1492
  outputPrice: 6,
1567
1493
  description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
@@ -1571,7 +1497,6 @@ var bedrockModels = {
1571
1497
  contextWindow: 128e3,
1572
1498
  supportsImages: false,
1573
1499
  supportsPromptCache: false,
1574
- supportsNativeTools: true,
1575
1500
  inputPrice: 0.72,
1576
1501
  outputPrice: 0.72,
1577
1502
  description: "Llama 3.3 Instruct (70B)"
@@ -1581,7 +1506,6 @@ var bedrockModels = {
1581
1506
  contextWindow: 128e3,
1582
1507
  supportsImages: true,
1583
1508
  supportsPromptCache: false,
1584
- supportsNativeTools: true,
1585
1509
  inputPrice: 0.72,
1586
1510
  outputPrice: 0.72,
1587
1511
  description: "Llama 3.2 Instruct (90B)"
@@ -1591,7 +1515,6 @@ var bedrockModels = {
1591
1515
  contextWindow: 128e3,
1592
1516
  supportsImages: true,
1593
1517
  supportsPromptCache: false,
1594
- supportsNativeTools: true,
1595
1518
  inputPrice: 0.16,
1596
1519
  outputPrice: 0.16,
1597
1520
  description: "Llama 3.2 Instruct (11B)"
@@ -1601,7 +1524,6 @@ var bedrockModels = {
1601
1524
  contextWindow: 128e3,
1602
1525
  supportsImages: false,
1603
1526
  supportsPromptCache: false,
1604
- supportsNativeTools: true,
1605
1527
  inputPrice: 0.15,
1606
1528
  outputPrice: 0.15,
1607
1529
  description: "Llama 3.2 Instruct (3B)"
@@ -1611,7 +1533,6 @@ var bedrockModels = {
1611
1533
  contextWindow: 128e3,
1612
1534
  supportsImages: false,
1613
1535
  supportsPromptCache: false,
1614
- supportsNativeTools: true,
1615
1536
  inputPrice: 0.1,
1616
1537
  outputPrice: 0.1,
1617
1538
  description: "Llama 3.2 Instruct (1B)"
@@ -1621,7 +1542,6 @@ var bedrockModels = {
1621
1542
  contextWindow: 128e3,
1622
1543
  supportsImages: false,
1623
1544
  supportsPromptCache: false,
1624
- supportsNativeTools: true,
1625
1545
  inputPrice: 2.4,
1626
1546
  outputPrice: 2.4,
1627
1547
  description: "Llama 3.1 Instruct (405B)"
@@ -1631,7 +1551,6 @@ var bedrockModels = {
1631
1551
  contextWindow: 128e3,
1632
1552
  supportsImages: false,
1633
1553
  supportsPromptCache: false,
1634
- supportsNativeTools: true,
1635
1554
  inputPrice: 0.72,
1636
1555
  outputPrice: 0.72,
1637
1556
  description: "Llama 3.1 Instruct (70B)"
@@ -1641,7 +1560,6 @@ var bedrockModels = {
1641
1560
  contextWindow: 128e3,
1642
1561
  supportsImages: false,
1643
1562
  supportsPromptCache: false,
1644
- supportsNativeTools: true,
1645
1563
  inputPrice: 0.9,
1646
1564
  outputPrice: 0.9,
1647
1565
  description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
@@ -1651,7 +1569,6 @@ var bedrockModels = {
1651
1569
  contextWindow: 8e3,
1652
1570
  supportsImages: false,
1653
1571
  supportsPromptCache: false,
1654
- supportsNativeTools: true,
1655
1572
  inputPrice: 0.22,
1656
1573
  outputPrice: 0.22,
1657
1574
  description: "Llama 3.1 Instruct (8B)"
@@ -1661,7 +1578,6 @@ var bedrockModels = {
1661
1578
  contextWindow: 8e3,
1662
1579
  supportsImages: false,
1663
1580
  supportsPromptCache: false,
1664
- supportsNativeTools: true,
1665
1581
  inputPrice: 2.65,
1666
1582
  outputPrice: 3.5
1667
1583
  },
@@ -1670,7 +1586,6 @@ var bedrockModels = {
1670
1586
  contextWindow: 4e3,
1671
1587
  supportsImages: false,
1672
1588
  supportsPromptCache: false,
1673
- supportsNativeTools: true,
1674
1589
  inputPrice: 0.3,
1675
1590
  outputPrice: 0.6
1676
1591
  },
@@ -1679,7 +1594,6 @@ var bedrockModels = {
1679
1594
  contextWindow: 8e3,
1680
1595
  supportsImages: false,
1681
1596
  supportsPromptCache: false,
1682
- supportsNativeTools: true,
1683
1597
  inputPrice: 0.15,
1684
1598
  outputPrice: 0.2,
1685
1599
  description: "Amazon Titan Text Lite"
@@ -1689,7 +1603,6 @@ var bedrockModels = {
1689
1603
  contextWindow: 8e3,
1690
1604
  supportsImages: false,
1691
1605
  supportsPromptCache: false,
1692
- supportsNativeTools: true,
1693
1606
  inputPrice: 0.2,
1694
1607
  outputPrice: 0.6,
1695
1608
  description: "Amazon Titan Text Express"
@@ -1699,8 +1612,6 @@ var bedrockModels = {
1699
1612
  contextWindow: 262144,
1700
1613
  supportsImages: false,
1701
1614
  supportsPromptCache: false,
1702
- supportsNativeTools: true,
1703
- defaultToolProtocol: "native",
1704
1615
  preserveReasoning: true,
1705
1616
  inputPrice: 0.6,
1706
1617
  outputPrice: 2.5,
@@ -1711,8 +1622,6 @@ var bedrockModels = {
1711
1622
  contextWindow: 196608,
1712
1623
  supportsImages: false,
1713
1624
  supportsPromptCache: false,
1714
- supportsNativeTools: true,
1715
- defaultToolProtocol: "native",
1716
1625
  preserveReasoning: true,
1717
1626
  inputPrice: 0.3,
1718
1627
  outputPrice: 1.2,
@@ -1723,8 +1632,6 @@ var bedrockModels = {
1723
1632
  contextWindow: 262144,
1724
1633
  supportsImages: false,
1725
1634
  supportsPromptCache: false,
1726
- supportsNativeTools: true,
1727
- defaultToolProtocol: "native",
1728
1635
  inputPrice: 0.15,
1729
1636
  outputPrice: 1.2,
1730
1637
  description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
@@ -1734,8 +1641,6 @@ var bedrockModels = {
1734
1641
  contextWindow: 262144,
1735
1642
  supportsImages: false,
1736
1643
  supportsPromptCache: false,
1737
- supportsNativeTools: true,
1738
- defaultToolProtocol: "native",
1739
1644
  inputPrice: 0.45,
1740
1645
  outputPrice: 1.8,
1741
1646
  description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
@@ -1826,26 +1731,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
1826
1731
  // src/providers/cerebras.ts
1827
1732
  var cerebrasDefaultModelId = "gpt-oss-120b";
1828
1733
  var cerebrasModels = {
1829
- "zai-glm-4.6": {
1830
- maxTokens: 16384,
1831
- // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1832
- contextWindow: 131072,
1833
- supportsImages: false,
1834
- supportsPromptCache: false,
1835
- supportsNativeTools: true,
1836
- defaultToolProtocol: "native",
1837
- inputPrice: 0,
1838
- outputPrice: 0,
1839
- description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
1840
- },
1841
1734
  "zai-glm-4.7": {
1842
1735
  maxTokens: 16384,
1843
1736
  // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1844
1737
  contextWindow: 131072,
1845
1738
  supportsImages: false,
1846
- supportsPromptCache: false,
1847
- supportsNativeTools: true,
1848
- defaultToolProtocol: "native",
1739
+ supportsPromptCache: true,
1740
+ supportsTemperature: true,
1741
+ defaultTemperature: 1,
1849
1742
  inputPrice: 0,
1850
1743
  outputPrice: 0,
1851
1744
  description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
@@ -1856,8 +1749,6 @@ var cerebrasModels = {
1856
1749
  contextWindow: 64e3,
1857
1750
  supportsImages: false,
1858
1751
  supportsPromptCache: false,
1859
- supportsNativeTools: true,
1860
- defaultToolProtocol: "native",
1861
1752
  inputPrice: 0,
1862
1753
  outputPrice: 0,
1863
1754
  description: "Intelligent model with ~1400 tokens/s"
@@ -1868,8 +1759,6 @@ var cerebrasModels = {
1868
1759
  contextWindow: 64e3,
1869
1760
  supportsImages: false,
1870
1761
  supportsPromptCache: false,
1871
- supportsNativeTools: true,
1872
- defaultToolProtocol: "native",
1873
1762
  inputPrice: 0,
1874
1763
  outputPrice: 0,
1875
1764
  description: "Powerful model with ~2600 tokens/s"
@@ -1880,8 +1769,6 @@ var cerebrasModels = {
1880
1769
  contextWindow: 64e3,
1881
1770
  supportsImages: false,
1882
1771
  supportsPromptCache: false,
1883
- supportsNativeTools: true,
1884
- defaultToolProtocol: "native",
1885
1772
  inputPrice: 0,
1886
1773
  outputPrice: 0,
1887
1774
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1892,8 +1779,6 @@ var cerebrasModels = {
1892
1779
  contextWindow: 64e3,
1893
1780
  supportsImages: false,
1894
1781
  supportsPromptCache: false,
1895
- supportsNativeTools: true,
1896
- defaultToolProtocol: "native",
1897
1782
  inputPrice: 0,
1898
1783
  outputPrice: 0,
1899
1784
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1908,8 +1793,6 @@ var chutesModels = {
1908
1793
  contextWindow: 163840,
1909
1794
  supportsImages: false,
1910
1795
  supportsPromptCache: false,
1911
- supportsNativeTools: true,
1912
- defaultToolProtocol: "native",
1913
1796
  inputPrice: 0,
1914
1797
  outputPrice: 0,
1915
1798
  description: "DeepSeek R1 0528 model."
@@ -1919,8 +1802,6 @@ var chutesModels = {
1919
1802
  contextWindow: 163840,
1920
1803
  supportsImages: false,
1921
1804
  supportsPromptCache: false,
1922
- supportsNativeTools: true,
1923
- defaultToolProtocol: "native",
1924
1805
  inputPrice: 0,
1925
1806
  outputPrice: 0,
1926
1807
  description: "DeepSeek R1 model."
@@ -1930,8 +1811,6 @@ var chutesModels = {
1930
1811
  contextWindow: 163840,
1931
1812
  supportsImages: false,
1932
1813
  supportsPromptCache: false,
1933
- supportsNativeTools: true,
1934
- defaultToolProtocol: "native",
1935
1814
  inputPrice: 0,
1936
1815
  outputPrice: 0,
1937
1816
  description: "DeepSeek V3 model."
@@ -1941,8 +1820,6 @@ var chutesModels = {
1941
1820
  contextWindow: 163840,
1942
1821
  supportsImages: false,
1943
1822
  supportsPromptCache: false,
1944
- supportsNativeTools: true,
1945
- defaultToolProtocol: "native",
1946
1823
  inputPrice: 0,
1947
1824
  outputPrice: 0,
1948
1825
  description: "DeepSeek V3.1 model."
@@ -1952,8 +1829,6 @@ var chutesModels = {
1952
1829
  contextWindow: 163840,
1953
1830
  supportsImages: false,
1954
1831
  supportsPromptCache: false,
1955
- supportsNativeTools: true,
1956
- defaultToolProtocol: "native",
1957
1832
  inputPrice: 0.23,
1958
1833
  outputPrice: 0.9,
1959
1834
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1963,8 +1838,6 @@ var chutesModels = {
1963
1838
  contextWindow: 163840,
1964
1839
  supportsImages: false,
1965
1840
  supportsPromptCache: false,
1966
- supportsNativeTools: true,
1967
- defaultToolProtocol: "native",
1968
1841
  inputPrice: 1,
1969
1842
  outputPrice: 3,
1970
1843
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1974,8 +1847,6 @@ var chutesModels = {
1974
1847
  contextWindow: 163840,
1975
1848
  supportsImages: false,
1976
1849
  supportsPromptCache: false,
1977
- supportsNativeTools: true,
1978
- defaultToolProtocol: "native",
1979
1850
  inputPrice: 0.25,
1980
1851
  outputPrice: 0.35,
1981
1852
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1987,8 +1858,6 @@ var chutesModels = {
1987
1858
  // From Groq
1988
1859
  supportsImages: false,
1989
1860
  supportsPromptCache: false,
1990
- supportsNativeTools: true,
1991
- defaultToolProtocol: "native",
1992
1861
  inputPrice: 0,
1993
1862
  outputPrice: 0,
1994
1863
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1998,8 +1867,6 @@ var chutesModels = {
1998
1867
  contextWindow: 512e3,
1999
1868
  supportsImages: false,
2000
1869
  supportsPromptCache: false,
2001
- supportsNativeTools: true,
2002
- defaultToolProtocol: "native",
2003
1870
  inputPrice: 0,
2004
1871
  outputPrice: 0,
2005
1872
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -2009,8 +1876,6 @@ var chutesModels = {
2009
1876
  contextWindow: 128e3,
2010
1877
  supportsImages: false,
2011
1878
  supportsPromptCache: false,
2012
- supportsNativeTools: true,
2013
- defaultToolProtocol: "native",
2014
1879
  inputPrice: 0,
2015
1880
  outputPrice: 0,
2016
1881
  description: "Unsloth Mistral Nemo Instruct model."
@@ -2020,8 +1885,6 @@ var chutesModels = {
2020
1885
  contextWindow: 131072,
2021
1886
  supportsImages: false,
2022
1887
  supportsPromptCache: false,
2023
- supportsNativeTools: true,
2024
- defaultToolProtocol: "native",
2025
1888
  inputPrice: 0,
2026
1889
  outputPrice: 0,
2027
1890
  description: "Unsloth Gemma 3 12B IT model."
@@ -2031,8 +1894,6 @@ var chutesModels = {
2031
1894
  contextWindow: 131072,
2032
1895
  supportsImages: false,
2033
1896
  supportsPromptCache: false,
2034
- supportsNativeTools: true,
2035
- defaultToolProtocol: "native",
2036
1897
  inputPrice: 0,
2037
1898
  outputPrice: 0,
2038
1899
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -2042,8 +1903,6 @@ var chutesModels = {
2042
1903
  contextWindow: 131072,
2043
1904
  supportsImages: false,
2044
1905
  supportsPromptCache: false,
2045
- supportsNativeTools: true,
2046
- defaultToolProtocol: "native",
2047
1906
  inputPrice: 0,
2048
1907
  outputPrice: 0,
2049
1908
  description: "Unsloth Gemma 3 4B IT model."
@@ -2053,8 +1912,6 @@ var chutesModels = {
2053
1912
  contextWindow: 131072,
2054
1913
  supportsImages: false,
2055
1914
  supportsPromptCache: false,
2056
- supportsNativeTools: true,
2057
- defaultToolProtocol: "native",
2058
1915
  inputPrice: 0,
2059
1916
  outputPrice: 0,
2060
1917
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -2064,8 +1921,6 @@ var chutesModels = {
2064
1921
  contextWindow: 131072,
2065
1922
  supportsImages: false,
2066
1923
  supportsPromptCache: false,
2067
- supportsNativeTools: true,
2068
- defaultToolProtocol: "native",
2069
1924
  inputPrice: 0,
2070
1925
  outputPrice: 0,
2071
1926
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -2075,8 +1930,6 @@ var chutesModels = {
2075
1930
  contextWindow: 256e3,
2076
1931
  supportsImages: false,
2077
1932
  supportsPromptCache: false,
2078
- supportsNativeTools: true,
2079
- defaultToolProtocol: "native",
2080
1933
  inputPrice: 0,
2081
1934
  outputPrice: 0,
2082
1935
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -2086,8 +1939,6 @@ var chutesModels = {
2086
1939
  contextWindow: 163840,
2087
1940
  supportsImages: false,
2088
1941
  supportsPromptCache: false,
2089
- supportsNativeTools: true,
2090
- defaultToolProtocol: "native",
2091
1942
  inputPrice: 0,
2092
1943
  outputPrice: 0,
2093
1944
  description: "DeepSeek V3 Base model."
@@ -2097,8 +1948,6 @@ var chutesModels = {
2097
1948
  contextWindow: 163840,
2098
1949
  supportsImages: false,
2099
1950
  supportsPromptCache: false,
2100
- supportsNativeTools: true,
2101
- defaultToolProtocol: "native",
2102
1951
  inputPrice: 0,
2103
1952
  outputPrice: 0,
2104
1953
  description: "DeepSeek R1 Zero model."
@@ -2108,8 +1957,6 @@ var chutesModels = {
2108
1957
  contextWindow: 163840,
2109
1958
  supportsImages: false,
2110
1959
  supportsPromptCache: false,
2111
- supportsNativeTools: true,
2112
- defaultToolProtocol: "native",
2113
1960
  inputPrice: 0,
2114
1961
  outputPrice: 0,
2115
1962
  description: "DeepSeek V3 (0324) model."
@@ -2119,8 +1966,6 @@ var chutesModels = {
2119
1966
  contextWindow: 262144,
2120
1967
  supportsImages: false,
2121
1968
  supportsPromptCache: false,
2122
- supportsNativeTools: true,
2123
- defaultToolProtocol: "native",
2124
1969
  inputPrice: 0,
2125
1970
  outputPrice: 0,
2126
1971
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -2130,8 +1975,6 @@ var chutesModels = {
2130
1975
  contextWindow: 40960,
2131
1976
  supportsImages: false,
2132
1977
  supportsPromptCache: false,
2133
- supportsNativeTools: true,
2134
- defaultToolProtocol: "native",
2135
1978
  inputPrice: 0,
2136
1979
  outputPrice: 0,
2137
1980
  description: "Qwen3 235B A22B model."
@@ -2141,8 +1984,6 @@ var chutesModels = {
2141
1984
  contextWindow: 40960,
2142
1985
  supportsImages: false,
2143
1986
  supportsPromptCache: false,
2144
- supportsNativeTools: true,
2145
- defaultToolProtocol: "native",
2146
1987
  inputPrice: 0,
2147
1988
  outputPrice: 0,
2148
1989
  description: "Qwen3 32B model."
@@ -2152,8 +1993,6 @@ var chutesModels = {
2152
1993
  contextWindow: 40960,
2153
1994
  supportsImages: false,
2154
1995
  supportsPromptCache: false,
2155
- supportsNativeTools: true,
2156
- defaultToolProtocol: "native",
2157
1996
  inputPrice: 0,
2158
1997
  outputPrice: 0,
2159
1998
  description: "Qwen3 30B A3B model."
@@ -2163,8 +2002,6 @@ var chutesModels = {
2163
2002
  contextWindow: 40960,
2164
2003
  supportsImages: false,
2165
2004
  supportsPromptCache: false,
2166
- supportsNativeTools: true,
2167
- defaultToolProtocol: "native",
2168
2005
  inputPrice: 0,
2169
2006
  outputPrice: 0,
2170
2007
  description: "Qwen3 14B model."
@@ -2174,8 +2011,6 @@ var chutesModels = {
2174
2011
  contextWindow: 40960,
2175
2012
  supportsImages: false,
2176
2013
  supportsPromptCache: false,
2177
- supportsNativeTools: true,
2178
- defaultToolProtocol: "native",
2179
2014
  inputPrice: 0,
2180
2015
  outputPrice: 0,
2181
2016
  description: "Qwen3 8B model."
@@ -2185,8 +2020,6 @@ var chutesModels = {
2185
2020
  contextWindow: 163840,
2186
2021
  supportsImages: false,
2187
2022
  supportsPromptCache: false,
2188
- supportsNativeTools: true,
2189
- defaultToolProtocol: "native",
2190
2023
  inputPrice: 0,
2191
2024
  outputPrice: 0,
2192
2025
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -2196,8 +2029,6 @@ var chutesModels = {
2196
2029
  contextWindow: 163840,
2197
2030
  supportsImages: false,
2198
2031
  supportsPromptCache: false,
2199
- supportsNativeTools: true,
2200
- defaultToolProtocol: "native",
2201
2032
  inputPrice: 0,
2202
2033
  outputPrice: 0,
2203
2034
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -2207,8 +2038,6 @@ var chutesModels = {
2207
2038
  contextWindow: 151329,
2208
2039
  supportsImages: false,
2209
2040
  supportsPromptCache: false,
2210
- supportsNativeTools: true,
2211
- defaultToolProtocol: "native",
2212
2041
  inputPrice: 0,
2213
2042
  outputPrice: 0,
2214
2043
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -2218,8 +2047,6 @@ var chutesModels = {
2218
2047
  contextWindow: 131072,
2219
2048
  supportsImages: false,
2220
2049
  supportsPromptCache: false,
2221
- supportsNativeTools: true,
2222
- defaultToolProtocol: "native",
2223
2050
  inputPrice: 0,
2224
2051
  outputPrice: 0,
2225
2052
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -2229,8 +2056,6 @@ var chutesModels = {
2229
2056
  contextWindow: 131072,
2230
2057
  supportsImages: false,
2231
2058
  supportsPromptCache: false,
2232
- supportsNativeTools: true,
2233
- defaultToolProtocol: "native",
2234
2059
  inputPrice: 1,
2235
2060
  outputPrice: 3,
2236
2061
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -2240,8 +2065,6 @@ var chutesModels = {
2240
2065
  contextWindow: 202752,
2241
2066
  supportsImages: false,
2242
2067
  supportsPromptCache: false,
2243
- supportsNativeTools: true,
2244
- defaultToolProtocol: "native",
2245
2068
  inputPrice: 0,
2246
2069
  outputPrice: 0,
2247
2070
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -2252,8 +2075,6 @@ var chutesModels = {
2252
2075
  contextWindow: 202752,
2253
2076
  supportsImages: false,
2254
2077
  supportsPromptCache: false,
2255
- supportsNativeTools: true,
2256
- defaultToolProtocol: "native",
2257
2078
  inputPrice: 1.15,
2258
2079
  outputPrice: 3.25,
2259
2080
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -2263,8 +2084,6 @@ var chutesModels = {
2263
2084
  contextWindow: 128e3,
2264
2085
  supportsImages: false,
2265
2086
  supportsPromptCache: false,
2266
- supportsNativeTools: true,
2267
- defaultToolProtocol: "native",
2268
2087
  inputPrice: 0,
2269
2088
  outputPrice: 0,
2270
2089
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -2274,8 +2093,6 @@ var chutesModels = {
2274
2093
  contextWindow: 262144,
2275
2094
  supportsImages: false,
2276
2095
  supportsPromptCache: false,
2277
- supportsNativeTools: true,
2278
- defaultToolProtocol: "native",
2279
2096
  inputPrice: 0,
2280
2097
  outputPrice: 0,
2281
2098
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -2285,8 +2102,6 @@ var chutesModels = {
2285
2102
  contextWindow: 75e3,
2286
2103
  supportsImages: false,
2287
2104
  supportsPromptCache: false,
2288
- supportsNativeTools: true,
2289
- defaultToolProtocol: "native",
2290
2105
  inputPrice: 0.1481,
2291
2106
  outputPrice: 0.5926,
2292
2107
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -2296,8 +2111,6 @@ var chutesModels = {
2296
2111
  contextWindow: 262144,
2297
2112
  supportsImages: false,
2298
2113
  supportsPromptCache: false,
2299
- supportsNativeTools: true,
2300
- defaultToolProtocol: "native",
2301
2114
  inputPrice: 0.1999,
2302
2115
  outputPrice: 0.8001,
2303
2116
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -2307,8 +2120,6 @@ var chutesModels = {
2307
2120
  contextWindow: 262144,
2308
2121
  supportsImages: false,
2309
2122
  supportsPromptCache: false,
2310
- supportsNativeTools: true,
2311
- defaultToolProtocol: "native",
2312
2123
  inputPrice: 0.077968332,
2313
2124
  outputPrice: 0.31202496,
2314
2125
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -2318,8 +2129,6 @@ var chutesModels = {
2318
2129
  contextWindow: 131072,
2319
2130
  supportsImages: false,
2320
2131
  supportsPromptCache: false,
2321
- supportsNativeTools: true,
2322
- defaultToolProtocol: "native",
2323
2132
  inputPrice: 0,
2324
2133
  outputPrice: 0,
2325
2134
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -2329,8 +2138,6 @@ var chutesModels = {
2329
2138
  contextWindow: 131072,
2330
2139
  supportsImages: false,
2331
2140
  supportsPromptCache: false,
2332
- supportsNativeTools: true,
2333
- defaultToolProtocol: "native",
2334
2141
  inputPrice: 0,
2335
2142
  outputPrice: 0,
2336
2143
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -2340,8 +2147,6 @@ var chutesModels = {
2340
2147
  contextWindow: 262144,
2341
2148
  supportsImages: true,
2342
2149
  supportsPromptCache: false,
2343
- supportsNativeTools: true,
2344
- defaultToolProtocol: "native",
2345
2150
  inputPrice: 0.16,
2346
2151
  outputPrice: 0.65,
2347
2152
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -2349,73 +2154,6 @@ var chutesModels = {
2349
2154
  };
2350
2155
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
2351
2156
 
2352
- // src/providers/claude-code.ts
2353
- var DATE_SUFFIX_PATTERN = /-\d{8}$/;
2354
- var claudeCodeModels = {
2355
- "claude-haiku-4-5": {
2356
- maxTokens: 32768,
2357
- contextWindow: 2e5,
2358
- supportsImages: true,
2359
- supportsPromptCache: true,
2360
- supportsNativeTools: true,
2361
- defaultToolProtocol: "native",
2362
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2363
- reasoningEffort: "medium",
2364
- description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2365
- },
2366
- "claude-sonnet-4-5": {
2367
- maxTokens: 32768,
2368
- contextWindow: 2e5,
2369
- supportsImages: true,
2370
- supportsPromptCache: true,
2371
- supportsNativeTools: true,
2372
- defaultToolProtocol: "native",
2373
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2374
- reasoningEffort: "medium",
2375
- description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2376
- },
2377
- "claude-opus-4-5": {
2378
- maxTokens: 32768,
2379
- contextWindow: 2e5,
2380
- supportsImages: true,
2381
- supportsPromptCache: true,
2382
- supportsNativeTools: true,
2383
- defaultToolProtocol: "native",
2384
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2385
- reasoningEffort: "medium",
2386
- description: "Claude Opus 4.5 - Most capable with thinking"
2387
- }
2388
- };
2389
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2390
- var MODEL_FAMILY_PATTERNS = [
2391
- // Opus models (any version) → claude-opus-4-5
2392
- { pattern: /opus/i, target: "claude-opus-4-5" },
2393
- // Haiku models (any version) → claude-haiku-4-5
2394
- { pattern: /haiku/i, target: "claude-haiku-4-5" },
2395
- // Sonnet models (any version) → claude-sonnet-4-5
2396
- { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2397
- ];
2398
- function normalizeClaudeCodeModelId(modelId) {
2399
- if (Object.hasOwn(claudeCodeModels, modelId)) {
2400
- return modelId;
2401
- }
2402
- const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2403
- if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2404
- return withoutDate;
2405
- }
2406
- for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2407
- if (pattern.test(modelId)) {
2408
- return target;
2409
- }
2410
- }
2411
- return claudeCodeDefaultModelId;
2412
- }
2413
- var claudeCodeReasoningConfig = {
2414
- low: { budgetTokens: 16e3 },
2415
- medium: { budgetTokens: 32e3 },
2416
- high: { budgetTokens: 64e3 }
2417
- };
2418
-
2419
2157
  // src/providers/deepseek.ts
2420
2158
  var deepSeekDefaultModelId = "deepseek-chat";
2421
2159
  var deepSeekModels = {
@@ -2425,8 +2163,6 @@ var deepSeekModels = {
2425
2163
  contextWindow: 128e3,
2426
2164
  supportsImages: false,
2427
2165
  supportsPromptCache: true,
2428
- supportsNativeTools: true,
2429
- defaultToolProtocol: "native",
2430
2166
  inputPrice: 0.28,
2431
2167
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2432
2168
  outputPrice: 0.42,
@@ -2443,8 +2179,6 @@ var deepSeekModels = {
2443
2179
  contextWindow: 128e3,
2444
2180
  supportsImages: false,
2445
2181
  supportsPromptCache: true,
2446
- supportsNativeTools: true,
2447
- defaultToolProtocol: "native",
2448
2182
  preserveReasoning: true,
2449
2183
  inputPrice: 0.28,
2450
2184
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
@@ -2467,8 +2201,6 @@ var doubaoModels = {
2467
2201
  contextWindow: 128e3,
2468
2202
  supportsImages: true,
2469
2203
  supportsPromptCache: true,
2470
- supportsNativeTools: true,
2471
- defaultToolProtocol: "native",
2472
2204
  inputPrice: 1e-4,
2473
2205
  // $0.0001 per million tokens (cache miss)
2474
2206
  outputPrice: 4e-4,
@@ -2484,8 +2216,6 @@ var doubaoModels = {
2484
2216
  contextWindow: 128e3,
2485
2217
  supportsImages: true,
2486
2218
  supportsPromptCache: true,
2487
- supportsNativeTools: true,
2488
- defaultToolProtocol: "native",
2489
2219
  inputPrice: 2e-4,
2490
2220
  // $0.0002 per million tokens
2491
2221
  outputPrice: 8e-4,
@@ -2501,8 +2231,6 @@ var doubaoModels = {
2501
2231
  contextWindow: 128e3,
2502
2232
  supportsImages: true,
2503
2233
  supportsPromptCache: true,
2504
- supportsNativeTools: true,
2505
- defaultToolProtocol: "native",
2506
2234
  inputPrice: 15e-5,
2507
2235
  // $0.00015 per million tokens
2508
2236
  outputPrice: 6e-4,
@@ -2525,7 +2253,6 @@ var featherlessModels = {
2525
2253
  contextWindow: 32678,
2526
2254
  supportsImages: false,
2527
2255
  supportsPromptCache: false,
2528
- supportsNativeTools: true,
2529
2256
  inputPrice: 0,
2530
2257
  outputPrice: 0,
2531
2258
  description: "DeepSeek V3 0324 model."
@@ -2535,7 +2262,6 @@ var featherlessModels = {
2535
2262
  contextWindow: 32678,
2536
2263
  supportsImages: false,
2537
2264
  supportsPromptCache: false,
2538
- supportsNativeTools: true,
2539
2265
  inputPrice: 0,
2540
2266
  outputPrice: 0,
2541
2267
  description: "DeepSeek R1 0528 model."
@@ -2545,7 +2271,6 @@ var featherlessModels = {
2545
2271
  contextWindow: 32678,
2546
2272
  supportsImages: false,
2547
2273
  supportsPromptCache: false,
2548
- supportsNativeTools: true,
2549
2274
  inputPrice: 0,
2550
2275
  outputPrice: 0,
2551
2276
  description: "Kimi K2 Instruct model."
@@ -2555,7 +2280,6 @@ var featherlessModels = {
2555
2280
  contextWindow: 32678,
2556
2281
  supportsImages: false,
2557
2282
  supportsPromptCache: false,
2558
- supportsNativeTools: true,
2559
2283
  inputPrice: 0,
2560
2284
  outputPrice: 0,
2561
2285
  description: "GPT-OSS 120B model."
@@ -2565,7 +2289,6 @@ var featherlessModels = {
2565
2289
  contextWindow: 32678,
2566
2290
  supportsImages: false,
2567
2291
  supportsPromptCache: false,
2568
- supportsNativeTools: true,
2569
2292
  inputPrice: 0,
2570
2293
  outputPrice: 0,
2571
2294
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -2581,8 +2304,6 @@ var fireworksModels = {
2581
2304
  contextWindow: 262144,
2582
2305
  supportsImages: false,
2583
2306
  supportsPromptCache: true,
2584
- supportsNativeTools: true,
2585
- defaultToolProtocol: "native",
2586
2307
  inputPrice: 0.6,
2587
2308
  outputPrice: 2.5,
2588
2309
  cacheReadsPrice: 0.15,
@@ -2593,8 +2314,6 @@ var fireworksModels = {
2593
2314
  contextWindow: 128e3,
2594
2315
  supportsImages: false,
2595
2316
  supportsPromptCache: false,
2596
- supportsNativeTools: true,
2597
- defaultToolProtocol: "native",
2598
2317
  inputPrice: 0.6,
2599
2318
  outputPrice: 2.5,
2600
2319
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2604,7 +2323,6 @@ var fireworksModels = {
2604
2323
  contextWindow: 256e3,
2605
2324
  supportsImages: false,
2606
2325
  supportsPromptCache: true,
2607
- supportsNativeTools: true,
2608
2326
  supportsTemperature: true,
2609
2327
  preserveReasoning: true,
2610
2328
  defaultTemperature: 1,
@@ -2618,8 +2336,6 @@ var fireworksModels = {
2618
2336
  contextWindow: 204800,
2619
2337
  supportsImages: false,
2620
2338
  supportsPromptCache: false,
2621
- supportsNativeTools: true,
2622
- defaultToolProtocol: "native",
2623
2339
  inputPrice: 0.3,
2624
2340
  outputPrice: 1.2,
2625
2341
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2629,8 +2345,6 @@ var fireworksModels = {
2629
2345
  contextWindow: 256e3,
2630
2346
  supportsImages: false,
2631
2347
  supportsPromptCache: false,
2632
- supportsNativeTools: true,
2633
- defaultToolProtocol: "native",
2634
2348
  inputPrice: 0.22,
2635
2349
  outputPrice: 0.88,
2636
2350
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2640,8 +2354,6 @@ var fireworksModels = {
2640
2354
  contextWindow: 256e3,
2641
2355
  supportsImages: false,
2642
2356
  supportsPromptCache: false,
2643
- supportsNativeTools: true,
2644
- defaultToolProtocol: "native",
2645
2357
  inputPrice: 0.45,
2646
2358
  outputPrice: 1.8,
2647
2359
  description: "Qwen3's most agentic code model to date."
@@ -2651,8 +2363,6 @@ var fireworksModels = {
2651
2363
  contextWindow: 16e4,
2652
2364
  supportsImages: false,
2653
2365
  supportsPromptCache: false,
2654
- supportsNativeTools: true,
2655
- defaultToolProtocol: "native",
2656
2366
  inputPrice: 3,
2657
2367
  outputPrice: 8,
2658
2368
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2662,8 +2372,6 @@ var fireworksModels = {
2662
2372
  contextWindow: 128e3,
2663
2373
  supportsImages: false,
2664
2374
  supportsPromptCache: false,
2665
- supportsNativeTools: true,
2666
- defaultToolProtocol: "native",
2667
2375
  inputPrice: 0.9,
2668
2376
  outputPrice: 0.9,
2669
2377
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2673,8 +2381,6 @@ var fireworksModels = {
2673
2381
  contextWindow: 163840,
2674
2382
  supportsImages: false,
2675
2383
  supportsPromptCache: false,
2676
- supportsNativeTools: true,
2677
- defaultToolProtocol: "native",
2678
2384
  inputPrice: 0.56,
2679
2385
  outputPrice: 1.68,
2680
2386
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2684,8 +2390,6 @@ var fireworksModels = {
2684
2390
  contextWindow: 128e3,
2685
2391
  supportsImages: false,
2686
2392
  supportsPromptCache: false,
2687
- supportsNativeTools: true,
2688
- defaultToolProtocol: "native",
2689
2393
  inputPrice: 0.55,
2690
2394
  outputPrice: 2.19,
2691
2395
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2695,8 +2399,6 @@ var fireworksModels = {
2695
2399
  contextWindow: 128e3,
2696
2400
  supportsImages: false,
2697
2401
  supportsPromptCache: false,
2698
- supportsNativeTools: true,
2699
- defaultToolProtocol: "native",
2700
2402
  inputPrice: 0.55,
2701
2403
  outputPrice: 2.19,
2702
2404
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2706,8 +2408,6 @@ var fireworksModels = {
2706
2408
  contextWindow: 198e3,
2707
2409
  supportsImages: false,
2708
2410
  supportsPromptCache: false,
2709
- supportsNativeTools: true,
2710
- defaultToolProtocol: "native",
2711
2411
  inputPrice: 0.55,
2712
2412
  outputPrice: 2.19,
2713
2413
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2717,8 +2417,6 @@ var fireworksModels = {
2717
2417
  contextWindow: 128e3,
2718
2418
  supportsImages: false,
2719
2419
  supportsPromptCache: false,
2720
- supportsNativeTools: true,
2721
- defaultToolProtocol: "native",
2722
2420
  inputPrice: 0.07,
2723
2421
  outputPrice: 0.3,
2724
2422
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2728,11 +2426,63 @@ var fireworksModels = {
2728
2426
  contextWindow: 128e3,
2729
2427
  supportsImages: false,
2730
2428
  supportsPromptCache: false,
2731
- supportsNativeTools: true,
2732
- defaultToolProtocol: "native",
2733
2429
  inputPrice: 0.15,
2734
2430
  outputPrice: 0.6,
2735
2431
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
2432
+ },
2433
+ "accounts/fireworks/models/minimax-m2p1": {
2434
+ maxTokens: 4096,
2435
+ contextWindow: 204800,
2436
+ supportsImages: false,
2437
+ supportsPromptCache: false,
2438
+ inputPrice: 0.3,
2439
+ outputPrice: 1.2,
2440
+ description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
2441
+ },
2442
+ "accounts/fireworks/models/deepseek-v3p2": {
2443
+ maxTokens: 16384,
2444
+ contextWindow: 163840,
2445
+ supportsImages: false,
2446
+ supportsPromptCache: false,
2447
+ inputPrice: 0.56,
2448
+ outputPrice: 1.68,
2449
+ description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
2450
+ },
2451
+ "accounts/fireworks/models/glm-4p7": {
2452
+ maxTokens: 25344,
2453
+ contextWindow: 198e3,
2454
+ supportsImages: false,
2455
+ supportsPromptCache: false,
2456
+ inputPrice: 0.55,
2457
+ outputPrice: 2.19,
2458
+ description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
2459
+ },
2460
+ "accounts/fireworks/models/llama-v3p3-70b-instruct": {
2461
+ maxTokens: 16384,
2462
+ contextWindow: 131072,
2463
+ supportsImages: false,
2464
+ supportsPromptCache: false,
2465
+ inputPrice: 0.9,
2466
+ outputPrice: 0.9,
2467
+ description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
2468
+ },
2469
+ "accounts/fireworks/models/llama4-maverick-instruct-basic": {
2470
+ maxTokens: 16384,
2471
+ contextWindow: 131072,
2472
+ supportsImages: true,
2473
+ supportsPromptCache: false,
2474
+ inputPrice: 0.22,
2475
+ outputPrice: 0.88,
2476
+ description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
2477
+ },
2478
+ "accounts/fireworks/models/llama4-scout-instruct-basic": {
2479
+ maxTokens: 16384,
2480
+ contextWindow: 131072,
2481
+ supportsImages: true,
2482
+ supportsPromptCache: false,
2483
+ inputPrice: 0.15,
2484
+ outputPrice: 0.6,
2485
+ description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
2736
2486
  }
2737
2487
  };
2738
2488
 
@@ -2743,8 +2493,6 @@ var geminiModels = {
2743
2493
  maxTokens: 65536,
2744
2494
  contextWindow: 1048576,
2745
2495
  supportsImages: true,
2746
- supportsNativeTools: true,
2747
- defaultToolProtocol: "native",
2748
2496
  supportsPromptCache: true,
2749
2497
  supportsReasoningEffort: ["low", "high"],
2750
2498
  reasoningEffort: "low",
@@ -2752,16 +2500,19 @@ var geminiModels = {
2752
2500
  defaultTemperature: 1,
2753
2501
  inputPrice: 4,
2754
2502
  outputPrice: 18,
2503
+ cacheReadsPrice: 0.4,
2755
2504
  tiers: [
2756
2505
  {
2757
2506
  contextWindow: 2e5,
2758
2507
  inputPrice: 2,
2759
- outputPrice: 12
2508
+ outputPrice: 12,
2509
+ cacheReadsPrice: 0.2
2760
2510
  },
2761
2511
  {
2762
2512
  contextWindow: Infinity,
2763
2513
  inputPrice: 4,
2764
- outputPrice: 18
2514
+ outputPrice: 18,
2515
+ cacheReadsPrice: 0.4
2765
2516
  }
2766
2517
  ]
2767
2518
  },
@@ -2769,25 +2520,20 @@ var geminiModels = {
2769
2520
  maxTokens: 65536,
2770
2521
  contextWindow: 1048576,
2771
2522
  supportsImages: true,
2772
- supportsNativeTools: true,
2773
- defaultToolProtocol: "native",
2774
2523
  supportsPromptCache: true,
2775
2524
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2776
2525
  reasoningEffort: "medium",
2777
2526
  supportsTemperature: true,
2778
2527
  defaultTemperature: 1,
2779
- inputPrice: 0.3,
2780
- outputPrice: 2.5,
2781
- cacheReadsPrice: 0.075,
2782
- cacheWritesPrice: 1
2528
+ inputPrice: 0.5,
2529
+ outputPrice: 3,
2530
+ cacheReadsPrice: 0.05
2783
2531
  },
2784
2532
  // 2.5 Pro models
2785
2533
  "gemini-2.5-pro": {
2786
2534
  maxTokens: 64e3,
2787
2535
  contextWindow: 1048576,
2788
2536
  supportsImages: true,
2789
- supportsNativeTools: true,
2790
- defaultToolProtocol: "native",
2791
2537
  supportsPromptCache: true,
2792
2538
  inputPrice: 2.5,
2793
2539
  // This is the pricing for prompts above 200k tokens.
@@ -2816,8 +2562,6 @@ var geminiModels = {
2816
2562
  maxTokens: 65535,
2817
2563
  contextWindow: 1048576,
2818
2564
  supportsImages: true,
2819
- supportsNativeTools: true,
2820
- defaultToolProtocol: "native",
2821
2565
  supportsPromptCache: true,
2822
2566
  inputPrice: 2.5,
2823
2567
  // This is the pricing for prompts above 200k tokens.
@@ -2845,8 +2589,6 @@ var geminiModels = {
2845
2589
  maxTokens: 65535,
2846
2590
  contextWindow: 1048576,
2847
2591
  supportsImages: true,
2848
- supportsNativeTools: true,
2849
- defaultToolProtocol: "native",
2850
2592
  supportsPromptCache: true,
2851
2593
  inputPrice: 2.5,
2852
2594
  // This is the pricing for prompts above 200k tokens.
@@ -2872,8 +2614,6 @@ var geminiModels = {
2872
2614
  maxTokens: 65535,
2873
2615
  contextWindow: 1048576,
2874
2616
  supportsImages: true,
2875
- supportsNativeTools: true,
2876
- defaultToolProtocol: "native",
2877
2617
  supportsPromptCache: true,
2878
2618
  inputPrice: 2.5,
2879
2619
  // This is the pricing for prompts above 200k tokens.
@@ -2902,8 +2642,6 @@ var geminiModels = {
2902
2642
  maxTokens: 65536,
2903
2643
  contextWindow: 1048576,
2904
2644
  supportsImages: true,
2905
- supportsNativeTools: true,
2906
- defaultToolProtocol: "native",
2907
2645
  supportsPromptCache: true,
2908
2646
  inputPrice: 0.3,
2909
2647
  outputPrice: 2.5,
@@ -2916,8 +2654,6 @@ var geminiModels = {
2916
2654
  maxTokens: 65536,
2917
2655
  contextWindow: 1048576,
2918
2656
  supportsImages: true,
2919
- supportsNativeTools: true,
2920
- defaultToolProtocol: "native",
2921
2657
  supportsPromptCache: true,
2922
2658
  inputPrice: 0.3,
2923
2659
  outputPrice: 2.5,
@@ -2930,8 +2666,6 @@ var geminiModels = {
2930
2666
  maxTokens: 64e3,
2931
2667
  contextWindow: 1048576,
2932
2668
  supportsImages: true,
2933
- supportsNativeTools: true,
2934
- defaultToolProtocol: "native",
2935
2669
  supportsPromptCache: true,
2936
2670
  inputPrice: 0.3,
2937
2671
  outputPrice: 2.5,
@@ -2945,8 +2679,6 @@ var geminiModels = {
2945
2679
  maxTokens: 65536,
2946
2680
  contextWindow: 1048576,
2947
2681
  supportsImages: true,
2948
- supportsNativeTools: true,
2949
- defaultToolProtocol: "native",
2950
2682
  supportsPromptCache: true,
2951
2683
  inputPrice: 0.1,
2952
2684
  outputPrice: 0.4,
@@ -2959,8 +2691,6 @@ var geminiModels = {
2959
2691
  maxTokens: 65536,
2960
2692
  contextWindow: 1048576,
2961
2693
  supportsImages: true,
2962
- supportsNativeTools: true,
2963
- defaultToolProtocol: "native",
2964
2694
  supportsPromptCache: true,
2965
2695
  inputPrice: 0.1,
2966
2696
  outputPrice: 0.4,
@@ -2980,8 +2710,6 @@ var groqModels = {
2980
2710
  contextWindow: 131072,
2981
2711
  supportsImages: false,
2982
2712
  supportsPromptCache: false,
2983
- supportsNativeTools: true,
2984
- defaultToolProtocol: "native",
2985
2713
  inputPrice: 0.05,
2986
2714
  outputPrice: 0.08,
2987
2715
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2991,8 +2719,6 @@ var groqModels = {
2991
2719
  contextWindow: 131072,
2992
2720
  supportsImages: false,
2993
2721
  supportsPromptCache: false,
2994
- supportsNativeTools: true,
2995
- defaultToolProtocol: "native",
2996
2722
  inputPrice: 0.59,
2997
2723
  outputPrice: 0.79,
2998
2724
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -3002,8 +2728,6 @@ var groqModels = {
3002
2728
  contextWindow: 131072,
3003
2729
  supportsImages: false,
3004
2730
  supportsPromptCache: false,
3005
- supportsNativeTools: true,
3006
- defaultToolProtocol: "native",
3007
2731
  inputPrice: 0.11,
3008
2732
  outputPrice: 0.34,
3009
2733
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -3013,8 +2737,6 @@ var groqModels = {
3013
2737
  contextWindow: 131072,
3014
2738
  supportsImages: false,
3015
2739
  supportsPromptCache: false,
3016
- supportsNativeTools: true,
3017
- defaultToolProtocol: "native",
3018
2740
  inputPrice: 0.29,
3019
2741
  outputPrice: 0.59,
3020
2742
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -3024,8 +2746,6 @@ var groqModels = {
3024
2746
  contextWindow: 262144,
3025
2747
  supportsImages: false,
3026
2748
  supportsPromptCache: true,
3027
- supportsNativeTools: true,
3028
- defaultToolProtocol: "native",
3029
2749
  inputPrice: 0.6,
3030
2750
  outputPrice: 2.5,
3031
2751
  cacheReadsPrice: 0.15,
@@ -3036,8 +2756,6 @@ var groqModels = {
3036
2756
  contextWindow: 131072,
3037
2757
  supportsImages: false,
3038
2758
  supportsPromptCache: false,
3039
- supportsNativeTools: true,
3040
- defaultToolProtocol: "native",
3041
2759
  inputPrice: 0.15,
3042
2760
  outputPrice: 0.75,
3043
2761
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -3047,8 +2765,6 @@ var groqModels = {
3047
2765
  contextWindow: 131072,
3048
2766
  supportsImages: false,
3049
2767
  supportsPromptCache: false,
3050
- supportsNativeTools: true,
3051
- defaultToolProtocol: "native",
3052
2768
  inputPrice: 0.1,
3053
2769
  outputPrice: 0.5,
3054
2770
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -3075,7 +2791,6 @@ var ioIntelligenceModels = {
3075
2791
  contextWindow: 128e3,
3076
2792
  supportsImages: false,
3077
2793
  supportsPromptCache: false,
3078
- supportsNativeTools: true,
3079
2794
  description: "DeepSeek R1 reasoning model"
3080
2795
  },
3081
2796
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -3083,7 +2798,6 @@ var ioIntelligenceModels = {
3083
2798
  contextWindow: 43e4,
3084
2799
  supportsImages: true,
3085
2800
  supportsPromptCache: false,
3086
- supportsNativeTools: true,
3087
2801
  description: "Llama 4 Maverick 17B model"
3088
2802
  },
3089
2803
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -3091,7 +2805,6 @@ var ioIntelligenceModels = {
3091
2805
  contextWindow: 106e3,
3092
2806
  supportsImages: false,
3093
2807
  supportsPromptCache: false,
3094
- supportsNativeTools: true,
3095
2808
  description: "Qwen3 Coder 480B specialized for coding"
3096
2809
  },
3097
2810
  "openai/gpt-oss-120b": {
@@ -3099,7 +2812,6 @@ var ioIntelligenceModels = {
3099
2812
  contextWindow: 131072,
3100
2813
  supportsImages: false,
3101
2814
  supportsPromptCache: false,
3102
- supportsNativeTools: true,
3103
2815
  description: "OpenAI GPT-OSS 120B model"
3104
2816
  }
3105
2817
  };
@@ -3111,8 +2823,6 @@ var litellmDefaultModelInfo = {
3111
2823
  contextWindow: 2e5,
3112
2824
  supportsImages: true,
3113
2825
  supportsPromptCache: true,
3114
- supportsNativeTools: true,
3115
- defaultToolProtocol: "native",
3116
2826
  inputPrice: 3,
3117
2827
  outputPrice: 15,
3118
2828
  cacheWritesPrice: 3.75,
@@ -3127,8 +2837,6 @@ var lMStudioDefaultModelInfo = {
3127
2837
  contextWindow: 2e5,
3128
2838
  supportsImages: true,
3129
2839
  supportsPromptCache: true,
3130
- supportsNativeTools: true,
3131
- defaultToolProtocol: "native",
3132
2840
  inputPrice: 0,
3133
2841
  outputPrice: 0,
3134
2842
  cacheWritesPrice: 0,
@@ -3144,8 +2852,6 @@ var mistralModels = {
3144
2852
  contextWindow: 128e3,
3145
2853
  supportsImages: true,
3146
2854
  supportsPromptCache: false,
3147
- supportsNativeTools: true,
3148
- defaultToolProtocol: "native",
3149
2855
  inputPrice: 2,
3150
2856
  outputPrice: 5
3151
2857
  },
@@ -3154,8 +2860,6 @@ var mistralModels = {
3154
2860
  contextWindow: 131e3,
3155
2861
  supportsImages: true,
3156
2862
  supportsPromptCache: false,
3157
- supportsNativeTools: true,
3158
- defaultToolProtocol: "native",
3159
2863
  inputPrice: 0.4,
3160
2864
  outputPrice: 2
3161
2865
  },
@@ -3164,8 +2868,6 @@ var mistralModels = {
3164
2868
  contextWindow: 131e3,
3165
2869
  supportsImages: true,
3166
2870
  supportsPromptCache: false,
3167
- supportsNativeTools: true,
3168
- defaultToolProtocol: "native",
3169
2871
  inputPrice: 0.4,
3170
2872
  outputPrice: 2
3171
2873
  },
@@ -3174,8 +2876,6 @@ var mistralModels = {
3174
2876
  contextWindow: 256e3,
3175
2877
  supportsImages: false,
3176
2878
  supportsPromptCache: false,
3177
- supportsNativeTools: true,
3178
- defaultToolProtocol: "native",
3179
2879
  inputPrice: 0.3,
3180
2880
  outputPrice: 0.9
3181
2881
  },
@@ -3184,8 +2884,6 @@ var mistralModels = {
3184
2884
  contextWindow: 131e3,
3185
2885
  supportsImages: false,
3186
2886
  supportsPromptCache: false,
3187
- supportsNativeTools: true,
3188
- defaultToolProtocol: "native",
3189
2887
  inputPrice: 2,
3190
2888
  outputPrice: 6
3191
2889
  },
@@ -3194,8 +2892,6 @@ var mistralModels = {
3194
2892
  contextWindow: 131e3,
3195
2893
  supportsImages: false,
3196
2894
  supportsPromptCache: false,
3197
- supportsNativeTools: true,
3198
- defaultToolProtocol: "native",
3199
2895
  inputPrice: 0.1,
3200
2896
  outputPrice: 0.1
3201
2897
  },
@@ -3204,8 +2900,6 @@ var mistralModels = {
3204
2900
  contextWindow: 131e3,
3205
2901
  supportsImages: false,
3206
2902
  supportsPromptCache: false,
3207
- supportsNativeTools: true,
3208
- defaultToolProtocol: "native",
3209
2903
  inputPrice: 0.04,
3210
2904
  outputPrice: 0.04
3211
2905
  },
@@ -3214,8 +2908,6 @@ var mistralModels = {
3214
2908
  contextWindow: 32e3,
3215
2909
  supportsImages: false,
3216
2910
  supportsPromptCache: false,
3217
- supportsNativeTools: true,
3218
- defaultToolProtocol: "native",
3219
2911
  inputPrice: 0.2,
3220
2912
  outputPrice: 0.6
3221
2913
  },
@@ -3224,8 +2916,6 @@ var mistralModels = {
3224
2916
  contextWindow: 131e3,
3225
2917
  supportsImages: true,
3226
2918
  supportsPromptCache: false,
3227
- supportsNativeTools: true,
3228
- defaultToolProtocol: "native",
3229
2919
  inputPrice: 2,
3230
2920
  outputPrice: 6
3231
2921
  }
@@ -3240,8 +2930,6 @@ var moonshotModels = {
3240
2930
  contextWindow: 131072,
3241
2931
  supportsImages: false,
3242
2932
  supportsPromptCache: true,
3243
- supportsNativeTools: true,
3244
- defaultToolProtocol: "native",
3245
2933
  inputPrice: 0.6,
3246
2934
  // $0.60 per million tokens (cache miss)
3247
2935
  outputPrice: 2.5,
@@ -3257,8 +2945,6 @@ var moonshotModels = {
3257
2945
  contextWindow: 262144,
3258
2946
  supportsImages: false,
3259
2947
  supportsPromptCache: true,
3260
- supportsNativeTools: true,
3261
- defaultToolProtocol: "native",
3262
2948
  inputPrice: 0.6,
3263
2949
  outputPrice: 2.5,
3264
2950
  cacheReadsPrice: 0.15,
@@ -3269,8 +2955,6 @@ var moonshotModels = {
3269
2955
  contextWindow: 262144,
3270
2956
  supportsImages: false,
3271
2957
  supportsPromptCache: true,
3272
- supportsNativeTools: true,
3273
- defaultToolProtocol: "native",
3274
2958
  inputPrice: 2.4,
3275
2959
  // $2.40 per million tokens (cache miss)
3276
2960
  outputPrice: 10,
@@ -3289,8 +2973,6 @@ var moonshotModels = {
3289
2973
  supportsImages: false,
3290
2974
  // Text-only (no image/vision support)
3291
2975
  supportsPromptCache: true,
3292
- supportsNativeTools: true,
3293
- defaultToolProtocol: "native",
3294
2976
  inputPrice: 0.6,
3295
2977
  // $0.60 per million tokens (cache miss)
3296
2978
  outputPrice: 2.5,
@@ -3304,6 +2986,21 @@ var moonshotModels = {
3304
2986
  preserveReasoning: true,
3305
2987
  defaultTemperature: 1,
3306
2988
  description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
2989
+ },
2990
+ "kimi-k2.5": {
2991
+ maxTokens: 16384,
2992
+ contextWindow: 262144,
2993
+ supportsImages: false,
2994
+ supportsPromptCache: true,
2995
+ inputPrice: 0.6,
2996
+ // $0.60 per million tokens (cache miss)
2997
+ outputPrice: 3,
2998
+ // $3.00 per million tokens
2999
+ cacheReadsPrice: 0.1,
3000
+ // $0.10 per million tokens (cache hit)
3001
+ supportsTemperature: true,
3002
+ defaultTemperature: 1,
3003
+ description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
3307
3004
  }
3308
3005
  };
3309
3006
  var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
@@ -3315,7 +3012,6 @@ var ollamaDefaultModelInfo = {
3315
3012
  contextWindow: 2e5,
3316
3013
  supportsImages: true,
3317
3014
  supportsPromptCache: true,
3318
- supportsNativeTools: true,
3319
3015
  inputPrice: 0,
3320
3016
  outputPrice: 0,
3321
3017
  cacheWritesPrice: 0,
@@ -3329,8 +3025,6 @@ var openAiNativeModels = {
3329
3025
  "gpt-5.1-codex-max": {
3330
3026
  maxTokens: 128e3,
3331
3027
  contextWindow: 4e5,
3332
- supportsNativeTools: true,
3333
- defaultToolProtocol: "native",
3334
3028
  includedTools: ["apply_patch"],
3335
3029
  excludedTools: ["apply_diff", "write_to_file"],
3336
3030
  supportsImages: true,
@@ -3348,8 +3042,6 @@ var openAiNativeModels = {
3348
3042
  "gpt-5.2": {
3349
3043
  maxTokens: 128e3,
3350
3044
  contextWindow: 4e5,
3351
- supportsNativeTools: true,
3352
- defaultToolProtocol: "native",
3353
3045
  includedTools: ["apply_patch"],
3354
3046
  excludedTools: ["apply_diff", "write_to_file"],
3355
3047
  supportsImages: true,
@@ -3368,11 +3060,26 @@ var openAiNativeModels = {
3368
3060
  ],
3369
3061
  description: "GPT-5.2: Our flagship model for coding and agentic tasks across industries"
3370
3062
  },
3063
+ "gpt-5.2-codex": {
3064
+ maxTokens: 128e3,
3065
+ contextWindow: 4e5,
3066
+ includedTools: ["apply_patch"],
3067
+ excludedTools: ["apply_diff", "write_to_file"],
3068
+ supportsImages: true,
3069
+ supportsPromptCache: true,
3070
+ promptCacheRetention: "24h",
3071
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3072
+ reasoningEffort: "medium",
3073
+ inputPrice: 1.75,
3074
+ outputPrice: 14,
3075
+ cacheReadsPrice: 0.175,
3076
+ supportsTemperature: false,
3077
+ tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 3.5, outputPrice: 28, cacheReadsPrice: 0.35 }],
3078
+ description: "GPT-5.2 Codex: Our most intelligent coding model optimized for long-horizon, agentic coding tasks"
3079
+ },
3371
3080
  "gpt-5.2-chat-latest": {
3372
3081
  maxTokens: 16384,
3373
3082
  contextWindow: 128e3,
3374
- supportsNativeTools: true,
3375
- defaultToolProtocol: "native",
3376
3083
  includedTools: ["apply_patch"],
3377
3084
  excludedTools: ["apply_diff", "write_to_file"],
3378
3085
  supportsImages: true,
@@ -3385,8 +3092,6 @@ var openAiNativeModels = {
3385
3092
  "gpt-5.1": {
3386
3093
  maxTokens: 128e3,
3387
3094
  contextWindow: 4e5,
3388
- supportsNativeTools: true,
3389
- defaultToolProtocol: "native",
3390
3095
  includedTools: ["apply_patch"],
3391
3096
  excludedTools: ["apply_diff", "write_to_file"],
3392
3097
  supportsImages: true,
@@ -3408,8 +3113,6 @@ var openAiNativeModels = {
3408
3113
  "gpt-5.1-codex": {
3409
3114
  maxTokens: 128e3,
3410
3115
  contextWindow: 4e5,
3411
- supportsNativeTools: true,
3412
- defaultToolProtocol: "native",
3413
3116
  includedTools: ["apply_patch"],
3414
3117
  excludedTools: ["apply_diff", "write_to_file"],
3415
3118
  supportsImages: true,
@@ -3427,8 +3130,6 @@ var openAiNativeModels = {
3427
3130
  "gpt-5.1-codex-mini": {
3428
3131
  maxTokens: 128e3,
3429
3132
  contextWindow: 4e5,
3430
- supportsNativeTools: true,
3431
- defaultToolProtocol: "native",
3432
3133
  includedTools: ["apply_patch"],
3433
3134
  excludedTools: ["apply_diff", "write_to_file"],
3434
3135
  supportsImages: true,
@@ -3445,8 +3146,6 @@ var openAiNativeModels = {
3445
3146
  "gpt-5": {
3446
3147
  maxTokens: 128e3,
3447
3148
  contextWindow: 4e5,
3448
- supportsNativeTools: true,
3449
- defaultToolProtocol: "native",
3450
3149
  includedTools: ["apply_patch"],
3451
3150
  excludedTools: ["apply_diff", "write_to_file"],
3452
3151
  supportsImages: true,
@@ -3467,8 +3166,6 @@ var openAiNativeModels = {
3467
3166
  "gpt-5-mini": {
3468
3167
  maxTokens: 128e3,
3469
3168
  contextWindow: 4e5,
3470
- supportsNativeTools: true,
3471
- defaultToolProtocol: "native",
3472
3169
  includedTools: ["apply_patch"],
3473
3170
  excludedTools: ["apply_diff", "write_to_file"],
3474
3171
  supportsImages: true,
@@ -3489,8 +3186,6 @@ var openAiNativeModels = {
3489
3186
  "gpt-5-codex": {
3490
3187
  maxTokens: 128e3,
3491
3188
  contextWindow: 4e5,
3492
- supportsNativeTools: true,
3493
- defaultToolProtocol: "native",
3494
3189
  includedTools: ["apply_patch"],
3495
3190
  excludedTools: ["apply_diff", "write_to_file"],
3496
3191
  supportsImages: true,
@@ -3507,8 +3202,6 @@ var openAiNativeModels = {
3507
3202
  "gpt-5-nano": {
3508
3203
  maxTokens: 128e3,
3509
3204
  contextWindow: 4e5,
3510
- supportsNativeTools: true,
3511
- defaultToolProtocol: "native",
3512
3205
  includedTools: ["apply_patch"],
3513
3206
  excludedTools: ["apply_diff", "write_to_file"],
3514
3207
  supportsImages: true,
@@ -3526,8 +3219,6 @@ var openAiNativeModels = {
3526
3219
  "gpt-5-chat-latest": {
3527
3220
  maxTokens: 128e3,
3528
3221
  contextWindow: 4e5,
3529
- supportsNativeTools: true,
3530
- defaultToolProtocol: "native",
3531
3222
  includedTools: ["apply_patch"],
3532
3223
  excludedTools: ["apply_diff", "write_to_file"],
3533
3224
  supportsImages: true,
@@ -3540,8 +3231,6 @@ var openAiNativeModels = {
3540
3231
  "gpt-4.1": {
3541
3232
  maxTokens: 32768,
3542
3233
  contextWindow: 1047576,
3543
- supportsNativeTools: true,
3544
- defaultToolProtocol: "native",
3545
3234
  includedTools: ["apply_patch"],
3546
3235
  excludedTools: ["apply_diff", "write_to_file"],
3547
3236
  supportsImages: true,
@@ -3557,8 +3246,6 @@ var openAiNativeModels = {
3557
3246
  "gpt-4.1-mini": {
3558
3247
  maxTokens: 32768,
3559
3248
  contextWindow: 1047576,
3560
- supportsNativeTools: true,
3561
- defaultToolProtocol: "native",
3562
3249
  includedTools: ["apply_patch"],
3563
3250
  excludedTools: ["apply_diff", "write_to_file"],
3564
3251
  supportsImages: true,
@@ -3574,8 +3261,6 @@ var openAiNativeModels = {
3574
3261
  "gpt-4.1-nano": {
3575
3262
  maxTokens: 32768,
3576
3263
  contextWindow: 1047576,
3577
- supportsNativeTools: true,
3578
- defaultToolProtocol: "native",
3579
3264
  includedTools: ["apply_patch"],
3580
3265
  excludedTools: ["apply_diff", "write_to_file"],
3581
3266
  supportsImages: true,
@@ -3591,8 +3276,6 @@ var openAiNativeModels = {
3591
3276
  o3: {
3592
3277
  maxTokens: 1e5,
3593
3278
  contextWindow: 2e5,
3594
- supportsNativeTools: true,
3595
- defaultToolProtocol: "native",
3596
3279
  supportsImages: true,
3597
3280
  supportsPromptCache: true,
3598
3281
  inputPrice: 2,
@@ -3609,8 +3292,6 @@ var openAiNativeModels = {
3609
3292
  "o3-high": {
3610
3293
  maxTokens: 1e5,
3611
3294
  contextWindow: 2e5,
3612
- supportsNativeTools: true,
3613
- defaultToolProtocol: "native",
3614
3295
  supportsImages: true,
3615
3296
  supportsPromptCache: true,
3616
3297
  inputPrice: 2,
@@ -3622,8 +3303,6 @@ var openAiNativeModels = {
3622
3303
  "o3-low": {
3623
3304
  maxTokens: 1e5,
3624
3305
  contextWindow: 2e5,
3625
- supportsNativeTools: true,
3626
- defaultToolProtocol: "native",
3627
3306
  supportsImages: true,
3628
3307
  supportsPromptCache: true,
3629
3308
  inputPrice: 2,
@@ -3635,8 +3314,6 @@ var openAiNativeModels = {
3635
3314
  "o4-mini": {
3636
3315
  maxTokens: 1e5,
3637
3316
  contextWindow: 2e5,
3638
- supportsNativeTools: true,
3639
- defaultToolProtocol: "native",
3640
3317
  supportsImages: true,
3641
3318
  supportsPromptCache: true,
3642
3319
  inputPrice: 1.1,
@@ -3653,8 +3330,6 @@ var openAiNativeModels = {
3653
3330
  "o4-mini-high": {
3654
3331
  maxTokens: 1e5,
3655
3332
  contextWindow: 2e5,
3656
- supportsNativeTools: true,
3657
- defaultToolProtocol: "native",
3658
3333
  supportsImages: true,
3659
3334
  supportsPromptCache: true,
3660
3335
  inputPrice: 1.1,
@@ -3666,8 +3341,6 @@ var openAiNativeModels = {
3666
3341
  "o4-mini-low": {
3667
3342
  maxTokens: 1e5,
3668
3343
  contextWindow: 2e5,
3669
- supportsNativeTools: true,
3670
- defaultToolProtocol: "native",
3671
3344
  supportsImages: true,
3672
3345
  supportsPromptCache: true,
3673
3346
  inputPrice: 1.1,
@@ -3679,8 +3352,6 @@ var openAiNativeModels = {
3679
3352
  "o3-mini": {
3680
3353
  maxTokens: 1e5,
3681
3354
  contextWindow: 2e5,
3682
- supportsNativeTools: true,
3683
- defaultToolProtocol: "native",
3684
3355
  supportsImages: false,
3685
3356
  supportsPromptCache: true,
3686
3357
  inputPrice: 1.1,
@@ -3693,8 +3364,6 @@ var openAiNativeModels = {
3693
3364
  "o3-mini-high": {
3694
3365
  maxTokens: 1e5,
3695
3366
  contextWindow: 2e5,
3696
- supportsNativeTools: true,
3697
- defaultToolProtocol: "native",
3698
3367
  supportsImages: false,
3699
3368
  supportsPromptCache: true,
3700
3369
  inputPrice: 1.1,
@@ -3706,8 +3375,6 @@ var openAiNativeModels = {
3706
3375
  "o3-mini-low": {
3707
3376
  maxTokens: 1e5,
3708
3377
  contextWindow: 2e5,
3709
- supportsNativeTools: true,
3710
- defaultToolProtocol: "native",
3711
3378
  supportsImages: false,
3712
3379
  supportsPromptCache: true,
3713
3380
  inputPrice: 1.1,
@@ -3719,8 +3386,6 @@ var openAiNativeModels = {
3719
3386
  o1: {
3720
3387
  maxTokens: 1e5,
3721
3388
  contextWindow: 2e5,
3722
- supportsNativeTools: true,
3723
- defaultToolProtocol: "native",
3724
3389
  supportsImages: true,
3725
3390
  supportsPromptCache: true,
3726
3391
  inputPrice: 15,
@@ -3731,8 +3396,6 @@ var openAiNativeModels = {
3731
3396
  "o1-preview": {
3732
3397
  maxTokens: 32768,
3733
3398
  contextWindow: 128e3,
3734
- supportsNativeTools: true,
3735
- defaultToolProtocol: "native",
3736
3399
  supportsImages: true,
3737
3400
  supportsPromptCache: true,
3738
3401
  inputPrice: 15,
@@ -3743,8 +3406,6 @@ var openAiNativeModels = {
3743
3406
  "o1-mini": {
3744
3407
  maxTokens: 65536,
3745
3408
  contextWindow: 128e3,
3746
- supportsNativeTools: true,
3747
- defaultToolProtocol: "native",
3748
3409
  supportsImages: true,
3749
3410
  supportsPromptCache: true,
3750
3411
  inputPrice: 1.1,
@@ -3755,8 +3416,6 @@ var openAiNativeModels = {
3755
3416
  "gpt-4o": {
3756
3417
  maxTokens: 16384,
3757
3418
  contextWindow: 128e3,
3758
- supportsNativeTools: true,
3759
- defaultToolProtocol: "native",
3760
3419
  supportsImages: true,
3761
3420
  supportsPromptCache: true,
3762
3421
  inputPrice: 2.5,
@@ -3770,8 +3429,6 @@ var openAiNativeModels = {
3770
3429
  "gpt-4o-mini": {
3771
3430
  maxTokens: 16384,
3772
3431
  contextWindow: 128e3,
3773
- supportsNativeTools: true,
3774
- defaultToolProtocol: "native",
3775
3432
  supportsImages: true,
3776
3433
  supportsPromptCache: true,
3777
3434
  inputPrice: 0.15,
@@ -3785,8 +3442,6 @@ var openAiNativeModels = {
3785
3442
  "codex-mini-latest": {
3786
3443
  maxTokens: 16384,
3787
3444
  contextWindow: 2e5,
3788
- supportsNativeTools: true,
3789
- defaultToolProtocol: "native",
3790
3445
  supportsImages: false,
3791
3446
  supportsPromptCache: false,
3792
3447
  inputPrice: 1.5,
@@ -3799,8 +3454,6 @@ var openAiNativeModels = {
3799
3454
  "gpt-5-2025-08-07": {
3800
3455
  maxTokens: 128e3,
3801
3456
  contextWindow: 4e5,
3802
- supportsNativeTools: true,
3803
- defaultToolProtocol: "native",
3804
3457
  includedTools: ["apply_patch"],
3805
3458
  excludedTools: ["apply_diff", "write_to_file"],
3806
3459
  supportsImages: true,
@@ -3821,8 +3474,6 @@ var openAiNativeModels = {
3821
3474
  "gpt-5-mini-2025-08-07": {
3822
3475
  maxTokens: 128e3,
3823
3476
  contextWindow: 4e5,
3824
- supportsNativeTools: true,
3825
- defaultToolProtocol: "native",
3826
3477
  includedTools: ["apply_patch"],
3827
3478
  excludedTools: ["apply_diff", "write_to_file"],
3828
3479
  supportsImages: true,
@@ -3843,8 +3494,6 @@ var openAiNativeModels = {
3843
3494
  "gpt-5-nano-2025-08-07": {
3844
3495
  maxTokens: 128e3,
3845
3496
  contextWindow: 4e5,
3846
- supportsNativeTools: true,
3847
- defaultToolProtocol: "native",
3848
3497
  includedTools: ["apply_patch"],
3849
3498
  excludedTools: ["apply_diff", "write_to_file"],
3850
3499
  supportsImages: true,
@@ -3866,14 +3515,151 @@ var openAiModelInfoSaneDefaults = {
3866
3515
  supportsImages: true,
3867
3516
  supportsPromptCache: false,
3868
3517
  inputPrice: 0,
3869
- outputPrice: 0,
3870
- supportsNativeTools: true,
3871
- defaultToolProtocol: "native"
3518
+ outputPrice: 0
3872
3519
  };
3873
3520
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3874
3521
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3875
3522
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3876
3523
 
3524
+ // src/providers/openai-codex.ts
3525
+ var openAiCodexDefaultModelId = "gpt-5.2-codex";
3526
+ var openAiCodexModels = {
3527
+ "gpt-5.1-codex-max": {
3528
+ maxTokens: 128e3,
3529
+ contextWindow: 4e5,
3530
+ includedTools: ["apply_patch"],
3531
+ excludedTools: ["apply_diff", "write_to_file"],
3532
+ supportsImages: true,
3533
+ supportsPromptCache: true,
3534
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3535
+ reasoningEffort: "xhigh",
3536
+ // Subscription-based: no per-token costs
3537
+ inputPrice: 0,
3538
+ outputPrice: 0,
3539
+ supportsTemperature: false,
3540
+ description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
3541
+ },
3542
+ "gpt-5.1-codex": {
3543
+ maxTokens: 128e3,
3544
+ contextWindow: 4e5,
3545
+ includedTools: ["apply_patch"],
3546
+ excludedTools: ["apply_diff", "write_to_file"],
3547
+ supportsImages: true,
3548
+ supportsPromptCache: true,
3549
+ supportsReasoningEffort: ["low", "medium", "high"],
3550
+ reasoningEffort: "medium",
3551
+ // Subscription-based: no per-token costs
3552
+ inputPrice: 0,
3553
+ outputPrice: 0,
3554
+ supportsTemperature: false,
3555
+ description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
3556
+ },
3557
+ "gpt-5.2-codex": {
3558
+ maxTokens: 128e3,
3559
+ contextWindow: 4e5,
3560
+ includedTools: ["apply_patch"],
3561
+ excludedTools: ["apply_diff", "write_to_file"],
3562
+ supportsImages: true,
3563
+ supportsPromptCache: true,
3564
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3565
+ reasoningEffort: "medium",
3566
+ inputPrice: 0,
3567
+ outputPrice: 0,
3568
+ supportsTemperature: false,
3569
+ description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
3570
+ },
3571
+ "gpt-5.1": {
3572
+ maxTokens: 128e3,
3573
+ contextWindow: 4e5,
3574
+ includedTools: ["apply_patch"],
3575
+ excludedTools: ["apply_diff", "write_to_file"],
3576
+ supportsImages: true,
3577
+ supportsPromptCache: true,
3578
+ supportsReasoningEffort: ["none", "low", "medium", "high"],
3579
+ reasoningEffort: "medium",
3580
+ // Subscription-based: no per-token costs
3581
+ inputPrice: 0,
3582
+ outputPrice: 0,
3583
+ supportsVerbosity: true,
3584
+ supportsTemperature: false,
3585
+ description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
3586
+ },
3587
+ "gpt-5": {
3588
+ maxTokens: 128e3,
3589
+ contextWindow: 4e5,
3590
+ includedTools: ["apply_patch"],
3591
+ excludedTools: ["apply_diff", "write_to_file"],
3592
+ supportsImages: true,
3593
+ supportsPromptCache: true,
3594
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3595
+ reasoningEffort: "medium",
3596
+ // Subscription-based: no per-token costs
3597
+ inputPrice: 0,
3598
+ outputPrice: 0,
3599
+ supportsVerbosity: true,
3600
+ supportsTemperature: false,
3601
+ description: "GPT-5: General GPT-5 model via ChatGPT subscription"
3602
+ },
3603
+ "gpt-5-codex": {
3604
+ maxTokens: 128e3,
3605
+ contextWindow: 4e5,
3606
+ includedTools: ["apply_patch"],
3607
+ excludedTools: ["apply_diff", "write_to_file"],
3608
+ supportsImages: true,
3609
+ supportsPromptCache: true,
3610
+ supportsReasoningEffort: ["low", "medium", "high"],
3611
+ reasoningEffort: "medium",
3612
+ // Subscription-based: no per-token costs
3613
+ inputPrice: 0,
3614
+ outputPrice: 0,
3615
+ supportsTemperature: false,
3616
+ description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
3617
+ },
3618
+ "gpt-5-codex-mini": {
3619
+ maxTokens: 128e3,
3620
+ contextWindow: 4e5,
3621
+ includedTools: ["apply_patch"],
3622
+ excludedTools: ["apply_diff", "write_to_file"],
3623
+ supportsImages: true,
3624
+ supportsPromptCache: true,
3625
+ supportsReasoningEffort: ["low", "medium", "high"],
3626
+ reasoningEffort: "medium",
3627
+ // Subscription-based: no per-token costs
3628
+ inputPrice: 0,
3629
+ outputPrice: 0,
3630
+ supportsTemperature: false,
3631
+ description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
3632
+ },
3633
+ "gpt-5.1-codex-mini": {
3634
+ maxTokens: 128e3,
3635
+ contextWindow: 4e5,
3636
+ includedTools: ["apply_patch"],
3637
+ excludedTools: ["apply_diff", "write_to_file"],
3638
+ supportsImages: true,
3639
+ supportsPromptCache: true,
3640
+ supportsReasoningEffort: ["low", "medium", "high"],
3641
+ reasoningEffort: "medium",
3642
+ inputPrice: 0,
3643
+ outputPrice: 0,
3644
+ supportsTemperature: false,
3645
+ description: "GPT-5.1 Codex Mini: Faster version for coding tasks via ChatGPT subscription"
3646
+ },
3647
+ "gpt-5.2": {
3648
+ maxTokens: 128e3,
3649
+ contextWindow: 4e5,
3650
+ includedTools: ["apply_patch"],
3651
+ excludedTools: ["apply_diff", "write_to_file"],
3652
+ supportsImages: true,
3653
+ supportsPromptCache: true,
3654
+ supportsReasoningEffort: ["none", "low", "medium", "high", "xhigh"],
3655
+ reasoningEffort: "medium",
3656
+ inputPrice: 0,
3657
+ outputPrice: 0,
3658
+ supportsTemperature: false,
3659
+ description: "GPT-5.2: Latest GPT model via ChatGPT subscription"
3660
+ }
3661
+ };
3662
+
3877
3663
  // src/providers/openrouter.ts
3878
3664
  var openRouterDefaultModelId = "anthropic/claude-sonnet-4.5";
3879
3665
  var openRouterDefaultModelInfo = {
@@ -3881,7 +3667,6 @@ var openRouterDefaultModelInfo = {
3881
3667
  contextWindow: 2e5,
3882
3668
  supportsImages: true,
3883
3669
  supportsPromptCache: true,
3884
- supportsNativeTools: true,
3885
3670
  inputPrice: 3,
3886
3671
  outputPrice: 15,
3887
3672
  cacheWritesPrice: 3.75,
@@ -3955,8 +3740,6 @@ var qwenCodeModels = {
3955
3740
  contextWindow: 1e6,
3956
3741
  supportsImages: false,
3957
3742
  supportsPromptCache: false,
3958
- supportsNativeTools: true,
3959
- defaultToolProtocol: "native",
3960
3743
  inputPrice: 0,
3961
3744
  outputPrice: 0,
3962
3745
  cacheWritesPrice: 0,
@@ -3968,8 +3751,6 @@ var qwenCodeModels = {
3968
3751
  contextWindow: 1e6,
3969
3752
  supportsImages: false,
3970
3753
  supportsPromptCache: false,
3971
- supportsNativeTools: true,
3972
- defaultToolProtocol: "native",
3973
3754
  inputPrice: 0,
3974
3755
  outputPrice: 0,
3975
3756
  cacheWritesPrice: 0,
@@ -3985,8 +3766,6 @@ var requestyDefaultModelInfo = {
3985
3766
  contextWindow: 2e5,
3986
3767
  supportsImages: true,
3987
3768
  supportsPromptCache: true,
3988
- supportsNativeTools: true,
3989
- defaultToolProtocol: "native",
3990
3769
  inputPrice: 3,
3991
3770
  outputPrice: 15,
3992
3771
  cacheWritesPrice: 3.75,
@@ -4042,8 +3821,6 @@ var sambaNovaModels = {
4042
3821
  contextWindow: 16384,
4043
3822
  supportsImages: false,
4044
3823
  supportsPromptCache: false,
4045
- supportsNativeTools: true,
4046
- defaultToolProtocol: "native",
4047
3824
  inputPrice: 0.1,
4048
3825
  outputPrice: 0.2,
4049
3826
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -4053,8 +3830,6 @@ var sambaNovaModels = {
4053
3830
  contextWindow: 131072,
4054
3831
  supportsImages: false,
4055
3832
  supportsPromptCache: false,
4056
- supportsNativeTools: true,
4057
- defaultToolProtocol: "native",
4058
3833
  inputPrice: 0.6,
4059
3834
  outputPrice: 1.2,
4060
3835
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -4065,8 +3840,6 @@ var sambaNovaModels = {
4065
3840
  supportsImages: false,
4066
3841
  supportsPromptCache: false,
4067
3842
  supportsReasoningBudget: true,
4068
- supportsNativeTools: true,
4069
- defaultToolProtocol: "native",
4070
3843
  inputPrice: 5,
4071
3844
  outputPrice: 7,
4072
3845
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -4076,8 +3849,6 @@ var sambaNovaModels = {
4076
3849
  contextWindow: 32768,
4077
3850
  supportsImages: false,
4078
3851
  supportsPromptCache: false,
4079
- supportsNativeTools: true,
4080
- defaultToolProtocol: "native",
4081
3852
  inputPrice: 3,
4082
3853
  outputPrice: 4.5,
4083
3854
  description: "DeepSeek V3 model with 32K context window."
@@ -4087,8 +3858,6 @@ var sambaNovaModels = {
4087
3858
  contextWindow: 32768,
4088
3859
  supportsImages: false,
4089
3860
  supportsPromptCache: false,
4090
- supportsNativeTools: true,
4091
- defaultToolProtocol: "native",
4092
3861
  inputPrice: 3,
4093
3862
  outputPrice: 4.5,
4094
3863
  description: "DeepSeek V3.1 model with 32K context window."
@@ -4098,8 +3867,6 @@ var sambaNovaModels = {
4098
3867
  contextWindow: 131072,
4099
3868
  supportsImages: true,
4100
3869
  supportsPromptCache: false,
4101
- supportsNativeTools: true,
4102
- defaultToolProtocol: "native",
4103
3870
  inputPrice: 0.63,
4104
3871
  outputPrice: 1.8,
4105
3872
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -4109,8 +3876,6 @@ var sambaNovaModels = {
4109
3876
  contextWindow: 8192,
4110
3877
  supportsImages: false,
4111
3878
  supportsPromptCache: false,
4112
- supportsNativeTools: true,
4113
- defaultToolProtocol: "native",
4114
3879
  inputPrice: 0.4,
4115
3880
  outputPrice: 0.8,
4116
3881
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -4120,8 +3885,6 @@ var sambaNovaModels = {
4120
3885
  contextWindow: 131072,
4121
3886
  supportsImages: false,
4122
3887
  supportsPromptCache: false,
4123
- supportsNativeTools: true,
4124
- defaultToolProtocol: "native",
4125
3888
  inputPrice: 0.22,
4126
3889
  outputPrice: 0.59,
4127
3890
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -4135,7 +3898,6 @@ var unboundDefaultModelInfo = {
4135
3898
  contextWindow: 2e5,
4136
3899
  supportsImages: true,
4137
3900
  supportsPromptCache: true,
4138
- supportsNativeTools: true,
4139
3901
  inputPrice: 3,
4140
3902
  outputPrice: 15,
4141
3903
  cacheWritesPrice: 3.75,
@@ -4149,8 +3911,6 @@ var vertexModels = {
4149
3911
  maxTokens: 65536,
4150
3912
  contextWindow: 1048576,
4151
3913
  supportsImages: true,
4152
- supportsNativeTools: true,
4153
- defaultToolProtocol: "native",
4154
3914
  supportsPromptCache: true,
4155
3915
  supportsReasoningEffort: ["low", "high"],
4156
3916
  reasoningEffort: "low",
@@ -4158,16 +3918,19 @@ var vertexModels = {
4158
3918
  defaultTemperature: 1,
4159
3919
  inputPrice: 4,
4160
3920
  outputPrice: 18,
3921
+ cacheReadsPrice: 0.4,
4161
3922
  tiers: [
4162
3923
  {
4163
3924
  contextWindow: 2e5,
4164
3925
  inputPrice: 2,
4165
- outputPrice: 12
3926
+ outputPrice: 12,
3927
+ cacheReadsPrice: 0.2
4166
3928
  },
4167
3929
  {
4168
3930
  contextWindow: Infinity,
4169
3931
  inputPrice: 4,
4170
- outputPrice: 18
3932
+ outputPrice: 18,
3933
+ cacheReadsPrice: 0.4
4171
3934
  }
4172
3935
  ]
4173
3936
  },
@@ -4175,24 +3938,19 @@ var vertexModels = {
4175
3938
  maxTokens: 65536,
4176
3939
  contextWindow: 1048576,
4177
3940
  supportsImages: true,
4178
- supportsNativeTools: true,
4179
- defaultToolProtocol: "native",
4180
3941
  supportsPromptCache: true,
4181
3942
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
4182
3943
  reasoningEffort: "medium",
4183
3944
  supportsTemperature: true,
4184
3945
  defaultTemperature: 1,
4185
- inputPrice: 0.3,
4186
- outputPrice: 2.5,
4187
- cacheReadsPrice: 0.075,
4188
- cacheWritesPrice: 1
3946
+ inputPrice: 0.5,
3947
+ outputPrice: 3,
3948
+ cacheReadsPrice: 0.05
4189
3949
  },
4190
3950
  "gemini-2.5-flash-preview-05-20:thinking": {
4191
3951
  maxTokens: 65535,
4192
3952
  contextWindow: 1048576,
4193
3953
  supportsImages: true,
4194
- supportsNativeTools: true,
4195
- defaultToolProtocol: "native",
4196
3954
  supportsPromptCache: true,
4197
3955
  inputPrice: 0.15,
4198
3956
  outputPrice: 3.5,
@@ -4204,8 +3962,6 @@ var vertexModels = {
4204
3962
  maxTokens: 65535,
4205
3963
  contextWindow: 1048576,
4206
3964
  supportsImages: true,
4207
- supportsNativeTools: true,
4208
- defaultToolProtocol: "native",
4209
3965
  supportsPromptCache: true,
4210
3966
  inputPrice: 0.15,
4211
3967
  outputPrice: 0.6
@@ -4214,8 +3970,6 @@ var vertexModels = {
4214
3970
  maxTokens: 64e3,
4215
3971
  contextWindow: 1048576,
4216
3972
  supportsImages: true,
4217
- supportsNativeTools: true,
4218
- defaultToolProtocol: "native",
4219
3973
  supportsPromptCache: true,
4220
3974
  inputPrice: 0.3,
4221
3975
  outputPrice: 2.5,
@@ -4228,8 +3982,6 @@ var vertexModels = {
4228
3982
  maxTokens: 65535,
4229
3983
  contextWindow: 1048576,
4230
3984
  supportsImages: true,
4231
- supportsNativeTools: true,
4232
- defaultToolProtocol: "native",
4233
3985
  supportsPromptCache: false,
4234
3986
  inputPrice: 0.15,
4235
3987
  outputPrice: 3.5,
@@ -4241,8 +3993,6 @@ var vertexModels = {
4241
3993
  maxTokens: 65535,
4242
3994
  contextWindow: 1048576,
4243
3995
  supportsImages: true,
4244
- supportsNativeTools: true,
4245
- defaultToolProtocol: "native",
4246
3996
  supportsPromptCache: false,
4247
3997
  inputPrice: 0.15,
4248
3998
  outputPrice: 0.6
@@ -4251,8 +4001,6 @@ var vertexModels = {
4251
4001
  maxTokens: 65535,
4252
4002
  contextWindow: 1048576,
4253
4003
  supportsImages: true,
4254
- supportsNativeTools: true,
4255
- defaultToolProtocol: "native",
4256
4004
  supportsPromptCache: true,
4257
4005
  inputPrice: 2.5,
4258
4006
  outputPrice: 15
@@ -4261,8 +4009,6 @@ var vertexModels = {
4261
4009
  maxTokens: 65535,
4262
4010
  contextWindow: 1048576,
4263
4011
  supportsImages: true,
4264
- supportsNativeTools: true,
4265
- defaultToolProtocol: "native",
4266
4012
  supportsPromptCache: true,
4267
4013
  inputPrice: 2.5,
4268
4014
  outputPrice: 15
@@ -4271,8 +4017,6 @@ var vertexModels = {
4271
4017
  maxTokens: 65535,
4272
4018
  contextWindow: 1048576,
4273
4019
  supportsImages: true,
4274
- supportsNativeTools: true,
4275
- defaultToolProtocol: "native",
4276
4020
  supportsPromptCache: true,
4277
4021
  inputPrice: 2.5,
4278
4022
  outputPrice: 15,
@@ -4283,8 +4027,6 @@ var vertexModels = {
4283
4027
  maxTokens: 64e3,
4284
4028
  contextWindow: 1048576,
4285
4029
  supportsImages: true,
4286
- supportsNativeTools: true,
4287
- defaultToolProtocol: "native",
4288
4030
  supportsPromptCache: true,
4289
4031
  inputPrice: 2.5,
4290
4032
  outputPrice: 15,
@@ -4310,8 +4052,6 @@ var vertexModels = {
4310
4052
  maxTokens: 65535,
4311
4053
  contextWindow: 1048576,
4312
4054
  supportsImages: true,
4313
- supportsNativeTools: true,
4314
- defaultToolProtocol: "native",
4315
4055
  supportsPromptCache: false,
4316
4056
  inputPrice: 0,
4317
4057
  outputPrice: 0
@@ -4320,8 +4060,6 @@ var vertexModels = {
4320
4060
  maxTokens: 8192,
4321
4061
  contextWindow: 2097152,
4322
4062
  supportsImages: true,
4323
- supportsNativeTools: true,
4324
- defaultToolProtocol: "native",
4325
4063
  supportsPromptCache: false,
4326
4064
  inputPrice: 0,
4327
4065
  outputPrice: 0
@@ -4330,8 +4068,6 @@ var vertexModels = {
4330
4068
  maxTokens: 8192,
4331
4069
  contextWindow: 1048576,
4332
4070
  supportsImages: true,
4333
- supportsNativeTools: true,
4334
- defaultToolProtocol: "native",
4335
4071
  supportsPromptCache: true,
4336
4072
  inputPrice: 0.15,
4337
4073
  outputPrice: 0.6
@@ -4340,8 +4076,6 @@ var vertexModels = {
4340
4076
  maxTokens: 8192,
4341
4077
  contextWindow: 1048576,
4342
4078
  supportsImages: true,
4343
- supportsNativeTools: true,
4344
- defaultToolProtocol: "native",
4345
4079
  supportsPromptCache: false,
4346
4080
  inputPrice: 0.075,
4347
4081
  outputPrice: 0.3
@@ -4350,8 +4084,6 @@ var vertexModels = {
4350
4084
  maxTokens: 8192,
4351
4085
  contextWindow: 32768,
4352
4086
  supportsImages: true,
4353
- supportsNativeTools: true,
4354
- defaultToolProtocol: "native",
4355
4087
  supportsPromptCache: false,
4356
4088
  inputPrice: 0,
4357
4089
  outputPrice: 0
@@ -4360,8 +4092,6 @@ var vertexModels = {
4360
4092
  maxTokens: 8192,
4361
4093
  contextWindow: 1048576,
4362
4094
  supportsImages: true,
4363
- supportsNativeTools: true,
4364
- defaultToolProtocol: "native",
4365
4095
  supportsPromptCache: true,
4366
4096
  inputPrice: 0.075,
4367
4097
  outputPrice: 0.3
@@ -4370,8 +4100,6 @@ var vertexModels = {
4370
4100
  maxTokens: 8192,
4371
4101
  contextWindow: 2097152,
4372
4102
  supportsImages: true,
4373
- supportsNativeTools: true,
4374
- defaultToolProtocol: "native",
4375
4103
  supportsPromptCache: false,
4376
4104
  inputPrice: 1.25,
4377
4105
  outputPrice: 5
@@ -4382,8 +4110,6 @@ var vertexModels = {
4382
4110
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4383
4111
  supportsImages: true,
4384
4112
  supportsPromptCache: true,
4385
- supportsNativeTools: true,
4386
- defaultToolProtocol: "native",
4387
4113
  inputPrice: 3,
4388
4114
  // $3 per million input tokens (≤200K context)
4389
4115
  outputPrice: 15,
@@ -4415,8 +4141,6 @@ var vertexModels = {
4415
4141
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4416
4142
  supportsImages: true,
4417
4143
  supportsPromptCache: true,
4418
- supportsNativeTools: true,
4419
- defaultToolProtocol: "native",
4420
4144
  inputPrice: 3,
4421
4145
  // $3 per million input tokens (≤200K context)
4422
4146
  outputPrice: 15,
@@ -4447,8 +4171,6 @@ var vertexModels = {
4447
4171
  contextWindow: 2e5,
4448
4172
  supportsImages: true,
4449
4173
  supportsPromptCache: true,
4450
- supportsNativeTools: true,
4451
- defaultToolProtocol: "native",
4452
4174
  inputPrice: 1,
4453
4175
  outputPrice: 5,
4454
4176
  cacheWritesPrice: 1.25,
@@ -4460,8 +4182,6 @@ var vertexModels = {
4460
4182
  contextWindow: 2e5,
4461
4183
  supportsImages: true,
4462
4184
  supportsPromptCache: true,
4463
- supportsNativeTools: true,
4464
- defaultToolProtocol: "native",
4465
4185
  inputPrice: 5,
4466
4186
  outputPrice: 25,
4467
4187
  cacheWritesPrice: 6.25,
@@ -4473,8 +4193,6 @@ var vertexModels = {
4473
4193
  contextWindow: 2e5,
4474
4194
  supportsImages: true,
4475
4195
  supportsPromptCache: true,
4476
- supportsNativeTools: true,
4477
- defaultToolProtocol: "native",
4478
4196
  inputPrice: 15,
4479
4197
  outputPrice: 75,
4480
4198
  cacheWritesPrice: 18.75,
@@ -4486,8 +4204,6 @@ var vertexModels = {
4486
4204
  contextWindow: 2e5,
4487
4205
  supportsImages: true,
4488
4206
  supportsPromptCache: true,
4489
- supportsNativeTools: true,
4490
- defaultToolProtocol: "native",
4491
4207
  inputPrice: 15,
4492
4208
  outputPrice: 75,
4493
4209
  cacheWritesPrice: 18.75,
@@ -4498,8 +4214,6 @@ var vertexModels = {
4498
4214
  contextWindow: 2e5,
4499
4215
  supportsImages: true,
4500
4216
  supportsPromptCache: true,
4501
- supportsNativeTools: true,
4502
- defaultToolProtocol: "native",
4503
4217
  inputPrice: 3,
4504
4218
  outputPrice: 15,
4505
4219
  cacheWritesPrice: 3.75,
@@ -4512,8 +4226,6 @@ var vertexModels = {
4512
4226
  contextWindow: 2e5,
4513
4227
  supportsImages: true,
4514
4228
  supportsPromptCache: true,
4515
- supportsNativeTools: true,
4516
- defaultToolProtocol: "native",
4517
4229
  inputPrice: 3,
4518
4230
  outputPrice: 15,
4519
4231
  cacheWritesPrice: 3.75,
@@ -4524,8 +4236,6 @@ var vertexModels = {
4524
4236
  contextWindow: 2e5,
4525
4237
  supportsImages: true,
4526
4238
  supportsPromptCache: true,
4527
- supportsNativeTools: true,
4528
- defaultToolProtocol: "native",
4529
4239
  inputPrice: 3,
4530
4240
  outputPrice: 15,
4531
4241
  cacheWritesPrice: 3.75,
@@ -4536,8 +4246,6 @@ var vertexModels = {
4536
4246
  contextWindow: 2e5,
4537
4247
  supportsImages: true,
4538
4248
  supportsPromptCache: true,
4539
- supportsNativeTools: true,
4540
- defaultToolProtocol: "native",
4541
4249
  inputPrice: 3,
4542
4250
  outputPrice: 15,
4543
4251
  cacheWritesPrice: 3.75,
@@ -4548,8 +4256,6 @@ var vertexModels = {
4548
4256
  contextWindow: 2e5,
4549
4257
  supportsImages: false,
4550
4258
  supportsPromptCache: true,
4551
- supportsNativeTools: true,
4552
- defaultToolProtocol: "native",
4553
4259
  inputPrice: 1,
4554
4260
  outputPrice: 5,
4555
4261
  cacheWritesPrice: 1.25,
@@ -4560,8 +4266,6 @@ var vertexModels = {
4560
4266
  contextWindow: 2e5,
4561
4267
  supportsImages: true,
4562
4268
  supportsPromptCache: true,
4563
- supportsNativeTools: true,
4564
- defaultToolProtocol: "native",
4565
4269
  inputPrice: 15,
4566
4270
  outputPrice: 75,
4567
4271
  cacheWritesPrice: 18.75,
@@ -4572,8 +4276,6 @@ var vertexModels = {
4572
4276
  contextWindow: 2e5,
4573
4277
  supportsImages: true,
4574
4278
  supportsPromptCache: true,
4575
- supportsNativeTools: true,
4576
- defaultToolProtocol: "native",
4577
4279
  inputPrice: 0.25,
4578
4280
  outputPrice: 1.25,
4579
4281
  cacheWritesPrice: 0.3,
@@ -4583,8 +4285,6 @@ var vertexModels = {
4583
4285
  maxTokens: 64e3,
4584
4286
  contextWindow: 1048576,
4585
4287
  supportsImages: true,
4586
- supportsNativeTools: true,
4587
- defaultToolProtocol: "native",
4588
4288
  supportsPromptCache: true,
4589
4289
  inputPrice: 0.1,
4590
4290
  outputPrice: 0.4,
@@ -4598,7 +4298,6 @@ var vertexModels = {
4598
4298
  contextWindow: 131072,
4599
4299
  supportsImages: false,
4600
4300
  supportsPromptCache: false,
4601
- supportsNativeTools: true,
4602
4301
  inputPrice: 0.35,
4603
4302
  outputPrice: 1.15,
4604
4303
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
@@ -4608,7 +4307,6 @@ var vertexModels = {
4608
4307
  contextWindow: 163840,
4609
4308
  supportsImages: false,
4610
4309
  supportsPromptCache: false,
4611
- supportsNativeTools: true,
4612
4310
  inputPrice: 1.35,
4613
4311
  outputPrice: 5.4,
4614
4312
  description: "DeepSeek R1 (0528). Available in us-central1"
@@ -4618,7 +4316,6 @@ var vertexModels = {
4618
4316
  contextWindow: 163840,
4619
4317
  supportsImages: false,
4620
4318
  supportsPromptCache: false,
4621
- supportsNativeTools: true,
4622
4319
  inputPrice: 0.6,
4623
4320
  outputPrice: 1.7,
4624
4321
  description: "DeepSeek V3.1. Available in us-west2"
@@ -4628,7 +4325,6 @@ var vertexModels = {
4628
4325
  contextWindow: 131072,
4629
4326
  supportsImages: false,
4630
4327
  supportsPromptCache: false,
4631
- supportsNativeTools: true,
4632
4328
  inputPrice: 0.15,
4633
4329
  outputPrice: 0.6,
4634
4330
  description: "OpenAI gpt-oss 120B. Available in us-central1"
@@ -4638,7 +4334,6 @@ var vertexModels = {
4638
4334
  contextWindow: 131072,
4639
4335
  supportsImages: false,
4640
4336
  supportsPromptCache: false,
4641
- supportsNativeTools: true,
4642
4337
  inputPrice: 0.075,
4643
4338
  outputPrice: 0.3,
4644
4339
  description: "OpenAI gpt-oss 20B. Available in us-central1"
@@ -4648,7 +4343,6 @@ var vertexModels = {
4648
4343
  contextWindow: 262144,
4649
4344
  supportsImages: false,
4650
4345
  supportsPromptCache: false,
4651
- supportsNativeTools: true,
4652
4346
  inputPrice: 1,
4653
4347
  outputPrice: 4,
4654
4348
  description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
@@ -4658,10 +4352,18 @@ var vertexModels = {
4658
4352
  contextWindow: 262144,
4659
4353
  supportsImages: false,
4660
4354
  supportsPromptCache: false,
4661
- supportsNativeTools: true,
4662
4355
  inputPrice: 0.25,
4663
4356
  outputPrice: 1,
4664
4357
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
4358
+ },
4359
+ "moonshotai/kimi-k2-thinking-maas": {
4360
+ maxTokens: 16384,
4361
+ contextWindow: 262144,
4362
+ supportsPromptCache: false,
4363
+ supportsImages: false,
4364
+ inputPrice: 0.6,
4365
+ outputPrice: 2.5,
4366
+ description: "Kimi K2 Thinking Model with 256K context window."
4665
4367
  }
4666
4368
  };
4667
4369
  var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
@@ -4894,8 +4596,6 @@ var xaiModels = {
4894
4596
  contextWindow: 256e3,
4895
4597
  supportsImages: true,
4896
4598
  supportsPromptCache: true,
4897
- supportsNativeTools: true,
4898
- defaultToolProtocol: "native",
4899
4599
  inputPrice: 0.2,
4900
4600
  outputPrice: 1.5,
4901
4601
  cacheWritesPrice: 0.02,
@@ -4909,8 +4609,6 @@ var xaiModels = {
4909
4609
  contextWindow: 2e6,
4910
4610
  supportsImages: true,
4911
4611
  supportsPromptCache: true,
4912
- supportsNativeTools: true,
4913
- defaultToolProtocol: "native",
4914
4612
  inputPrice: 0.2,
4915
4613
  outputPrice: 0.5,
4916
4614
  cacheWritesPrice: 0.05,
@@ -4924,8 +4622,6 @@ var xaiModels = {
4924
4622
  contextWindow: 2e6,
4925
4623
  supportsImages: true,
4926
4624
  supportsPromptCache: true,
4927
- supportsNativeTools: true,
4928
- defaultToolProtocol: "native",
4929
4625
  inputPrice: 0.2,
4930
4626
  outputPrice: 0.5,
4931
4627
  cacheWritesPrice: 0.05,
@@ -4939,8 +4635,6 @@ var xaiModels = {
4939
4635
  contextWindow: 2e6,
4940
4636
  supportsImages: true,
4941
4637
  supportsPromptCache: true,
4942
- supportsNativeTools: true,
4943
- defaultToolProtocol: "native",
4944
4638
  inputPrice: 0.2,
4945
4639
  outputPrice: 0.5,
4946
4640
  cacheWritesPrice: 0.05,
@@ -4954,8 +4648,6 @@ var xaiModels = {
4954
4648
  contextWindow: 2e6,
4955
4649
  supportsImages: true,
4956
4650
  supportsPromptCache: true,
4957
- supportsNativeTools: true,
4958
- defaultToolProtocol: "native",
4959
4651
  inputPrice: 0.2,
4960
4652
  outputPrice: 0.5,
4961
4653
  cacheWritesPrice: 0.05,
@@ -4969,8 +4661,6 @@ var xaiModels = {
4969
4661
  contextWindow: 256e3,
4970
4662
  supportsImages: true,
4971
4663
  supportsPromptCache: true,
4972
- supportsNativeTools: true,
4973
- defaultToolProtocol: "native",
4974
4664
  inputPrice: 3,
4975
4665
  outputPrice: 15,
4976
4666
  cacheWritesPrice: 0.75,
@@ -4984,8 +4674,6 @@ var xaiModels = {
4984
4674
  contextWindow: 131072,
4985
4675
  supportsImages: true,
4986
4676
  supportsPromptCache: true,
4987
- supportsNativeTools: true,
4988
- defaultToolProtocol: "native",
4989
4677
  inputPrice: 0.3,
4990
4678
  outputPrice: 0.5,
4991
4679
  cacheWritesPrice: 0.07,
@@ -5001,8 +4689,6 @@ var xaiModels = {
5001
4689
  contextWindow: 131072,
5002
4690
  supportsImages: true,
5003
4691
  supportsPromptCache: true,
5004
- supportsNativeTools: true,
5005
- defaultToolProtocol: "native",
5006
4692
  inputPrice: 3,
5007
4693
  outputPrice: 15,
5008
4694
  cacheWritesPrice: 0.75,
@@ -5099,7 +4785,6 @@ var vercelAiGatewayDefaultModelInfo = {
5099
4785
  contextWindow: 2e5,
5100
4786
  supportsImages: true,
5101
4787
  supportsPromptCache: true,
5102
- supportsNativeTools: true,
5103
4788
  inputPrice: 3,
5104
4789
  outputPrice: 15,
5105
4790
  cacheWritesPrice: 3.75,
@@ -5116,8 +4801,6 @@ var internationalZAiModels = {
5116
4801
  contextWindow: 131072,
5117
4802
  supportsImages: false,
5118
4803
  supportsPromptCache: true,
5119
- supportsNativeTools: true,
5120
- defaultToolProtocol: "native",
5121
4804
  inputPrice: 0.6,
5122
4805
  outputPrice: 2.2,
5123
4806
  cacheWritesPrice: 0,
@@ -5129,8 +4812,6 @@ var internationalZAiModels = {
5129
4812
  contextWindow: 131072,
5130
4813
  supportsImages: false,
5131
4814
  supportsPromptCache: true,
5132
- supportsNativeTools: true,
5133
- defaultToolProtocol: "native",
5134
4815
  inputPrice: 0.2,
5135
4816
  outputPrice: 1.1,
5136
4817
  cacheWritesPrice: 0,
@@ -5142,8 +4823,6 @@ var internationalZAiModels = {
5142
4823
  contextWindow: 131072,
5143
4824
  supportsImages: false,
5144
4825
  supportsPromptCache: true,
5145
- supportsNativeTools: true,
5146
- defaultToolProtocol: "native",
5147
4826
  inputPrice: 2.2,
5148
4827
  outputPrice: 8.9,
5149
4828
  cacheWritesPrice: 0,
@@ -5155,8 +4834,6 @@ var internationalZAiModels = {
5155
4834
  contextWindow: 131072,
5156
4835
  supportsImages: false,
5157
4836
  supportsPromptCache: true,
5158
- supportsNativeTools: true,
5159
- defaultToolProtocol: "native",
5160
4837
  inputPrice: 1.1,
5161
4838
  outputPrice: 4.5,
5162
4839
  cacheWritesPrice: 0,
@@ -5168,8 +4845,6 @@ var internationalZAiModels = {
5168
4845
  contextWindow: 131072,
5169
4846
  supportsImages: false,
5170
4847
  supportsPromptCache: true,
5171
- supportsNativeTools: true,
5172
- defaultToolProtocol: "native",
5173
4848
  inputPrice: 0,
5174
4849
  outputPrice: 0,
5175
4850
  cacheWritesPrice: 0,
@@ -5181,21 +4856,28 @@ var internationalZAiModels = {
5181
4856
  contextWindow: 131072,
5182
4857
  supportsImages: true,
5183
4858
  supportsPromptCache: true,
5184
- supportsNativeTools: true,
5185
- defaultToolProtocol: "native",
5186
4859
  inputPrice: 0.6,
5187
4860
  outputPrice: 1.8,
5188
4861
  cacheWritesPrice: 0,
5189
4862
  cacheReadsPrice: 0.11,
5190
4863
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
5191
4864
  },
4865
+ "glm-4.6v": {
4866
+ maxTokens: 16384,
4867
+ contextWindow: 131072,
4868
+ supportsImages: true,
4869
+ supportsPromptCache: true,
4870
+ inputPrice: 0.3,
4871
+ outputPrice: 0.9,
4872
+ cacheWritesPrice: 0,
4873
+ cacheReadsPrice: 0.05,
4874
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4875
+ },
5192
4876
  "glm-4.6": {
5193
4877
  maxTokens: 16384,
5194
4878
  contextWindow: 2e5,
5195
4879
  supportsImages: false,
5196
4880
  supportsPromptCache: true,
5197
- supportsNativeTools: true,
5198
- defaultToolProtocol: "native",
5199
4881
  inputPrice: 0.6,
5200
4882
  outputPrice: 2.2,
5201
4883
  cacheWritesPrice: 0,
@@ -5207,8 +4889,6 @@ var internationalZAiModels = {
5207
4889
  contextWindow: 2e5,
5208
4890
  supportsImages: false,
5209
4891
  supportsPromptCache: true,
5210
- supportsNativeTools: true,
5211
- defaultToolProtocol: "native",
5212
4892
  supportsReasoningEffort: ["disable", "medium"],
5213
4893
  reasoningEffort: "medium",
5214
4894
  preserveReasoning: true,
@@ -5218,13 +4898,55 @@ var internationalZAiModels = {
5218
4898
  cacheReadsPrice: 0.11,
5219
4899
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
5220
4900
  },
4901
+ "glm-4.7-flash": {
4902
+ maxTokens: 16384,
4903
+ contextWindow: 2e5,
4904
+ supportsImages: false,
4905
+ supportsPromptCache: true,
4906
+ inputPrice: 0,
4907
+ outputPrice: 0,
4908
+ cacheWritesPrice: 0,
4909
+ cacheReadsPrice: 0,
4910
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4911
+ },
4912
+ "glm-4.7-flashx": {
4913
+ maxTokens: 16384,
4914
+ contextWindow: 2e5,
4915
+ supportsImages: false,
4916
+ supportsPromptCache: true,
4917
+ inputPrice: 0.07,
4918
+ outputPrice: 0.4,
4919
+ cacheWritesPrice: 0,
4920
+ cacheReadsPrice: 0.01,
4921
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4922
+ },
4923
+ "glm-4.6v-flash": {
4924
+ maxTokens: 16384,
4925
+ contextWindow: 131072,
4926
+ supportsImages: true,
4927
+ supportsPromptCache: true,
4928
+ inputPrice: 0,
4929
+ outputPrice: 0,
4930
+ cacheWritesPrice: 0,
4931
+ cacheReadsPrice: 0,
4932
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4933
+ },
4934
+ "glm-4.6v-flashx": {
4935
+ maxTokens: 16384,
4936
+ contextWindow: 131072,
4937
+ supportsImages: true,
4938
+ supportsPromptCache: true,
4939
+ inputPrice: 0.04,
4940
+ outputPrice: 0.4,
4941
+ cacheWritesPrice: 0,
4942
+ cacheReadsPrice: 4e-3,
4943
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
4944
+ },
5221
4945
  "glm-4-32b-0414-128k": {
5222
4946
  maxTokens: 16384,
5223
4947
  contextWindow: 131072,
5224
4948
  supportsImages: false,
5225
4949
  supportsPromptCache: false,
5226
- supportsNativeTools: true,
5227
- defaultToolProtocol: "native",
5228
4950
  inputPrice: 0.1,
5229
4951
  outputPrice: 0.1,
5230
4952
  cacheWritesPrice: 0,
@@ -5239,8 +4961,6 @@ var mainlandZAiModels = {
5239
4961
  contextWindow: 131072,
5240
4962
  supportsImages: false,
5241
4963
  supportsPromptCache: true,
5242
- supportsNativeTools: true,
5243
- defaultToolProtocol: "native",
5244
4964
  inputPrice: 0.29,
5245
4965
  outputPrice: 1.14,
5246
4966
  cacheWritesPrice: 0,
@@ -5252,8 +4972,6 @@ var mainlandZAiModels = {
5252
4972
  contextWindow: 131072,
5253
4973
  supportsImages: false,
5254
4974
  supportsPromptCache: true,
5255
- supportsNativeTools: true,
5256
- defaultToolProtocol: "native",
5257
4975
  inputPrice: 0.1,
5258
4976
  outputPrice: 0.6,
5259
4977
  cacheWritesPrice: 0,
@@ -5265,8 +4983,6 @@ var mainlandZAiModels = {
5265
4983
  contextWindow: 131072,
5266
4984
  supportsImages: false,
5267
4985
  supportsPromptCache: true,
5268
- supportsNativeTools: true,
5269
- defaultToolProtocol: "native",
5270
4986
  inputPrice: 0.29,
5271
4987
  outputPrice: 1.14,
5272
4988
  cacheWritesPrice: 0,
@@ -5278,8 +4994,6 @@ var mainlandZAiModels = {
5278
4994
  contextWindow: 131072,
5279
4995
  supportsImages: false,
5280
4996
  supportsPromptCache: true,
5281
- supportsNativeTools: true,
5282
- defaultToolProtocol: "native",
5283
4997
  inputPrice: 0.1,
5284
4998
  outputPrice: 0.6,
5285
4999
  cacheWritesPrice: 0,
@@ -5291,8 +5005,6 @@ var mainlandZAiModels = {
5291
5005
  contextWindow: 131072,
5292
5006
  supportsImages: false,
5293
5007
  supportsPromptCache: true,
5294
- supportsNativeTools: true,
5295
- defaultToolProtocol: "native",
5296
5008
  inputPrice: 0,
5297
5009
  outputPrice: 0,
5298
5010
  cacheWritesPrice: 0,
@@ -5304,8 +5016,6 @@ var mainlandZAiModels = {
5304
5016
  contextWindow: 131072,
5305
5017
  supportsImages: true,
5306
5018
  supportsPromptCache: true,
5307
- supportsNativeTools: true,
5308
- defaultToolProtocol: "native",
5309
5019
  inputPrice: 0.29,
5310
5020
  outputPrice: 0.93,
5311
5021
  cacheWritesPrice: 0,
@@ -5317,8 +5027,6 @@ var mainlandZAiModels = {
5317
5027
  contextWindow: 204800,
5318
5028
  supportsImages: false,
5319
5029
  supportsPromptCache: true,
5320
- supportsNativeTools: true,
5321
- defaultToolProtocol: "native",
5322
5030
  inputPrice: 0.29,
5323
5031
  outputPrice: 1.14,
5324
5032
  cacheWritesPrice: 0,
@@ -5330,8 +5038,6 @@ var mainlandZAiModels = {
5330
5038
  contextWindow: 204800,
5331
5039
  supportsImages: false,
5332
5040
  supportsPromptCache: true,
5333
- supportsNativeTools: true,
5334
- defaultToolProtocol: "native",
5335
5041
  supportsReasoningEffort: ["disable", "medium"],
5336
5042
  reasoningEffort: "medium",
5337
5043
  preserveReasoning: true,
@@ -5340,6 +5046,61 @@ var mainlandZAiModels = {
5340
5046
  cacheWritesPrice: 0,
5341
5047
  cacheReadsPrice: 0.057,
5342
5048
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
5049
+ },
5050
+ "glm-4.7-flash": {
5051
+ maxTokens: 16384,
5052
+ contextWindow: 204800,
5053
+ supportsImages: false,
5054
+ supportsPromptCache: true,
5055
+ inputPrice: 0,
5056
+ outputPrice: 0,
5057
+ cacheWritesPrice: 0,
5058
+ cacheReadsPrice: 0,
5059
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
5060
+ },
5061
+ "glm-4.7-flashx": {
5062
+ maxTokens: 16384,
5063
+ contextWindow: 204800,
5064
+ supportsImages: false,
5065
+ supportsPromptCache: true,
5066
+ inputPrice: 0.035,
5067
+ outputPrice: 0.2,
5068
+ cacheWritesPrice: 0,
5069
+ cacheReadsPrice: 5e-3,
5070
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
5071
+ },
5072
+ "glm-4.6v": {
5073
+ maxTokens: 16384,
5074
+ contextWindow: 131072,
5075
+ supportsImages: true,
5076
+ supportsPromptCache: true,
5077
+ inputPrice: 0.15,
5078
+ outputPrice: 0.45,
5079
+ cacheWritesPrice: 0,
5080
+ cacheReadsPrice: 0.025,
5081
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
5082
+ },
5083
+ "glm-4.6v-flash": {
5084
+ maxTokens: 16384,
5085
+ contextWindow: 131072,
5086
+ supportsImages: true,
5087
+ supportsPromptCache: true,
5088
+ inputPrice: 0,
5089
+ outputPrice: 0,
5090
+ cacheWritesPrice: 0,
5091
+ cacheReadsPrice: 0,
5092
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
5093
+ },
5094
+ "glm-4.6v-flashx": {
5095
+ maxTokens: 16384,
5096
+ contextWindow: 131072,
5097
+ supportsImages: true,
5098
+ supportsPromptCache: true,
5099
+ inputPrice: 0.02,
5100
+ outputPrice: 0.2,
5101
+ cacheWritesPrice: 0,
5102
+ cacheReadsPrice: 2e-3,
5103
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
5343
5104
  }
5344
5105
  };
5345
5106
  var ZAI_DEFAULT_TEMPERATURE = 0.6;
@@ -5373,7 +5134,6 @@ var deepInfraDefaultModelInfo = {
5373
5134
  contextWindow: 262144,
5374
5135
  supportsImages: false,
5375
5136
  supportsPromptCache: false,
5376
- supportsNativeTools: true,
5377
5137
  inputPrice: 0.3,
5378
5138
  outputPrice: 1.2,
5379
5139
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -5387,8 +5147,6 @@ var minimaxModels = {
5387
5147
  contextWindow: 192e3,
5388
5148
  supportsImages: false,
5389
5149
  supportsPromptCache: true,
5390
- supportsNativeTools: true,
5391
- defaultToolProtocol: "native",
5392
5150
  includedTools: ["search_and_replace"],
5393
5151
  excludedTools: ["apply_diff"],
5394
5152
  preserveReasoning: true,
@@ -5403,8 +5161,6 @@ var minimaxModels = {
5403
5161
  contextWindow: 192e3,
5404
5162
  supportsImages: false,
5405
5163
  supportsPromptCache: true,
5406
- supportsNativeTools: true,
5407
- defaultToolProtocol: "native",
5408
5164
  includedTools: ["search_and_replace"],
5409
5165
  excludedTools: ["apply_diff"],
5410
5166
  preserveReasoning: true,
@@ -5419,8 +5175,6 @@ var minimaxModels = {
5419
5175
  contextWindow: 192e3,
5420
5176
  supportsImages: false,
5421
5177
  supportsPromptCache: true,
5422
- supportsNativeTools: true,
5423
- defaultToolProtocol: "native",
5424
5178
  includedTools: ["search_and_replace"],
5425
5179
  excludedTools: ["apply_diff"],
5426
5180
  preserveReasoning: true,
@@ -5475,6 +5229,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
5475
5229
  case "openai-native":
5476
5230
  return "gpt-4o";
5477
5231
  // Based on openai-native patterns
5232
+ case "openai-codex":
5233
+ return openAiCodexDefaultModelId;
5478
5234
  case "mistral":
5479
5235
  return mistralDefaultModelId;
5480
5236
  case "openai":
@@ -5490,8 +5246,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
5490
5246
  return deepInfraDefaultModelId;
5491
5247
  case "vscode-lm":
5492
5248
  return vscodeLlmDefaultModelId;
5493
- case "claude-code":
5494
- return claudeCodeDefaultModelId;
5495
5249
  case "cerebras":
5496
5250
  return cerebrasDefaultModelId;
5497
5251
  case "sambanova":
@@ -5549,7 +5303,6 @@ var providerNames = [
5549
5303
  "bedrock",
5550
5304
  "baseten",
5551
5305
  "cerebras",
5552
- "claude-code",
5553
5306
  "doubao",
5554
5307
  "deepseek",
5555
5308
  "featherless",
@@ -5560,6 +5313,7 @@ var providerNames = [
5560
5313
  "mistral",
5561
5314
  "moonshot",
5562
5315
  "minimax",
5316
+ "openai-codex",
5563
5317
  "openai-native",
5564
5318
  "qwen-code",
5565
5319
  "roo",
@@ -5578,9 +5332,7 @@ var providerSettingsEntrySchema = import_zod8.z.object({
5578
5332
  });
5579
5333
  var baseProviderSettingsSchema = import_zod8.z.object({
5580
5334
  includeMaxTokens: import_zod8.z.boolean().optional(),
5581
- diffEnabled: import_zod8.z.boolean().optional(),
5582
5335
  todoListEnabled: import_zod8.z.boolean().optional(),
5583
- fuzzyMatchThreshold: import_zod8.z.number().optional(),
5584
5336
  modelTemperature: import_zod8.z.number().nullish(),
5585
5337
  rateLimitSeconds: import_zod8.z.number().optional(),
5586
5338
  consecutiveMistakeLimit: import_zod8.z.number().min(0).optional(),
@@ -5590,9 +5342,7 @@ var baseProviderSettingsSchema = import_zod8.z.object({
5590
5342
  modelMaxTokens: import_zod8.z.number().optional(),
5591
5343
  modelMaxThinkingTokens: import_zod8.z.number().optional(),
5592
5344
  // Model verbosity.
5593
- verbosity: verbosityLevelsSchema.optional(),
5594
- // Tool protocol override for this profile.
5595
- toolProtocol: import_zod8.z.enum(["xml", "native"]).optional()
5345
+ verbosity: verbosityLevelsSchema.optional()
5596
5346
  });
5597
5347
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
5598
5348
  apiModelId: import_zod8.z.string().optional()
@@ -5604,7 +5354,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
5604
5354
  anthropicBeta1MContext: import_zod8.z.boolean().optional()
5605
5355
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5606
5356
  });
5607
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5608
5357
  var openRouterSchema = baseProviderSettingsSchema.extend({
5609
5358
  openRouterApiKey: import_zod8.z.string().optional(),
5610
5359
  openRouterModelId: import_zod8.z.string().optional(),
@@ -5686,6 +5435,9 @@ var geminiCliSchema = apiModelIdProviderModelSchema.extend({
5686
5435
  geminiCliOAuthPath: import_zod8.z.string().optional(),
5687
5436
  geminiCliProjectId: import_zod8.z.string().optional()
5688
5437
  });
5438
+ var openAiCodexSchema = apiModelIdProviderModelSchema.extend({
5439
+ // No additional settings needed - uses OAuth authentication
5440
+ });
5689
5441
  var openAiNativeSchema = apiModelIdProviderModelSchema.extend({
5690
5442
  openAiNativeApiKey: import_zod8.z.string().optional(),
5691
5443
  openAiNativeBaseUrl: import_zod8.z.string().optional(),
@@ -5790,7 +5542,6 @@ var defaultSchema = import_zod8.z.object({
5790
5542
  });
5791
5543
  var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiProvider", [
5792
5544
  anthropicSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("anthropic") })),
5793
- claudeCodeSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("claude-code") })),
5794
5545
  openRouterSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("openrouter") })),
5795
5546
  bedrockSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("bedrock") })),
5796
5547
  vertexSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("vertex") })),
@@ -5800,6 +5551,7 @@ var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiP
5800
5551
  lmStudioSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("lmstudio") })),
5801
5552
  geminiSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("gemini") })),
5802
5553
  geminiCliSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("gemini-cli") })),
5554
+ openAiCodexSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("openai-codex") })),
5803
5555
  openAiNativeSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("openai-native") })),
5804
5556
  mistralSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("mistral") })),
5805
5557
  deepSeekSchema.merge(import_zod8.z.object({ apiProvider: import_zod8.z.literal("deepseek") })),
@@ -5830,7 +5582,6 @@ var providerSettingsSchemaDiscriminated = import_zod8.z.discriminatedUnion("apiP
5830
5582
  var providerSettingsSchema = import_zod8.z.object({
5831
5583
  apiProvider: providerNamesSchema.optional(),
5832
5584
  ...anthropicSchema.shape,
5833
- ...claudeCodeSchema.shape,
5834
5585
  ...openRouterSchema.shape,
5835
5586
  ...bedrockSchema.shape,
5836
5587
  ...vertexSchema.shape,
@@ -5840,6 +5591,7 @@ var providerSettingsSchema = import_zod8.z.object({
5840
5591
  ...lmStudioSchema.shape,
5841
5592
  ...geminiSchema.shape,
5842
5593
  ...geminiCliSchema.shape,
5594
+ ...openAiCodexSchema.shape,
5843
5595
  ...openAiNativeSchema.shape,
5844
5596
  ...mistralSchema.shape,
5845
5597
  ...deepSeekSchema.shape,
@@ -5894,10 +5646,10 @@ var getModelId = (settings) => {
5894
5646
  var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
5895
5647
  var modelIdKeysByProvider = {
5896
5648
  anthropic: "apiModelId",
5897
- "claude-code": "apiModelId",
5898
5649
  openrouter: "openRouterModelId",
5899
5650
  bedrock: "apiModelId",
5900
5651
  vertex: "apiModelId",
5652
+ "openai-codex": "apiModelId",
5901
5653
  "openai-native": "openAiModelId",
5902
5654
  ollama: "ollamaModelId",
5903
5655
  lmstudio: "lmStudioModelId",
@@ -5927,7 +5679,7 @@ var modelIdKeysByProvider = {
5927
5679
  roo: "apiModelId",
5928
5680
  "vercel-ai-gateway": "vercelAiGatewayModelId"
5929
5681
  };
5930
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
5682
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
5931
5683
  var getApiProtocol = (provider, modelId) => {
5932
5684
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
5933
5685
  return "anthropic";
@@ -5956,7 +5708,6 @@ var MODELS_BY_PROVIDER = {
5956
5708
  label: "Cerebras",
5957
5709
  models: Object.keys(cerebrasModels)
5958
5710
  },
5959
- "claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
5960
5711
  deepseek: {
5961
5712
  id: "deepseek",
5962
5713
  label: "DeepSeek",
@@ -5999,6 +5750,11 @@ var MODELS_BY_PROVIDER = {
5999
5750
  label: "MiniMax",
6000
5751
  models: Object.keys(minimaxModels)
6001
5752
  },
5753
+ "openai-codex": {
5754
+ id: "openai-codex",
5755
+ label: "OpenAI - ChatGPT Plus/Pro",
5756
+ models: Object.keys(openAiCodexModels)
5757
+ },
6002
5758
  "openai-native": {
6003
5759
  id: "openai-native",
6004
5760
  label: "OpenAI",
@@ -6055,16 +5811,6 @@ var historyItemSchema = import_zod9.z.object({
6055
5811
  size: import_zod9.z.number().optional(),
6056
5812
  workspace: import_zod9.z.string().optional(),
6057
5813
  mode: import_zod9.z.string().optional(),
6058
- /**
6059
- * The tool protocol used by this task. Once a task uses tools with a specific
6060
- * protocol (XML or Native), it is permanently locked to that protocol.
6061
- *
6062
- * - "xml": Tool calls are parsed from XML text (no tool IDs)
6063
- * - "native": Tool calls come as tool_call chunks with IDs
6064
- *
6065
- * This ensures task resumption works correctly even when NTC settings change.
6066
- */
6067
- toolProtocol: import_zod9.z.enum(["xml", "native"]).optional(),
6068
5814
  apiConfigName: import_zod9.z.string().optional(),
6069
5815
  // Provider profile name for sticky profile feature
6070
5816
  status: import_zod9.z.enum(["active", "completed", "delegated"]).optional(),
@@ -6082,23 +5828,12 @@ var historyItemSchema = import_zod9.z.object({
6082
5828
 
6083
5829
  // src/experiment.ts
6084
5830
  var import_zod10 = require("zod");
6085
- var experimentIds = [
6086
- "powerSteering",
6087
- "multiFileApplyDiff",
6088
- "preventFocusDisruption",
6089
- "imageGeneration",
6090
- "runSlashCommand",
6091
- "multipleNativeToolCalls",
6092
- "customTools"
6093
- ];
5831
+ var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
6094
5832
  var experimentIdsSchema = import_zod10.z.enum(experimentIds);
6095
5833
  var experimentsSchema = import_zod10.z.object({
6096
- powerSteering: import_zod10.z.boolean().optional(),
6097
- multiFileApplyDiff: import_zod10.z.boolean().optional(),
6098
5834
  preventFocusDisruption: import_zod10.z.boolean().optional(),
6099
5835
  imageGeneration: import_zod10.z.boolean().optional(),
6100
5836
  runSlashCommand: import_zod10.z.boolean().optional(),
6101
- multipleNativeToolCalls: import_zod10.z.boolean().optional(),
6102
5837
  customTools: import_zod10.z.boolean().optional()
6103
5838
  });
6104
5839
 
@@ -6150,6 +5885,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
6150
5885
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
6151
5886
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
6152
5887
  TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5888
+ TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
6153
5889
  return TelemetryEventName2;
6154
5890
  })(TelemetryEventName || {});
6155
5891
  var staticAppPropertiesSchema = import_zod11.z.object({
@@ -6238,7 +5974,8 @@ var rooCodeTelemetryEventSchema = import_zod11.z.discriminatedUnion("type", [
6238
5974
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
6239
5975
  "Tab Shown" /* TAB_SHOWN */,
6240
5976
  "Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
6241
- "Custom Mode Created" /* CUSTOM_MODE_CREATED */
5977
+ "Custom Mode Created" /* CUSTOM_MODE_CREATED */,
5978
+ "Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
6242
5979
  ]),
6243
5980
  properties: telemetryPropertiesSchema
6244
5981
  }),
@@ -6552,7 +6289,15 @@ var isLanguage = (value) => languages.includes(value);
6552
6289
 
6553
6290
  // src/global-settings.ts
6554
6291
  var DEFAULT_WRITE_DELAY_MS = 1e3;
6555
- var DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT = 5e4;
6292
+ var TERMINAL_PREVIEW_BYTES = {
6293
+ small: 5 * 1024,
6294
+ // 5KB
6295
+ medium: 10 * 1024,
6296
+ // 10KB
6297
+ large: 20 * 1024
6298
+ // 20KB
6299
+ };
6300
+ var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
6556
6301
  var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
6557
6302
  var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
6558
6303
  var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
@@ -6568,7 +6313,6 @@ var globalSettingsSchema = import_zod14.z.object({
6568
6313
  imageGenerationProvider: import_zod14.z.enum(["openrouter", "roo"]).optional(),
6569
6314
  openRouterImageApiKey: import_zod14.z.string().optional(),
6570
6315
  openRouterImageGenerationSelectedModel: import_zod14.z.string().optional(),
6571
- condensingApiConfigId: import_zod14.z.string().optional(),
6572
6316
  customCondensingPrompt: import_zod14.z.string().optional(),
6573
6317
  autoApprovalEnabled: import_zod14.z.boolean().optional(),
6574
6318
  alwaysAllowReadOnly: import_zod14.z.boolean().optional(),
@@ -6594,7 +6338,6 @@ var globalSettingsSchema = import_zod14.z.object({
6594
6338
  allowedMaxCost: import_zod14.z.number().nullish(),
6595
6339
  autoCondenseContext: import_zod14.z.boolean().optional(),
6596
6340
  autoCondenseContextPercent: import_zod14.z.number().optional(),
6597
- maxConcurrentFileReads: import_zod14.z.number().optional(),
6598
6341
  /**
6599
6342
  * Whether to include current time in the environment details
6600
6343
  * @default true
@@ -6637,11 +6380,9 @@ var globalSettingsSchema = import_zod14.z.object({
6637
6380
  maxWorkspaceFiles: import_zod14.z.number().optional(),
6638
6381
  showRooIgnoredFiles: import_zod14.z.boolean().optional(),
6639
6382
  enableSubfolderRules: import_zod14.z.boolean().optional(),
6640
- maxReadFileLine: import_zod14.z.number().optional(),
6641
6383
  maxImageFileSize: import_zod14.z.number().optional(),
6642
6384
  maxTotalImageSize: import_zod14.z.number().optional(),
6643
- terminalOutputLineLimit: import_zod14.z.number().optional(),
6644
- terminalOutputCharacterLimit: import_zod14.z.number().optional(),
6385
+ terminalOutputPreviewSize: import_zod14.z.enum(["small", "medium", "large"]).optional(),
6645
6386
  terminalShellIntegrationTimeout: import_zod14.z.number().optional(),
6646
6387
  terminalShellIntegrationDisabled: import_zod14.z.boolean().optional(),
6647
6388
  terminalCommandDelay: import_zod14.z.number().optional(),
@@ -6650,18 +6391,14 @@ var globalSettingsSchema = import_zod14.z.object({
6650
6391
  terminalZshOhMy: import_zod14.z.boolean().optional(),
6651
6392
  terminalZshP10k: import_zod14.z.boolean().optional(),
6652
6393
  terminalZdotdir: import_zod14.z.boolean().optional(),
6653
- terminalCompressProgressBar: import_zod14.z.boolean().optional(),
6654
6394
  diagnosticsEnabled: import_zod14.z.boolean().optional(),
6655
6395
  rateLimitSeconds: import_zod14.z.number().optional(),
6656
- diffEnabled: import_zod14.z.boolean().optional(),
6657
- fuzzyMatchThreshold: import_zod14.z.number().optional(),
6658
6396
  experiments: experimentsSchema.optional(),
6659
6397
  codebaseIndexModels: codebaseIndexModelsSchema.optional(),
6660
6398
  codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
6661
6399
  language: languagesSchema.optional(),
6662
6400
  telemetrySetting: telemetrySettingsSchema.optional(),
6663
6401
  mcpEnabled: import_zod14.z.boolean().optional(),
6664
- enableMcpServerCreation: import_zod14.z.boolean().optional(),
6665
6402
  mode: import_zod14.z.string().optional(),
6666
6403
  modeApiConfigs: import_zod14.z.record(import_zod14.z.string(), import_zod14.z.string()).optional(),
6667
6404
  customModes: import_zod14.z.array(modeConfigSchema).optional(),
@@ -6681,7 +6418,20 @@ var globalSettingsSchema = import_zod14.z.object({
6681
6418
  profileThresholds: import_zod14.z.record(import_zod14.z.string(), import_zod14.z.number()).optional(),
6682
6419
  hasOpenedModeSelector: import_zod14.z.boolean().optional(),
6683
6420
  lastModeExportPath: import_zod14.z.string().optional(),
6684
- lastModeImportPath: import_zod14.z.string().optional()
6421
+ lastModeImportPath: import_zod14.z.string().optional(),
6422
+ lastSettingsExportPath: import_zod14.z.string().optional(),
6423
+ lastTaskExportPath: import_zod14.z.string().optional(),
6424
+ lastImageSavePath: import_zod14.z.string().optional(),
6425
+ /**
6426
+ * Path to worktree to auto-open after switching workspaces.
6427
+ * Used by the worktree feature to open the Roo Code sidebar in a new window.
6428
+ */
6429
+ worktreeAutoOpenPath: import_zod14.z.string().optional(),
6430
+ /**
6431
+ * Whether to show the worktree selector in the home screen.
6432
+ * @default true
6433
+ */
6434
+ showWorktreesInHomeScreen: import_zod14.z.boolean().optional()
6685
6435
  });
6686
6436
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
6687
6437
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -6765,8 +6515,6 @@ var EVALS_SETTINGS = {
6765
6515
  ttsSpeed: 1,
6766
6516
  soundEnabled: false,
6767
6517
  soundVolume: 0.5,
6768
- terminalOutputLineLimit: 500,
6769
- terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6770
6518
  terminalShellIntegrationTimeout: 3e4,
6771
6519
  terminalCommandDelay: 0,
6772
6520
  terminalPowershellCounter: false,
@@ -6774,19 +6522,14 @@ var EVALS_SETTINGS = {
6774
6522
  terminalZshClearEolMark: true,
6775
6523
  terminalZshP10k: false,
6776
6524
  terminalZdotdir: true,
6777
- terminalCompressProgressBar: true,
6778
6525
  terminalShellIntegrationDisabled: true,
6779
6526
  diagnosticsEnabled: true,
6780
- diffEnabled: true,
6781
- fuzzyMatchThreshold: 1,
6782
6527
  enableCheckpoints: false,
6783
6528
  rateLimitSeconds: 0,
6784
6529
  maxOpenTabsContext: 20,
6785
6530
  maxWorkspaceFiles: 200,
6786
6531
  maxGitStatusFiles: 20,
6787
6532
  showRooIgnoredFiles: true,
6788
- maxReadFileLine: -1,
6789
- // -1 to enable full file reading.
6790
6533
  includeDiagnosticMessages: true,
6791
6534
  maxDiagnosticMessages: 50,
6792
6535
  language: "en",
@@ -6860,24 +6603,18 @@ var organizationAllowListSchema = import_zod16.z.object({
6860
6603
  });
6861
6604
  var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6862
6605
  enableCheckpoints: true,
6863
- fuzzyMatchThreshold: true,
6864
6606
  maxOpenTabsContext: true,
6865
- maxReadFileLine: true,
6866
6607
  maxWorkspaceFiles: true,
6867
6608
  showRooIgnoredFiles: true,
6868
6609
  terminalCommandDelay: true,
6869
- terminalCompressProgressBar: true,
6870
- terminalOutputLineLimit: true,
6871
6610
  terminalShellIntegrationDisabled: true,
6872
6611
  terminalShellIntegrationTimeout: true,
6873
6612
  terminalZshClearEolMark: true
6874
6613
  }).merge(
6875
6614
  import_zod16.z.object({
6876
6615
  maxOpenTabsContext: import_zod16.z.number().int().nonnegative().optional(),
6877
- maxReadFileLine: import_zod16.z.number().int().gte(-1).optional(),
6878
6616
  maxWorkspaceFiles: import_zod16.z.number().int().nonnegative().optional(),
6879
6617
  terminalCommandDelay: import_zod16.z.number().int().nonnegative().optional(),
6880
- terminalOutputLineLimit: import_zod16.z.number().int().nonnegative().optional(),
6881
6618
  terminalShellIntegrationTimeout: import_zod16.z.number().int().nonnegative().optional()
6882
6619
  })
6883
6620
  );
@@ -6910,7 +6647,8 @@ var userFeaturesSchema = import_zod16.z.object({
6910
6647
  });
6911
6648
  var userSettingsConfigSchema = import_zod16.z.object({
6912
6649
  extensionBridgeEnabled: import_zod16.z.boolean().optional(),
6913
- taskSyncEnabled: import_zod16.z.boolean().optional()
6650
+ taskSyncEnabled: import_zod16.z.boolean().optional(),
6651
+ llmEnhancedFeaturesEnabled: import_zod16.z.boolean().optional()
6914
6652
  });
6915
6653
  var userSettingsDataSchema = import_zod16.z.object({
6916
6654
  features: userFeaturesSchema,
@@ -7351,12 +7089,10 @@ var taskCommandSchema = import_zod18.z.discriminatedUnion("commandName", [
7351
7089
  })
7352
7090
  }),
7353
7091
  import_zod18.z.object({
7354
- commandName: import_zod18.z.literal("CancelTask" /* CancelTask */),
7355
- data: import_zod18.z.string()
7092
+ commandName: import_zod18.z.literal("CancelTask" /* CancelTask */)
7356
7093
  }),
7357
7094
  import_zod18.z.object({
7358
- commandName: import_zod18.z.literal("CloseTask" /* CloseTask */),
7359
- data: import_zod18.z.string()
7095
+ commandName: import_zod18.z.literal("CloseTask" /* CloseTask */)
7360
7096
  }),
7361
7097
  import_zod18.z.object({
7362
7098
  commandName: import_zod18.z.literal("ResumeTask" /* ResumeTask */),
@@ -7392,6 +7128,7 @@ var ipcMessageSchema = import_zod18.z.discriminatedUnion("type", [
7392
7128
 
7393
7129
  // src/mcp.ts
7394
7130
  var import_zod19 = require("zod");
7131
+ var MAX_MCP_TOOLS_THRESHOLD = 60;
7395
7132
  var mcpExecutionStatusSchema = import_zod19.z.discriminatedUnion("status", [
7396
7133
  import_zod19.z.object({
7397
7134
  executionId: import_zod19.z.string(),
@@ -7415,6 +7152,46 @@ var mcpExecutionStatusSchema = import_zod19.z.discriminatedUnion("status", [
7415
7152
  error: import_zod19.z.string().optional()
7416
7153
  })
7417
7154
  ]);
7155
+ function countEnabledMcpTools(servers) {
7156
+ let serverCount = 0;
7157
+ let toolCount = 0;
7158
+ for (const server of servers) {
7159
+ if (server.disabled) continue;
7160
+ if (server.status !== "connected") continue;
7161
+ serverCount++;
7162
+ if (server.tools) {
7163
+ for (const tool of server.tools) {
7164
+ if (tool.enabledForPrompt !== false) {
7165
+ toolCount++;
7166
+ }
7167
+ }
7168
+ }
7169
+ }
7170
+ return { enabledToolCount: toolCount, enabledServerCount: serverCount };
7171
+ }
7172
+
7173
+ // src/skills.ts
7174
+ var SKILL_NAME_MIN_LENGTH = 1;
7175
+ var SKILL_NAME_MAX_LENGTH = 64;
7176
+ var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
7177
+ var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
7178
+ SkillNameValidationError2["Empty"] = "empty";
7179
+ SkillNameValidationError2["TooLong"] = "too_long";
7180
+ SkillNameValidationError2["InvalidFormat"] = "invalid_format";
7181
+ return SkillNameValidationError2;
7182
+ })(SkillNameValidationError || {});
7183
+ function validateSkillName(name) {
7184
+ if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
7185
+ return { valid: false, error: "empty" /* Empty */ };
7186
+ }
7187
+ if (name.length > SKILL_NAME_MAX_LENGTH) {
7188
+ return { valid: false, error: "too_long" /* TooLong */ };
7189
+ }
7190
+ if (!SKILL_NAME_REGEX.test(name)) {
7191
+ return { valid: false, error: "invalid_format" /* InvalidFormat */ };
7192
+ }
7193
+ return { valid: true };
7194
+ }
7418
7195
 
7419
7196
  // src/todo.ts
7420
7197
  var import_zod20 = require("zod");
@@ -7454,6 +7231,11 @@ var commandExecutionStatusSchema = import_zod21.z.discriminatedUnion("status", [
7454
7231
  })
7455
7232
  ]);
7456
7233
 
7234
+ // src/tool-params.ts
7235
+ function isLegacyReadFileParams(params) {
7236
+ return "_legacyFormat" in params && params._legacyFormat === true;
7237
+ }
7238
+
7457
7239
  // src/vscode-extension-host.ts
7458
7240
  var import_zod22 = require("zod");
7459
7241
  var checkoutDiffPayloadSchema = import_zod22.z.object({
@@ -7507,7 +7289,7 @@ var browserActions = [
7507
7289
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
7508
7290
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
7509
7291
  DEFAULT_MODES,
7510
- DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
7292
+ DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
7511
7293
  DEFAULT_WRITE_DELAY_MS,
7512
7294
  DOUBAO_API_BASE_URL,
7513
7295
  DOUBAO_API_CHAT_PATH,
@@ -7537,13 +7319,13 @@ var browserActions = [
7537
7319
  IpcOrigin,
7538
7320
  LMSTUDIO_DEFAULT_TEMPERATURE,
7539
7321
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
7322
+ MAX_MCP_TOOLS_THRESHOLD,
7540
7323
  MINIMAX_DEFAULT_MAX_TOKENS,
7541
7324
  MINIMAX_DEFAULT_TEMPERATURE,
7542
7325
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
7543
7326
  MISTRAL_DEFAULT_TEMPERATURE,
7544
7327
  MODELS_BY_PROVIDER,
7545
7328
  MOONSHOT_DEFAULT_TEMPERATURE,
7546
- NATIVE_TOOL_DEFAULTS,
7547
7329
  OPENAI_AZURE_AI_INFERENCE_PATH,
7548
7330
  OPENAI_NATIVE_DEFAULT_TEMPERATURE,
7549
7331
  OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -7558,7 +7340,11 @@ var browserActions = [
7558
7340
  RooModelsResponseSchema,
7559
7341
  RooPricingSchema,
7560
7342
  SECRET_STATE_KEYS,
7561
- TOOL_PROTOCOL,
7343
+ SKILL_NAME_MAX_LENGTH,
7344
+ SKILL_NAME_MIN_LENGTH,
7345
+ SKILL_NAME_REGEX,
7346
+ SkillNameValidationError,
7347
+ TERMINAL_PREVIEW_BYTES,
7562
7348
  TaskBridgeCommandName,
7563
7349
  TaskBridgeEventName,
7564
7350
  TaskCommandName,
@@ -7590,9 +7376,6 @@ var browserActions = [
7590
7376
  chutesDefaultModelId,
7591
7377
  chutesDefaultModelInfo,
7592
7378
  chutesModels,
7593
- claudeCodeDefaultModelId,
7594
- claudeCodeModels,
7595
- claudeCodeReasoningConfig,
7596
7379
  clineAskSchema,
7597
7380
  clineAsks,
7598
7381
  clineMessageSchema,
@@ -7607,6 +7390,7 @@ var browserActions = [
7607
7390
  commandIds,
7608
7391
  contextCondenseSchema,
7609
7392
  contextTruncationSchema,
7393
+ countEnabledMcpTools,
7610
7394
  customModePromptsSchema,
7611
7395
  customModesSettingsSchema,
7612
7396
  customProviders,
@@ -7640,7 +7424,6 @@ var browserActions = [
7640
7424
  geminiDefaultModelId,
7641
7425
  geminiModels,
7642
7426
  getApiProtocol,
7643
- getEffectiveProtocol,
7644
7427
  getErrorMessage,
7645
7428
  getErrorStatusCode,
7646
7429
  getImageGenerationProvider,
@@ -7675,9 +7458,9 @@ var browserActions = [
7675
7458
  isInteractiveAsk,
7676
7459
  isInternalProvider,
7677
7460
  isLanguage,
7461
+ isLegacyReadFileParams,
7678
7462
  isLocalProvider,
7679
7463
  isModelParameter,
7680
- isNativeProtocol,
7681
7464
  isNonBlockingAsk,
7682
7465
  isProviderName,
7683
7466
  isResumableAsk,
@@ -7713,9 +7496,10 @@ var browserActions = [
7713
7496
  moonshotDefaultModelId,
7714
7497
  moonshotModels,
7715
7498
  nonBlockingAsks,
7716
- normalizeClaudeCodeModelId,
7717
7499
  ollamaDefaultModelId,
7718
7500
  ollamaDefaultModelInfo,
7501
+ openAiCodexDefaultModelId,
7502
+ openAiCodexModels,
7719
7503
  openAiModelInfoSaneDefaults,
7720
7504
  openAiNativeDefaultModelId,
7721
7505
  openAiNativeModels,
@@ -7785,6 +7569,7 @@ var browserActions = [
7785
7569
  userFeaturesSchema,
7786
7570
  userSettingsConfigSchema,
7787
7571
  userSettingsDataSchema,
7572
+ validateSkillName,
7788
7573
  verbosityLevels,
7789
7574
  verbosityLevelsSchema,
7790
7575
  vercelAiGatewayDefaultModelId,