@roo-code/types 1.106.0 → 1.107.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -78,7 +78,9 @@ var clineSays = [
78
78
  "condense_context_error",
79
79
  "sliding_window_truncation",
80
80
  "codebase_search_result",
81
- "user_edit_todos"
81
+ "user_edit_todos",
82
+ "too_many_tools_warning",
83
+ "tool"
82
84
  ];
83
85
  var clineSaySchema = z.enum(clineSays);
84
86
  var toolProgressStatusSchema = z.object({
@@ -146,6 +148,7 @@ var toolGroupsSchema = z2.enum(toolGroups);
146
148
  var toolNames = [
147
149
  "execute_command",
148
150
  "read_file",
151
+ "read_command_output",
149
152
  "write_to_file",
150
153
  "apply_diff",
151
154
  "search_and_replace",
@@ -161,10 +164,10 @@ var toolNames = [
161
164
  "attempt_completion",
162
165
  "switch_mode",
163
166
  "new_task",
164
- "fetch_instructions",
165
167
  "codebase_search",
166
168
  "update_todo_list",
167
169
  "run_slash_command",
170
+ "skill",
168
171
  "generate_image",
169
172
  "custom_tool"
170
173
  ];
@@ -176,20 +179,6 @@ var toolUsageSchema = z2.record(
176
179
  failures: z2.number()
177
180
  })
178
181
  );
179
- var TOOL_PROTOCOL = {
180
- XML: "xml",
181
- NATIVE: "native"
182
- };
183
- var NATIVE_TOOL_DEFAULTS = {
184
- supportsNativeTools: true,
185
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
186
- };
187
- function isNativeProtocol(protocol) {
188
- return protocol === TOOL_PROTOCOL.NATIVE;
189
- }
190
- function getEffectiveProtocol(toolProtocol) {
191
- return toolProtocol || TOOL_PROTOCOL.XML;
192
- }
193
182
 
194
183
  // src/events.ts
195
184
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
@@ -213,6 +202,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
213
202
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
214
203
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
215
204
  RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
205
+ RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
216
206
  RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
217
207
  RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
218
208
  RooCodeEventName2["ModeChanged"] = "modeChanged";
@@ -272,6 +262,7 @@ var rooCodeEventsSchema = z3.object({
272
262
  ["taskModeSwitched" /* TaskModeSwitched */]: z3.tuple([z3.string(), z3.string()]),
273
263
  ["taskAskResponded" /* TaskAskResponded */]: z3.tuple([z3.string()]),
274
264
  ["taskUserMessage" /* TaskUserMessage */]: z3.tuple([z3.string()]),
265
+ ["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: z3.tuple([z3.string(), z3.array(queuedMessageSchema)]),
275
266
  ["taskToolFailed" /* TaskToolFailed */]: z3.tuple([z3.string(), toolNamesSchema, z3.string()]),
276
267
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z3.tuple([z3.string(), tokenUsageSchema, toolUsageSchema]),
277
268
  ["modeChanged" /* ModeChanged */]: z3.tuple([z3.string()]),
@@ -377,6 +368,11 @@ var taskEventSchema = z3.discriminatedUnion("eventName", [
377
368
  payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
378
369
  taskId: z3.number().optional()
379
370
  }),
371
+ z3.object({
372
+ eventName: z3.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
373
+ payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
374
+ taskId: z3.number().optional()
375
+ }),
380
376
  // Task Analytics
381
377
  z3.object({
382
378
  eventName: z3.literal("taskToolFailed" /* TaskToolFailed */),
@@ -477,10 +473,6 @@ var modelInfoSchema = z5.object({
477
473
  isStealthModel: z5.boolean().optional(),
478
474
  // Flag to indicate if the model is free (no cost)
479
475
  isFree: z5.boolean().optional(),
480
- // Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
481
- supportsNativeTools: z5.boolean().optional(),
482
- // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
483
- defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
484
476
  // Exclude specific native tools from being available (only applies to native protocol)
485
477
  // These tools will be removed from the set of tools available to the model
486
478
  excludedTools: z5.array(z5.string()).optional(),
@@ -577,8 +569,6 @@ var anthropicModels = {
577
569
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
578
570
  supportsImages: true,
579
571
  supportsPromptCache: true,
580
- supportsNativeTools: true,
581
- defaultToolProtocol: "native",
582
572
  inputPrice: 3,
583
573
  // $3 per million input tokens (≤200K context)
584
574
  outputPrice: 15,
@@ -611,8 +601,6 @@ var anthropicModels = {
611
601
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
612
602
  supportsImages: true,
613
603
  supportsPromptCache: true,
614
- supportsNativeTools: true,
615
- defaultToolProtocol: "native",
616
604
  inputPrice: 3,
617
605
  // $3 per million input tokens (≤200K context)
618
606
  outputPrice: 15,
@@ -644,8 +632,6 @@ var anthropicModels = {
644
632
  contextWindow: 2e5,
645
633
  supportsImages: true,
646
634
  supportsPromptCache: true,
647
- supportsNativeTools: true,
648
- defaultToolProtocol: "native",
649
635
  inputPrice: 5,
650
636
  // $5 per million input tokens
651
637
  outputPrice: 25,
@@ -662,8 +648,6 @@ var anthropicModels = {
662
648
  contextWindow: 2e5,
663
649
  supportsImages: true,
664
650
  supportsPromptCache: true,
665
- supportsNativeTools: true,
666
- defaultToolProtocol: "native",
667
651
  inputPrice: 15,
668
652
  // $15 per million input tokens
669
653
  outputPrice: 75,
@@ -680,8 +664,6 @@ var anthropicModels = {
680
664
  contextWindow: 2e5,
681
665
  supportsImages: true,
682
666
  supportsPromptCache: true,
683
- supportsNativeTools: true,
684
- defaultToolProtocol: "native",
685
667
  inputPrice: 15,
686
668
  // $15 per million input tokens
687
669
  outputPrice: 75,
@@ -698,8 +680,6 @@ var anthropicModels = {
698
680
  contextWindow: 2e5,
699
681
  supportsImages: true,
700
682
  supportsPromptCache: true,
701
- supportsNativeTools: true,
702
- defaultToolProtocol: "native",
703
683
  inputPrice: 3,
704
684
  // $3 per million input tokens
705
685
  outputPrice: 15,
@@ -717,8 +697,6 @@ var anthropicModels = {
717
697
  contextWindow: 2e5,
718
698
  supportsImages: true,
719
699
  supportsPromptCache: true,
720
- supportsNativeTools: true,
721
- defaultToolProtocol: "native",
722
700
  inputPrice: 3,
723
701
  // $3 per million input tokens
724
702
  outputPrice: 15,
@@ -733,8 +711,6 @@ var anthropicModels = {
733
711
  contextWindow: 2e5,
734
712
  supportsImages: true,
735
713
  supportsPromptCache: true,
736
- supportsNativeTools: true,
737
- defaultToolProtocol: "native",
738
714
  inputPrice: 3,
739
715
  // $3 per million input tokens
740
716
  outputPrice: 15,
@@ -749,8 +725,6 @@ var anthropicModels = {
749
725
  contextWindow: 2e5,
750
726
  supportsImages: false,
751
727
  supportsPromptCache: true,
752
- supportsNativeTools: true,
753
- defaultToolProtocol: "native",
754
728
  inputPrice: 1,
755
729
  outputPrice: 5,
756
730
  cacheWritesPrice: 1.25,
@@ -761,8 +735,6 @@ var anthropicModels = {
761
735
  contextWindow: 2e5,
762
736
  supportsImages: true,
763
737
  supportsPromptCache: true,
764
- supportsNativeTools: true,
765
- defaultToolProtocol: "native",
766
738
  inputPrice: 15,
767
739
  outputPrice: 75,
768
740
  cacheWritesPrice: 18.75,
@@ -773,8 +745,6 @@ var anthropicModels = {
773
745
  contextWindow: 2e5,
774
746
  supportsImages: true,
775
747
  supportsPromptCache: true,
776
- supportsNativeTools: true,
777
- defaultToolProtocol: "native",
778
748
  inputPrice: 0.25,
779
749
  outputPrice: 1.25,
780
750
  cacheWritesPrice: 0.3,
@@ -785,8 +755,6 @@ var anthropicModels = {
785
755
  contextWindow: 2e5,
786
756
  supportsImages: true,
787
757
  supportsPromptCache: true,
788
- supportsNativeTools: true,
789
- defaultToolProtocol: "native",
790
758
  inputPrice: 1,
791
759
  outputPrice: 5,
792
760
  cacheWritesPrice: 1.25,
@@ -804,7 +772,6 @@ var basetenModels = {
804
772
  contextWindow: 262e3,
805
773
  supportsImages: false,
806
774
  supportsPromptCache: false,
807
- supportsNativeTools: true,
808
775
  inputPrice: 0.6,
809
776
  outputPrice: 2.5,
810
777
  cacheWritesPrice: 0,
@@ -816,7 +783,6 @@ var basetenModels = {
816
783
  contextWindow: 2e5,
817
784
  supportsImages: false,
818
785
  supportsPromptCache: false,
819
- supportsNativeTools: true,
820
786
  inputPrice: 0.6,
821
787
  outputPrice: 2.2,
822
788
  cacheWritesPrice: 0,
@@ -828,7 +794,6 @@ var basetenModels = {
828
794
  contextWindow: 163840,
829
795
  supportsImages: false,
830
796
  supportsPromptCache: false,
831
- supportsNativeTools: true,
832
797
  inputPrice: 2.55,
833
798
  outputPrice: 5.95,
834
799
  cacheWritesPrice: 0,
@@ -840,7 +805,6 @@ var basetenModels = {
840
805
  contextWindow: 163840,
841
806
  supportsImages: false,
842
807
  supportsPromptCache: false,
843
- supportsNativeTools: true,
844
808
  inputPrice: 2.55,
845
809
  outputPrice: 5.95,
846
810
  cacheWritesPrice: 0,
@@ -852,7 +816,6 @@ var basetenModels = {
852
816
  contextWindow: 163840,
853
817
  supportsImages: false,
854
818
  supportsPromptCache: false,
855
- supportsNativeTools: true,
856
819
  inputPrice: 0.77,
857
820
  outputPrice: 0.77,
858
821
  cacheWritesPrice: 0,
@@ -864,7 +827,6 @@ var basetenModels = {
864
827
  contextWindow: 163840,
865
828
  supportsImages: false,
866
829
  supportsPromptCache: false,
867
- supportsNativeTools: true,
868
830
  inputPrice: 0.5,
869
831
  outputPrice: 1.5,
870
832
  cacheWritesPrice: 0,
@@ -876,7 +838,6 @@ var basetenModels = {
876
838
  contextWindow: 163840,
877
839
  supportsImages: false,
878
840
  supportsPromptCache: false,
879
- supportsNativeTools: true,
880
841
  inputPrice: 0.3,
881
842
  outputPrice: 0.45,
882
843
  cacheWritesPrice: 0,
@@ -888,7 +849,6 @@ var basetenModels = {
888
849
  contextWindow: 128072,
889
850
  supportsImages: false,
890
851
  supportsPromptCache: false,
891
- supportsNativeTools: true,
892
852
  inputPrice: 0.1,
893
853
  outputPrice: 0.5,
894
854
  cacheWritesPrice: 0,
@@ -900,7 +860,6 @@ var basetenModels = {
900
860
  contextWindow: 262144,
901
861
  supportsImages: false,
902
862
  supportsPromptCache: false,
903
- supportsNativeTools: true,
904
863
  inputPrice: 0.22,
905
864
  outputPrice: 0.8,
906
865
  cacheWritesPrice: 0,
@@ -912,7 +871,6 @@ var basetenModels = {
912
871
  contextWindow: 262144,
913
872
  supportsImages: false,
914
873
  supportsPromptCache: false,
915
- supportsNativeTools: true,
916
874
  inputPrice: 0.38,
917
875
  outputPrice: 1.53,
918
876
  cacheWritesPrice: 0,
@@ -924,7 +882,6 @@ var basetenModels = {
924
882
  contextWindow: 262e3,
925
883
  supportsImages: false,
926
884
  supportsPromptCache: false,
927
- supportsNativeTools: true,
928
885
  inputPrice: 0.6,
929
886
  outputPrice: 2.5,
930
887
  cacheWritesPrice: 0,
@@ -944,8 +901,6 @@ var bedrockModels = {
944
901
  supportsImages: true,
945
902
  supportsPromptCache: true,
946
903
  supportsReasoningBudget: true,
947
- supportsNativeTools: true,
948
- defaultToolProtocol: "native",
949
904
  inputPrice: 3,
950
905
  outputPrice: 15,
951
906
  cacheWritesPrice: 3.75,
@@ -959,7 +914,6 @@ var bedrockModels = {
959
914
  contextWindow: 3e5,
960
915
  supportsImages: true,
961
916
  supportsPromptCache: true,
962
- supportsNativeTools: true,
963
917
  inputPrice: 0.8,
964
918
  outputPrice: 3.2,
965
919
  cacheWritesPrice: 0.8,
@@ -975,7 +929,6 @@ var bedrockModels = {
975
929
  contextWindow: 3e5,
976
930
  supportsImages: true,
977
931
  supportsPromptCache: false,
978
- supportsNativeTools: true,
979
932
  inputPrice: 1,
980
933
  outputPrice: 4,
981
934
  cacheWritesPrice: 1,
@@ -989,7 +942,6 @@ var bedrockModels = {
989
942
  contextWindow: 3e5,
990
943
  supportsImages: true,
991
944
  supportsPromptCache: true,
992
- supportsNativeTools: true,
993
945
  inputPrice: 0.06,
994
946
  outputPrice: 0.24,
995
947
  cacheWritesPrice: 0.06,
@@ -1005,7 +957,6 @@ var bedrockModels = {
1005
957
  contextWindow: 1e6,
1006
958
  supportsImages: true,
1007
959
  supportsPromptCache: true,
1008
- supportsNativeTools: true,
1009
960
  inputPrice: 0.33,
1010
961
  outputPrice: 2.75,
1011
962
  cacheWritesPrice: 0,
@@ -1021,7 +972,6 @@ var bedrockModels = {
1021
972
  contextWindow: 128e3,
1022
973
  supportsImages: false,
1023
974
  supportsPromptCache: true,
1024
- supportsNativeTools: true,
1025
975
  inputPrice: 0.035,
1026
976
  outputPrice: 0.14,
1027
977
  cacheWritesPrice: 0.035,
@@ -1038,8 +988,6 @@ var bedrockModels = {
1038
988
  supportsImages: true,
1039
989
  supportsPromptCache: true,
1040
990
  supportsReasoningBudget: true,
1041
- supportsNativeTools: true,
1042
- defaultToolProtocol: "native",
1043
991
  inputPrice: 3,
1044
992
  outputPrice: 15,
1045
993
  cacheWritesPrice: 3.75,
@@ -1054,8 +1002,6 @@ var bedrockModels = {
1054
1002
  supportsImages: true,
1055
1003
  supportsPromptCache: true,
1056
1004
  supportsReasoningBudget: true,
1057
- supportsNativeTools: true,
1058
- defaultToolProtocol: "native",
1059
1005
  inputPrice: 15,
1060
1006
  outputPrice: 75,
1061
1007
  cacheWritesPrice: 18.75,
@@ -1070,8 +1016,6 @@ var bedrockModels = {
1070
1016
  supportsImages: true,
1071
1017
  supportsPromptCache: true,
1072
1018
  supportsReasoningBudget: true,
1073
- supportsNativeTools: true,
1074
- defaultToolProtocol: "native",
1075
1019
  inputPrice: 5,
1076
1020
  outputPrice: 25,
1077
1021
  cacheWritesPrice: 6.25,
@@ -1086,8 +1030,6 @@ var bedrockModels = {
1086
1030
  supportsImages: true,
1087
1031
  supportsPromptCache: true,
1088
1032
  supportsReasoningBudget: true,
1089
- supportsNativeTools: true,
1090
- defaultToolProtocol: "native",
1091
1033
  inputPrice: 15,
1092
1034
  outputPrice: 75,
1093
1035
  cacheWritesPrice: 18.75,
@@ -1102,8 +1044,6 @@ var bedrockModels = {
1102
1044
  supportsImages: true,
1103
1045
  supportsPromptCache: true,
1104
1046
  supportsReasoningBudget: true,
1105
- supportsNativeTools: true,
1106
- defaultToolProtocol: "native",
1107
1047
  inputPrice: 3,
1108
1048
  outputPrice: 15,
1109
1049
  cacheWritesPrice: 3.75,
@@ -1117,8 +1057,6 @@ var bedrockModels = {
1117
1057
  contextWindow: 2e5,
1118
1058
  supportsImages: true,
1119
1059
  supportsPromptCache: true,
1120
- supportsNativeTools: true,
1121
- defaultToolProtocol: "native",
1122
1060
  inputPrice: 3,
1123
1061
  outputPrice: 15,
1124
1062
  cacheWritesPrice: 3.75,
@@ -1132,8 +1070,6 @@ var bedrockModels = {
1132
1070
  contextWindow: 2e5,
1133
1071
  supportsImages: false,
1134
1072
  supportsPromptCache: true,
1135
- supportsNativeTools: true,
1136
- defaultToolProtocol: "native",
1137
1073
  inputPrice: 0.8,
1138
1074
  outputPrice: 4,
1139
1075
  cacheWritesPrice: 1,
@@ -1148,8 +1084,6 @@ var bedrockModels = {
1148
1084
  supportsImages: true,
1149
1085
  supportsPromptCache: true,
1150
1086
  supportsReasoningBudget: true,
1151
- supportsNativeTools: true,
1152
- defaultToolProtocol: "native",
1153
1087
  inputPrice: 1,
1154
1088
  outputPrice: 5,
1155
1089
  cacheWritesPrice: 1.25,
@@ -1165,8 +1099,6 @@ var bedrockModels = {
1165
1099
  contextWindow: 2e5,
1166
1100
  supportsImages: true,
1167
1101
  supportsPromptCache: false,
1168
- supportsNativeTools: true,
1169
- defaultToolProtocol: "native",
1170
1102
  inputPrice: 3,
1171
1103
  outputPrice: 15
1172
1104
  },
@@ -1175,8 +1107,6 @@ var bedrockModels = {
1175
1107
  contextWindow: 2e5,
1176
1108
  supportsImages: true,
1177
1109
  supportsPromptCache: false,
1178
- supportsNativeTools: true,
1179
- defaultToolProtocol: "native",
1180
1110
  inputPrice: 15,
1181
1111
  outputPrice: 75
1182
1112
  },
@@ -1185,8 +1115,6 @@ var bedrockModels = {
1185
1115
  contextWindow: 2e5,
1186
1116
  supportsImages: true,
1187
1117
  supportsPromptCache: false,
1188
- supportsNativeTools: true,
1189
- defaultToolProtocol: "native",
1190
1118
  inputPrice: 3,
1191
1119
  outputPrice: 15
1192
1120
  },
@@ -1195,8 +1123,6 @@ var bedrockModels = {
1195
1123
  contextWindow: 2e5,
1196
1124
  supportsImages: true,
1197
1125
  supportsPromptCache: false,
1198
- supportsNativeTools: true,
1199
- defaultToolProtocol: "native",
1200
1126
  inputPrice: 0.25,
1201
1127
  outputPrice: 1.25
1202
1128
  },
@@ -1205,7 +1131,6 @@ var bedrockModels = {
1205
1131
  contextWindow: 128e3,
1206
1132
  supportsImages: false,
1207
1133
  supportsPromptCache: false,
1208
- supportsNativeTools: true,
1209
1134
  inputPrice: 1.35,
1210
1135
  outputPrice: 5.4
1211
1136
  },
@@ -1214,7 +1139,6 @@ var bedrockModels = {
1214
1139
  contextWindow: 128e3,
1215
1140
  supportsImages: false,
1216
1141
  supportsPromptCache: false,
1217
- supportsNativeTools: true,
1218
1142
  inputPrice: 0.5,
1219
1143
  outputPrice: 1.5,
1220
1144
  description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
@@ -1224,7 +1148,6 @@ var bedrockModels = {
1224
1148
  contextWindow: 128e3,
1225
1149
  supportsImages: false,
1226
1150
  supportsPromptCache: false,
1227
- supportsNativeTools: true,
1228
1151
  inputPrice: 2,
1229
1152
  outputPrice: 6,
1230
1153
  description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
@@ -1234,7 +1157,6 @@ var bedrockModels = {
1234
1157
  contextWindow: 128e3,
1235
1158
  supportsImages: false,
1236
1159
  supportsPromptCache: false,
1237
- supportsNativeTools: true,
1238
1160
  inputPrice: 0.72,
1239
1161
  outputPrice: 0.72,
1240
1162
  description: "Llama 3.3 Instruct (70B)"
@@ -1244,7 +1166,6 @@ var bedrockModels = {
1244
1166
  contextWindow: 128e3,
1245
1167
  supportsImages: true,
1246
1168
  supportsPromptCache: false,
1247
- supportsNativeTools: true,
1248
1169
  inputPrice: 0.72,
1249
1170
  outputPrice: 0.72,
1250
1171
  description: "Llama 3.2 Instruct (90B)"
@@ -1254,7 +1175,6 @@ var bedrockModels = {
1254
1175
  contextWindow: 128e3,
1255
1176
  supportsImages: true,
1256
1177
  supportsPromptCache: false,
1257
- supportsNativeTools: true,
1258
1178
  inputPrice: 0.16,
1259
1179
  outputPrice: 0.16,
1260
1180
  description: "Llama 3.2 Instruct (11B)"
@@ -1264,7 +1184,6 @@ var bedrockModels = {
1264
1184
  contextWindow: 128e3,
1265
1185
  supportsImages: false,
1266
1186
  supportsPromptCache: false,
1267
- supportsNativeTools: true,
1268
1187
  inputPrice: 0.15,
1269
1188
  outputPrice: 0.15,
1270
1189
  description: "Llama 3.2 Instruct (3B)"
@@ -1274,7 +1193,6 @@ var bedrockModels = {
1274
1193
  contextWindow: 128e3,
1275
1194
  supportsImages: false,
1276
1195
  supportsPromptCache: false,
1277
- supportsNativeTools: true,
1278
1196
  inputPrice: 0.1,
1279
1197
  outputPrice: 0.1,
1280
1198
  description: "Llama 3.2 Instruct (1B)"
@@ -1284,7 +1202,6 @@ var bedrockModels = {
1284
1202
  contextWindow: 128e3,
1285
1203
  supportsImages: false,
1286
1204
  supportsPromptCache: false,
1287
- supportsNativeTools: true,
1288
1205
  inputPrice: 2.4,
1289
1206
  outputPrice: 2.4,
1290
1207
  description: "Llama 3.1 Instruct (405B)"
@@ -1294,7 +1211,6 @@ var bedrockModels = {
1294
1211
  contextWindow: 128e3,
1295
1212
  supportsImages: false,
1296
1213
  supportsPromptCache: false,
1297
- supportsNativeTools: true,
1298
1214
  inputPrice: 0.72,
1299
1215
  outputPrice: 0.72,
1300
1216
  description: "Llama 3.1 Instruct (70B)"
@@ -1304,7 +1220,6 @@ var bedrockModels = {
1304
1220
  contextWindow: 128e3,
1305
1221
  supportsImages: false,
1306
1222
  supportsPromptCache: false,
1307
- supportsNativeTools: true,
1308
1223
  inputPrice: 0.9,
1309
1224
  outputPrice: 0.9,
1310
1225
  description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
@@ -1314,7 +1229,6 @@ var bedrockModels = {
1314
1229
  contextWindow: 8e3,
1315
1230
  supportsImages: false,
1316
1231
  supportsPromptCache: false,
1317
- supportsNativeTools: true,
1318
1232
  inputPrice: 0.22,
1319
1233
  outputPrice: 0.22,
1320
1234
  description: "Llama 3.1 Instruct (8B)"
@@ -1324,7 +1238,6 @@ var bedrockModels = {
1324
1238
  contextWindow: 8e3,
1325
1239
  supportsImages: false,
1326
1240
  supportsPromptCache: false,
1327
- supportsNativeTools: true,
1328
1241
  inputPrice: 2.65,
1329
1242
  outputPrice: 3.5
1330
1243
  },
@@ -1333,7 +1246,6 @@ var bedrockModels = {
1333
1246
  contextWindow: 4e3,
1334
1247
  supportsImages: false,
1335
1248
  supportsPromptCache: false,
1336
- supportsNativeTools: true,
1337
1249
  inputPrice: 0.3,
1338
1250
  outputPrice: 0.6
1339
1251
  },
@@ -1342,7 +1254,6 @@ var bedrockModels = {
1342
1254
  contextWindow: 8e3,
1343
1255
  supportsImages: false,
1344
1256
  supportsPromptCache: false,
1345
- supportsNativeTools: true,
1346
1257
  inputPrice: 0.15,
1347
1258
  outputPrice: 0.2,
1348
1259
  description: "Amazon Titan Text Lite"
@@ -1352,7 +1263,6 @@ var bedrockModels = {
1352
1263
  contextWindow: 8e3,
1353
1264
  supportsImages: false,
1354
1265
  supportsPromptCache: false,
1355
- supportsNativeTools: true,
1356
1266
  inputPrice: 0.2,
1357
1267
  outputPrice: 0.6,
1358
1268
  description: "Amazon Titan Text Express"
@@ -1362,8 +1272,6 @@ var bedrockModels = {
1362
1272
  contextWindow: 262144,
1363
1273
  supportsImages: false,
1364
1274
  supportsPromptCache: false,
1365
- supportsNativeTools: true,
1366
- defaultToolProtocol: "native",
1367
1275
  preserveReasoning: true,
1368
1276
  inputPrice: 0.6,
1369
1277
  outputPrice: 2.5,
@@ -1374,8 +1282,6 @@ var bedrockModels = {
1374
1282
  contextWindow: 196608,
1375
1283
  supportsImages: false,
1376
1284
  supportsPromptCache: false,
1377
- supportsNativeTools: true,
1378
- defaultToolProtocol: "native",
1379
1285
  preserveReasoning: true,
1380
1286
  inputPrice: 0.3,
1381
1287
  outputPrice: 1.2,
@@ -1386,8 +1292,6 @@ var bedrockModels = {
1386
1292
  contextWindow: 262144,
1387
1293
  supportsImages: false,
1388
1294
  supportsPromptCache: false,
1389
- supportsNativeTools: true,
1390
- defaultToolProtocol: "native",
1391
1295
  inputPrice: 0.15,
1392
1296
  outputPrice: 1.2,
1393
1297
  description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
@@ -1397,8 +1301,6 @@ var bedrockModels = {
1397
1301
  contextWindow: 262144,
1398
1302
  supportsImages: false,
1399
1303
  supportsPromptCache: false,
1400
- supportsNativeTools: true,
1401
- defaultToolProtocol: "native",
1402
1304
  inputPrice: 0.45,
1403
1305
  outputPrice: 1.8,
1404
1306
  description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
@@ -1489,26 +1391,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
1489
1391
  // src/providers/cerebras.ts
1490
1392
  var cerebrasDefaultModelId = "gpt-oss-120b";
1491
1393
  var cerebrasModels = {
1492
- "zai-glm-4.6": {
1493
- maxTokens: 16384,
1494
- // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1495
- contextWindow: 131072,
1496
- supportsImages: false,
1497
- supportsPromptCache: false,
1498
- supportsNativeTools: true,
1499
- defaultToolProtocol: "native",
1500
- inputPrice: 0,
1501
- outputPrice: 0,
1502
- description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
1503
- },
1504
1394
  "zai-glm-4.7": {
1505
1395
  maxTokens: 16384,
1506
1396
  // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1507
1397
  contextWindow: 131072,
1508
1398
  supportsImages: false,
1509
- supportsPromptCache: false,
1510
- supportsNativeTools: true,
1511
- defaultToolProtocol: "native",
1399
+ supportsPromptCache: true,
1400
+ supportsTemperature: true,
1401
+ defaultTemperature: 1,
1512
1402
  inputPrice: 0,
1513
1403
  outputPrice: 0,
1514
1404
  description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
@@ -1519,8 +1409,6 @@ var cerebrasModels = {
1519
1409
  contextWindow: 64e3,
1520
1410
  supportsImages: false,
1521
1411
  supportsPromptCache: false,
1522
- supportsNativeTools: true,
1523
- defaultToolProtocol: "native",
1524
1412
  inputPrice: 0,
1525
1413
  outputPrice: 0,
1526
1414
  description: "Intelligent model with ~1400 tokens/s"
@@ -1531,8 +1419,6 @@ var cerebrasModels = {
1531
1419
  contextWindow: 64e3,
1532
1420
  supportsImages: false,
1533
1421
  supportsPromptCache: false,
1534
- supportsNativeTools: true,
1535
- defaultToolProtocol: "native",
1536
1422
  inputPrice: 0,
1537
1423
  outputPrice: 0,
1538
1424
  description: "Powerful model with ~2600 tokens/s"
@@ -1543,8 +1429,6 @@ var cerebrasModels = {
1543
1429
  contextWindow: 64e3,
1544
1430
  supportsImages: false,
1545
1431
  supportsPromptCache: false,
1546
- supportsNativeTools: true,
1547
- defaultToolProtocol: "native",
1548
1432
  inputPrice: 0,
1549
1433
  outputPrice: 0,
1550
1434
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1555,8 +1439,6 @@ var cerebrasModels = {
1555
1439
  contextWindow: 64e3,
1556
1440
  supportsImages: false,
1557
1441
  supportsPromptCache: false,
1558
- supportsNativeTools: true,
1559
- defaultToolProtocol: "native",
1560
1442
  inputPrice: 0,
1561
1443
  outputPrice: 0,
1562
1444
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1571,8 +1453,6 @@ var chutesModels = {
1571
1453
  contextWindow: 163840,
1572
1454
  supportsImages: false,
1573
1455
  supportsPromptCache: false,
1574
- supportsNativeTools: true,
1575
- defaultToolProtocol: "native",
1576
1456
  inputPrice: 0,
1577
1457
  outputPrice: 0,
1578
1458
  description: "DeepSeek R1 0528 model."
@@ -1582,8 +1462,6 @@ var chutesModels = {
1582
1462
  contextWindow: 163840,
1583
1463
  supportsImages: false,
1584
1464
  supportsPromptCache: false,
1585
- supportsNativeTools: true,
1586
- defaultToolProtocol: "native",
1587
1465
  inputPrice: 0,
1588
1466
  outputPrice: 0,
1589
1467
  description: "DeepSeek R1 model."
@@ -1593,8 +1471,6 @@ var chutesModels = {
1593
1471
  contextWindow: 163840,
1594
1472
  supportsImages: false,
1595
1473
  supportsPromptCache: false,
1596
- supportsNativeTools: true,
1597
- defaultToolProtocol: "native",
1598
1474
  inputPrice: 0,
1599
1475
  outputPrice: 0,
1600
1476
  description: "DeepSeek V3 model."
@@ -1604,8 +1480,6 @@ var chutesModels = {
1604
1480
  contextWindow: 163840,
1605
1481
  supportsImages: false,
1606
1482
  supportsPromptCache: false,
1607
- supportsNativeTools: true,
1608
- defaultToolProtocol: "native",
1609
1483
  inputPrice: 0,
1610
1484
  outputPrice: 0,
1611
1485
  description: "DeepSeek V3.1 model."
@@ -1615,8 +1489,6 @@ var chutesModels = {
1615
1489
  contextWindow: 163840,
1616
1490
  supportsImages: false,
1617
1491
  supportsPromptCache: false,
1618
- supportsNativeTools: true,
1619
- defaultToolProtocol: "native",
1620
1492
  inputPrice: 0.23,
1621
1493
  outputPrice: 0.9,
1622
1494
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1626,8 +1498,6 @@ var chutesModels = {
1626
1498
  contextWindow: 163840,
1627
1499
  supportsImages: false,
1628
1500
  supportsPromptCache: false,
1629
- supportsNativeTools: true,
1630
- defaultToolProtocol: "native",
1631
1501
  inputPrice: 1,
1632
1502
  outputPrice: 3,
1633
1503
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1637,8 +1507,6 @@ var chutesModels = {
1637
1507
  contextWindow: 163840,
1638
1508
  supportsImages: false,
1639
1509
  supportsPromptCache: false,
1640
- supportsNativeTools: true,
1641
- defaultToolProtocol: "native",
1642
1510
  inputPrice: 0.25,
1643
1511
  outputPrice: 0.35,
1644
1512
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1650,8 +1518,6 @@ var chutesModels = {
1650
1518
  // From Groq
1651
1519
  supportsImages: false,
1652
1520
  supportsPromptCache: false,
1653
- supportsNativeTools: true,
1654
- defaultToolProtocol: "native",
1655
1521
  inputPrice: 0,
1656
1522
  outputPrice: 0,
1657
1523
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1661,8 +1527,6 @@ var chutesModels = {
1661
1527
  contextWindow: 512e3,
1662
1528
  supportsImages: false,
1663
1529
  supportsPromptCache: false,
1664
- supportsNativeTools: true,
1665
- defaultToolProtocol: "native",
1666
1530
  inputPrice: 0,
1667
1531
  outputPrice: 0,
1668
1532
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1672,8 +1536,6 @@ var chutesModels = {
1672
1536
  contextWindow: 128e3,
1673
1537
  supportsImages: false,
1674
1538
  supportsPromptCache: false,
1675
- supportsNativeTools: true,
1676
- defaultToolProtocol: "native",
1677
1539
  inputPrice: 0,
1678
1540
  outputPrice: 0,
1679
1541
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1683,8 +1545,6 @@ var chutesModels = {
1683
1545
  contextWindow: 131072,
1684
1546
  supportsImages: false,
1685
1547
  supportsPromptCache: false,
1686
- supportsNativeTools: true,
1687
- defaultToolProtocol: "native",
1688
1548
  inputPrice: 0,
1689
1549
  outputPrice: 0,
1690
1550
  description: "Unsloth Gemma 3 12B IT model."
@@ -1694,8 +1554,6 @@ var chutesModels = {
1694
1554
  contextWindow: 131072,
1695
1555
  supportsImages: false,
1696
1556
  supportsPromptCache: false,
1697
- supportsNativeTools: true,
1698
- defaultToolProtocol: "native",
1699
1557
  inputPrice: 0,
1700
1558
  outputPrice: 0,
1701
1559
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1705,8 +1563,6 @@ var chutesModels = {
1705
1563
  contextWindow: 131072,
1706
1564
  supportsImages: false,
1707
1565
  supportsPromptCache: false,
1708
- supportsNativeTools: true,
1709
- defaultToolProtocol: "native",
1710
1566
  inputPrice: 0,
1711
1567
  outputPrice: 0,
1712
1568
  description: "Unsloth Gemma 3 4B IT model."
@@ -1716,8 +1572,6 @@ var chutesModels = {
1716
1572
  contextWindow: 131072,
1717
1573
  supportsImages: false,
1718
1574
  supportsPromptCache: false,
1719
- supportsNativeTools: true,
1720
- defaultToolProtocol: "native",
1721
1575
  inputPrice: 0,
1722
1576
  outputPrice: 0,
1723
1577
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -1727,8 +1581,6 @@ var chutesModels = {
1727
1581
  contextWindow: 131072,
1728
1582
  supportsImages: false,
1729
1583
  supportsPromptCache: false,
1730
- supportsNativeTools: true,
1731
- defaultToolProtocol: "native",
1732
1584
  inputPrice: 0,
1733
1585
  outputPrice: 0,
1734
1586
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -1738,8 +1590,6 @@ var chutesModels = {
1738
1590
  contextWindow: 256e3,
1739
1591
  supportsImages: false,
1740
1592
  supportsPromptCache: false,
1741
- supportsNativeTools: true,
1742
- defaultToolProtocol: "native",
1743
1593
  inputPrice: 0,
1744
1594
  outputPrice: 0,
1745
1595
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -1749,8 +1599,6 @@ var chutesModels = {
1749
1599
  contextWindow: 163840,
1750
1600
  supportsImages: false,
1751
1601
  supportsPromptCache: false,
1752
- supportsNativeTools: true,
1753
- defaultToolProtocol: "native",
1754
1602
  inputPrice: 0,
1755
1603
  outputPrice: 0,
1756
1604
  description: "DeepSeek V3 Base model."
@@ -1760,8 +1608,6 @@ var chutesModels = {
1760
1608
  contextWindow: 163840,
1761
1609
  supportsImages: false,
1762
1610
  supportsPromptCache: false,
1763
- supportsNativeTools: true,
1764
- defaultToolProtocol: "native",
1765
1611
  inputPrice: 0,
1766
1612
  outputPrice: 0,
1767
1613
  description: "DeepSeek R1 Zero model."
@@ -1771,8 +1617,6 @@ var chutesModels = {
1771
1617
  contextWindow: 163840,
1772
1618
  supportsImages: false,
1773
1619
  supportsPromptCache: false,
1774
- supportsNativeTools: true,
1775
- defaultToolProtocol: "native",
1776
1620
  inputPrice: 0,
1777
1621
  outputPrice: 0,
1778
1622
  description: "DeepSeek V3 (0324) model."
@@ -1782,8 +1626,6 @@ var chutesModels = {
1782
1626
  contextWindow: 262144,
1783
1627
  supportsImages: false,
1784
1628
  supportsPromptCache: false,
1785
- supportsNativeTools: true,
1786
- defaultToolProtocol: "native",
1787
1629
  inputPrice: 0,
1788
1630
  outputPrice: 0,
1789
1631
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -1793,8 +1635,6 @@ var chutesModels = {
1793
1635
  contextWindow: 40960,
1794
1636
  supportsImages: false,
1795
1637
  supportsPromptCache: false,
1796
- supportsNativeTools: true,
1797
- defaultToolProtocol: "native",
1798
1638
  inputPrice: 0,
1799
1639
  outputPrice: 0,
1800
1640
  description: "Qwen3 235B A22B model."
@@ -1804,8 +1644,6 @@ var chutesModels = {
1804
1644
  contextWindow: 40960,
1805
1645
  supportsImages: false,
1806
1646
  supportsPromptCache: false,
1807
- supportsNativeTools: true,
1808
- defaultToolProtocol: "native",
1809
1647
  inputPrice: 0,
1810
1648
  outputPrice: 0,
1811
1649
  description: "Qwen3 32B model."
@@ -1815,8 +1653,6 @@ var chutesModels = {
1815
1653
  contextWindow: 40960,
1816
1654
  supportsImages: false,
1817
1655
  supportsPromptCache: false,
1818
- supportsNativeTools: true,
1819
- defaultToolProtocol: "native",
1820
1656
  inputPrice: 0,
1821
1657
  outputPrice: 0,
1822
1658
  description: "Qwen3 30B A3B model."
@@ -1826,8 +1662,6 @@ var chutesModels = {
1826
1662
  contextWindow: 40960,
1827
1663
  supportsImages: false,
1828
1664
  supportsPromptCache: false,
1829
- supportsNativeTools: true,
1830
- defaultToolProtocol: "native",
1831
1665
  inputPrice: 0,
1832
1666
  outputPrice: 0,
1833
1667
  description: "Qwen3 14B model."
@@ -1837,8 +1671,6 @@ var chutesModels = {
1837
1671
  contextWindow: 40960,
1838
1672
  supportsImages: false,
1839
1673
  supportsPromptCache: false,
1840
- supportsNativeTools: true,
1841
- defaultToolProtocol: "native",
1842
1674
  inputPrice: 0,
1843
1675
  outputPrice: 0,
1844
1676
  description: "Qwen3 8B model."
@@ -1848,8 +1680,6 @@ var chutesModels = {
1848
1680
  contextWindow: 163840,
1849
1681
  supportsImages: false,
1850
1682
  supportsPromptCache: false,
1851
- supportsNativeTools: true,
1852
- defaultToolProtocol: "native",
1853
1683
  inputPrice: 0,
1854
1684
  outputPrice: 0,
1855
1685
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -1859,8 +1689,6 @@ var chutesModels = {
1859
1689
  contextWindow: 163840,
1860
1690
  supportsImages: false,
1861
1691
  supportsPromptCache: false,
1862
- supportsNativeTools: true,
1863
- defaultToolProtocol: "native",
1864
1692
  inputPrice: 0,
1865
1693
  outputPrice: 0,
1866
1694
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -1870,8 +1698,6 @@ var chutesModels = {
1870
1698
  contextWindow: 151329,
1871
1699
  supportsImages: false,
1872
1700
  supportsPromptCache: false,
1873
- supportsNativeTools: true,
1874
- defaultToolProtocol: "native",
1875
1701
  inputPrice: 0,
1876
1702
  outputPrice: 0,
1877
1703
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -1881,8 +1707,6 @@ var chutesModels = {
1881
1707
  contextWindow: 131072,
1882
1708
  supportsImages: false,
1883
1709
  supportsPromptCache: false,
1884
- supportsNativeTools: true,
1885
- defaultToolProtocol: "native",
1886
1710
  inputPrice: 0,
1887
1711
  outputPrice: 0,
1888
1712
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -1892,8 +1716,6 @@ var chutesModels = {
1892
1716
  contextWindow: 131072,
1893
1717
  supportsImages: false,
1894
1718
  supportsPromptCache: false,
1895
- supportsNativeTools: true,
1896
- defaultToolProtocol: "native",
1897
1719
  inputPrice: 1,
1898
1720
  outputPrice: 3,
1899
1721
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -1903,8 +1725,6 @@ var chutesModels = {
1903
1725
  contextWindow: 202752,
1904
1726
  supportsImages: false,
1905
1727
  supportsPromptCache: false,
1906
- supportsNativeTools: true,
1907
- defaultToolProtocol: "native",
1908
1728
  inputPrice: 0,
1909
1729
  outputPrice: 0,
1910
1730
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -1915,8 +1735,6 @@ var chutesModels = {
1915
1735
  contextWindow: 202752,
1916
1736
  supportsImages: false,
1917
1737
  supportsPromptCache: false,
1918
- supportsNativeTools: true,
1919
- defaultToolProtocol: "native",
1920
1738
  inputPrice: 1.15,
1921
1739
  outputPrice: 3.25,
1922
1740
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -1926,8 +1744,6 @@ var chutesModels = {
1926
1744
  contextWindow: 128e3,
1927
1745
  supportsImages: false,
1928
1746
  supportsPromptCache: false,
1929
- supportsNativeTools: true,
1930
- defaultToolProtocol: "native",
1931
1747
  inputPrice: 0,
1932
1748
  outputPrice: 0,
1933
1749
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -1937,8 +1753,6 @@ var chutesModels = {
1937
1753
  contextWindow: 262144,
1938
1754
  supportsImages: false,
1939
1755
  supportsPromptCache: false,
1940
- supportsNativeTools: true,
1941
- defaultToolProtocol: "native",
1942
1756
  inputPrice: 0,
1943
1757
  outputPrice: 0,
1944
1758
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -1948,8 +1762,6 @@ var chutesModels = {
1948
1762
  contextWindow: 75e3,
1949
1763
  supportsImages: false,
1950
1764
  supportsPromptCache: false,
1951
- supportsNativeTools: true,
1952
- defaultToolProtocol: "native",
1953
1765
  inputPrice: 0.1481,
1954
1766
  outputPrice: 0.5926,
1955
1767
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -1959,8 +1771,6 @@ var chutesModels = {
1959
1771
  contextWindow: 262144,
1960
1772
  supportsImages: false,
1961
1773
  supportsPromptCache: false,
1962
- supportsNativeTools: true,
1963
- defaultToolProtocol: "native",
1964
1774
  inputPrice: 0.1999,
1965
1775
  outputPrice: 0.8001,
1966
1776
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -1970,8 +1780,6 @@ var chutesModels = {
1970
1780
  contextWindow: 262144,
1971
1781
  supportsImages: false,
1972
1782
  supportsPromptCache: false,
1973
- supportsNativeTools: true,
1974
- defaultToolProtocol: "native",
1975
1783
  inputPrice: 0.077968332,
1976
1784
  outputPrice: 0.31202496,
1977
1785
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -1981,8 +1789,6 @@ var chutesModels = {
1981
1789
  contextWindow: 131072,
1982
1790
  supportsImages: false,
1983
1791
  supportsPromptCache: false,
1984
- supportsNativeTools: true,
1985
- defaultToolProtocol: "native",
1986
1792
  inputPrice: 0,
1987
1793
  outputPrice: 0,
1988
1794
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -1992,8 +1798,6 @@ var chutesModels = {
1992
1798
  contextWindow: 131072,
1993
1799
  supportsImages: false,
1994
1800
  supportsPromptCache: false,
1995
- supportsNativeTools: true,
1996
- defaultToolProtocol: "native",
1997
1801
  inputPrice: 0,
1998
1802
  outputPrice: 0,
1999
1803
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -2003,8 +1807,6 @@ var chutesModels = {
2003
1807
  contextWindow: 262144,
2004
1808
  supportsImages: true,
2005
1809
  supportsPromptCache: false,
2006
- supportsNativeTools: true,
2007
- defaultToolProtocol: "native",
2008
1810
  inputPrice: 0.16,
2009
1811
  outputPrice: 0.65,
2010
1812
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -2012,73 +1814,6 @@ var chutesModels = {
2012
1814
  };
2013
1815
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
2014
1816
 
2015
- // src/providers/claude-code.ts
2016
- var DATE_SUFFIX_PATTERN = /-\d{8}$/;
2017
- var claudeCodeModels = {
2018
- "claude-haiku-4-5": {
2019
- maxTokens: 32768,
2020
- contextWindow: 2e5,
2021
- supportsImages: true,
2022
- supportsPromptCache: true,
2023
- supportsNativeTools: true,
2024
- defaultToolProtocol: "native",
2025
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2026
- reasoningEffort: "medium",
2027
- description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2028
- },
2029
- "claude-sonnet-4-5": {
2030
- maxTokens: 32768,
2031
- contextWindow: 2e5,
2032
- supportsImages: true,
2033
- supportsPromptCache: true,
2034
- supportsNativeTools: true,
2035
- defaultToolProtocol: "native",
2036
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2037
- reasoningEffort: "medium",
2038
- description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2039
- },
2040
- "claude-opus-4-5": {
2041
- maxTokens: 32768,
2042
- contextWindow: 2e5,
2043
- supportsImages: true,
2044
- supportsPromptCache: true,
2045
- supportsNativeTools: true,
2046
- defaultToolProtocol: "native",
2047
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2048
- reasoningEffort: "medium",
2049
- description: "Claude Opus 4.5 - Most capable with thinking"
2050
- }
2051
- };
2052
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2053
- var MODEL_FAMILY_PATTERNS = [
2054
- // Opus models (any version) → claude-opus-4-5
2055
- { pattern: /opus/i, target: "claude-opus-4-5" },
2056
- // Haiku models (any version) → claude-haiku-4-5
2057
- { pattern: /haiku/i, target: "claude-haiku-4-5" },
2058
- // Sonnet models (any version) → claude-sonnet-4-5
2059
- { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2060
- ];
2061
- function normalizeClaudeCodeModelId(modelId) {
2062
- if (Object.hasOwn(claudeCodeModels, modelId)) {
2063
- return modelId;
2064
- }
2065
- const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2066
- if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2067
- return withoutDate;
2068
- }
2069
- for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2070
- if (pattern.test(modelId)) {
2071
- return target;
2072
- }
2073
- }
2074
- return claudeCodeDefaultModelId;
2075
- }
2076
- var claudeCodeReasoningConfig = {
2077
- low: { budgetTokens: 16e3 },
2078
- medium: { budgetTokens: 32e3 },
2079
- high: { budgetTokens: 64e3 }
2080
- };
2081
-
2082
1817
  // src/providers/deepseek.ts
2083
1818
  var deepSeekDefaultModelId = "deepseek-chat";
2084
1819
  var deepSeekModels = {
@@ -2088,8 +1823,6 @@ var deepSeekModels = {
2088
1823
  contextWindow: 128e3,
2089
1824
  supportsImages: false,
2090
1825
  supportsPromptCache: true,
2091
- supportsNativeTools: true,
2092
- defaultToolProtocol: "native",
2093
1826
  inputPrice: 0.28,
2094
1827
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2095
1828
  outputPrice: 0.42,
@@ -2106,8 +1839,6 @@ var deepSeekModels = {
2106
1839
  contextWindow: 128e3,
2107
1840
  supportsImages: false,
2108
1841
  supportsPromptCache: true,
2109
- supportsNativeTools: true,
2110
- defaultToolProtocol: "native",
2111
1842
  preserveReasoning: true,
2112
1843
  inputPrice: 0.28,
2113
1844
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
@@ -2130,8 +1861,6 @@ var doubaoModels = {
2130
1861
  contextWindow: 128e3,
2131
1862
  supportsImages: true,
2132
1863
  supportsPromptCache: true,
2133
- supportsNativeTools: true,
2134
- defaultToolProtocol: "native",
2135
1864
  inputPrice: 1e-4,
2136
1865
  // $0.0001 per million tokens (cache miss)
2137
1866
  outputPrice: 4e-4,
@@ -2147,8 +1876,6 @@ var doubaoModels = {
2147
1876
  contextWindow: 128e3,
2148
1877
  supportsImages: true,
2149
1878
  supportsPromptCache: true,
2150
- supportsNativeTools: true,
2151
- defaultToolProtocol: "native",
2152
1879
  inputPrice: 2e-4,
2153
1880
  // $0.0002 per million tokens
2154
1881
  outputPrice: 8e-4,
@@ -2164,8 +1891,6 @@ var doubaoModels = {
2164
1891
  contextWindow: 128e3,
2165
1892
  supportsImages: true,
2166
1893
  supportsPromptCache: true,
2167
- supportsNativeTools: true,
2168
- defaultToolProtocol: "native",
2169
1894
  inputPrice: 15e-5,
2170
1895
  // $0.00015 per million tokens
2171
1896
  outputPrice: 6e-4,
@@ -2188,7 +1913,6 @@ var featherlessModels = {
2188
1913
  contextWindow: 32678,
2189
1914
  supportsImages: false,
2190
1915
  supportsPromptCache: false,
2191
- supportsNativeTools: true,
2192
1916
  inputPrice: 0,
2193
1917
  outputPrice: 0,
2194
1918
  description: "DeepSeek V3 0324 model."
@@ -2198,7 +1922,6 @@ var featherlessModels = {
2198
1922
  contextWindow: 32678,
2199
1923
  supportsImages: false,
2200
1924
  supportsPromptCache: false,
2201
- supportsNativeTools: true,
2202
1925
  inputPrice: 0,
2203
1926
  outputPrice: 0,
2204
1927
  description: "DeepSeek R1 0528 model."
@@ -2208,7 +1931,6 @@ var featherlessModels = {
2208
1931
  contextWindow: 32678,
2209
1932
  supportsImages: false,
2210
1933
  supportsPromptCache: false,
2211
- supportsNativeTools: true,
2212
1934
  inputPrice: 0,
2213
1935
  outputPrice: 0,
2214
1936
  description: "Kimi K2 Instruct model."
@@ -2218,7 +1940,6 @@ var featherlessModels = {
2218
1940
  contextWindow: 32678,
2219
1941
  supportsImages: false,
2220
1942
  supportsPromptCache: false,
2221
- supportsNativeTools: true,
2222
1943
  inputPrice: 0,
2223
1944
  outputPrice: 0,
2224
1945
  description: "GPT-OSS 120B model."
@@ -2228,7 +1949,6 @@ var featherlessModels = {
2228
1949
  contextWindow: 32678,
2229
1950
  supportsImages: false,
2230
1951
  supportsPromptCache: false,
2231
- supportsNativeTools: true,
2232
1952
  inputPrice: 0,
2233
1953
  outputPrice: 0,
2234
1954
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -2244,8 +1964,6 @@ var fireworksModels = {
2244
1964
  contextWindow: 262144,
2245
1965
  supportsImages: false,
2246
1966
  supportsPromptCache: true,
2247
- supportsNativeTools: true,
2248
- defaultToolProtocol: "native",
2249
1967
  inputPrice: 0.6,
2250
1968
  outputPrice: 2.5,
2251
1969
  cacheReadsPrice: 0.15,
@@ -2256,8 +1974,6 @@ var fireworksModels = {
2256
1974
  contextWindow: 128e3,
2257
1975
  supportsImages: false,
2258
1976
  supportsPromptCache: false,
2259
- supportsNativeTools: true,
2260
- defaultToolProtocol: "native",
2261
1977
  inputPrice: 0.6,
2262
1978
  outputPrice: 2.5,
2263
1979
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2267,7 +1983,6 @@ var fireworksModels = {
2267
1983
  contextWindow: 256e3,
2268
1984
  supportsImages: false,
2269
1985
  supportsPromptCache: true,
2270
- supportsNativeTools: true,
2271
1986
  supportsTemperature: true,
2272
1987
  preserveReasoning: true,
2273
1988
  defaultTemperature: 1,
@@ -2281,8 +1996,6 @@ var fireworksModels = {
2281
1996
  contextWindow: 204800,
2282
1997
  supportsImages: false,
2283
1998
  supportsPromptCache: false,
2284
- supportsNativeTools: true,
2285
- defaultToolProtocol: "native",
2286
1999
  inputPrice: 0.3,
2287
2000
  outputPrice: 1.2,
2288
2001
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2292,8 +2005,6 @@ var fireworksModels = {
2292
2005
  contextWindow: 256e3,
2293
2006
  supportsImages: false,
2294
2007
  supportsPromptCache: false,
2295
- supportsNativeTools: true,
2296
- defaultToolProtocol: "native",
2297
2008
  inputPrice: 0.22,
2298
2009
  outputPrice: 0.88,
2299
2010
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2303,8 +2014,6 @@ var fireworksModels = {
2303
2014
  contextWindow: 256e3,
2304
2015
  supportsImages: false,
2305
2016
  supportsPromptCache: false,
2306
- supportsNativeTools: true,
2307
- defaultToolProtocol: "native",
2308
2017
  inputPrice: 0.45,
2309
2018
  outputPrice: 1.8,
2310
2019
  description: "Qwen3's most agentic code model to date."
@@ -2314,8 +2023,6 @@ var fireworksModels = {
2314
2023
  contextWindow: 16e4,
2315
2024
  supportsImages: false,
2316
2025
  supportsPromptCache: false,
2317
- supportsNativeTools: true,
2318
- defaultToolProtocol: "native",
2319
2026
  inputPrice: 3,
2320
2027
  outputPrice: 8,
2321
2028
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2325,8 +2032,6 @@ var fireworksModels = {
2325
2032
  contextWindow: 128e3,
2326
2033
  supportsImages: false,
2327
2034
  supportsPromptCache: false,
2328
- supportsNativeTools: true,
2329
- defaultToolProtocol: "native",
2330
2035
  inputPrice: 0.9,
2331
2036
  outputPrice: 0.9,
2332
2037
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2336,8 +2041,6 @@ var fireworksModels = {
2336
2041
  contextWindow: 163840,
2337
2042
  supportsImages: false,
2338
2043
  supportsPromptCache: false,
2339
- supportsNativeTools: true,
2340
- defaultToolProtocol: "native",
2341
2044
  inputPrice: 0.56,
2342
2045
  outputPrice: 1.68,
2343
2046
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2347,8 +2050,6 @@ var fireworksModels = {
2347
2050
  contextWindow: 128e3,
2348
2051
  supportsImages: false,
2349
2052
  supportsPromptCache: false,
2350
- supportsNativeTools: true,
2351
- defaultToolProtocol: "native",
2352
2053
  inputPrice: 0.55,
2353
2054
  outputPrice: 2.19,
2354
2055
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2358,8 +2059,6 @@ var fireworksModels = {
2358
2059
  contextWindow: 128e3,
2359
2060
  supportsImages: false,
2360
2061
  supportsPromptCache: false,
2361
- supportsNativeTools: true,
2362
- defaultToolProtocol: "native",
2363
2062
  inputPrice: 0.55,
2364
2063
  outputPrice: 2.19,
2365
2064
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2369,8 +2068,6 @@ var fireworksModels = {
2369
2068
  contextWindow: 198e3,
2370
2069
  supportsImages: false,
2371
2070
  supportsPromptCache: false,
2372
- supportsNativeTools: true,
2373
- defaultToolProtocol: "native",
2374
2071
  inputPrice: 0.55,
2375
2072
  outputPrice: 2.19,
2376
2073
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2380,8 +2077,6 @@ var fireworksModels = {
2380
2077
  contextWindow: 128e3,
2381
2078
  supportsImages: false,
2382
2079
  supportsPromptCache: false,
2383
- supportsNativeTools: true,
2384
- defaultToolProtocol: "native",
2385
2080
  inputPrice: 0.07,
2386
2081
  outputPrice: 0.3,
2387
2082
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2391,11 +2086,63 @@ var fireworksModels = {
2391
2086
  contextWindow: 128e3,
2392
2087
  supportsImages: false,
2393
2088
  supportsPromptCache: false,
2394
- supportsNativeTools: true,
2395
- defaultToolProtocol: "native",
2396
2089
  inputPrice: 0.15,
2397
2090
  outputPrice: 0.6,
2398
2091
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
2092
+ },
2093
+ "accounts/fireworks/models/minimax-m2p1": {
2094
+ maxTokens: 4096,
2095
+ contextWindow: 204800,
2096
+ supportsImages: false,
2097
+ supportsPromptCache: false,
2098
+ inputPrice: 0.3,
2099
+ outputPrice: 1.2,
2100
+ description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
2101
+ },
2102
+ "accounts/fireworks/models/deepseek-v3p2": {
2103
+ maxTokens: 16384,
2104
+ contextWindow: 163840,
2105
+ supportsImages: false,
2106
+ supportsPromptCache: false,
2107
+ inputPrice: 0.56,
2108
+ outputPrice: 1.68,
2109
+ description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
2110
+ },
2111
+ "accounts/fireworks/models/glm-4p7": {
2112
+ maxTokens: 25344,
2113
+ contextWindow: 198e3,
2114
+ supportsImages: false,
2115
+ supportsPromptCache: false,
2116
+ inputPrice: 0.55,
2117
+ outputPrice: 2.19,
2118
+ description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
2119
+ },
2120
+ "accounts/fireworks/models/llama-v3p3-70b-instruct": {
2121
+ maxTokens: 16384,
2122
+ contextWindow: 131072,
2123
+ supportsImages: false,
2124
+ supportsPromptCache: false,
2125
+ inputPrice: 0.9,
2126
+ outputPrice: 0.9,
2127
+ description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
2128
+ },
2129
+ "accounts/fireworks/models/llama4-maverick-instruct-basic": {
2130
+ maxTokens: 16384,
2131
+ contextWindow: 131072,
2132
+ supportsImages: true,
2133
+ supportsPromptCache: false,
2134
+ inputPrice: 0.22,
2135
+ outputPrice: 0.88,
2136
+ description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
2137
+ },
2138
+ "accounts/fireworks/models/llama4-scout-instruct-basic": {
2139
+ maxTokens: 16384,
2140
+ contextWindow: 131072,
2141
+ supportsImages: true,
2142
+ supportsPromptCache: false,
2143
+ inputPrice: 0.15,
2144
+ outputPrice: 0.6,
2145
+ description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
2399
2146
  }
2400
2147
  };
2401
2148
 
@@ -2406,8 +2153,6 @@ var geminiModels = {
2406
2153
  maxTokens: 65536,
2407
2154
  contextWindow: 1048576,
2408
2155
  supportsImages: true,
2409
- supportsNativeTools: true,
2410
- defaultToolProtocol: "native",
2411
2156
  supportsPromptCache: true,
2412
2157
  supportsReasoningEffort: ["low", "high"],
2413
2158
  reasoningEffort: "low",
@@ -2415,16 +2160,19 @@ var geminiModels = {
2415
2160
  defaultTemperature: 1,
2416
2161
  inputPrice: 4,
2417
2162
  outputPrice: 18,
2163
+ cacheReadsPrice: 0.4,
2418
2164
  tiers: [
2419
2165
  {
2420
2166
  contextWindow: 2e5,
2421
2167
  inputPrice: 2,
2422
- outputPrice: 12
2168
+ outputPrice: 12,
2169
+ cacheReadsPrice: 0.2
2423
2170
  },
2424
2171
  {
2425
2172
  contextWindow: Infinity,
2426
2173
  inputPrice: 4,
2427
- outputPrice: 18
2174
+ outputPrice: 18,
2175
+ cacheReadsPrice: 0.4
2428
2176
  }
2429
2177
  ]
2430
2178
  },
@@ -2432,25 +2180,20 @@ var geminiModels = {
2432
2180
  maxTokens: 65536,
2433
2181
  contextWindow: 1048576,
2434
2182
  supportsImages: true,
2435
- supportsNativeTools: true,
2436
- defaultToolProtocol: "native",
2437
2183
  supportsPromptCache: true,
2438
2184
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2439
2185
  reasoningEffort: "medium",
2440
2186
  supportsTemperature: true,
2441
2187
  defaultTemperature: 1,
2442
- inputPrice: 0.3,
2443
- outputPrice: 2.5,
2444
- cacheReadsPrice: 0.075,
2445
- cacheWritesPrice: 1
2188
+ inputPrice: 0.5,
2189
+ outputPrice: 3,
2190
+ cacheReadsPrice: 0.05
2446
2191
  },
2447
2192
  // 2.5 Pro models
2448
2193
  "gemini-2.5-pro": {
2449
2194
  maxTokens: 64e3,
2450
2195
  contextWindow: 1048576,
2451
2196
  supportsImages: true,
2452
- supportsNativeTools: true,
2453
- defaultToolProtocol: "native",
2454
2197
  supportsPromptCache: true,
2455
2198
  inputPrice: 2.5,
2456
2199
  // This is the pricing for prompts above 200k tokens.
@@ -2479,8 +2222,6 @@ var geminiModels = {
2479
2222
  maxTokens: 65535,
2480
2223
  contextWindow: 1048576,
2481
2224
  supportsImages: true,
2482
- supportsNativeTools: true,
2483
- defaultToolProtocol: "native",
2484
2225
  supportsPromptCache: true,
2485
2226
  inputPrice: 2.5,
2486
2227
  // This is the pricing for prompts above 200k tokens.
@@ -2508,8 +2249,6 @@ var geminiModels = {
2508
2249
  maxTokens: 65535,
2509
2250
  contextWindow: 1048576,
2510
2251
  supportsImages: true,
2511
- supportsNativeTools: true,
2512
- defaultToolProtocol: "native",
2513
2252
  supportsPromptCache: true,
2514
2253
  inputPrice: 2.5,
2515
2254
  // This is the pricing for prompts above 200k tokens.
@@ -2535,8 +2274,6 @@ var geminiModels = {
2535
2274
  maxTokens: 65535,
2536
2275
  contextWindow: 1048576,
2537
2276
  supportsImages: true,
2538
- supportsNativeTools: true,
2539
- defaultToolProtocol: "native",
2540
2277
  supportsPromptCache: true,
2541
2278
  inputPrice: 2.5,
2542
2279
  // This is the pricing for prompts above 200k tokens.
@@ -2565,8 +2302,6 @@ var geminiModels = {
2565
2302
  maxTokens: 65536,
2566
2303
  contextWindow: 1048576,
2567
2304
  supportsImages: true,
2568
- supportsNativeTools: true,
2569
- defaultToolProtocol: "native",
2570
2305
  supportsPromptCache: true,
2571
2306
  inputPrice: 0.3,
2572
2307
  outputPrice: 2.5,
@@ -2579,8 +2314,6 @@ var geminiModels = {
2579
2314
  maxTokens: 65536,
2580
2315
  contextWindow: 1048576,
2581
2316
  supportsImages: true,
2582
- supportsNativeTools: true,
2583
- defaultToolProtocol: "native",
2584
2317
  supportsPromptCache: true,
2585
2318
  inputPrice: 0.3,
2586
2319
  outputPrice: 2.5,
@@ -2593,8 +2326,6 @@ var geminiModels = {
2593
2326
  maxTokens: 64e3,
2594
2327
  contextWindow: 1048576,
2595
2328
  supportsImages: true,
2596
- supportsNativeTools: true,
2597
- defaultToolProtocol: "native",
2598
2329
  supportsPromptCache: true,
2599
2330
  inputPrice: 0.3,
2600
2331
  outputPrice: 2.5,
@@ -2608,8 +2339,6 @@ var geminiModels = {
2608
2339
  maxTokens: 65536,
2609
2340
  contextWindow: 1048576,
2610
2341
  supportsImages: true,
2611
- supportsNativeTools: true,
2612
- defaultToolProtocol: "native",
2613
2342
  supportsPromptCache: true,
2614
2343
  inputPrice: 0.1,
2615
2344
  outputPrice: 0.4,
@@ -2622,8 +2351,6 @@ var geminiModels = {
2622
2351
  maxTokens: 65536,
2623
2352
  contextWindow: 1048576,
2624
2353
  supportsImages: true,
2625
- supportsNativeTools: true,
2626
- defaultToolProtocol: "native",
2627
2354
  supportsPromptCache: true,
2628
2355
  inputPrice: 0.1,
2629
2356
  outputPrice: 0.4,
@@ -2643,8 +2370,6 @@ var groqModels = {
2643
2370
  contextWindow: 131072,
2644
2371
  supportsImages: false,
2645
2372
  supportsPromptCache: false,
2646
- supportsNativeTools: true,
2647
- defaultToolProtocol: "native",
2648
2373
  inputPrice: 0.05,
2649
2374
  outputPrice: 0.08,
2650
2375
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2654,8 +2379,6 @@ var groqModels = {
2654
2379
  contextWindow: 131072,
2655
2380
  supportsImages: false,
2656
2381
  supportsPromptCache: false,
2657
- supportsNativeTools: true,
2658
- defaultToolProtocol: "native",
2659
2382
  inputPrice: 0.59,
2660
2383
  outputPrice: 0.79,
2661
2384
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2665,8 +2388,6 @@ var groqModels = {
2665
2388
  contextWindow: 131072,
2666
2389
  supportsImages: false,
2667
2390
  supportsPromptCache: false,
2668
- supportsNativeTools: true,
2669
- defaultToolProtocol: "native",
2670
2391
  inputPrice: 0.11,
2671
2392
  outputPrice: 0.34,
2672
2393
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2676,8 +2397,6 @@ var groqModels = {
2676
2397
  contextWindow: 131072,
2677
2398
  supportsImages: false,
2678
2399
  supportsPromptCache: false,
2679
- supportsNativeTools: true,
2680
- defaultToolProtocol: "native",
2681
2400
  inputPrice: 0.29,
2682
2401
  outputPrice: 0.59,
2683
2402
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2687,8 +2406,6 @@ var groqModels = {
2687
2406
  contextWindow: 262144,
2688
2407
  supportsImages: false,
2689
2408
  supportsPromptCache: true,
2690
- supportsNativeTools: true,
2691
- defaultToolProtocol: "native",
2692
2409
  inputPrice: 0.6,
2693
2410
  outputPrice: 2.5,
2694
2411
  cacheReadsPrice: 0.15,
@@ -2699,8 +2416,6 @@ var groqModels = {
2699
2416
  contextWindow: 131072,
2700
2417
  supportsImages: false,
2701
2418
  supportsPromptCache: false,
2702
- supportsNativeTools: true,
2703
- defaultToolProtocol: "native",
2704
2419
  inputPrice: 0.15,
2705
2420
  outputPrice: 0.75,
2706
2421
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2710,8 +2425,6 @@ var groqModels = {
2710
2425
  contextWindow: 131072,
2711
2426
  supportsImages: false,
2712
2427
  supportsPromptCache: false,
2713
- supportsNativeTools: true,
2714
- defaultToolProtocol: "native",
2715
2428
  inputPrice: 0.1,
2716
2429
  outputPrice: 0.5,
2717
2430
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2738,7 +2451,6 @@ var ioIntelligenceModels = {
2738
2451
  contextWindow: 128e3,
2739
2452
  supportsImages: false,
2740
2453
  supportsPromptCache: false,
2741
- supportsNativeTools: true,
2742
2454
  description: "DeepSeek R1 reasoning model"
2743
2455
  },
2744
2456
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -2746,7 +2458,6 @@ var ioIntelligenceModels = {
2746
2458
  contextWindow: 43e4,
2747
2459
  supportsImages: true,
2748
2460
  supportsPromptCache: false,
2749
- supportsNativeTools: true,
2750
2461
  description: "Llama 4 Maverick 17B model"
2751
2462
  },
2752
2463
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -2754,7 +2465,6 @@ var ioIntelligenceModels = {
2754
2465
  contextWindow: 106e3,
2755
2466
  supportsImages: false,
2756
2467
  supportsPromptCache: false,
2757
- supportsNativeTools: true,
2758
2468
  description: "Qwen3 Coder 480B specialized for coding"
2759
2469
  },
2760
2470
  "openai/gpt-oss-120b": {
@@ -2762,7 +2472,6 @@ var ioIntelligenceModels = {
2762
2472
  contextWindow: 131072,
2763
2473
  supportsImages: false,
2764
2474
  supportsPromptCache: false,
2765
- supportsNativeTools: true,
2766
2475
  description: "OpenAI GPT-OSS 120B model"
2767
2476
  }
2768
2477
  };
@@ -2774,8 +2483,6 @@ var litellmDefaultModelInfo = {
2774
2483
  contextWindow: 2e5,
2775
2484
  supportsImages: true,
2776
2485
  supportsPromptCache: true,
2777
- supportsNativeTools: true,
2778
- defaultToolProtocol: "native",
2779
2486
  inputPrice: 3,
2780
2487
  outputPrice: 15,
2781
2488
  cacheWritesPrice: 3.75,
@@ -2790,8 +2497,6 @@ var lMStudioDefaultModelInfo = {
2790
2497
  contextWindow: 2e5,
2791
2498
  supportsImages: true,
2792
2499
  supportsPromptCache: true,
2793
- supportsNativeTools: true,
2794
- defaultToolProtocol: "native",
2795
2500
  inputPrice: 0,
2796
2501
  outputPrice: 0,
2797
2502
  cacheWritesPrice: 0,
@@ -2807,8 +2512,6 @@ var mistralModels = {
2807
2512
  contextWindow: 128e3,
2808
2513
  supportsImages: true,
2809
2514
  supportsPromptCache: false,
2810
- supportsNativeTools: true,
2811
- defaultToolProtocol: "native",
2812
2515
  inputPrice: 2,
2813
2516
  outputPrice: 5
2814
2517
  },
@@ -2817,8 +2520,6 @@ var mistralModels = {
2817
2520
  contextWindow: 131e3,
2818
2521
  supportsImages: true,
2819
2522
  supportsPromptCache: false,
2820
- supportsNativeTools: true,
2821
- defaultToolProtocol: "native",
2822
2523
  inputPrice: 0.4,
2823
2524
  outputPrice: 2
2824
2525
  },
@@ -2827,8 +2528,6 @@ var mistralModels = {
2827
2528
  contextWindow: 131e3,
2828
2529
  supportsImages: true,
2829
2530
  supportsPromptCache: false,
2830
- supportsNativeTools: true,
2831
- defaultToolProtocol: "native",
2832
2531
  inputPrice: 0.4,
2833
2532
  outputPrice: 2
2834
2533
  },
@@ -2837,8 +2536,6 @@ var mistralModels = {
2837
2536
  contextWindow: 256e3,
2838
2537
  supportsImages: false,
2839
2538
  supportsPromptCache: false,
2840
- supportsNativeTools: true,
2841
- defaultToolProtocol: "native",
2842
2539
  inputPrice: 0.3,
2843
2540
  outputPrice: 0.9
2844
2541
  },
@@ -2847,8 +2544,6 @@ var mistralModels = {
2847
2544
  contextWindow: 131e3,
2848
2545
  supportsImages: false,
2849
2546
  supportsPromptCache: false,
2850
- supportsNativeTools: true,
2851
- defaultToolProtocol: "native",
2852
2547
  inputPrice: 2,
2853
2548
  outputPrice: 6
2854
2549
  },
@@ -2857,8 +2552,6 @@ var mistralModels = {
2857
2552
  contextWindow: 131e3,
2858
2553
  supportsImages: false,
2859
2554
  supportsPromptCache: false,
2860
- supportsNativeTools: true,
2861
- defaultToolProtocol: "native",
2862
2555
  inputPrice: 0.1,
2863
2556
  outputPrice: 0.1
2864
2557
  },
@@ -2867,8 +2560,6 @@ var mistralModels = {
2867
2560
  contextWindow: 131e3,
2868
2561
  supportsImages: false,
2869
2562
  supportsPromptCache: false,
2870
- supportsNativeTools: true,
2871
- defaultToolProtocol: "native",
2872
2563
  inputPrice: 0.04,
2873
2564
  outputPrice: 0.04
2874
2565
  },
@@ -2877,8 +2568,6 @@ var mistralModels = {
2877
2568
  contextWindow: 32e3,
2878
2569
  supportsImages: false,
2879
2570
  supportsPromptCache: false,
2880
- supportsNativeTools: true,
2881
- defaultToolProtocol: "native",
2882
2571
  inputPrice: 0.2,
2883
2572
  outputPrice: 0.6
2884
2573
  },
@@ -2887,8 +2576,6 @@ var mistralModels = {
2887
2576
  contextWindow: 131e3,
2888
2577
  supportsImages: true,
2889
2578
  supportsPromptCache: false,
2890
- supportsNativeTools: true,
2891
- defaultToolProtocol: "native",
2892
2579
  inputPrice: 2,
2893
2580
  outputPrice: 6
2894
2581
  }
@@ -2903,8 +2590,6 @@ var moonshotModels = {
2903
2590
  contextWindow: 131072,
2904
2591
  supportsImages: false,
2905
2592
  supportsPromptCache: true,
2906
- supportsNativeTools: true,
2907
- defaultToolProtocol: "native",
2908
2593
  inputPrice: 0.6,
2909
2594
  // $0.60 per million tokens (cache miss)
2910
2595
  outputPrice: 2.5,
@@ -2920,8 +2605,6 @@ var moonshotModels = {
2920
2605
  contextWindow: 262144,
2921
2606
  supportsImages: false,
2922
2607
  supportsPromptCache: true,
2923
- supportsNativeTools: true,
2924
- defaultToolProtocol: "native",
2925
2608
  inputPrice: 0.6,
2926
2609
  outputPrice: 2.5,
2927
2610
  cacheReadsPrice: 0.15,
@@ -2932,8 +2615,6 @@ var moonshotModels = {
2932
2615
  contextWindow: 262144,
2933
2616
  supportsImages: false,
2934
2617
  supportsPromptCache: true,
2935
- supportsNativeTools: true,
2936
- defaultToolProtocol: "native",
2937
2618
  inputPrice: 2.4,
2938
2619
  // $2.40 per million tokens (cache miss)
2939
2620
  outputPrice: 10,
@@ -2952,8 +2633,6 @@ var moonshotModels = {
2952
2633
  supportsImages: false,
2953
2634
  // Text-only (no image/vision support)
2954
2635
  supportsPromptCache: true,
2955
- supportsNativeTools: true,
2956
- defaultToolProtocol: "native",
2957
2636
  inputPrice: 0.6,
2958
2637
  // $0.60 per million tokens (cache miss)
2959
2638
  outputPrice: 2.5,
@@ -2967,6 +2646,21 @@ var moonshotModels = {
2967
2646
  preserveReasoning: true,
2968
2647
  defaultTemperature: 1,
2969
2648
  description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
2649
+ },
2650
+ "kimi-k2.5": {
2651
+ maxTokens: 16384,
2652
+ contextWindow: 262144,
2653
+ supportsImages: false,
2654
+ supportsPromptCache: true,
2655
+ inputPrice: 0.6,
2656
+ // $0.60 per million tokens (cache miss)
2657
+ outputPrice: 3,
2658
+ // $3.00 per million tokens
2659
+ cacheReadsPrice: 0.1,
2660
+ // $0.10 per million tokens (cache hit)
2661
+ supportsTemperature: true,
2662
+ defaultTemperature: 1,
2663
+ description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
2970
2664
  }
2971
2665
  };
2972
2666
  var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
@@ -2978,7 +2672,6 @@ var ollamaDefaultModelInfo = {
2978
2672
  contextWindow: 2e5,
2979
2673
  supportsImages: true,
2980
2674
  supportsPromptCache: true,
2981
- supportsNativeTools: true,
2982
2675
  inputPrice: 0,
2983
2676
  outputPrice: 0,
2984
2677
  cacheWritesPrice: 0,
@@ -2992,8 +2685,6 @@ var openAiNativeModels = {
2992
2685
  "gpt-5.1-codex-max": {
2993
2686
  maxTokens: 128e3,
2994
2687
  contextWindow: 4e5,
2995
- supportsNativeTools: true,
2996
- defaultToolProtocol: "native",
2997
2688
  includedTools: ["apply_patch"],
2998
2689
  excludedTools: ["apply_diff", "write_to_file"],
2999
2690
  supportsImages: true,
@@ -3011,8 +2702,6 @@ var openAiNativeModels = {
3011
2702
  "gpt-5.2": {
3012
2703
  maxTokens: 128e3,
3013
2704
  contextWindow: 4e5,
3014
- supportsNativeTools: true,
3015
- defaultToolProtocol: "native",
3016
2705
  includedTools: ["apply_patch"],
3017
2706
  excludedTools: ["apply_diff", "write_to_file"],
3018
2707
  supportsImages: true,
@@ -3034,8 +2723,6 @@ var openAiNativeModels = {
3034
2723
  "gpt-5.2-codex": {
3035
2724
  maxTokens: 128e3,
3036
2725
  contextWindow: 4e5,
3037
- supportsNativeTools: true,
3038
- defaultToolProtocol: "native",
3039
2726
  includedTools: ["apply_patch"],
3040
2727
  excludedTools: ["apply_diff", "write_to_file"],
3041
2728
  supportsImages: true,
@@ -3053,8 +2740,6 @@ var openAiNativeModels = {
3053
2740
  "gpt-5.2-chat-latest": {
3054
2741
  maxTokens: 16384,
3055
2742
  contextWindow: 128e3,
3056
- supportsNativeTools: true,
3057
- defaultToolProtocol: "native",
3058
2743
  includedTools: ["apply_patch"],
3059
2744
  excludedTools: ["apply_diff", "write_to_file"],
3060
2745
  supportsImages: true,
@@ -3067,8 +2752,6 @@ var openAiNativeModels = {
3067
2752
  "gpt-5.1": {
3068
2753
  maxTokens: 128e3,
3069
2754
  contextWindow: 4e5,
3070
- supportsNativeTools: true,
3071
- defaultToolProtocol: "native",
3072
2755
  includedTools: ["apply_patch"],
3073
2756
  excludedTools: ["apply_diff", "write_to_file"],
3074
2757
  supportsImages: true,
@@ -3090,8 +2773,6 @@ var openAiNativeModels = {
3090
2773
  "gpt-5.1-codex": {
3091
2774
  maxTokens: 128e3,
3092
2775
  contextWindow: 4e5,
3093
- supportsNativeTools: true,
3094
- defaultToolProtocol: "native",
3095
2776
  includedTools: ["apply_patch"],
3096
2777
  excludedTools: ["apply_diff", "write_to_file"],
3097
2778
  supportsImages: true,
@@ -3109,8 +2790,6 @@ var openAiNativeModels = {
3109
2790
  "gpt-5.1-codex-mini": {
3110
2791
  maxTokens: 128e3,
3111
2792
  contextWindow: 4e5,
3112
- supportsNativeTools: true,
3113
- defaultToolProtocol: "native",
3114
2793
  includedTools: ["apply_patch"],
3115
2794
  excludedTools: ["apply_diff", "write_to_file"],
3116
2795
  supportsImages: true,
@@ -3127,8 +2806,6 @@ var openAiNativeModels = {
3127
2806
  "gpt-5": {
3128
2807
  maxTokens: 128e3,
3129
2808
  contextWindow: 4e5,
3130
- supportsNativeTools: true,
3131
- defaultToolProtocol: "native",
3132
2809
  includedTools: ["apply_patch"],
3133
2810
  excludedTools: ["apply_diff", "write_to_file"],
3134
2811
  supportsImages: true,
@@ -3149,8 +2826,6 @@ var openAiNativeModels = {
3149
2826
  "gpt-5-mini": {
3150
2827
  maxTokens: 128e3,
3151
2828
  contextWindow: 4e5,
3152
- supportsNativeTools: true,
3153
- defaultToolProtocol: "native",
3154
2829
  includedTools: ["apply_patch"],
3155
2830
  excludedTools: ["apply_diff", "write_to_file"],
3156
2831
  supportsImages: true,
@@ -3171,8 +2846,6 @@ var openAiNativeModels = {
3171
2846
  "gpt-5-codex": {
3172
2847
  maxTokens: 128e3,
3173
2848
  contextWindow: 4e5,
3174
- supportsNativeTools: true,
3175
- defaultToolProtocol: "native",
3176
2849
  includedTools: ["apply_patch"],
3177
2850
  excludedTools: ["apply_diff", "write_to_file"],
3178
2851
  supportsImages: true,
@@ -3189,8 +2862,6 @@ var openAiNativeModels = {
3189
2862
  "gpt-5-nano": {
3190
2863
  maxTokens: 128e3,
3191
2864
  contextWindow: 4e5,
3192
- supportsNativeTools: true,
3193
- defaultToolProtocol: "native",
3194
2865
  includedTools: ["apply_patch"],
3195
2866
  excludedTools: ["apply_diff", "write_to_file"],
3196
2867
  supportsImages: true,
@@ -3208,8 +2879,6 @@ var openAiNativeModels = {
3208
2879
  "gpt-5-chat-latest": {
3209
2880
  maxTokens: 128e3,
3210
2881
  contextWindow: 4e5,
3211
- supportsNativeTools: true,
3212
- defaultToolProtocol: "native",
3213
2882
  includedTools: ["apply_patch"],
3214
2883
  excludedTools: ["apply_diff", "write_to_file"],
3215
2884
  supportsImages: true,
@@ -3222,8 +2891,6 @@ var openAiNativeModels = {
3222
2891
  "gpt-4.1": {
3223
2892
  maxTokens: 32768,
3224
2893
  contextWindow: 1047576,
3225
- supportsNativeTools: true,
3226
- defaultToolProtocol: "native",
3227
2894
  includedTools: ["apply_patch"],
3228
2895
  excludedTools: ["apply_diff", "write_to_file"],
3229
2896
  supportsImages: true,
@@ -3239,8 +2906,6 @@ var openAiNativeModels = {
3239
2906
  "gpt-4.1-mini": {
3240
2907
  maxTokens: 32768,
3241
2908
  contextWindow: 1047576,
3242
- supportsNativeTools: true,
3243
- defaultToolProtocol: "native",
3244
2909
  includedTools: ["apply_patch"],
3245
2910
  excludedTools: ["apply_diff", "write_to_file"],
3246
2911
  supportsImages: true,
@@ -3256,8 +2921,6 @@ var openAiNativeModels = {
3256
2921
  "gpt-4.1-nano": {
3257
2922
  maxTokens: 32768,
3258
2923
  contextWindow: 1047576,
3259
- supportsNativeTools: true,
3260
- defaultToolProtocol: "native",
3261
2924
  includedTools: ["apply_patch"],
3262
2925
  excludedTools: ["apply_diff", "write_to_file"],
3263
2926
  supportsImages: true,
@@ -3273,8 +2936,6 @@ var openAiNativeModels = {
3273
2936
  o3: {
3274
2937
  maxTokens: 1e5,
3275
2938
  contextWindow: 2e5,
3276
- supportsNativeTools: true,
3277
- defaultToolProtocol: "native",
3278
2939
  supportsImages: true,
3279
2940
  supportsPromptCache: true,
3280
2941
  inputPrice: 2,
@@ -3291,8 +2952,6 @@ var openAiNativeModels = {
3291
2952
  "o3-high": {
3292
2953
  maxTokens: 1e5,
3293
2954
  contextWindow: 2e5,
3294
- supportsNativeTools: true,
3295
- defaultToolProtocol: "native",
3296
2955
  supportsImages: true,
3297
2956
  supportsPromptCache: true,
3298
2957
  inputPrice: 2,
@@ -3304,8 +2963,6 @@ var openAiNativeModels = {
3304
2963
  "o3-low": {
3305
2964
  maxTokens: 1e5,
3306
2965
  contextWindow: 2e5,
3307
- supportsNativeTools: true,
3308
- defaultToolProtocol: "native",
3309
2966
  supportsImages: true,
3310
2967
  supportsPromptCache: true,
3311
2968
  inputPrice: 2,
@@ -3317,8 +2974,6 @@ var openAiNativeModels = {
3317
2974
  "o4-mini": {
3318
2975
  maxTokens: 1e5,
3319
2976
  contextWindow: 2e5,
3320
- supportsNativeTools: true,
3321
- defaultToolProtocol: "native",
3322
2977
  supportsImages: true,
3323
2978
  supportsPromptCache: true,
3324
2979
  inputPrice: 1.1,
@@ -3335,8 +2990,6 @@ var openAiNativeModels = {
3335
2990
  "o4-mini-high": {
3336
2991
  maxTokens: 1e5,
3337
2992
  contextWindow: 2e5,
3338
- supportsNativeTools: true,
3339
- defaultToolProtocol: "native",
3340
2993
  supportsImages: true,
3341
2994
  supportsPromptCache: true,
3342
2995
  inputPrice: 1.1,
@@ -3348,8 +3001,6 @@ var openAiNativeModels = {
3348
3001
  "o4-mini-low": {
3349
3002
  maxTokens: 1e5,
3350
3003
  contextWindow: 2e5,
3351
- supportsNativeTools: true,
3352
- defaultToolProtocol: "native",
3353
3004
  supportsImages: true,
3354
3005
  supportsPromptCache: true,
3355
3006
  inputPrice: 1.1,
@@ -3361,8 +3012,6 @@ var openAiNativeModels = {
3361
3012
  "o3-mini": {
3362
3013
  maxTokens: 1e5,
3363
3014
  contextWindow: 2e5,
3364
- supportsNativeTools: true,
3365
- defaultToolProtocol: "native",
3366
3015
  supportsImages: false,
3367
3016
  supportsPromptCache: true,
3368
3017
  inputPrice: 1.1,
@@ -3375,8 +3024,6 @@ var openAiNativeModels = {
3375
3024
  "o3-mini-high": {
3376
3025
  maxTokens: 1e5,
3377
3026
  contextWindow: 2e5,
3378
- supportsNativeTools: true,
3379
- defaultToolProtocol: "native",
3380
3027
  supportsImages: false,
3381
3028
  supportsPromptCache: true,
3382
3029
  inputPrice: 1.1,
@@ -3388,8 +3035,6 @@ var openAiNativeModels = {
3388
3035
  "o3-mini-low": {
3389
3036
  maxTokens: 1e5,
3390
3037
  contextWindow: 2e5,
3391
- supportsNativeTools: true,
3392
- defaultToolProtocol: "native",
3393
3038
  supportsImages: false,
3394
3039
  supportsPromptCache: true,
3395
3040
  inputPrice: 1.1,
@@ -3401,8 +3046,6 @@ var openAiNativeModels = {
3401
3046
  o1: {
3402
3047
  maxTokens: 1e5,
3403
3048
  contextWindow: 2e5,
3404
- supportsNativeTools: true,
3405
- defaultToolProtocol: "native",
3406
3049
  supportsImages: true,
3407
3050
  supportsPromptCache: true,
3408
3051
  inputPrice: 15,
@@ -3413,8 +3056,6 @@ var openAiNativeModels = {
3413
3056
  "o1-preview": {
3414
3057
  maxTokens: 32768,
3415
3058
  contextWindow: 128e3,
3416
- supportsNativeTools: true,
3417
- defaultToolProtocol: "native",
3418
3059
  supportsImages: true,
3419
3060
  supportsPromptCache: true,
3420
3061
  inputPrice: 15,
@@ -3425,8 +3066,6 @@ var openAiNativeModels = {
3425
3066
  "o1-mini": {
3426
3067
  maxTokens: 65536,
3427
3068
  contextWindow: 128e3,
3428
- supportsNativeTools: true,
3429
- defaultToolProtocol: "native",
3430
3069
  supportsImages: true,
3431
3070
  supportsPromptCache: true,
3432
3071
  inputPrice: 1.1,
@@ -3437,8 +3076,6 @@ var openAiNativeModels = {
3437
3076
  "gpt-4o": {
3438
3077
  maxTokens: 16384,
3439
3078
  contextWindow: 128e3,
3440
- supportsNativeTools: true,
3441
- defaultToolProtocol: "native",
3442
3079
  supportsImages: true,
3443
3080
  supportsPromptCache: true,
3444
3081
  inputPrice: 2.5,
@@ -3452,8 +3089,6 @@ var openAiNativeModels = {
3452
3089
  "gpt-4o-mini": {
3453
3090
  maxTokens: 16384,
3454
3091
  contextWindow: 128e3,
3455
- supportsNativeTools: true,
3456
- defaultToolProtocol: "native",
3457
3092
  supportsImages: true,
3458
3093
  supportsPromptCache: true,
3459
3094
  inputPrice: 0.15,
@@ -3467,8 +3102,6 @@ var openAiNativeModels = {
3467
3102
  "codex-mini-latest": {
3468
3103
  maxTokens: 16384,
3469
3104
  contextWindow: 2e5,
3470
- supportsNativeTools: true,
3471
- defaultToolProtocol: "native",
3472
3105
  supportsImages: false,
3473
3106
  supportsPromptCache: false,
3474
3107
  inputPrice: 1.5,
@@ -3481,8 +3114,6 @@ var openAiNativeModels = {
3481
3114
  "gpt-5-2025-08-07": {
3482
3115
  maxTokens: 128e3,
3483
3116
  contextWindow: 4e5,
3484
- supportsNativeTools: true,
3485
- defaultToolProtocol: "native",
3486
3117
  includedTools: ["apply_patch"],
3487
3118
  excludedTools: ["apply_diff", "write_to_file"],
3488
3119
  supportsImages: true,
@@ -3503,8 +3134,6 @@ var openAiNativeModels = {
3503
3134
  "gpt-5-mini-2025-08-07": {
3504
3135
  maxTokens: 128e3,
3505
3136
  contextWindow: 4e5,
3506
- supportsNativeTools: true,
3507
- defaultToolProtocol: "native",
3508
3137
  includedTools: ["apply_patch"],
3509
3138
  excludedTools: ["apply_diff", "write_to_file"],
3510
3139
  supportsImages: true,
@@ -3525,8 +3154,6 @@ var openAiNativeModels = {
3525
3154
  "gpt-5-nano-2025-08-07": {
3526
3155
  maxTokens: 128e3,
3527
3156
  contextWindow: 4e5,
3528
- supportsNativeTools: true,
3529
- defaultToolProtocol: "native",
3530
3157
  includedTools: ["apply_patch"],
3531
3158
  excludedTools: ["apply_diff", "write_to_file"],
3532
3159
  supportsImages: true,
@@ -3548,9 +3175,7 @@ var openAiModelInfoSaneDefaults = {
3548
3175
  supportsImages: true,
3549
3176
  supportsPromptCache: false,
3550
3177
  inputPrice: 0,
3551
- outputPrice: 0,
3552
- supportsNativeTools: true,
3553
- defaultToolProtocol: "native"
3178
+ outputPrice: 0
3554
3179
  };
3555
3180
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3556
3181
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
@@ -3562,8 +3187,6 @@ var openAiCodexModels = {
3562
3187
  "gpt-5.1-codex-max": {
3563
3188
  maxTokens: 128e3,
3564
3189
  contextWindow: 4e5,
3565
- supportsNativeTools: true,
3566
- defaultToolProtocol: "native",
3567
3190
  includedTools: ["apply_patch"],
3568
3191
  excludedTools: ["apply_diff", "write_to_file"],
3569
3192
  supportsImages: true,
@@ -3576,11 +3199,24 @@ var openAiCodexModels = {
3576
3199
  supportsTemperature: false,
3577
3200
  description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
3578
3201
  },
3202
+ "gpt-5.1-codex": {
3203
+ maxTokens: 128e3,
3204
+ contextWindow: 4e5,
3205
+ includedTools: ["apply_patch"],
3206
+ excludedTools: ["apply_diff", "write_to_file"],
3207
+ supportsImages: true,
3208
+ supportsPromptCache: true,
3209
+ supportsReasoningEffort: ["low", "medium", "high"],
3210
+ reasoningEffort: "medium",
3211
+ // Subscription-based: no per-token costs
3212
+ inputPrice: 0,
3213
+ outputPrice: 0,
3214
+ supportsTemperature: false,
3215
+ description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
3216
+ },
3579
3217
  "gpt-5.2-codex": {
3580
3218
  maxTokens: 128e3,
3581
3219
  contextWindow: 4e5,
3582
- supportsNativeTools: true,
3583
- defaultToolProtocol: "native",
3584
3220
  includedTools: ["apply_patch"],
3585
3221
  excludedTools: ["apply_diff", "write_to_file"],
3586
3222
  supportsImages: true,
@@ -3592,11 +3228,71 @@ var openAiCodexModels = {
3592
3228
  supportsTemperature: false,
3593
3229
  description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
3594
3230
  },
3231
+ "gpt-5.1": {
3232
+ maxTokens: 128e3,
3233
+ contextWindow: 4e5,
3234
+ includedTools: ["apply_patch"],
3235
+ excludedTools: ["apply_diff", "write_to_file"],
3236
+ supportsImages: true,
3237
+ supportsPromptCache: true,
3238
+ supportsReasoningEffort: ["none", "low", "medium", "high"],
3239
+ reasoningEffort: "medium",
3240
+ // Subscription-based: no per-token costs
3241
+ inputPrice: 0,
3242
+ outputPrice: 0,
3243
+ supportsVerbosity: true,
3244
+ supportsTemperature: false,
3245
+ description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
3246
+ },
3247
+ "gpt-5": {
3248
+ maxTokens: 128e3,
3249
+ contextWindow: 4e5,
3250
+ includedTools: ["apply_patch"],
3251
+ excludedTools: ["apply_diff", "write_to_file"],
3252
+ supportsImages: true,
3253
+ supportsPromptCache: true,
3254
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3255
+ reasoningEffort: "medium",
3256
+ // Subscription-based: no per-token costs
3257
+ inputPrice: 0,
3258
+ outputPrice: 0,
3259
+ supportsVerbosity: true,
3260
+ supportsTemperature: false,
3261
+ description: "GPT-5: General GPT-5 model via ChatGPT subscription"
3262
+ },
3263
+ "gpt-5-codex": {
3264
+ maxTokens: 128e3,
3265
+ contextWindow: 4e5,
3266
+ includedTools: ["apply_patch"],
3267
+ excludedTools: ["apply_diff", "write_to_file"],
3268
+ supportsImages: true,
3269
+ supportsPromptCache: true,
3270
+ supportsReasoningEffort: ["low", "medium", "high"],
3271
+ reasoningEffort: "medium",
3272
+ // Subscription-based: no per-token costs
3273
+ inputPrice: 0,
3274
+ outputPrice: 0,
3275
+ supportsTemperature: false,
3276
+ description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
3277
+ },
3278
+ "gpt-5-codex-mini": {
3279
+ maxTokens: 128e3,
3280
+ contextWindow: 4e5,
3281
+ includedTools: ["apply_patch"],
3282
+ excludedTools: ["apply_diff", "write_to_file"],
3283
+ supportsImages: true,
3284
+ supportsPromptCache: true,
3285
+ supportsReasoningEffort: ["low", "medium", "high"],
3286
+ reasoningEffort: "medium",
3287
+ // Subscription-based: no per-token costs
3288
+ inputPrice: 0,
3289
+ outputPrice: 0,
3290
+ supportsTemperature: false,
3291
+ description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
3292
+ },
3595
3293
  "gpt-5.1-codex-mini": {
3596
3294
  maxTokens: 128e3,
3597
3295
  contextWindow: 4e5,
3598
- supportsNativeTools: true,
3599
- defaultToolProtocol: "native",
3600
3296
  includedTools: ["apply_patch"],
3601
3297
  excludedTools: ["apply_diff", "write_to_file"],
3602
3298
  supportsImages: true,
@@ -3611,8 +3307,6 @@ var openAiCodexModels = {
3611
3307
  "gpt-5.2": {
3612
3308
  maxTokens: 128e3,
3613
3309
  contextWindow: 4e5,
3614
- supportsNativeTools: true,
3615
- defaultToolProtocol: "native",
3616
3310
  includedTools: ["apply_patch"],
3617
3311
  excludedTools: ["apply_diff", "write_to_file"],
3618
3312
  supportsImages: true,
@@ -3633,7 +3327,6 @@ var openRouterDefaultModelInfo = {
3633
3327
  contextWindow: 2e5,
3634
3328
  supportsImages: true,
3635
3329
  supportsPromptCache: true,
3636
- supportsNativeTools: true,
3637
3330
  inputPrice: 3,
3638
3331
  outputPrice: 15,
3639
3332
  cacheWritesPrice: 3.75,
@@ -3707,8 +3400,6 @@ var qwenCodeModels = {
3707
3400
  contextWindow: 1e6,
3708
3401
  supportsImages: false,
3709
3402
  supportsPromptCache: false,
3710
- supportsNativeTools: true,
3711
- defaultToolProtocol: "native",
3712
3403
  inputPrice: 0,
3713
3404
  outputPrice: 0,
3714
3405
  cacheWritesPrice: 0,
@@ -3720,8 +3411,6 @@ var qwenCodeModels = {
3720
3411
  contextWindow: 1e6,
3721
3412
  supportsImages: false,
3722
3413
  supportsPromptCache: false,
3723
- supportsNativeTools: true,
3724
- defaultToolProtocol: "native",
3725
3414
  inputPrice: 0,
3726
3415
  outputPrice: 0,
3727
3416
  cacheWritesPrice: 0,
@@ -3737,8 +3426,6 @@ var requestyDefaultModelInfo = {
3737
3426
  contextWindow: 2e5,
3738
3427
  supportsImages: true,
3739
3428
  supportsPromptCache: true,
3740
- supportsNativeTools: true,
3741
- defaultToolProtocol: "native",
3742
3429
  inputPrice: 3,
3743
3430
  outputPrice: 15,
3744
3431
  cacheWritesPrice: 3.75,
@@ -3794,8 +3481,6 @@ var sambaNovaModels = {
3794
3481
  contextWindow: 16384,
3795
3482
  supportsImages: false,
3796
3483
  supportsPromptCache: false,
3797
- supportsNativeTools: true,
3798
- defaultToolProtocol: "native",
3799
3484
  inputPrice: 0.1,
3800
3485
  outputPrice: 0.2,
3801
3486
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3805,8 +3490,6 @@ var sambaNovaModels = {
3805
3490
  contextWindow: 131072,
3806
3491
  supportsImages: false,
3807
3492
  supportsPromptCache: false,
3808
- supportsNativeTools: true,
3809
- defaultToolProtocol: "native",
3810
3493
  inputPrice: 0.6,
3811
3494
  outputPrice: 1.2,
3812
3495
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3817,8 +3500,6 @@ var sambaNovaModels = {
3817
3500
  supportsImages: false,
3818
3501
  supportsPromptCache: false,
3819
3502
  supportsReasoningBudget: true,
3820
- supportsNativeTools: true,
3821
- defaultToolProtocol: "native",
3822
3503
  inputPrice: 5,
3823
3504
  outputPrice: 7,
3824
3505
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3828,8 +3509,6 @@ var sambaNovaModels = {
3828
3509
  contextWindow: 32768,
3829
3510
  supportsImages: false,
3830
3511
  supportsPromptCache: false,
3831
- supportsNativeTools: true,
3832
- defaultToolProtocol: "native",
3833
3512
  inputPrice: 3,
3834
3513
  outputPrice: 4.5,
3835
3514
  description: "DeepSeek V3 model with 32K context window."
@@ -3839,8 +3518,6 @@ var sambaNovaModels = {
3839
3518
  contextWindow: 32768,
3840
3519
  supportsImages: false,
3841
3520
  supportsPromptCache: false,
3842
- supportsNativeTools: true,
3843
- defaultToolProtocol: "native",
3844
3521
  inputPrice: 3,
3845
3522
  outputPrice: 4.5,
3846
3523
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3850,8 +3527,6 @@ var sambaNovaModels = {
3850
3527
  contextWindow: 131072,
3851
3528
  supportsImages: true,
3852
3529
  supportsPromptCache: false,
3853
- supportsNativeTools: true,
3854
- defaultToolProtocol: "native",
3855
3530
  inputPrice: 0.63,
3856
3531
  outputPrice: 1.8,
3857
3532
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3861,8 +3536,6 @@ var sambaNovaModels = {
3861
3536
  contextWindow: 8192,
3862
3537
  supportsImages: false,
3863
3538
  supportsPromptCache: false,
3864
- supportsNativeTools: true,
3865
- defaultToolProtocol: "native",
3866
3539
  inputPrice: 0.4,
3867
3540
  outputPrice: 0.8,
3868
3541
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3872,8 +3545,6 @@ var sambaNovaModels = {
3872
3545
  contextWindow: 131072,
3873
3546
  supportsImages: false,
3874
3547
  supportsPromptCache: false,
3875
- supportsNativeTools: true,
3876
- defaultToolProtocol: "native",
3877
3548
  inputPrice: 0.22,
3878
3549
  outputPrice: 0.59,
3879
3550
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3887,7 +3558,6 @@ var unboundDefaultModelInfo = {
3887
3558
  contextWindow: 2e5,
3888
3559
  supportsImages: true,
3889
3560
  supportsPromptCache: true,
3890
- supportsNativeTools: true,
3891
3561
  inputPrice: 3,
3892
3562
  outputPrice: 15,
3893
3563
  cacheWritesPrice: 3.75,
@@ -3901,8 +3571,6 @@ var vertexModels = {
3901
3571
  maxTokens: 65536,
3902
3572
  contextWindow: 1048576,
3903
3573
  supportsImages: true,
3904
- supportsNativeTools: true,
3905
- defaultToolProtocol: "native",
3906
3574
  supportsPromptCache: true,
3907
3575
  supportsReasoningEffort: ["low", "high"],
3908
3576
  reasoningEffort: "low",
@@ -3910,16 +3578,19 @@ var vertexModels = {
3910
3578
  defaultTemperature: 1,
3911
3579
  inputPrice: 4,
3912
3580
  outputPrice: 18,
3581
+ cacheReadsPrice: 0.4,
3913
3582
  tiers: [
3914
3583
  {
3915
3584
  contextWindow: 2e5,
3916
3585
  inputPrice: 2,
3917
- outputPrice: 12
3586
+ outputPrice: 12,
3587
+ cacheReadsPrice: 0.2
3918
3588
  },
3919
3589
  {
3920
3590
  contextWindow: Infinity,
3921
3591
  inputPrice: 4,
3922
- outputPrice: 18
3592
+ outputPrice: 18,
3593
+ cacheReadsPrice: 0.4
3923
3594
  }
3924
3595
  ]
3925
3596
  },
@@ -3927,24 +3598,19 @@ var vertexModels = {
3927
3598
  maxTokens: 65536,
3928
3599
  contextWindow: 1048576,
3929
3600
  supportsImages: true,
3930
- supportsNativeTools: true,
3931
- defaultToolProtocol: "native",
3932
3601
  supportsPromptCache: true,
3933
3602
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3934
3603
  reasoningEffort: "medium",
3935
3604
  supportsTemperature: true,
3936
3605
  defaultTemperature: 1,
3937
- inputPrice: 0.3,
3938
- outputPrice: 2.5,
3939
- cacheReadsPrice: 0.075,
3940
- cacheWritesPrice: 1
3606
+ inputPrice: 0.5,
3607
+ outputPrice: 3,
3608
+ cacheReadsPrice: 0.05
3941
3609
  },
3942
3610
  "gemini-2.5-flash-preview-05-20:thinking": {
3943
3611
  maxTokens: 65535,
3944
3612
  contextWindow: 1048576,
3945
3613
  supportsImages: true,
3946
- supportsNativeTools: true,
3947
- defaultToolProtocol: "native",
3948
3614
  supportsPromptCache: true,
3949
3615
  inputPrice: 0.15,
3950
3616
  outputPrice: 3.5,
@@ -3956,8 +3622,6 @@ var vertexModels = {
3956
3622
  maxTokens: 65535,
3957
3623
  contextWindow: 1048576,
3958
3624
  supportsImages: true,
3959
- supportsNativeTools: true,
3960
- defaultToolProtocol: "native",
3961
3625
  supportsPromptCache: true,
3962
3626
  inputPrice: 0.15,
3963
3627
  outputPrice: 0.6
@@ -3966,8 +3630,6 @@ var vertexModels = {
3966
3630
  maxTokens: 64e3,
3967
3631
  contextWindow: 1048576,
3968
3632
  supportsImages: true,
3969
- supportsNativeTools: true,
3970
- defaultToolProtocol: "native",
3971
3633
  supportsPromptCache: true,
3972
3634
  inputPrice: 0.3,
3973
3635
  outputPrice: 2.5,
@@ -3980,8 +3642,6 @@ var vertexModels = {
3980
3642
  maxTokens: 65535,
3981
3643
  contextWindow: 1048576,
3982
3644
  supportsImages: true,
3983
- supportsNativeTools: true,
3984
- defaultToolProtocol: "native",
3985
3645
  supportsPromptCache: false,
3986
3646
  inputPrice: 0.15,
3987
3647
  outputPrice: 3.5,
@@ -3993,8 +3653,6 @@ var vertexModels = {
3993
3653
  maxTokens: 65535,
3994
3654
  contextWindow: 1048576,
3995
3655
  supportsImages: true,
3996
- supportsNativeTools: true,
3997
- defaultToolProtocol: "native",
3998
3656
  supportsPromptCache: false,
3999
3657
  inputPrice: 0.15,
4000
3658
  outputPrice: 0.6
@@ -4003,8 +3661,6 @@ var vertexModels = {
4003
3661
  maxTokens: 65535,
4004
3662
  contextWindow: 1048576,
4005
3663
  supportsImages: true,
4006
- supportsNativeTools: true,
4007
- defaultToolProtocol: "native",
4008
3664
  supportsPromptCache: true,
4009
3665
  inputPrice: 2.5,
4010
3666
  outputPrice: 15
@@ -4013,8 +3669,6 @@ var vertexModels = {
4013
3669
  maxTokens: 65535,
4014
3670
  contextWindow: 1048576,
4015
3671
  supportsImages: true,
4016
- supportsNativeTools: true,
4017
- defaultToolProtocol: "native",
4018
3672
  supportsPromptCache: true,
4019
3673
  inputPrice: 2.5,
4020
3674
  outputPrice: 15
@@ -4023,8 +3677,6 @@ var vertexModels = {
4023
3677
  maxTokens: 65535,
4024
3678
  contextWindow: 1048576,
4025
3679
  supportsImages: true,
4026
- supportsNativeTools: true,
4027
- defaultToolProtocol: "native",
4028
3680
  supportsPromptCache: true,
4029
3681
  inputPrice: 2.5,
4030
3682
  outputPrice: 15,
@@ -4035,8 +3687,6 @@ var vertexModels = {
4035
3687
  maxTokens: 64e3,
4036
3688
  contextWindow: 1048576,
4037
3689
  supportsImages: true,
4038
- supportsNativeTools: true,
4039
- defaultToolProtocol: "native",
4040
3690
  supportsPromptCache: true,
4041
3691
  inputPrice: 2.5,
4042
3692
  outputPrice: 15,
@@ -4062,8 +3712,6 @@ var vertexModels = {
4062
3712
  maxTokens: 65535,
4063
3713
  contextWindow: 1048576,
4064
3714
  supportsImages: true,
4065
- supportsNativeTools: true,
4066
- defaultToolProtocol: "native",
4067
3715
  supportsPromptCache: false,
4068
3716
  inputPrice: 0,
4069
3717
  outputPrice: 0
@@ -4072,8 +3720,6 @@ var vertexModels = {
4072
3720
  maxTokens: 8192,
4073
3721
  contextWindow: 2097152,
4074
3722
  supportsImages: true,
4075
- supportsNativeTools: true,
4076
- defaultToolProtocol: "native",
4077
3723
  supportsPromptCache: false,
4078
3724
  inputPrice: 0,
4079
3725
  outputPrice: 0
@@ -4082,8 +3728,6 @@ var vertexModels = {
4082
3728
  maxTokens: 8192,
4083
3729
  contextWindow: 1048576,
4084
3730
  supportsImages: true,
4085
- supportsNativeTools: true,
4086
- defaultToolProtocol: "native",
4087
3731
  supportsPromptCache: true,
4088
3732
  inputPrice: 0.15,
4089
3733
  outputPrice: 0.6
@@ -4092,8 +3736,6 @@ var vertexModels = {
4092
3736
  maxTokens: 8192,
4093
3737
  contextWindow: 1048576,
4094
3738
  supportsImages: true,
4095
- supportsNativeTools: true,
4096
- defaultToolProtocol: "native",
4097
3739
  supportsPromptCache: false,
4098
3740
  inputPrice: 0.075,
4099
3741
  outputPrice: 0.3
@@ -4102,8 +3744,6 @@ var vertexModels = {
4102
3744
  maxTokens: 8192,
4103
3745
  contextWindow: 32768,
4104
3746
  supportsImages: true,
4105
- supportsNativeTools: true,
4106
- defaultToolProtocol: "native",
4107
3747
  supportsPromptCache: false,
4108
3748
  inputPrice: 0,
4109
3749
  outputPrice: 0
@@ -4112,8 +3752,6 @@ var vertexModels = {
4112
3752
  maxTokens: 8192,
4113
3753
  contextWindow: 1048576,
4114
3754
  supportsImages: true,
4115
- supportsNativeTools: true,
4116
- defaultToolProtocol: "native",
4117
3755
  supportsPromptCache: true,
4118
3756
  inputPrice: 0.075,
4119
3757
  outputPrice: 0.3
@@ -4122,8 +3760,6 @@ var vertexModels = {
4122
3760
  maxTokens: 8192,
4123
3761
  contextWindow: 2097152,
4124
3762
  supportsImages: true,
4125
- supportsNativeTools: true,
4126
- defaultToolProtocol: "native",
4127
3763
  supportsPromptCache: false,
4128
3764
  inputPrice: 1.25,
4129
3765
  outputPrice: 5
@@ -4134,8 +3770,6 @@ var vertexModels = {
4134
3770
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4135
3771
  supportsImages: true,
4136
3772
  supportsPromptCache: true,
4137
- supportsNativeTools: true,
4138
- defaultToolProtocol: "native",
4139
3773
  inputPrice: 3,
4140
3774
  // $3 per million input tokens (≤200K context)
4141
3775
  outputPrice: 15,
@@ -4167,8 +3801,6 @@ var vertexModels = {
4167
3801
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4168
3802
  supportsImages: true,
4169
3803
  supportsPromptCache: true,
4170
- supportsNativeTools: true,
4171
- defaultToolProtocol: "native",
4172
3804
  inputPrice: 3,
4173
3805
  // $3 per million input tokens (≤200K context)
4174
3806
  outputPrice: 15,
@@ -4199,8 +3831,6 @@ var vertexModels = {
4199
3831
  contextWindow: 2e5,
4200
3832
  supportsImages: true,
4201
3833
  supportsPromptCache: true,
4202
- supportsNativeTools: true,
4203
- defaultToolProtocol: "native",
4204
3834
  inputPrice: 1,
4205
3835
  outputPrice: 5,
4206
3836
  cacheWritesPrice: 1.25,
@@ -4212,8 +3842,6 @@ var vertexModels = {
4212
3842
  contextWindow: 2e5,
4213
3843
  supportsImages: true,
4214
3844
  supportsPromptCache: true,
4215
- supportsNativeTools: true,
4216
- defaultToolProtocol: "native",
4217
3845
  inputPrice: 5,
4218
3846
  outputPrice: 25,
4219
3847
  cacheWritesPrice: 6.25,
@@ -4225,8 +3853,6 @@ var vertexModels = {
4225
3853
  contextWindow: 2e5,
4226
3854
  supportsImages: true,
4227
3855
  supportsPromptCache: true,
4228
- supportsNativeTools: true,
4229
- defaultToolProtocol: "native",
4230
3856
  inputPrice: 15,
4231
3857
  outputPrice: 75,
4232
3858
  cacheWritesPrice: 18.75,
@@ -4238,8 +3864,6 @@ var vertexModels = {
4238
3864
  contextWindow: 2e5,
4239
3865
  supportsImages: true,
4240
3866
  supportsPromptCache: true,
4241
- supportsNativeTools: true,
4242
- defaultToolProtocol: "native",
4243
3867
  inputPrice: 15,
4244
3868
  outputPrice: 75,
4245
3869
  cacheWritesPrice: 18.75,
@@ -4250,8 +3874,6 @@ var vertexModels = {
4250
3874
  contextWindow: 2e5,
4251
3875
  supportsImages: true,
4252
3876
  supportsPromptCache: true,
4253
- supportsNativeTools: true,
4254
- defaultToolProtocol: "native",
4255
3877
  inputPrice: 3,
4256
3878
  outputPrice: 15,
4257
3879
  cacheWritesPrice: 3.75,
@@ -4264,8 +3886,6 @@ var vertexModels = {
4264
3886
  contextWindow: 2e5,
4265
3887
  supportsImages: true,
4266
3888
  supportsPromptCache: true,
4267
- supportsNativeTools: true,
4268
- defaultToolProtocol: "native",
4269
3889
  inputPrice: 3,
4270
3890
  outputPrice: 15,
4271
3891
  cacheWritesPrice: 3.75,
@@ -4276,8 +3896,6 @@ var vertexModels = {
4276
3896
  contextWindow: 2e5,
4277
3897
  supportsImages: true,
4278
3898
  supportsPromptCache: true,
4279
- supportsNativeTools: true,
4280
- defaultToolProtocol: "native",
4281
3899
  inputPrice: 3,
4282
3900
  outputPrice: 15,
4283
3901
  cacheWritesPrice: 3.75,
@@ -4288,8 +3906,6 @@ var vertexModels = {
4288
3906
  contextWindow: 2e5,
4289
3907
  supportsImages: true,
4290
3908
  supportsPromptCache: true,
4291
- supportsNativeTools: true,
4292
- defaultToolProtocol: "native",
4293
3909
  inputPrice: 3,
4294
3910
  outputPrice: 15,
4295
3911
  cacheWritesPrice: 3.75,
@@ -4300,8 +3916,6 @@ var vertexModels = {
4300
3916
  contextWindow: 2e5,
4301
3917
  supportsImages: false,
4302
3918
  supportsPromptCache: true,
4303
- supportsNativeTools: true,
4304
- defaultToolProtocol: "native",
4305
3919
  inputPrice: 1,
4306
3920
  outputPrice: 5,
4307
3921
  cacheWritesPrice: 1.25,
@@ -4312,8 +3926,6 @@ var vertexModels = {
4312
3926
  contextWindow: 2e5,
4313
3927
  supportsImages: true,
4314
3928
  supportsPromptCache: true,
4315
- supportsNativeTools: true,
4316
- defaultToolProtocol: "native",
4317
3929
  inputPrice: 15,
4318
3930
  outputPrice: 75,
4319
3931
  cacheWritesPrice: 18.75,
@@ -4324,8 +3936,6 @@ var vertexModels = {
4324
3936
  contextWindow: 2e5,
4325
3937
  supportsImages: true,
4326
3938
  supportsPromptCache: true,
4327
- supportsNativeTools: true,
4328
- defaultToolProtocol: "native",
4329
3939
  inputPrice: 0.25,
4330
3940
  outputPrice: 1.25,
4331
3941
  cacheWritesPrice: 0.3,
@@ -4335,8 +3945,6 @@ var vertexModels = {
4335
3945
  maxTokens: 64e3,
4336
3946
  contextWindow: 1048576,
4337
3947
  supportsImages: true,
4338
- supportsNativeTools: true,
4339
- defaultToolProtocol: "native",
4340
3948
  supportsPromptCache: true,
4341
3949
  inputPrice: 0.1,
4342
3950
  outputPrice: 0.4,
@@ -4350,7 +3958,6 @@ var vertexModels = {
4350
3958
  contextWindow: 131072,
4351
3959
  supportsImages: false,
4352
3960
  supportsPromptCache: false,
4353
- supportsNativeTools: true,
4354
3961
  inputPrice: 0.35,
4355
3962
  outputPrice: 1.15,
4356
3963
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
@@ -4360,7 +3967,6 @@ var vertexModels = {
4360
3967
  contextWindow: 163840,
4361
3968
  supportsImages: false,
4362
3969
  supportsPromptCache: false,
4363
- supportsNativeTools: true,
4364
3970
  inputPrice: 1.35,
4365
3971
  outputPrice: 5.4,
4366
3972
  description: "DeepSeek R1 (0528). Available in us-central1"
@@ -4370,7 +3976,6 @@ var vertexModels = {
4370
3976
  contextWindow: 163840,
4371
3977
  supportsImages: false,
4372
3978
  supportsPromptCache: false,
4373
- supportsNativeTools: true,
4374
3979
  inputPrice: 0.6,
4375
3980
  outputPrice: 1.7,
4376
3981
  description: "DeepSeek V3.1. Available in us-west2"
@@ -4380,7 +3985,6 @@ var vertexModels = {
4380
3985
  contextWindow: 131072,
4381
3986
  supportsImages: false,
4382
3987
  supportsPromptCache: false,
4383
- supportsNativeTools: true,
4384
3988
  inputPrice: 0.15,
4385
3989
  outputPrice: 0.6,
4386
3990
  description: "OpenAI gpt-oss 120B. Available in us-central1"
@@ -4390,7 +3994,6 @@ var vertexModels = {
4390
3994
  contextWindow: 131072,
4391
3995
  supportsImages: false,
4392
3996
  supportsPromptCache: false,
4393
- supportsNativeTools: true,
4394
3997
  inputPrice: 0.075,
4395
3998
  outputPrice: 0.3,
4396
3999
  description: "OpenAI gpt-oss 20B. Available in us-central1"
@@ -4400,7 +4003,6 @@ var vertexModels = {
4400
4003
  contextWindow: 262144,
4401
4004
  supportsImages: false,
4402
4005
  supportsPromptCache: false,
4403
- supportsNativeTools: true,
4404
4006
  inputPrice: 1,
4405
4007
  outputPrice: 4,
4406
4008
  description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
@@ -4410,10 +4012,18 @@ var vertexModels = {
4410
4012
  contextWindow: 262144,
4411
4013
  supportsImages: false,
4412
4014
  supportsPromptCache: false,
4413
- supportsNativeTools: true,
4414
4015
  inputPrice: 0.25,
4415
4016
  outputPrice: 1,
4416
4017
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
4018
+ },
4019
+ "moonshotai/kimi-k2-thinking-maas": {
4020
+ maxTokens: 16384,
4021
+ contextWindow: 262144,
4022
+ supportsPromptCache: false,
4023
+ supportsImages: false,
4024
+ inputPrice: 0.6,
4025
+ outputPrice: 2.5,
4026
+ description: "Kimi K2 Thinking Model with 256K context window."
4417
4027
  }
4418
4028
  };
4419
4029
  var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
@@ -4646,8 +4256,6 @@ var xaiModels = {
4646
4256
  contextWindow: 256e3,
4647
4257
  supportsImages: true,
4648
4258
  supportsPromptCache: true,
4649
- supportsNativeTools: true,
4650
- defaultToolProtocol: "native",
4651
4259
  inputPrice: 0.2,
4652
4260
  outputPrice: 1.5,
4653
4261
  cacheWritesPrice: 0.02,
@@ -4661,8 +4269,6 @@ var xaiModels = {
4661
4269
  contextWindow: 2e6,
4662
4270
  supportsImages: true,
4663
4271
  supportsPromptCache: true,
4664
- supportsNativeTools: true,
4665
- defaultToolProtocol: "native",
4666
4272
  inputPrice: 0.2,
4667
4273
  outputPrice: 0.5,
4668
4274
  cacheWritesPrice: 0.05,
@@ -4676,8 +4282,6 @@ var xaiModels = {
4676
4282
  contextWindow: 2e6,
4677
4283
  supportsImages: true,
4678
4284
  supportsPromptCache: true,
4679
- supportsNativeTools: true,
4680
- defaultToolProtocol: "native",
4681
4285
  inputPrice: 0.2,
4682
4286
  outputPrice: 0.5,
4683
4287
  cacheWritesPrice: 0.05,
@@ -4691,8 +4295,6 @@ var xaiModels = {
4691
4295
  contextWindow: 2e6,
4692
4296
  supportsImages: true,
4693
4297
  supportsPromptCache: true,
4694
- supportsNativeTools: true,
4695
- defaultToolProtocol: "native",
4696
4298
  inputPrice: 0.2,
4697
4299
  outputPrice: 0.5,
4698
4300
  cacheWritesPrice: 0.05,
@@ -4706,8 +4308,6 @@ var xaiModels = {
4706
4308
  contextWindow: 2e6,
4707
4309
  supportsImages: true,
4708
4310
  supportsPromptCache: true,
4709
- supportsNativeTools: true,
4710
- defaultToolProtocol: "native",
4711
4311
  inputPrice: 0.2,
4712
4312
  outputPrice: 0.5,
4713
4313
  cacheWritesPrice: 0.05,
@@ -4721,8 +4321,6 @@ var xaiModels = {
4721
4321
  contextWindow: 256e3,
4722
4322
  supportsImages: true,
4723
4323
  supportsPromptCache: true,
4724
- supportsNativeTools: true,
4725
- defaultToolProtocol: "native",
4726
4324
  inputPrice: 3,
4727
4325
  outputPrice: 15,
4728
4326
  cacheWritesPrice: 0.75,
@@ -4736,8 +4334,6 @@ var xaiModels = {
4736
4334
  contextWindow: 131072,
4737
4335
  supportsImages: true,
4738
4336
  supportsPromptCache: true,
4739
- supportsNativeTools: true,
4740
- defaultToolProtocol: "native",
4741
4337
  inputPrice: 0.3,
4742
4338
  outputPrice: 0.5,
4743
4339
  cacheWritesPrice: 0.07,
@@ -4753,8 +4349,6 @@ var xaiModels = {
4753
4349
  contextWindow: 131072,
4754
4350
  supportsImages: true,
4755
4351
  supportsPromptCache: true,
4756
- supportsNativeTools: true,
4757
- defaultToolProtocol: "native",
4758
4352
  inputPrice: 3,
4759
4353
  outputPrice: 15,
4760
4354
  cacheWritesPrice: 0.75,
@@ -4851,7 +4445,6 @@ var vercelAiGatewayDefaultModelInfo = {
4851
4445
  contextWindow: 2e5,
4852
4446
  supportsImages: true,
4853
4447
  supportsPromptCache: true,
4854
- supportsNativeTools: true,
4855
4448
  inputPrice: 3,
4856
4449
  outputPrice: 15,
4857
4450
  cacheWritesPrice: 3.75,
@@ -4868,8 +4461,6 @@ var internationalZAiModels = {
4868
4461
  contextWindow: 131072,
4869
4462
  supportsImages: false,
4870
4463
  supportsPromptCache: true,
4871
- supportsNativeTools: true,
4872
- defaultToolProtocol: "native",
4873
4464
  inputPrice: 0.6,
4874
4465
  outputPrice: 2.2,
4875
4466
  cacheWritesPrice: 0,
@@ -4881,8 +4472,6 @@ var internationalZAiModels = {
4881
4472
  contextWindow: 131072,
4882
4473
  supportsImages: false,
4883
4474
  supportsPromptCache: true,
4884
- supportsNativeTools: true,
4885
- defaultToolProtocol: "native",
4886
4475
  inputPrice: 0.2,
4887
4476
  outputPrice: 1.1,
4888
4477
  cacheWritesPrice: 0,
@@ -4894,8 +4483,6 @@ var internationalZAiModels = {
4894
4483
  contextWindow: 131072,
4895
4484
  supportsImages: false,
4896
4485
  supportsPromptCache: true,
4897
- supportsNativeTools: true,
4898
- defaultToolProtocol: "native",
4899
4486
  inputPrice: 2.2,
4900
4487
  outputPrice: 8.9,
4901
4488
  cacheWritesPrice: 0,
@@ -4907,8 +4494,6 @@ var internationalZAiModels = {
4907
4494
  contextWindow: 131072,
4908
4495
  supportsImages: false,
4909
4496
  supportsPromptCache: true,
4910
- supportsNativeTools: true,
4911
- defaultToolProtocol: "native",
4912
4497
  inputPrice: 1.1,
4913
4498
  outputPrice: 4.5,
4914
4499
  cacheWritesPrice: 0,
@@ -4920,8 +4505,6 @@ var internationalZAiModels = {
4920
4505
  contextWindow: 131072,
4921
4506
  supportsImages: false,
4922
4507
  supportsPromptCache: true,
4923
- supportsNativeTools: true,
4924
- defaultToolProtocol: "native",
4925
4508
  inputPrice: 0,
4926
4509
  outputPrice: 0,
4927
4510
  cacheWritesPrice: 0,
@@ -4933,21 +4516,28 @@ var internationalZAiModels = {
4933
4516
  contextWindow: 131072,
4934
4517
  supportsImages: true,
4935
4518
  supportsPromptCache: true,
4936
- supportsNativeTools: true,
4937
- defaultToolProtocol: "native",
4938
4519
  inputPrice: 0.6,
4939
4520
  outputPrice: 1.8,
4940
4521
  cacheWritesPrice: 0,
4941
4522
  cacheReadsPrice: 0.11,
4942
4523
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
4943
4524
  },
4525
+ "glm-4.6v": {
4526
+ maxTokens: 16384,
4527
+ contextWindow: 131072,
4528
+ supportsImages: true,
4529
+ supportsPromptCache: true,
4530
+ inputPrice: 0.3,
4531
+ outputPrice: 0.9,
4532
+ cacheWritesPrice: 0,
4533
+ cacheReadsPrice: 0.05,
4534
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4535
+ },
4944
4536
  "glm-4.6": {
4945
4537
  maxTokens: 16384,
4946
4538
  contextWindow: 2e5,
4947
4539
  supportsImages: false,
4948
4540
  supportsPromptCache: true,
4949
- supportsNativeTools: true,
4950
- defaultToolProtocol: "native",
4951
4541
  inputPrice: 0.6,
4952
4542
  outputPrice: 2.2,
4953
4543
  cacheWritesPrice: 0,
@@ -4959,8 +4549,6 @@ var internationalZAiModels = {
4959
4549
  contextWindow: 2e5,
4960
4550
  supportsImages: false,
4961
4551
  supportsPromptCache: true,
4962
- supportsNativeTools: true,
4963
- defaultToolProtocol: "native",
4964
4552
  supportsReasoningEffort: ["disable", "medium"],
4965
4553
  reasoningEffort: "medium",
4966
4554
  preserveReasoning: true,
@@ -4970,13 +4558,55 @@ var internationalZAiModels = {
4970
4558
  cacheReadsPrice: 0.11,
4971
4559
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4972
4560
  },
4561
+ "glm-4.7-flash": {
4562
+ maxTokens: 16384,
4563
+ contextWindow: 2e5,
4564
+ supportsImages: false,
4565
+ supportsPromptCache: true,
4566
+ inputPrice: 0,
4567
+ outputPrice: 0,
4568
+ cacheWritesPrice: 0,
4569
+ cacheReadsPrice: 0,
4570
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4571
+ },
4572
+ "glm-4.7-flashx": {
4573
+ maxTokens: 16384,
4574
+ contextWindow: 2e5,
4575
+ supportsImages: false,
4576
+ supportsPromptCache: true,
4577
+ inputPrice: 0.07,
4578
+ outputPrice: 0.4,
4579
+ cacheWritesPrice: 0,
4580
+ cacheReadsPrice: 0.01,
4581
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4582
+ },
4583
+ "glm-4.6v-flash": {
4584
+ maxTokens: 16384,
4585
+ contextWindow: 131072,
4586
+ supportsImages: true,
4587
+ supportsPromptCache: true,
4588
+ inputPrice: 0,
4589
+ outputPrice: 0,
4590
+ cacheWritesPrice: 0,
4591
+ cacheReadsPrice: 0,
4592
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4593
+ },
4594
+ "glm-4.6v-flashx": {
4595
+ maxTokens: 16384,
4596
+ contextWindow: 131072,
4597
+ supportsImages: true,
4598
+ supportsPromptCache: true,
4599
+ inputPrice: 0.04,
4600
+ outputPrice: 0.4,
4601
+ cacheWritesPrice: 0,
4602
+ cacheReadsPrice: 4e-3,
4603
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
4604
+ },
4973
4605
  "glm-4-32b-0414-128k": {
4974
4606
  maxTokens: 16384,
4975
4607
  contextWindow: 131072,
4976
4608
  supportsImages: false,
4977
4609
  supportsPromptCache: false,
4978
- supportsNativeTools: true,
4979
- defaultToolProtocol: "native",
4980
4610
  inputPrice: 0.1,
4981
4611
  outputPrice: 0.1,
4982
4612
  cacheWritesPrice: 0,
@@ -4991,8 +4621,6 @@ var mainlandZAiModels = {
4991
4621
  contextWindow: 131072,
4992
4622
  supportsImages: false,
4993
4623
  supportsPromptCache: true,
4994
- supportsNativeTools: true,
4995
- defaultToolProtocol: "native",
4996
4624
  inputPrice: 0.29,
4997
4625
  outputPrice: 1.14,
4998
4626
  cacheWritesPrice: 0,
@@ -5004,8 +4632,6 @@ var mainlandZAiModels = {
5004
4632
  contextWindow: 131072,
5005
4633
  supportsImages: false,
5006
4634
  supportsPromptCache: true,
5007
- supportsNativeTools: true,
5008
- defaultToolProtocol: "native",
5009
4635
  inputPrice: 0.1,
5010
4636
  outputPrice: 0.6,
5011
4637
  cacheWritesPrice: 0,
@@ -5017,8 +4643,6 @@ var mainlandZAiModels = {
5017
4643
  contextWindow: 131072,
5018
4644
  supportsImages: false,
5019
4645
  supportsPromptCache: true,
5020
- supportsNativeTools: true,
5021
- defaultToolProtocol: "native",
5022
4646
  inputPrice: 0.29,
5023
4647
  outputPrice: 1.14,
5024
4648
  cacheWritesPrice: 0,
@@ -5030,8 +4654,6 @@ var mainlandZAiModels = {
5030
4654
  contextWindow: 131072,
5031
4655
  supportsImages: false,
5032
4656
  supportsPromptCache: true,
5033
- supportsNativeTools: true,
5034
- defaultToolProtocol: "native",
5035
4657
  inputPrice: 0.1,
5036
4658
  outputPrice: 0.6,
5037
4659
  cacheWritesPrice: 0,
@@ -5043,8 +4665,6 @@ var mainlandZAiModels = {
5043
4665
  contextWindow: 131072,
5044
4666
  supportsImages: false,
5045
4667
  supportsPromptCache: true,
5046
- supportsNativeTools: true,
5047
- defaultToolProtocol: "native",
5048
4668
  inputPrice: 0,
5049
4669
  outputPrice: 0,
5050
4670
  cacheWritesPrice: 0,
@@ -5056,8 +4676,6 @@ var mainlandZAiModels = {
5056
4676
  contextWindow: 131072,
5057
4677
  supportsImages: true,
5058
4678
  supportsPromptCache: true,
5059
- supportsNativeTools: true,
5060
- defaultToolProtocol: "native",
5061
4679
  inputPrice: 0.29,
5062
4680
  outputPrice: 0.93,
5063
4681
  cacheWritesPrice: 0,
@@ -5069,8 +4687,6 @@ var mainlandZAiModels = {
5069
4687
  contextWindow: 204800,
5070
4688
  supportsImages: false,
5071
4689
  supportsPromptCache: true,
5072
- supportsNativeTools: true,
5073
- defaultToolProtocol: "native",
5074
4690
  inputPrice: 0.29,
5075
4691
  outputPrice: 1.14,
5076
4692
  cacheWritesPrice: 0,
@@ -5082,8 +4698,6 @@ var mainlandZAiModels = {
5082
4698
  contextWindow: 204800,
5083
4699
  supportsImages: false,
5084
4700
  supportsPromptCache: true,
5085
- supportsNativeTools: true,
5086
- defaultToolProtocol: "native",
5087
4701
  supportsReasoningEffort: ["disable", "medium"],
5088
4702
  reasoningEffort: "medium",
5089
4703
  preserveReasoning: true,
@@ -5092,6 +4706,61 @@ var mainlandZAiModels = {
5092
4706
  cacheWritesPrice: 0,
5093
4707
  cacheReadsPrice: 0.057,
5094
4708
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4709
+ },
4710
+ "glm-4.7-flash": {
4711
+ maxTokens: 16384,
4712
+ contextWindow: 204800,
4713
+ supportsImages: false,
4714
+ supportsPromptCache: true,
4715
+ inputPrice: 0,
4716
+ outputPrice: 0,
4717
+ cacheWritesPrice: 0,
4718
+ cacheReadsPrice: 0,
4719
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4720
+ },
4721
+ "glm-4.7-flashx": {
4722
+ maxTokens: 16384,
4723
+ contextWindow: 204800,
4724
+ supportsImages: false,
4725
+ supportsPromptCache: true,
4726
+ inputPrice: 0.035,
4727
+ outputPrice: 0.2,
4728
+ cacheWritesPrice: 0,
4729
+ cacheReadsPrice: 5e-3,
4730
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4731
+ },
4732
+ "glm-4.6v": {
4733
+ maxTokens: 16384,
4734
+ contextWindow: 131072,
4735
+ supportsImages: true,
4736
+ supportsPromptCache: true,
4737
+ inputPrice: 0.15,
4738
+ outputPrice: 0.45,
4739
+ cacheWritesPrice: 0,
4740
+ cacheReadsPrice: 0.025,
4741
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4742
+ },
4743
+ "glm-4.6v-flash": {
4744
+ maxTokens: 16384,
4745
+ contextWindow: 131072,
4746
+ supportsImages: true,
4747
+ supportsPromptCache: true,
4748
+ inputPrice: 0,
4749
+ outputPrice: 0,
4750
+ cacheWritesPrice: 0,
4751
+ cacheReadsPrice: 0,
4752
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4753
+ },
4754
+ "glm-4.6v-flashx": {
4755
+ maxTokens: 16384,
4756
+ contextWindow: 131072,
4757
+ supportsImages: true,
4758
+ supportsPromptCache: true,
4759
+ inputPrice: 0.02,
4760
+ outputPrice: 0.2,
4761
+ cacheWritesPrice: 0,
4762
+ cacheReadsPrice: 2e-3,
4763
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
5095
4764
  }
5096
4765
  };
5097
4766
  var ZAI_DEFAULT_TEMPERATURE = 0.6;
@@ -5125,7 +4794,6 @@ var deepInfraDefaultModelInfo = {
5125
4794
  contextWindow: 262144,
5126
4795
  supportsImages: false,
5127
4796
  supportsPromptCache: false,
5128
- supportsNativeTools: true,
5129
4797
  inputPrice: 0.3,
5130
4798
  outputPrice: 1.2,
5131
4799
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -5139,8 +4807,6 @@ var minimaxModels = {
5139
4807
  contextWindow: 192e3,
5140
4808
  supportsImages: false,
5141
4809
  supportsPromptCache: true,
5142
- supportsNativeTools: true,
5143
- defaultToolProtocol: "native",
5144
4810
  includedTools: ["search_and_replace"],
5145
4811
  excludedTools: ["apply_diff"],
5146
4812
  preserveReasoning: true,
@@ -5155,8 +4821,6 @@ var minimaxModels = {
5155
4821
  contextWindow: 192e3,
5156
4822
  supportsImages: false,
5157
4823
  supportsPromptCache: true,
5158
- supportsNativeTools: true,
5159
- defaultToolProtocol: "native",
5160
4824
  includedTools: ["search_and_replace"],
5161
4825
  excludedTools: ["apply_diff"],
5162
4826
  preserveReasoning: true,
@@ -5171,8 +4835,6 @@ var minimaxModels = {
5171
4835
  contextWindow: 192e3,
5172
4836
  supportsImages: false,
5173
4837
  supportsPromptCache: true,
5174
- supportsNativeTools: true,
5175
- defaultToolProtocol: "native",
5176
4838
  includedTools: ["search_and_replace"],
5177
4839
  excludedTools: ["apply_diff"],
5178
4840
  preserveReasoning: true,
@@ -5244,8 +4906,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
5244
4906
  return deepInfraDefaultModelId;
5245
4907
  case "vscode-lm":
5246
4908
  return vscodeLlmDefaultModelId;
5247
- case "claude-code":
5248
- return claudeCodeDefaultModelId;
5249
4909
  case "cerebras":
5250
4910
  return cerebrasDefaultModelId;
5251
4911
  case "sambanova":
@@ -5303,7 +4963,6 @@ var providerNames = [
5303
4963
  "bedrock",
5304
4964
  "baseten",
5305
4965
  "cerebras",
5306
- "claude-code",
5307
4966
  "doubao",
5308
4967
  "deepseek",
5309
4968
  "featherless",
@@ -5333,9 +4992,7 @@ var providerSettingsEntrySchema = z8.object({
5333
4992
  });
5334
4993
  var baseProviderSettingsSchema = z8.object({
5335
4994
  includeMaxTokens: z8.boolean().optional(),
5336
- diffEnabled: z8.boolean().optional(),
5337
4995
  todoListEnabled: z8.boolean().optional(),
5338
- fuzzyMatchThreshold: z8.number().optional(),
5339
4996
  modelTemperature: z8.number().nullish(),
5340
4997
  rateLimitSeconds: z8.number().optional(),
5341
4998
  consecutiveMistakeLimit: z8.number().min(0).optional(),
@@ -5345,9 +5002,7 @@ var baseProviderSettingsSchema = z8.object({
5345
5002
  modelMaxTokens: z8.number().optional(),
5346
5003
  modelMaxThinkingTokens: z8.number().optional(),
5347
5004
  // Model verbosity.
5348
- verbosity: verbosityLevelsSchema.optional(),
5349
- // Tool protocol override for this profile.
5350
- toolProtocol: z8.enum(["xml", "native"]).optional()
5005
+ verbosity: verbosityLevelsSchema.optional()
5351
5006
  });
5352
5007
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
5353
5008
  apiModelId: z8.string().optional()
@@ -5359,7 +5014,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
5359
5014
  anthropicBeta1MContext: z8.boolean().optional()
5360
5015
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5361
5016
  });
5362
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5363
5017
  var openRouterSchema = baseProviderSettingsSchema.extend({
5364
5018
  openRouterApiKey: z8.string().optional(),
5365
5019
  openRouterModelId: z8.string().optional(),
@@ -5548,7 +5202,6 @@ var defaultSchema = z8.object({
5548
5202
  });
5549
5203
  var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5550
5204
  anthropicSchema.merge(z8.object({ apiProvider: z8.literal("anthropic") })),
5551
- claudeCodeSchema.merge(z8.object({ apiProvider: z8.literal("claude-code") })),
5552
5205
  openRouterSchema.merge(z8.object({ apiProvider: z8.literal("openrouter") })),
5553
5206
  bedrockSchema.merge(z8.object({ apiProvider: z8.literal("bedrock") })),
5554
5207
  vertexSchema.merge(z8.object({ apiProvider: z8.literal("vertex") })),
@@ -5589,7 +5242,6 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5589
5242
  var providerSettingsSchema = z8.object({
5590
5243
  apiProvider: providerNamesSchema.optional(),
5591
5244
  ...anthropicSchema.shape,
5592
- ...claudeCodeSchema.shape,
5593
5245
  ...openRouterSchema.shape,
5594
5246
  ...bedrockSchema.shape,
5595
5247
  ...vertexSchema.shape,
@@ -5654,7 +5306,6 @@ var getModelId = (settings) => {
5654
5306
  var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
5655
5307
  var modelIdKeysByProvider = {
5656
5308
  anthropic: "apiModelId",
5657
- "claude-code": "apiModelId",
5658
5309
  openrouter: "openRouterModelId",
5659
5310
  bedrock: "apiModelId",
5660
5311
  vertex: "apiModelId",
@@ -5688,7 +5339,7 @@ var modelIdKeysByProvider = {
5688
5339
  roo: "apiModelId",
5689
5340
  "vercel-ai-gateway": "vercelAiGatewayModelId"
5690
5341
  };
5691
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
5342
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
5692
5343
  var getApiProtocol = (provider, modelId) => {
5693
5344
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
5694
5345
  return "anthropic";
@@ -5717,7 +5368,6 @@ var MODELS_BY_PROVIDER = {
5717
5368
  label: "Cerebras",
5718
5369
  models: Object.keys(cerebrasModels)
5719
5370
  },
5720
- "claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
5721
5371
  deepseek: {
5722
5372
  id: "deepseek",
5723
5373
  label: "DeepSeek",
@@ -5821,16 +5471,6 @@ var historyItemSchema = z9.object({
5821
5471
  size: z9.number().optional(),
5822
5472
  workspace: z9.string().optional(),
5823
5473
  mode: z9.string().optional(),
5824
- /**
5825
- * The tool protocol used by this task. Once a task uses tools with a specific
5826
- * protocol (XML or Native), it is permanently locked to that protocol.
5827
- *
5828
- * - "xml": Tool calls are parsed from XML text (no tool IDs)
5829
- * - "native": Tool calls come as tool_call chunks with IDs
5830
- *
5831
- * This ensures task resumption works correctly even when NTC settings change.
5832
- */
5833
- toolProtocol: z9.enum(["xml", "native"]).optional(),
5834
5474
  apiConfigName: z9.string().optional(),
5835
5475
  // Provider profile name for sticky profile feature
5836
5476
  status: z9.enum(["active", "completed", "delegated"]).optional(),
@@ -5848,23 +5488,12 @@ var historyItemSchema = z9.object({
5848
5488
 
5849
5489
  // src/experiment.ts
5850
5490
  import { z as z10 } from "zod";
5851
- var experimentIds = [
5852
- "powerSteering",
5853
- "multiFileApplyDiff",
5854
- "preventFocusDisruption",
5855
- "imageGeneration",
5856
- "runSlashCommand",
5857
- "multipleNativeToolCalls",
5858
- "customTools"
5859
- ];
5491
+ var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
5860
5492
  var experimentIdsSchema = z10.enum(experimentIds);
5861
5493
  var experimentsSchema = z10.object({
5862
- powerSteering: z10.boolean().optional(),
5863
- multiFileApplyDiff: z10.boolean().optional(),
5864
5494
  preventFocusDisruption: z10.boolean().optional(),
5865
5495
  imageGeneration: z10.boolean().optional(),
5866
5496
  runSlashCommand: z10.boolean().optional(),
5867
- multipleNativeToolCalls: z10.boolean().optional(),
5868
5497
  customTools: z10.boolean().optional()
5869
5498
  });
5870
5499
 
@@ -5916,6 +5545,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
5916
5545
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
5917
5546
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
5918
5547
  TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5548
+ TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
5919
5549
  return TelemetryEventName2;
5920
5550
  })(TelemetryEventName || {});
5921
5551
  var staticAppPropertiesSchema = z11.object({
@@ -6004,7 +5634,8 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
6004
5634
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
6005
5635
  "Tab Shown" /* TAB_SHOWN */,
6006
5636
  "Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
6007
- "Custom Mode Created" /* CUSTOM_MODE_CREATED */
5637
+ "Custom Mode Created" /* CUSTOM_MODE_CREATED */,
5638
+ "Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
6008
5639
  ]),
6009
5640
  properties: telemetryPropertiesSchema
6010
5641
  }),
@@ -6318,7 +5949,15 @@ var isLanguage = (value) => languages.includes(value);
6318
5949
 
6319
5950
  // src/global-settings.ts
6320
5951
  var DEFAULT_WRITE_DELAY_MS = 1e3;
6321
- var DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT = 5e4;
5952
+ var TERMINAL_PREVIEW_BYTES = {
5953
+ small: 5 * 1024,
5954
+ // 5KB
5955
+ medium: 10 * 1024,
5956
+ // 10KB
5957
+ large: 20 * 1024
5958
+ // 20KB
5959
+ };
5960
+ var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
6322
5961
  var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
6323
5962
  var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
6324
5963
  var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
@@ -6334,7 +5973,6 @@ var globalSettingsSchema = z14.object({
6334
5973
  imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
6335
5974
  openRouterImageApiKey: z14.string().optional(),
6336
5975
  openRouterImageGenerationSelectedModel: z14.string().optional(),
6337
- condensingApiConfigId: z14.string().optional(),
6338
5976
  customCondensingPrompt: z14.string().optional(),
6339
5977
  autoApprovalEnabled: z14.boolean().optional(),
6340
5978
  alwaysAllowReadOnly: z14.boolean().optional(),
@@ -6360,7 +5998,6 @@ var globalSettingsSchema = z14.object({
6360
5998
  allowedMaxCost: z14.number().nullish(),
6361
5999
  autoCondenseContext: z14.boolean().optional(),
6362
6000
  autoCondenseContextPercent: z14.number().optional(),
6363
- maxConcurrentFileReads: z14.number().optional(),
6364
6001
  /**
6365
6002
  * Whether to include current time in the environment details
6366
6003
  * @default true
@@ -6403,11 +6040,9 @@ var globalSettingsSchema = z14.object({
6403
6040
  maxWorkspaceFiles: z14.number().optional(),
6404
6041
  showRooIgnoredFiles: z14.boolean().optional(),
6405
6042
  enableSubfolderRules: z14.boolean().optional(),
6406
- maxReadFileLine: z14.number().optional(),
6407
6043
  maxImageFileSize: z14.number().optional(),
6408
6044
  maxTotalImageSize: z14.number().optional(),
6409
- terminalOutputLineLimit: z14.number().optional(),
6410
- terminalOutputCharacterLimit: z14.number().optional(),
6045
+ terminalOutputPreviewSize: z14.enum(["small", "medium", "large"]).optional(),
6411
6046
  terminalShellIntegrationTimeout: z14.number().optional(),
6412
6047
  terminalShellIntegrationDisabled: z14.boolean().optional(),
6413
6048
  terminalCommandDelay: z14.number().optional(),
@@ -6416,18 +6051,14 @@ var globalSettingsSchema = z14.object({
6416
6051
  terminalZshOhMy: z14.boolean().optional(),
6417
6052
  terminalZshP10k: z14.boolean().optional(),
6418
6053
  terminalZdotdir: z14.boolean().optional(),
6419
- terminalCompressProgressBar: z14.boolean().optional(),
6420
6054
  diagnosticsEnabled: z14.boolean().optional(),
6421
6055
  rateLimitSeconds: z14.number().optional(),
6422
- diffEnabled: z14.boolean().optional(),
6423
- fuzzyMatchThreshold: z14.number().optional(),
6424
6056
  experiments: experimentsSchema.optional(),
6425
6057
  codebaseIndexModels: codebaseIndexModelsSchema.optional(),
6426
6058
  codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
6427
6059
  language: languagesSchema.optional(),
6428
6060
  telemetrySetting: telemetrySettingsSchema.optional(),
6429
6061
  mcpEnabled: z14.boolean().optional(),
6430
- enableMcpServerCreation: z14.boolean().optional(),
6431
6062
  mode: z14.string().optional(),
6432
6063
  modeApiConfigs: z14.record(z14.string(), z14.string()).optional(),
6433
6064
  customModes: z14.array(modeConfigSchema).optional(),
@@ -6447,7 +6078,20 @@ var globalSettingsSchema = z14.object({
6447
6078
  profileThresholds: z14.record(z14.string(), z14.number()).optional(),
6448
6079
  hasOpenedModeSelector: z14.boolean().optional(),
6449
6080
  lastModeExportPath: z14.string().optional(),
6450
- lastModeImportPath: z14.string().optional()
6081
+ lastModeImportPath: z14.string().optional(),
6082
+ lastSettingsExportPath: z14.string().optional(),
6083
+ lastTaskExportPath: z14.string().optional(),
6084
+ lastImageSavePath: z14.string().optional(),
6085
+ /**
6086
+ * Path to worktree to auto-open after switching workspaces.
6087
+ * Used by the worktree feature to open the Roo Code sidebar in a new window.
6088
+ */
6089
+ worktreeAutoOpenPath: z14.string().optional(),
6090
+ /**
6091
+ * Whether to show the worktree selector in the home screen.
6092
+ * @default true
6093
+ */
6094
+ showWorktreesInHomeScreen: z14.boolean().optional()
6451
6095
  });
6452
6096
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
6453
6097
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -6531,8 +6175,6 @@ var EVALS_SETTINGS = {
6531
6175
  ttsSpeed: 1,
6532
6176
  soundEnabled: false,
6533
6177
  soundVolume: 0.5,
6534
- terminalOutputLineLimit: 500,
6535
- terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6536
6178
  terminalShellIntegrationTimeout: 3e4,
6537
6179
  terminalCommandDelay: 0,
6538
6180
  terminalPowershellCounter: false,
@@ -6540,19 +6182,14 @@ var EVALS_SETTINGS = {
6540
6182
  terminalZshClearEolMark: true,
6541
6183
  terminalZshP10k: false,
6542
6184
  terminalZdotdir: true,
6543
- terminalCompressProgressBar: true,
6544
6185
  terminalShellIntegrationDisabled: true,
6545
6186
  diagnosticsEnabled: true,
6546
- diffEnabled: true,
6547
- fuzzyMatchThreshold: 1,
6548
6187
  enableCheckpoints: false,
6549
6188
  rateLimitSeconds: 0,
6550
6189
  maxOpenTabsContext: 20,
6551
6190
  maxWorkspaceFiles: 200,
6552
6191
  maxGitStatusFiles: 20,
6553
6192
  showRooIgnoredFiles: true,
6554
- maxReadFileLine: -1,
6555
- // -1 to enable full file reading.
6556
6193
  includeDiagnosticMessages: true,
6557
6194
  maxDiagnosticMessages: 50,
6558
6195
  language: "en",
@@ -6626,24 +6263,18 @@ var organizationAllowListSchema = z16.object({
6626
6263
  });
6627
6264
  var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6628
6265
  enableCheckpoints: true,
6629
- fuzzyMatchThreshold: true,
6630
6266
  maxOpenTabsContext: true,
6631
- maxReadFileLine: true,
6632
6267
  maxWorkspaceFiles: true,
6633
6268
  showRooIgnoredFiles: true,
6634
6269
  terminalCommandDelay: true,
6635
- terminalCompressProgressBar: true,
6636
- terminalOutputLineLimit: true,
6637
6270
  terminalShellIntegrationDisabled: true,
6638
6271
  terminalShellIntegrationTimeout: true,
6639
6272
  terminalZshClearEolMark: true
6640
6273
  }).merge(
6641
6274
  z16.object({
6642
6275
  maxOpenTabsContext: z16.number().int().nonnegative().optional(),
6643
- maxReadFileLine: z16.number().int().gte(-1).optional(),
6644
6276
  maxWorkspaceFiles: z16.number().int().nonnegative().optional(),
6645
6277
  terminalCommandDelay: z16.number().int().nonnegative().optional(),
6646
- terminalOutputLineLimit: z16.number().int().nonnegative().optional(),
6647
6278
  terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
6648
6279
  })
6649
6280
  );
@@ -7118,12 +6749,10 @@ var taskCommandSchema = z19.discriminatedUnion("commandName", [
7118
6749
  })
7119
6750
  }),
7120
6751
  z19.object({
7121
- commandName: z19.literal("CancelTask" /* CancelTask */),
7122
- data: z19.string()
6752
+ commandName: z19.literal("CancelTask" /* CancelTask */)
7123
6753
  }),
7124
6754
  z19.object({
7125
- commandName: z19.literal("CloseTask" /* CloseTask */),
7126
- data: z19.string()
6755
+ commandName: z19.literal("CloseTask" /* CloseTask */)
7127
6756
  }),
7128
6757
  z19.object({
7129
6758
  commandName: z19.literal("ResumeTask" /* ResumeTask */),
@@ -7159,6 +6788,7 @@ var ipcMessageSchema = z19.discriminatedUnion("type", [
7159
6788
 
7160
6789
  // src/mcp.ts
7161
6790
  import { z as z20 } from "zod";
6791
+ var MAX_MCP_TOOLS_THRESHOLD = 60;
7162
6792
  var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7163
6793
  z20.object({
7164
6794
  executionId: z20.string(),
@@ -7182,6 +6812,46 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7182
6812
  error: z20.string().optional()
7183
6813
  })
7184
6814
  ]);
6815
+ function countEnabledMcpTools(servers) {
6816
+ let serverCount = 0;
6817
+ let toolCount = 0;
6818
+ for (const server of servers) {
6819
+ if (server.disabled) continue;
6820
+ if (server.status !== "connected") continue;
6821
+ serverCount++;
6822
+ if (server.tools) {
6823
+ for (const tool of server.tools) {
6824
+ if (tool.enabledForPrompt !== false) {
6825
+ toolCount++;
6826
+ }
6827
+ }
6828
+ }
6829
+ }
6830
+ return { enabledToolCount: toolCount, enabledServerCount: serverCount };
6831
+ }
6832
+
6833
+ // src/skills.ts
6834
+ var SKILL_NAME_MIN_LENGTH = 1;
6835
+ var SKILL_NAME_MAX_LENGTH = 64;
6836
+ var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
6837
+ var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
6838
+ SkillNameValidationError2["Empty"] = "empty";
6839
+ SkillNameValidationError2["TooLong"] = "too_long";
6840
+ SkillNameValidationError2["InvalidFormat"] = "invalid_format";
6841
+ return SkillNameValidationError2;
6842
+ })(SkillNameValidationError || {});
6843
+ function validateSkillName(name) {
6844
+ if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
6845
+ return { valid: false, error: "empty" /* Empty */ };
6846
+ }
6847
+ if (name.length > SKILL_NAME_MAX_LENGTH) {
6848
+ return { valid: false, error: "too_long" /* TooLong */ };
6849
+ }
6850
+ if (!SKILL_NAME_REGEX.test(name)) {
6851
+ return { valid: false, error: "invalid_format" /* InvalidFormat */ };
6852
+ }
6853
+ return { valid: true };
6854
+ }
7185
6855
 
7186
6856
  // src/todo.ts
7187
6857
  import { z as z21 } from "zod";
@@ -7221,6 +6891,11 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
7221
6891
  })
7222
6892
  ]);
7223
6893
 
6894
+ // src/tool-params.ts
6895
+ function isLegacyReadFileParams(params) {
6896
+ return "_legacyFormat" in params && params._legacyFormat === true;
6897
+ }
6898
+
7224
6899
  // src/vscode-extension-host.ts
7225
6900
  import { z as z23 } from "zod";
7226
6901
  var checkoutDiffPayloadSchema = z23.object({
@@ -7273,7 +6948,7 @@ export {
7273
6948
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
7274
6949
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
7275
6950
  DEFAULT_MODES,
7276
- DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6951
+ DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
7277
6952
  DEFAULT_WRITE_DELAY_MS,
7278
6953
  DOUBAO_API_BASE_URL,
7279
6954
  DOUBAO_API_CHAT_PATH,
@@ -7303,13 +6978,13 @@ export {
7303
6978
  IpcOrigin,
7304
6979
  LMSTUDIO_DEFAULT_TEMPERATURE,
7305
6980
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6981
+ MAX_MCP_TOOLS_THRESHOLD,
7306
6982
  MINIMAX_DEFAULT_MAX_TOKENS,
7307
6983
  MINIMAX_DEFAULT_TEMPERATURE,
7308
6984
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
7309
6985
  MISTRAL_DEFAULT_TEMPERATURE,
7310
6986
  MODELS_BY_PROVIDER,
7311
6987
  MOONSHOT_DEFAULT_TEMPERATURE,
7312
- NATIVE_TOOL_DEFAULTS,
7313
6988
  OPENAI_AZURE_AI_INFERENCE_PATH,
7314
6989
  OPENAI_NATIVE_DEFAULT_TEMPERATURE,
7315
6990
  OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -7324,7 +6999,11 @@ export {
7324
6999
  RooModelsResponseSchema,
7325
7000
  RooPricingSchema,
7326
7001
  SECRET_STATE_KEYS,
7327
- TOOL_PROTOCOL,
7002
+ SKILL_NAME_MAX_LENGTH,
7003
+ SKILL_NAME_MIN_LENGTH,
7004
+ SKILL_NAME_REGEX,
7005
+ SkillNameValidationError,
7006
+ TERMINAL_PREVIEW_BYTES,
7328
7007
  TaskBridgeCommandName,
7329
7008
  TaskBridgeEventName,
7330
7009
  TaskCommandName,
@@ -7356,9 +7035,6 @@ export {
7356
7035
  chutesDefaultModelId,
7357
7036
  chutesDefaultModelInfo,
7358
7037
  chutesModels,
7359
- claudeCodeDefaultModelId,
7360
- claudeCodeModels,
7361
- claudeCodeReasoningConfig,
7362
7038
  clineAskSchema,
7363
7039
  clineAsks,
7364
7040
  clineMessageSchema,
@@ -7373,6 +7049,7 @@ export {
7373
7049
  commandIds,
7374
7050
  contextCondenseSchema,
7375
7051
  contextTruncationSchema,
7052
+ countEnabledMcpTools,
7376
7053
  customModePromptsSchema,
7377
7054
  customModesSettingsSchema,
7378
7055
  customProviders,
@@ -7406,7 +7083,6 @@ export {
7406
7083
  geminiDefaultModelId,
7407
7084
  geminiModels,
7408
7085
  getApiProtocol,
7409
- getEffectiveProtocol,
7410
7086
  getErrorMessage,
7411
7087
  getErrorStatusCode,
7412
7088
  getImageGenerationProvider,
@@ -7441,9 +7117,9 @@ export {
7441
7117
  isInteractiveAsk,
7442
7118
  isInternalProvider,
7443
7119
  isLanguage,
7120
+ isLegacyReadFileParams,
7444
7121
  isLocalProvider,
7445
7122
  isModelParameter,
7446
- isNativeProtocol,
7447
7123
  isNonBlockingAsk,
7448
7124
  isProviderName,
7449
7125
  isResumableAsk,
@@ -7479,7 +7155,6 @@ export {
7479
7155
  moonshotDefaultModelId,
7480
7156
  moonshotModels,
7481
7157
  nonBlockingAsks,
7482
- normalizeClaudeCodeModelId,
7483
7158
  ollamaDefaultModelId,
7484
7159
  ollamaDefaultModelInfo,
7485
7160
  openAiCodexDefaultModelId,
@@ -7553,6 +7228,7 @@ export {
7553
7228
  userFeaturesSchema,
7554
7229
  userSettingsConfigSchema,
7555
7230
  userSettingsDataSchema,
7231
+ validateSkillName,
7556
7232
  verbosityLevels,
7557
7233
  verbosityLevelsSchema,
7558
7234
  vercelAiGatewayDefaultModelId,