@roo-code/types 1.105.0 → 1.107.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
package/dist/index.js CHANGED
@@ -78,7 +78,9 @@ var clineSays = [
78
78
  "condense_context_error",
79
79
  "sliding_window_truncation",
80
80
  "codebase_search_result",
81
- "user_edit_todos"
81
+ "user_edit_todos",
82
+ "too_many_tools_warning",
83
+ "tool"
82
84
  ];
83
85
  var clineSaySchema = z.enum(clineSays);
84
86
  var toolProgressStatusSchema = z.object({
@@ -146,6 +148,7 @@ var toolGroupsSchema = z2.enum(toolGroups);
146
148
  var toolNames = [
147
149
  "execute_command",
148
150
  "read_file",
151
+ "read_command_output",
149
152
  "write_to_file",
150
153
  "apply_diff",
151
154
  "search_and_replace",
@@ -161,10 +164,10 @@ var toolNames = [
161
164
  "attempt_completion",
162
165
  "switch_mode",
163
166
  "new_task",
164
- "fetch_instructions",
165
167
  "codebase_search",
166
168
  "update_todo_list",
167
169
  "run_slash_command",
170
+ "skill",
168
171
  "generate_image",
169
172
  "custom_tool"
170
173
  ];
@@ -176,20 +179,6 @@ var toolUsageSchema = z2.record(
176
179
  failures: z2.number()
177
180
  })
178
181
  );
179
- var TOOL_PROTOCOL = {
180
- XML: "xml",
181
- NATIVE: "native"
182
- };
183
- var NATIVE_TOOL_DEFAULTS = {
184
- supportsNativeTools: true,
185
- defaultToolProtocol: TOOL_PROTOCOL.NATIVE
186
- };
187
- function isNativeProtocol(protocol) {
188
- return protocol === TOOL_PROTOCOL.NATIVE;
189
- }
190
- function getEffectiveProtocol(toolProtocol) {
191
- return toolProtocol || TOOL_PROTOCOL.XML;
192
- }
193
182
 
194
183
  // src/events.ts
195
184
  var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
@@ -213,6 +202,7 @@ var RooCodeEventName = /* @__PURE__ */ ((RooCodeEventName2) => {
213
202
  RooCodeEventName2["TaskModeSwitched"] = "taskModeSwitched";
214
203
  RooCodeEventName2["TaskAskResponded"] = "taskAskResponded";
215
204
  RooCodeEventName2["TaskUserMessage"] = "taskUserMessage";
205
+ RooCodeEventName2["QueuedMessagesUpdated"] = "queuedMessagesUpdated";
216
206
  RooCodeEventName2["TaskTokenUsageUpdated"] = "taskTokenUsageUpdated";
217
207
  RooCodeEventName2["TaskToolFailed"] = "taskToolFailed";
218
208
  RooCodeEventName2["ModeChanged"] = "modeChanged";
@@ -272,6 +262,7 @@ var rooCodeEventsSchema = z3.object({
272
262
  ["taskModeSwitched" /* TaskModeSwitched */]: z3.tuple([z3.string(), z3.string()]),
273
263
  ["taskAskResponded" /* TaskAskResponded */]: z3.tuple([z3.string()]),
274
264
  ["taskUserMessage" /* TaskUserMessage */]: z3.tuple([z3.string()]),
265
+ ["queuedMessagesUpdated" /* QueuedMessagesUpdated */]: z3.tuple([z3.string(), z3.array(queuedMessageSchema)]),
275
266
  ["taskToolFailed" /* TaskToolFailed */]: z3.tuple([z3.string(), toolNamesSchema, z3.string()]),
276
267
  ["taskTokenUsageUpdated" /* TaskTokenUsageUpdated */]: z3.tuple([z3.string(), tokenUsageSchema, toolUsageSchema]),
277
268
  ["modeChanged" /* ModeChanged */]: z3.tuple([z3.string()]),
@@ -377,6 +368,11 @@ var taskEventSchema = z3.discriminatedUnion("eventName", [
377
368
  payload: rooCodeEventsSchema.shape["taskAskResponded" /* TaskAskResponded */],
378
369
  taskId: z3.number().optional()
379
370
  }),
371
+ z3.object({
372
+ eventName: z3.literal("queuedMessagesUpdated" /* QueuedMessagesUpdated */),
373
+ payload: rooCodeEventsSchema.shape["queuedMessagesUpdated" /* QueuedMessagesUpdated */],
374
+ taskId: z3.number().optional()
375
+ }),
380
376
  // Task Analytics
381
377
  z3.object({
382
378
  eventName: z3.literal("taskToolFailed" /* TaskToolFailed */),
@@ -477,10 +473,6 @@ var modelInfoSchema = z5.object({
477
473
  isStealthModel: z5.boolean().optional(),
478
474
  // Flag to indicate if the model is free (no cost)
479
475
  isFree: z5.boolean().optional(),
480
- // Flag to indicate if the model supports native tool calling (OpenAI-style function calling)
481
- supportsNativeTools: z5.boolean().optional(),
482
- // Default tool protocol preferred by this model (if not specified, falls back to capability/provider defaults)
483
- defaultToolProtocol: z5.enum(["xml", "native"]).optional(),
484
476
  // Exclude specific native tools from being available (only applies to native protocol)
485
477
  // These tools will be removed from the set of tools available to the model
486
478
  excludedTools: z5.array(z5.string()).optional(),
@@ -577,8 +569,6 @@ var anthropicModels = {
577
569
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
578
570
  supportsImages: true,
579
571
  supportsPromptCache: true,
580
- supportsNativeTools: true,
581
- defaultToolProtocol: "native",
582
572
  inputPrice: 3,
583
573
  // $3 per million input tokens (≤200K context)
584
574
  outputPrice: 15,
@@ -611,8 +601,6 @@ var anthropicModels = {
611
601
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
612
602
  supportsImages: true,
613
603
  supportsPromptCache: true,
614
- supportsNativeTools: true,
615
- defaultToolProtocol: "native",
616
604
  inputPrice: 3,
617
605
  // $3 per million input tokens (≤200K context)
618
606
  outputPrice: 15,
@@ -644,8 +632,6 @@ var anthropicModels = {
644
632
  contextWindow: 2e5,
645
633
  supportsImages: true,
646
634
  supportsPromptCache: true,
647
- supportsNativeTools: true,
648
- defaultToolProtocol: "native",
649
635
  inputPrice: 5,
650
636
  // $5 per million input tokens
651
637
  outputPrice: 25,
@@ -662,8 +648,6 @@ var anthropicModels = {
662
648
  contextWindow: 2e5,
663
649
  supportsImages: true,
664
650
  supportsPromptCache: true,
665
- supportsNativeTools: true,
666
- defaultToolProtocol: "native",
667
651
  inputPrice: 15,
668
652
  // $15 per million input tokens
669
653
  outputPrice: 75,
@@ -680,8 +664,6 @@ var anthropicModels = {
680
664
  contextWindow: 2e5,
681
665
  supportsImages: true,
682
666
  supportsPromptCache: true,
683
- supportsNativeTools: true,
684
- defaultToolProtocol: "native",
685
667
  inputPrice: 15,
686
668
  // $15 per million input tokens
687
669
  outputPrice: 75,
@@ -698,8 +680,6 @@ var anthropicModels = {
698
680
  contextWindow: 2e5,
699
681
  supportsImages: true,
700
682
  supportsPromptCache: true,
701
- supportsNativeTools: true,
702
- defaultToolProtocol: "native",
703
683
  inputPrice: 3,
704
684
  // $3 per million input tokens
705
685
  outputPrice: 15,
@@ -717,8 +697,6 @@ var anthropicModels = {
717
697
  contextWindow: 2e5,
718
698
  supportsImages: true,
719
699
  supportsPromptCache: true,
720
- supportsNativeTools: true,
721
- defaultToolProtocol: "native",
722
700
  inputPrice: 3,
723
701
  // $3 per million input tokens
724
702
  outputPrice: 15,
@@ -733,8 +711,6 @@ var anthropicModels = {
733
711
  contextWindow: 2e5,
734
712
  supportsImages: true,
735
713
  supportsPromptCache: true,
736
- supportsNativeTools: true,
737
- defaultToolProtocol: "native",
738
714
  inputPrice: 3,
739
715
  // $3 per million input tokens
740
716
  outputPrice: 15,
@@ -749,8 +725,6 @@ var anthropicModels = {
749
725
  contextWindow: 2e5,
750
726
  supportsImages: false,
751
727
  supportsPromptCache: true,
752
- supportsNativeTools: true,
753
- defaultToolProtocol: "native",
754
728
  inputPrice: 1,
755
729
  outputPrice: 5,
756
730
  cacheWritesPrice: 1.25,
@@ -761,8 +735,6 @@ var anthropicModels = {
761
735
  contextWindow: 2e5,
762
736
  supportsImages: true,
763
737
  supportsPromptCache: true,
764
- supportsNativeTools: true,
765
- defaultToolProtocol: "native",
766
738
  inputPrice: 15,
767
739
  outputPrice: 75,
768
740
  cacheWritesPrice: 18.75,
@@ -773,8 +745,6 @@ var anthropicModels = {
773
745
  contextWindow: 2e5,
774
746
  supportsImages: true,
775
747
  supportsPromptCache: true,
776
- supportsNativeTools: true,
777
- defaultToolProtocol: "native",
778
748
  inputPrice: 0.25,
779
749
  outputPrice: 1.25,
780
750
  cacheWritesPrice: 0.3,
@@ -785,8 +755,6 @@ var anthropicModels = {
785
755
  contextWindow: 2e5,
786
756
  supportsImages: true,
787
757
  supportsPromptCache: true,
788
- supportsNativeTools: true,
789
- defaultToolProtocol: "native",
790
758
  inputPrice: 1,
791
759
  outputPrice: 5,
792
760
  cacheWritesPrice: 1.25,
@@ -804,7 +772,6 @@ var basetenModels = {
804
772
  contextWindow: 262e3,
805
773
  supportsImages: false,
806
774
  supportsPromptCache: false,
807
- supportsNativeTools: true,
808
775
  inputPrice: 0.6,
809
776
  outputPrice: 2.5,
810
777
  cacheWritesPrice: 0,
@@ -816,7 +783,6 @@ var basetenModels = {
816
783
  contextWindow: 2e5,
817
784
  supportsImages: false,
818
785
  supportsPromptCache: false,
819
- supportsNativeTools: true,
820
786
  inputPrice: 0.6,
821
787
  outputPrice: 2.2,
822
788
  cacheWritesPrice: 0,
@@ -828,7 +794,6 @@ var basetenModels = {
828
794
  contextWindow: 163840,
829
795
  supportsImages: false,
830
796
  supportsPromptCache: false,
831
- supportsNativeTools: true,
832
797
  inputPrice: 2.55,
833
798
  outputPrice: 5.95,
834
799
  cacheWritesPrice: 0,
@@ -840,7 +805,6 @@ var basetenModels = {
840
805
  contextWindow: 163840,
841
806
  supportsImages: false,
842
807
  supportsPromptCache: false,
843
- supportsNativeTools: true,
844
808
  inputPrice: 2.55,
845
809
  outputPrice: 5.95,
846
810
  cacheWritesPrice: 0,
@@ -852,7 +816,6 @@ var basetenModels = {
852
816
  contextWindow: 163840,
853
817
  supportsImages: false,
854
818
  supportsPromptCache: false,
855
- supportsNativeTools: true,
856
819
  inputPrice: 0.77,
857
820
  outputPrice: 0.77,
858
821
  cacheWritesPrice: 0,
@@ -864,7 +827,6 @@ var basetenModels = {
864
827
  contextWindow: 163840,
865
828
  supportsImages: false,
866
829
  supportsPromptCache: false,
867
- supportsNativeTools: true,
868
830
  inputPrice: 0.5,
869
831
  outputPrice: 1.5,
870
832
  cacheWritesPrice: 0,
@@ -876,7 +838,6 @@ var basetenModels = {
876
838
  contextWindow: 163840,
877
839
  supportsImages: false,
878
840
  supportsPromptCache: false,
879
- supportsNativeTools: true,
880
841
  inputPrice: 0.3,
881
842
  outputPrice: 0.45,
882
843
  cacheWritesPrice: 0,
@@ -888,7 +849,6 @@ var basetenModels = {
888
849
  contextWindow: 128072,
889
850
  supportsImages: false,
890
851
  supportsPromptCache: false,
891
- supportsNativeTools: true,
892
852
  inputPrice: 0.1,
893
853
  outputPrice: 0.5,
894
854
  cacheWritesPrice: 0,
@@ -900,7 +860,6 @@ var basetenModels = {
900
860
  contextWindow: 262144,
901
861
  supportsImages: false,
902
862
  supportsPromptCache: false,
903
- supportsNativeTools: true,
904
863
  inputPrice: 0.22,
905
864
  outputPrice: 0.8,
906
865
  cacheWritesPrice: 0,
@@ -912,7 +871,6 @@ var basetenModels = {
912
871
  contextWindow: 262144,
913
872
  supportsImages: false,
914
873
  supportsPromptCache: false,
915
- supportsNativeTools: true,
916
874
  inputPrice: 0.38,
917
875
  outputPrice: 1.53,
918
876
  cacheWritesPrice: 0,
@@ -924,7 +882,6 @@ var basetenModels = {
924
882
  contextWindow: 262e3,
925
883
  supportsImages: false,
926
884
  supportsPromptCache: false,
927
- supportsNativeTools: true,
928
885
  inputPrice: 0.6,
929
886
  outputPrice: 2.5,
930
887
  cacheWritesPrice: 0,
@@ -944,8 +901,6 @@ var bedrockModels = {
944
901
  supportsImages: true,
945
902
  supportsPromptCache: true,
946
903
  supportsReasoningBudget: true,
947
- supportsNativeTools: true,
948
- defaultToolProtocol: "native",
949
904
  inputPrice: 3,
950
905
  outputPrice: 15,
951
906
  cacheWritesPrice: 3.75,
@@ -959,7 +914,6 @@ var bedrockModels = {
959
914
  contextWindow: 3e5,
960
915
  supportsImages: true,
961
916
  supportsPromptCache: true,
962
- supportsNativeTools: true,
963
917
  inputPrice: 0.8,
964
918
  outputPrice: 3.2,
965
919
  cacheWritesPrice: 0.8,
@@ -975,7 +929,6 @@ var bedrockModels = {
975
929
  contextWindow: 3e5,
976
930
  supportsImages: true,
977
931
  supportsPromptCache: false,
978
- supportsNativeTools: true,
979
932
  inputPrice: 1,
980
933
  outputPrice: 4,
981
934
  cacheWritesPrice: 1,
@@ -989,7 +942,6 @@ var bedrockModels = {
989
942
  contextWindow: 3e5,
990
943
  supportsImages: true,
991
944
  supportsPromptCache: true,
992
- supportsNativeTools: true,
993
945
  inputPrice: 0.06,
994
946
  outputPrice: 0.24,
995
947
  cacheWritesPrice: 0.06,
@@ -1005,7 +957,6 @@ var bedrockModels = {
1005
957
  contextWindow: 1e6,
1006
958
  supportsImages: true,
1007
959
  supportsPromptCache: true,
1008
- supportsNativeTools: true,
1009
960
  inputPrice: 0.33,
1010
961
  outputPrice: 2.75,
1011
962
  cacheWritesPrice: 0,
@@ -1021,7 +972,6 @@ var bedrockModels = {
1021
972
  contextWindow: 128e3,
1022
973
  supportsImages: false,
1023
974
  supportsPromptCache: true,
1024
- supportsNativeTools: true,
1025
975
  inputPrice: 0.035,
1026
976
  outputPrice: 0.14,
1027
977
  cacheWritesPrice: 0.035,
@@ -1038,8 +988,6 @@ var bedrockModels = {
1038
988
  supportsImages: true,
1039
989
  supportsPromptCache: true,
1040
990
  supportsReasoningBudget: true,
1041
- supportsNativeTools: true,
1042
- defaultToolProtocol: "native",
1043
991
  inputPrice: 3,
1044
992
  outputPrice: 15,
1045
993
  cacheWritesPrice: 3.75,
@@ -1054,8 +1002,6 @@ var bedrockModels = {
1054
1002
  supportsImages: true,
1055
1003
  supportsPromptCache: true,
1056
1004
  supportsReasoningBudget: true,
1057
- supportsNativeTools: true,
1058
- defaultToolProtocol: "native",
1059
1005
  inputPrice: 15,
1060
1006
  outputPrice: 75,
1061
1007
  cacheWritesPrice: 18.75,
@@ -1070,8 +1016,6 @@ var bedrockModels = {
1070
1016
  supportsImages: true,
1071
1017
  supportsPromptCache: true,
1072
1018
  supportsReasoningBudget: true,
1073
- supportsNativeTools: true,
1074
- defaultToolProtocol: "native",
1075
1019
  inputPrice: 5,
1076
1020
  outputPrice: 25,
1077
1021
  cacheWritesPrice: 6.25,
@@ -1086,8 +1030,6 @@ var bedrockModels = {
1086
1030
  supportsImages: true,
1087
1031
  supportsPromptCache: true,
1088
1032
  supportsReasoningBudget: true,
1089
- supportsNativeTools: true,
1090
- defaultToolProtocol: "native",
1091
1033
  inputPrice: 15,
1092
1034
  outputPrice: 75,
1093
1035
  cacheWritesPrice: 18.75,
@@ -1102,8 +1044,6 @@ var bedrockModels = {
1102
1044
  supportsImages: true,
1103
1045
  supportsPromptCache: true,
1104
1046
  supportsReasoningBudget: true,
1105
- supportsNativeTools: true,
1106
- defaultToolProtocol: "native",
1107
1047
  inputPrice: 3,
1108
1048
  outputPrice: 15,
1109
1049
  cacheWritesPrice: 3.75,
@@ -1117,8 +1057,6 @@ var bedrockModels = {
1117
1057
  contextWindow: 2e5,
1118
1058
  supportsImages: true,
1119
1059
  supportsPromptCache: true,
1120
- supportsNativeTools: true,
1121
- defaultToolProtocol: "native",
1122
1060
  inputPrice: 3,
1123
1061
  outputPrice: 15,
1124
1062
  cacheWritesPrice: 3.75,
@@ -1132,8 +1070,6 @@ var bedrockModels = {
1132
1070
  contextWindow: 2e5,
1133
1071
  supportsImages: false,
1134
1072
  supportsPromptCache: true,
1135
- supportsNativeTools: true,
1136
- defaultToolProtocol: "native",
1137
1073
  inputPrice: 0.8,
1138
1074
  outputPrice: 4,
1139
1075
  cacheWritesPrice: 1,
@@ -1148,8 +1084,6 @@ var bedrockModels = {
1148
1084
  supportsImages: true,
1149
1085
  supportsPromptCache: true,
1150
1086
  supportsReasoningBudget: true,
1151
- supportsNativeTools: true,
1152
- defaultToolProtocol: "native",
1153
1087
  inputPrice: 1,
1154
1088
  outputPrice: 5,
1155
1089
  cacheWritesPrice: 1.25,
@@ -1165,8 +1099,6 @@ var bedrockModels = {
1165
1099
  contextWindow: 2e5,
1166
1100
  supportsImages: true,
1167
1101
  supportsPromptCache: false,
1168
- supportsNativeTools: true,
1169
- defaultToolProtocol: "native",
1170
1102
  inputPrice: 3,
1171
1103
  outputPrice: 15
1172
1104
  },
@@ -1175,8 +1107,6 @@ var bedrockModels = {
1175
1107
  contextWindow: 2e5,
1176
1108
  supportsImages: true,
1177
1109
  supportsPromptCache: false,
1178
- supportsNativeTools: true,
1179
- defaultToolProtocol: "native",
1180
1110
  inputPrice: 15,
1181
1111
  outputPrice: 75
1182
1112
  },
@@ -1185,8 +1115,6 @@ var bedrockModels = {
1185
1115
  contextWindow: 2e5,
1186
1116
  supportsImages: true,
1187
1117
  supportsPromptCache: false,
1188
- supportsNativeTools: true,
1189
- defaultToolProtocol: "native",
1190
1118
  inputPrice: 3,
1191
1119
  outputPrice: 15
1192
1120
  },
@@ -1195,8 +1123,6 @@ var bedrockModels = {
1195
1123
  contextWindow: 2e5,
1196
1124
  supportsImages: true,
1197
1125
  supportsPromptCache: false,
1198
- supportsNativeTools: true,
1199
- defaultToolProtocol: "native",
1200
1126
  inputPrice: 0.25,
1201
1127
  outputPrice: 1.25
1202
1128
  },
@@ -1205,7 +1131,6 @@ var bedrockModels = {
1205
1131
  contextWindow: 128e3,
1206
1132
  supportsImages: false,
1207
1133
  supportsPromptCache: false,
1208
- supportsNativeTools: true,
1209
1134
  inputPrice: 1.35,
1210
1135
  outputPrice: 5.4
1211
1136
  },
@@ -1214,7 +1139,6 @@ var bedrockModels = {
1214
1139
  contextWindow: 128e3,
1215
1140
  supportsImages: false,
1216
1141
  supportsPromptCache: false,
1217
- supportsNativeTools: true,
1218
1142
  inputPrice: 0.5,
1219
1143
  outputPrice: 1.5,
1220
1144
  description: "GPT-OSS 20B - Optimized for low latency and local/specialized use cases"
@@ -1224,7 +1148,6 @@ var bedrockModels = {
1224
1148
  contextWindow: 128e3,
1225
1149
  supportsImages: false,
1226
1150
  supportsPromptCache: false,
1227
- supportsNativeTools: true,
1228
1151
  inputPrice: 2,
1229
1152
  outputPrice: 6,
1230
1153
  description: "GPT-OSS 120B - Production-ready, general-purpose, high-reasoning model"
@@ -1234,7 +1157,6 @@ var bedrockModels = {
1234
1157
  contextWindow: 128e3,
1235
1158
  supportsImages: false,
1236
1159
  supportsPromptCache: false,
1237
- supportsNativeTools: true,
1238
1160
  inputPrice: 0.72,
1239
1161
  outputPrice: 0.72,
1240
1162
  description: "Llama 3.3 Instruct (70B)"
@@ -1244,7 +1166,6 @@ var bedrockModels = {
1244
1166
  contextWindow: 128e3,
1245
1167
  supportsImages: true,
1246
1168
  supportsPromptCache: false,
1247
- supportsNativeTools: true,
1248
1169
  inputPrice: 0.72,
1249
1170
  outputPrice: 0.72,
1250
1171
  description: "Llama 3.2 Instruct (90B)"
@@ -1254,7 +1175,6 @@ var bedrockModels = {
1254
1175
  contextWindow: 128e3,
1255
1176
  supportsImages: true,
1256
1177
  supportsPromptCache: false,
1257
- supportsNativeTools: true,
1258
1178
  inputPrice: 0.16,
1259
1179
  outputPrice: 0.16,
1260
1180
  description: "Llama 3.2 Instruct (11B)"
@@ -1264,7 +1184,6 @@ var bedrockModels = {
1264
1184
  contextWindow: 128e3,
1265
1185
  supportsImages: false,
1266
1186
  supportsPromptCache: false,
1267
- supportsNativeTools: true,
1268
1187
  inputPrice: 0.15,
1269
1188
  outputPrice: 0.15,
1270
1189
  description: "Llama 3.2 Instruct (3B)"
@@ -1274,7 +1193,6 @@ var bedrockModels = {
1274
1193
  contextWindow: 128e3,
1275
1194
  supportsImages: false,
1276
1195
  supportsPromptCache: false,
1277
- supportsNativeTools: true,
1278
1196
  inputPrice: 0.1,
1279
1197
  outputPrice: 0.1,
1280
1198
  description: "Llama 3.2 Instruct (1B)"
@@ -1284,7 +1202,6 @@ var bedrockModels = {
1284
1202
  contextWindow: 128e3,
1285
1203
  supportsImages: false,
1286
1204
  supportsPromptCache: false,
1287
- supportsNativeTools: true,
1288
1205
  inputPrice: 2.4,
1289
1206
  outputPrice: 2.4,
1290
1207
  description: "Llama 3.1 Instruct (405B)"
@@ -1294,7 +1211,6 @@ var bedrockModels = {
1294
1211
  contextWindow: 128e3,
1295
1212
  supportsImages: false,
1296
1213
  supportsPromptCache: false,
1297
- supportsNativeTools: true,
1298
1214
  inputPrice: 0.72,
1299
1215
  outputPrice: 0.72,
1300
1216
  description: "Llama 3.1 Instruct (70B)"
@@ -1304,7 +1220,6 @@ var bedrockModels = {
1304
1220
  contextWindow: 128e3,
1305
1221
  supportsImages: false,
1306
1222
  supportsPromptCache: false,
1307
- supportsNativeTools: true,
1308
1223
  inputPrice: 0.9,
1309
1224
  outputPrice: 0.9,
1310
1225
  description: "Llama 3.1 Instruct (70B) (w/ latency optimized inference)"
@@ -1314,7 +1229,6 @@ var bedrockModels = {
1314
1229
  contextWindow: 8e3,
1315
1230
  supportsImages: false,
1316
1231
  supportsPromptCache: false,
1317
- supportsNativeTools: true,
1318
1232
  inputPrice: 0.22,
1319
1233
  outputPrice: 0.22,
1320
1234
  description: "Llama 3.1 Instruct (8B)"
@@ -1324,7 +1238,6 @@ var bedrockModels = {
1324
1238
  contextWindow: 8e3,
1325
1239
  supportsImages: false,
1326
1240
  supportsPromptCache: false,
1327
- supportsNativeTools: true,
1328
1241
  inputPrice: 2.65,
1329
1242
  outputPrice: 3.5
1330
1243
  },
@@ -1333,7 +1246,6 @@ var bedrockModels = {
1333
1246
  contextWindow: 4e3,
1334
1247
  supportsImages: false,
1335
1248
  supportsPromptCache: false,
1336
- supportsNativeTools: true,
1337
1249
  inputPrice: 0.3,
1338
1250
  outputPrice: 0.6
1339
1251
  },
@@ -1342,7 +1254,6 @@ var bedrockModels = {
1342
1254
  contextWindow: 8e3,
1343
1255
  supportsImages: false,
1344
1256
  supportsPromptCache: false,
1345
- supportsNativeTools: true,
1346
1257
  inputPrice: 0.15,
1347
1258
  outputPrice: 0.2,
1348
1259
  description: "Amazon Titan Text Lite"
@@ -1352,7 +1263,6 @@ var bedrockModels = {
1352
1263
  contextWindow: 8e3,
1353
1264
  supportsImages: false,
1354
1265
  supportsPromptCache: false,
1355
- supportsNativeTools: true,
1356
1266
  inputPrice: 0.2,
1357
1267
  outputPrice: 0.6,
1358
1268
  description: "Amazon Titan Text Express"
@@ -1362,8 +1272,6 @@ var bedrockModels = {
1362
1272
  contextWindow: 262144,
1363
1273
  supportsImages: false,
1364
1274
  supportsPromptCache: false,
1365
- supportsNativeTools: true,
1366
- defaultToolProtocol: "native",
1367
1275
  preserveReasoning: true,
1368
1276
  inputPrice: 0.6,
1369
1277
  outputPrice: 2.5,
@@ -1374,8 +1282,6 @@ var bedrockModels = {
1374
1282
  contextWindow: 196608,
1375
1283
  supportsImages: false,
1376
1284
  supportsPromptCache: false,
1377
- supportsNativeTools: true,
1378
- defaultToolProtocol: "native",
1379
1285
  preserveReasoning: true,
1380
1286
  inputPrice: 0.3,
1381
1287
  outputPrice: 1.2,
@@ -1386,8 +1292,6 @@ var bedrockModels = {
1386
1292
  contextWindow: 262144,
1387
1293
  supportsImages: false,
1388
1294
  supportsPromptCache: false,
1389
- supportsNativeTools: true,
1390
- defaultToolProtocol: "native",
1391
1295
  inputPrice: 0.15,
1392
1296
  outputPrice: 1.2,
1393
1297
  description: "Qwen3 Next 80B (MoE model with 3B active parameters)"
@@ -1397,8 +1301,6 @@ var bedrockModels = {
1397
1301
  contextWindow: 262144,
1398
1302
  supportsImages: false,
1399
1303
  supportsPromptCache: false,
1400
- supportsNativeTools: true,
1401
- defaultToolProtocol: "native",
1402
1304
  inputPrice: 0.45,
1403
1305
  outputPrice: 1.8,
1404
1306
  description: "Qwen3 Coder 480B (MoE model with 35B active parameters)"
@@ -1489,26 +1391,14 @@ var BEDROCK_SERVICE_TIER_PRICING = {
1489
1391
  // src/providers/cerebras.ts
1490
1392
  var cerebrasDefaultModelId = "gpt-oss-120b";
1491
1393
  var cerebrasModels = {
1492
- "zai-glm-4.6": {
1493
- maxTokens: 16384,
1494
- // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1495
- contextWindow: 131072,
1496
- supportsImages: false,
1497
- supportsPromptCache: false,
1498
- supportsNativeTools: true,
1499
- defaultToolProtocol: "native",
1500
- inputPrice: 0,
1501
- outputPrice: 0,
1502
- description: "Fast general-purpose model on Cerebras (up to 1,000 tokens/s). To be deprecated soon."
1503
- },
1504
1394
  "zai-glm-4.7": {
1505
1395
  maxTokens: 16384,
1506
1396
  // Conservative default to avoid premature rate limiting (Cerebras reserves quota upfront)
1507
1397
  contextWindow: 131072,
1508
1398
  supportsImages: false,
1509
- supportsPromptCache: false,
1510
- supportsNativeTools: true,
1511
- defaultToolProtocol: "native",
1399
+ supportsPromptCache: true,
1400
+ supportsTemperature: true,
1401
+ defaultTemperature: 1,
1512
1402
  inputPrice: 0,
1513
1403
  outputPrice: 0,
1514
1404
  description: "Highly capable general-purpose model on Cerebras (up to 1,000 tokens/s), competitive with leading proprietary models on coding tasks."
@@ -1519,8 +1409,6 @@ var cerebrasModels = {
1519
1409
  contextWindow: 64e3,
1520
1410
  supportsImages: false,
1521
1411
  supportsPromptCache: false,
1522
- supportsNativeTools: true,
1523
- defaultToolProtocol: "native",
1524
1412
  inputPrice: 0,
1525
1413
  outputPrice: 0,
1526
1414
  description: "Intelligent model with ~1400 tokens/s"
@@ -1531,8 +1419,6 @@ var cerebrasModels = {
1531
1419
  contextWindow: 64e3,
1532
1420
  supportsImages: false,
1533
1421
  supportsPromptCache: false,
1534
- supportsNativeTools: true,
1535
- defaultToolProtocol: "native",
1536
1422
  inputPrice: 0,
1537
1423
  outputPrice: 0,
1538
1424
  description: "Powerful model with ~2600 tokens/s"
@@ -1543,8 +1429,6 @@ var cerebrasModels = {
1543
1429
  contextWindow: 64e3,
1544
1430
  supportsImages: false,
1545
1431
  supportsPromptCache: false,
1546
- supportsNativeTools: true,
1547
- defaultToolProtocol: "native",
1548
1432
  inputPrice: 0,
1549
1433
  outputPrice: 0,
1550
1434
  description: "SOTA coding performance with ~2500 tokens/s"
@@ -1555,8 +1439,6 @@ var cerebrasModels = {
1555
1439
  contextWindow: 64e3,
1556
1440
  supportsImages: false,
1557
1441
  supportsPromptCache: false,
1558
- supportsNativeTools: true,
1559
- defaultToolProtocol: "native",
1560
1442
  inputPrice: 0,
1561
1443
  outputPrice: 0,
1562
1444
  description: "OpenAI GPT OSS model with ~2800 tokens/s\n\n\u2022 64K context window\n\u2022 Excels at efficient reasoning across science, math, and coding"
@@ -1571,8 +1453,6 @@ var chutesModels = {
1571
1453
  contextWindow: 163840,
1572
1454
  supportsImages: false,
1573
1455
  supportsPromptCache: false,
1574
- supportsNativeTools: true,
1575
- defaultToolProtocol: "native",
1576
1456
  inputPrice: 0,
1577
1457
  outputPrice: 0,
1578
1458
  description: "DeepSeek R1 0528 model."
@@ -1582,8 +1462,6 @@ var chutesModels = {
1582
1462
  contextWindow: 163840,
1583
1463
  supportsImages: false,
1584
1464
  supportsPromptCache: false,
1585
- supportsNativeTools: true,
1586
- defaultToolProtocol: "native",
1587
1465
  inputPrice: 0,
1588
1466
  outputPrice: 0,
1589
1467
  description: "DeepSeek R1 model."
@@ -1593,8 +1471,6 @@ var chutesModels = {
1593
1471
  contextWindow: 163840,
1594
1472
  supportsImages: false,
1595
1473
  supportsPromptCache: false,
1596
- supportsNativeTools: true,
1597
- defaultToolProtocol: "native",
1598
1474
  inputPrice: 0,
1599
1475
  outputPrice: 0,
1600
1476
  description: "DeepSeek V3 model."
@@ -1604,8 +1480,6 @@ var chutesModels = {
1604
1480
  contextWindow: 163840,
1605
1481
  supportsImages: false,
1606
1482
  supportsPromptCache: false,
1607
- supportsNativeTools: true,
1608
- defaultToolProtocol: "native",
1609
1483
  inputPrice: 0,
1610
1484
  outputPrice: 0,
1611
1485
  description: "DeepSeek V3.1 model."
@@ -1615,8 +1489,6 @@ var chutesModels = {
1615
1489
  contextWindow: 163840,
1616
1490
  supportsImages: false,
1617
1491
  supportsPromptCache: false,
1618
- supportsNativeTools: true,
1619
- defaultToolProtocol: "native",
1620
1492
  inputPrice: 0.23,
1621
1493
  outputPrice: 0.9,
1622
1494
  description: "DeepSeek\u2011V3.1\u2011Terminus is an update to V3.1 that improves language consistency by reducing CN/EN mix\u2011ups and eliminating random characters, while strengthening agent capabilities with notably better Code Agent and Search Agent performance."
@@ -1626,8 +1498,6 @@ var chutesModels = {
1626
1498
  contextWindow: 163840,
1627
1499
  supportsImages: false,
1628
1500
  supportsPromptCache: false,
1629
- supportsNativeTools: true,
1630
- defaultToolProtocol: "native",
1631
1501
  inputPrice: 1,
1632
1502
  outputPrice: 3,
1633
1503
  description: "DeepSeek-V3.1-turbo is an FP8, speculative-decoding turbo variant optimized for ultra-fast single-shot queries (~200 TPS), with outputs close to the originals and solid function calling/reasoning/structured output, priced at $1/M input and $3/M output tokens, using 2\xD7 quota per request and not intended for bulk workloads."
@@ -1637,8 +1507,6 @@ var chutesModels = {
1637
1507
  contextWindow: 163840,
1638
1508
  supportsImages: false,
1639
1509
  supportsPromptCache: false,
1640
- supportsNativeTools: true,
1641
- defaultToolProtocol: "native",
1642
1510
  inputPrice: 0.25,
1643
1511
  outputPrice: 0.35,
1644
1512
  description: "DeepSeek-V3.2-Exp is an experimental LLM that introduces DeepSeek Sparse Attention to improve long\u2011context training and inference efficiency while maintaining performance comparable to V3.1\u2011Terminus."
@@ -1650,8 +1518,6 @@ var chutesModels = {
1650
1518
  // From Groq
1651
1519
  supportsImages: false,
1652
1520
  supportsPromptCache: false,
1653
- supportsNativeTools: true,
1654
- defaultToolProtocol: "native",
1655
1521
  inputPrice: 0,
1656
1522
  outputPrice: 0,
1657
1523
  description: "Unsloth Llama 3.3 70B Instruct model."
@@ -1661,8 +1527,6 @@ var chutesModels = {
1661
1527
  contextWindow: 512e3,
1662
1528
  supportsImages: false,
1663
1529
  supportsPromptCache: false,
1664
- supportsNativeTools: true,
1665
- defaultToolProtocol: "native",
1666
1530
  inputPrice: 0,
1667
1531
  outputPrice: 0,
1668
1532
  description: "ChutesAI Llama 4 Scout 17B Instruct model, 512K context."
@@ -1672,8 +1536,6 @@ var chutesModels = {
1672
1536
  contextWindow: 128e3,
1673
1537
  supportsImages: false,
1674
1538
  supportsPromptCache: false,
1675
- supportsNativeTools: true,
1676
- defaultToolProtocol: "native",
1677
1539
  inputPrice: 0,
1678
1540
  outputPrice: 0,
1679
1541
  description: "Unsloth Mistral Nemo Instruct model."
@@ -1683,8 +1545,6 @@ var chutesModels = {
1683
1545
  contextWindow: 131072,
1684
1546
  supportsImages: false,
1685
1547
  supportsPromptCache: false,
1686
- supportsNativeTools: true,
1687
- defaultToolProtocol: "native",
1688
1548
  inputPrice: 0,
1689
1549
  outputPrice: 0,
1690
1550
  description: "Unsloth Gemma 3 12B IT model."
@@ -1694,8 +1554,6 @@ var chutesModels = {
1694
1554
  contextWindow: 131072,
1695
1555
  supportsImages: false,
1696
1556
  supportsPromptCache: false,
1697
- supportsNativeTools: true,
1698
- defaultToolProtocol: "native",
1699
1557
  inputPrice: 0,
1700
1558
  outputPrice: 0,
1701
1559
  description: "Nous DeepHermes 3 Llama 3 8B Preview model."
@@ -1705,8 +1563,6 @@ var chutesModels = {
1705
1563
  contextWindow: 131072,
1706
1564
  supportsImages: false,
1707
1565
  supportsPromptCache: false,
1708
- supportsNativeTools: true,
1709
- defaultToolProtocol: "native",
1710
1566
  inputPrice: 0,
1711
1567
  outputPrice: 0,
1712
1568
  description: "Unsloth Gemma 3 4B IT model."
@@ -1716,8 +1572,6 @@ var chutesModels = {
1716
1572
  contextWindow: 131072,
1717
1573
  supportsImages: false,
1718
1574
  supportsPromptCache: false,
1719
- supportsNativeTools: true,
1720
- defaultToolProtocol: "native",
1721
1575
  inputPrice: 0,
1722
1576
  outputPrice: 0,
1723
1577
  description: "Nvidia Llama 3.3 Nemotron Super 49B model."
@@ -1727,8 +1581,6 @@ var chutesModels = {
1727
1581
  contextWindow: 131072,
1728
1582
  supportsImages: false,
1729
1583
  supportsPromptCache: false,
1730
- supportsNativeTools: true,
1731
- defaultToolProtocol: "native",
1732
1584
  inputPrice: 0,
1733
1585
  outputPrice: 0,
1734
1586
  description: "Nvidia Llama 3.1 Nemotron Ultra 253B model."
@@ -1738,8 +1590,6 @@ var chutesModels = {
1738
1590
  contextWindow: 256e3,
1739
1591
  supportsImages: false,
1740
1592
  supportsPromptCache: false,
1741
- supportsNativeTools: true,
1742
- defaultToolProtocol: "native",
1743
1593
  inputPrice: 0,
1744
1594
  outputPrice: 0,
1745
1595
  description: "ChutesAI Llama 4 Maverick 17B Instruct FP8 model."
@@ -1749,8 +1599,6 @@ var chutesModels = {
1749
1599
  contextWindow: 163840,
1750
1600
  supportsImages: false,
1751
1601
  supportsPromptCache: false,
1752
- supportsNativeTools: true,
1753
- defaultToolProtocol: "native",
1754
1602
  inputPrice: 0,
1755
1603
  outputPrice: 0,
1756
1604
  description: "DeepSeek V3 Base model."
@@ -1760,8 +1608,6 @@ var chutesModels = {
1760
1608
  contextWindow: 163840,
1761
1609
  supportsImages: false,
1762
1610
  supportsPromptCache: false,
1763
- supportsNativeTools: true,
1764
- defaultToolProtocol: "native",
1765
1611
  inputPrice: 0,
1766
1612
  outputPrice: 0,
1767
1613
  description: "DeepSeek R1 Zero model."
@@ -1771,8 +1617,6 @@ var chutesModels = {
1771
1617
  contextWindow: 163840,
1772
1618
  supportsImages: false,
1773
1619
  supportsPromptCache: false,
1774
- supportsNativeTools: true,
1775
- defaultToolProtocol: "native",
1776
1620
  inputPrice: 0,
1777
1621
  outputPrice: 0,
1778
1622
  description: "DeepSeek V3 (0324) model."
@@ -1782,8 +1626,6 @@ var chutesModels = {
1782
1626
  contextWindow: 262144,
1783
1627
  supportsImages: false,
1784
1628
  supportsPromptCache: false,
1785
- supportsNativeTools: true,
1786
- defaultToolProtocol: "native",
1787
1629
  inputPrice: 0,
1788
1630
  outputPrice: 0,
1789
1631
  description: "Qwen3 235B A22B Instruct 2507 model with 262K context window."
@@ -1793,8 +1635,6 @@ var chutesModels = {
1793
1635
  contextWindow: 40960,
1794
1636
  supportsImages: false,
1795
1637
  supportsPromptCache: false,
1796
- supportsNativeTools: true,
1797
- defaultToolProtocol: "native",
1798
1638
  inputPrice: 0,
1799
1639
  outputPrice: 0,
1800
1640
  description: "Qwen3 235B A22B model."
@@ -1804,8 +1644,6 @@ var chutesModels = {
1804
1644
  contextWindow: 40960,
1805
1645
  supportsImages: false,
1806
1646
  supportsPromptCache: false,
1807
- supportsNativeTools: true,
1808
- defaultToolProtocol: "native",
1809
1647
  inputPrice: 0,
1810
1648
  outputPrice: 0,
1811
1649
  description: "Qwen3 32B model."
@@ -1815,8 +1653,6 @@ var chutesModels = {
1815
1653
  contextWindow: 40960,
1816
1654
  supportsImages: false,
1817
1655
  supportsPromptCache: false,
1818
- supportsNativeTools: true,
1819
- defaultToolProtocol: "native",
1820
1656
  inputPrice: 0,
1821
1657
  outputPrice: 0,
1822
1658
  description: "Qwen3 30B A3B model."
@@ -1826,8 +1662,6 @@ var chutesModels = {
1826
1662
  contextWindow: 40960,
1827
1663
  supportsImages: false,
1828
1664
  supportsPromptCache: false,
1829
- supportsNativeTools: true,
1830
- defaultToolProtocol: "native",
1831
1665
  inputPrice: 0,
1832
1666
  outputPrice: 0,
1833
1667
  description: "Qwen3 14B model."
@@ -1837,8 +1671,6 @@ var chutesModels = {
1837
1671
  contextWindow: 40960,
1838
1672
  supportsImages: false,
1839
1673
  supportsPromptCache: false,
1840
- supportsNativeTools: true,
1841
- defaultToolProtocol: "native",
1842
1674
  inputPrice: 0,
1843
1675
  outputPrice: 0,
1844
1676
  description: "Qwen3 8B model."
@@ -1848,8 +1680,6 @@ var chutesModels = {
1848
1680
  contextWindow: 163840,
1849
1681
  supportsImages: false,
1850
1682
  supportsPromptCache: false,
1851
- supportsNativeTools: true,
1852
- defaultToolProtocol: "native",
1853
1683
  inputPrice: 0,
1854
1684
  outputPrice: 0,
1855
1685
  description: "Microsoft MAI-DS-R1 FP8 model."
@@ -1859,8 +1689,6 @@ var chutesModels = {
1859
1689
  contextWindow: 163840,
1860
1690
  supportsImages: false,
1861
1691
  supportsPromptCache: false,
1862
- supportsNativeTools: true,
1863
- defaultToolProtocol: "native",
1864
1692
  inputPrice: 0,
1865
1693
  outputPrice: 0,
1866
1694
  description: "TNGTech DeepSeek R1T Chimera model."
@@ -1870,8 +1698,6 @@ var chutesModels = {
1870
1698
  contextWindow: 151329,
1871
1699
  supportsImages: false,
1872
1700
  supportsPromptCache: false,
1873
- supportsNativeTools: true,
1874
- defaultToolProtocol: "native",
1875
1701
  inputPrice: 0,
1876
1702
  outputPrice: 0,
1877
1703
  description: "GLM-4.5-Air model with 151,329 token context window and 106B total parameters with 12B activated."
@@ -1881,8 +1707,6 @@ var chutesModels = {
1881
1707
  contextWindow: 131072,
1882
1708
  supportsImages: false,
1883
1709
  supportsPromptCache: false,
1884
- supportsNativeTools: true,
1885
- defaultToolProtocol: "native",
1886
1710
  inputPrice: 0,
1887
1711
  outputPrice: 0,
1888
1712
  description: "GLM-4.5-FP8 model with 128k token context window, optimized for agent-based applications with MoE architecture."
@@ -1892,8 +1716,6 @@ var chutesModels = {
1892
1716
  contextWindow: 131072,
1893
1717
  supportsImages: false,
1894
1718
  supportsPromptCache: false,
1895
- supportsNativeTools: true,
1896
- defaultToolProtocol: "native",
1897
1719
  inputPrice: 1,
1898
1720
  outputPrice: 3,
1899
1721
  description: "GLM-4.5-turbo model with 128K token context window, optimized for fast inference."
@@ -1903,8 +1725,6 @@ var chutesModels = {
1903
1725
  contextWindow: 202752,
1904
1726
  supportsImages: false,
1905
1727
  supportsPromptCache: false,
1906
- supportsNativeTools: true,
1907
- defaultToolProtocol: "native",
1908
1728
  inputPrice: 0,
1909
1729
  outputPrice: 0,
1910
1730
  description: "GLM-4.6 introduces major upgrades over GLM-4.5, including a longer 200K-token context window for complex tasks, stronger coding performance in benchmarks and real-world tools (such as Claude Code, Cline, Roo Code, and Kilo Code), improved reasoning with tool use during inference, more capable and efficient agent integration, and refined writing that better matches human style, readability, and natural role-play scenarios."
@@ -1915,8 +1735,6 @@ var chutesModels = {
1915
1735
  contextWindow: 202752,
1916
1736
  supportsImages: false,
1917
1737
  supportsPromptCache: false,
1918
- supportsNativeTools: true,
1919
- defaultToolProtocol: "native",
1920
1738
  inputPrice: 1.15,
1921
1739
  outputPrice: 3.25,
1922
1740
  description: "GLM-4.6-turbo model with 200K-token context window, optimized for fast inference."
@@ -1926,8 +1744,6 @@ var chutesModels = {
1926
1744
  contextWindow: 128e3,
1927
1745
  supportsImages: false,
1928
1746
  supportsPromptCache: false,
1929
- supportsNativeTools: true,
1930
- defaultToolProtocol: "native",
1931
1747
  inputPrice: 0,
1932
1748
  outputPrice: 0,
1933
1749
  description: "LongCat Flash Thinking FP8 model with 128K context window, optimized for complex reasoning and coding tasks."
@@ -1937,8 +1753,6 @@ var chutesModels = {
1937
1753
  contextWindow: 262144,
1938
1754
  supportsImages: false,
1939
1755
  supportsPromptCache: false,
1940
- supportsNativeTools: true,
1941
- defaultToolProtocol: "native",
1942
1756
  inputPrice: 0,
1943
1757
  outputPrice: 0,
1944
1758
  description: "Qwen3 Coder 480B A35B Instruct FP8 model, optimized for coding tasks."
@@ -1948,8 +1762,6 @@ var chutesModels = {
1948
1762
  contextWindow: 75e3,
1949
1763
  supportsImages: false,
1950
1764
  supportsPromptCache: false,
1951
- supportsNativeTools: true,
1952
- defaultToolProtocol: "native",
1953
1765
  inputPrice: 0.1481,
1954
1766
  outputPrice: 0.5926,
1955
1767
  description: "Moonshot AI Kimi K2 Instruct model with 75k context window."
@@ -1959,8 +1771,6 @@ var chutesModels = {
1959
1771
  contextWindow: 262144,
1960
1772
  supportsImages: false,
1961
1773
  supportsPromptCache: false,
1962
- supportsNativeTools: true,
1963
- defaultToolProtocol: "native",
1964
1774
  inputPrice: 0.1999,
1965
1775
  outputPrice: 0.8001,
1966
1776
  description: "Moonshot AI Kimi K2 Instruct 0905 model with 256k context window."
@@ -1970,8 +1780,6 @@ var chutesModels = {
1970
1780
  contextWindow: 262144,
1971
1781
  supportsImages: false,
1972
1782
  supportsPromptCache: false,
1973
- supportsNativeTools: true,
1974
- defaultToolProtocol: "native",
1975
1783
  inputPrice: 0.077968332,
1976
1784
  outputPrice: 0.31202496,
1977
1785
  description: "Qwen3 235B A22B Thinking 2507 model with 262K context window."
@@ -1981,8 +1789,6 @@ var chutesModels = {
1981
1789
  contextWindow: 131072,
1982
1790
  supportsImages: false,
1983
1791
  supportsPromptCache: false,
1984
- supportsNativeTools: true,
1985
- defaultToolProtocol: "native",
1986
1792
  inputPrice: 0,
1987
1793
  outputPrice: 0,
1988
1794
  description: "Fast, stable instruction-tuned model optimized for complex tasks, RAG, and tool use without thinking traces."
@@ -1992,8 +1798,6 @@ var chutesModels = {
1992
1798
  contextWindow: 131072,
1993
1799
  supportsImages: false,
1994
1800
  supportsPromptCache: false,
1995
- supportsNativeTools: true,
1996
- defaultToolProtocol: "native",
1997
1801
  inputPrice: 0,
1998
1802
  outputPrice: 0,
1999
1803
  description: "Reasoning-first model with structured thinking traces for multi-step problems, math proofs, and code synthesis."
@@ -2003,8 +1807,6 @@ var chutesModels = {
2003
1807
  contextWindow: 262144,
2004
1808
  supportsImages: true,
2005
1809
  supportsPromptCache: false,
2006
- supportsNativeTools: true,
2007
- defaultToolProtocol: "native",
2008
1810
  inputPrice: 0.16,
2009
1811
  outputPrice: 0.65,
2010
1812
  description: "Qwen3\u2011VL\u2011235B\u2011A22B\u2011Thinking is an open\u2011weight MoE vision\u2011language model (235B total, ~22B activated) optimized for deliberate multi\u2011step reasoning with strong text\u2011image\u2011video understanding and long\u2011context capabilities."
@@ -2012,73 +1814,6 @@ var chutesModels = {
2012
1814
  };
2013
1815
  var chutesDefaultModelInfo = chutesModels[chutesDefaultModelId];
2014
1816
 
2015
- // src/providers/claude-code.ts
2016
- var DATE_SUFFIX_PATTERN = /-\d{8}$/;
2017
- var claudeCodeModels = {
2018
- "claude-haiku-4-5": {
2019
- maxTokens: 32768,
2020
- contextWindow: 2e5,
2021
- supportsImages: true,
2022
- supportsPromptCache: true,
2023
- supportsNativeTools: true,
2024
- defaultToolProtocol: "native",
2025
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2026
- reasoningEffort: "medium",
2027
- description: "Claude Haiku 4.5 - Fast and efficient with thinking"
2028
- },
2029
- "claude-sonnet-4-5": {
2030
- maxTokens: 32768,
2031
- contextWindow: 2e5,
2032
- supportsImages: true,
2033
- supportsPromptCache: true,
2034
- supportsNativeTools: true,
2035
- defaultToolProtocol: "native",
2036
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2037
- reasoningEffort: "medium",
2038
- description: "Claude Sonnet 4.5 - Balanced performance with thinking"
2039
- },
2040
- "claude-opus-4-5": {
2041
- maxTokens: 32768,
2042
- contextWindow: 2e5,
2043
- supportsImages: true,
2044
- supportsPromptCache: true,
2045
- supportsNativeTools: true,
2046
- defaultToolProtocol: "native",
2047
- supportsReasoningEffort: ["disable", "low", "medium", "high"],
2048
- reasoningEffort: "medium",
2049
- description: "Claude Opus 4.5 - Most capable with thinking"
2050
- }
2051
- };
2052
- var claudeCodeDefaultModelId = "claude-sonnet-4-5";
2053
- var MODEL_FAMILY_PATTERNS = [
2054
- // Opus models (any version) → claude-opus-4-5
2055
- { pattern: /opus/i, target: "claude-opus-4-5" },
2056
- // Haiku models (any version) → claude-haiku-4-5
2057
- { pattern: /haiku/i, target: "claude-haiku-4-5" },
2058
- // Sonnet models (any version) → claude-sonnet-4-5
2059
- { pattern: /sonnet/i, target: "claude-sonnet-4-5" }
2060
- ];
2061
- function normalizeClaudeCodeModelId(modelId) {
2062
- if (Object.hasOwn(claudeCodeModels, modelId)) {
2063
- return modelId;
2064
- }
2065
- const withoutDate = modelId.replace(DATE_SUFFIX_PATTERN, "");
2066
- if (Object.hasOwn(claudeCodeModels, withoutDate)) {
2067
- return withoutDate;
2068
- }
2069
- for (const { pattern, target } of MODEL_FAMILY_PATTERNS) {
2070
- if (pattern.test(modelId)) {
2071
- return target;
2072
- }
2073
- }
2074
- return claudeCodeDefaultModelId;
2075
- }
2076
- var claudeCodeReasoningConfig = {
2077
- low: { budgetTokens: 16e3 },
2078
- medium: { budgetTokens: 32e3 },
2079
- high: { budgetTokens: 64e3 }
2080
- };
2081
-
2082
1817
  // src/providers/deepseek.ts
2083
1818
  var deepSeekDefaultModelId = "deepseek-chat";
2084
1819
  var deepSeekModels = {
@@ -2088,8 +1823,6 @@ var deepSeekModels = {
2088
1823
  contextWindow: 128e3,
2089
1824
  supportsImages: false,
2090
1825
  supportsPromptCache: true,
2091
- supportsNativeTools: true,
2092
- defaultToolProtocol: "native",
2093
1826
  inputPrice: 0.28,
2094
1827
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
2095
1828
  outputPrice: 0.42,
@@ -2106,8 +1839,6 @@ var deepSeekModels = {
2106
1839
  contextWindow: 128e3,
2107
1840
  supportsImages: false,
2108
1841
  supportsPromptCache: true,
2109
- supportsNativeTools: true,
2110
- defaultToolProtocol: "native",
2111
1842
  preserveReasoning: true,
2112
1843
  inputPrice: 0.28,
2113
1844
  // $0.28 per million tokens (cache miss) - Updated Dec 9, 2025
@@ -2130,8 +1861,6 @@ var doubaoModels = {
2130
1861
  contextWindow: 128e3,
2131
1862
  supportsImages: true,
2132
1863
  supportsPromptCache: true,
2133
- supportsNativeTools: true,
2134
- defaultToolProtocol: "native",
2135
1864
  inputPrice: 1e-4,
2136
1865
  // $0.0001 per million tokens (cache miss)
2137
1866
  outputPrice: 4e-4,
@@ -2147,8 +1876,6 @@ var doubaoModels = {
2147
1876
  contextWindow: 128e3,
2148
1877
  supportsImages: true,
2149
1878
  supportsPromptCache: true,
2150
- supportsNativeTools: true,
2151
- defaultToolProtocol: "native",
2152
1879
  inputPrice: 2e-4,
2153
1880
  // $0.0002 per million tokens
2154
1881
  outputPrice: 8e-4,
@@ -2164,8 +1891,6 @@ var doubaoModels = {
2164
1891
  contextWindow: 128e3,
2165
1892
  supportsImages: true,
2166
1893
  supportsPromptCache: true,
2167
- supportsNativeTools: true,
2168
- defaultToolProtocol: "native",
2169
1894
  inputPrice: 15e-5,
2170
1895
  // $0.00015 per million tokens
2171
1896
  outputPrice: 6e-4,
@@ -2188,7 +1913,6 @@ var featherlessModels = {
2188
1913
  contextWindow: 32678,
2189
1914
  supportsImages: false,
2190
1915
  supportsPromptCache: false,
2191
- supportsNativeTools: true,
2192
1916
  inputPrice: 0,
2193
1917
  outputPrice: 0,
2194
1918
  description: "DeepSeek V3 0324 model."
@@ -2198,7 +1922,6 @@ var featherlessModels = {
2198
1922
  contextWindow: 32678,
2199
1923
  supportsImages: false,
2200
1924
  supportsPromptCache: false,
2201
- supportsNativeTools: true,
2202
1925
  inputPrice: 0,
2203
1926
  outputPrice: 0,
2204
1927
  description: "DeepSeek R1 0528 model."
@@ -2208,7 +1931,6 @@ var featherlessModels = {
2208
1931
  contextWindow: 32678,
2209
1932
  supportsImages: false,
2210
1933
  supportsPromptCache: false,
2211
- supportsNativeTools: true,
2212
1934
  inputPrice: 0,
2213
1935
  outputPrice: 0,
2214
1936
  description: "Kimi K2 Instruct model."
@@ -2218,7 +1940,6 @@ var featherlessModels = {
2218
1940
  contextWindow: 32678,
2219
1941
  supportsImages: false,
2220
1942
  supportsPromptCache: false,
2221
- supportsNativeTools: true,
2222
1943
  inputPrice: 0,
2223
1944
  outputPrice: 0,
2224
1945
  description: "GPT-OSS 120B model."
@@ -2228,7 +1949,6 @@ var featherlessModels = {
2228
1949
  contextWindow: 32678,
2229
1950
  supportsImages: false,
2230
1951
  supportsPromptCache: false,
2231
- supportsNativeTools: true,
2232
1952
  inputPrice: 0,
2233
1953
  outputPrice: 0,
2234
1954
  description: "Qwen3 Coder 480B A35B Instruct model."
@@ -2244,8 +1964,6 @@ var fireworksModels = {
2244
1964
  contextWindow: 262144,
2245
1965
  supportsImages: false,
2246
1966
  supportsPromptCache: true,
2247
- supportsNativeTools: true,
2248
- defaultToolProtocol: "native",
2249
1967
  inputPrice: 0.6,
2250
1968
  outputPrice: 2.5,
2251
1969
  cacheReadsPrice: 0.15,
@@ -2256,8 +1974,6 @@ var fireworksModels = {
2256
1974
  contextWindow: 128e3,
2257
1975
  supportsImages: false,
2258
1976
  supportsPromptCache: false,
2259
- supportsNativeTools: true,
2260
- defaultToolProtocol: "native",
2261
1977
  inputPrice: 0.6,
2262
1978
  outputPrice: 2.5,
2263
1979
  description: "Kimi K2 is a state-of-the-art mixture-of-experts (MoE) language model with 32 billion activated parameters and 1 trillion total parameters. Trained with the Muon optimizer, Kimi K2 achieves exceptional performance across frontier knowledge, reasoning, and coding tasks while being meticulously optimized for agentic capabilities."
@@ -2267,7 +1983,6 @@ var fireworksModels = {
2267
1983
  contextWindow: 256e3,
2268
1984
  supportsImages: false,
2269
1985
  supportsPromptCache: true,
2270
- supportsNativeTools: true,
2271
1986
  supportsTemperature: true,
2272
1987
  preserveReasoning: true,
2273
1988
  defaultTemperature: 1,
@@ -2281,8 +1996,6 @@ var fireworksModels = {
2281
1996
  contextWindow: 204800,
2282
1997
  supportsImages: false,
2283
1998
  supportsPromptCache: false,
2284
- supportsNativeTools: true,
2285
- defaultToolProtocol: "native",
2286
1999
  inputPrice: 0.3,
2287
2000
  outputPrice: 1.2,
2288
2001
  description: "MiniMax M2 is a high-performance language model with 204.8K context window, optimized for long-context understanding and generation tasks."
@@ -2292,8 +2005,6 @@ var fireworksModels = {
2292
2005
  contextWindow: 256e3,
2293
2006
  supportsImages: false,
2294
2007
  supportsPromptCache: false,
2295
- supportsNativeTools: true,
2296
- defaultToolProtocol: "native",
2297
2008
  inputPrice: 0.22,
2298
2009
  outputPrice: 0.88,
2299
2010
  description: "Latest Qwen3 thinking model, competitive against the best closed source models in Jul 2025."
@@ -2303,8 +2014,6 @@ var fireworksModels = {
2303
2014
  contextWindow: 256e3,
2304
2015
  supportsImages: false,
2305
2016
  supportsPromptCache: false,
2306
- supportsNativeTools: true,
2307
- defaultToolProtocol: "native",
2308
2017
  inputPrice: 0.45,
2309
2018
  outputPrice: 1.8,
2310
2019
  description: "Qwen3's most agentic code model to date."
@@ -2314,8 +2023,6 @@ var fireworksModels = {
2314
2023
  contextWindow: 16e4,
2315
2024
  supportsImages: false,
2316
2025
  supportsPromptCache: false,
2317
- supportsNativeTools: true,
2318
- defaultToolProtocol: "native",
2319
2026
  inputPrice: 3,
2320
2027
  outputPrice: 8,
2321
2028
  description: "05/28 updated checkpoint of Deepseek R1. Its overall performance is now approaching that of leading models, such as O3 and Gemini 2.5 Pro. Compared to the previous version, the upgraded model shows significant improvements in handling complex reasoning tasks, and this version also offers a reduced hallucination rate, enhanced support for function calling, and better experience for vibe coding. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2325,8 +2032,6 @@ var fireworksModels = {
2325
2032
  contextWindow: 128e3,
2326
2033
  supportsImages: false,
2327
2034
  supportsPromptCache: false,
2328
- supportsNativeTools: true,
2329
- defaultToolProtocol: "native",
2330
2035
  inputPrice: 0.9,
2331
2036
  outputPrice: 0.9,
2332
2037
  description: "A strong Mixture-of-Experts (MoE) language model with 671B total parameters with 37B activated for each token from Deepseek. Note that fine-tuning for this model is only available through contacting fireworks at https://fireworks.ai/company/contact-us."
@@ -2336,8 +2041,6 @@ var fireworksModels = {
2336
2041
  contextWindow: 163840,
2337
2042
  supportsImages: false,
2338
2043
  supportsPromptCache: false,
2339
- supportsNativeTools: true,
2340
- defaultToolProtocol: "native",
2341
2044
  inputPrice: 0.56,
2342
2045
  outputPrice: 1.68,
2343
2046
  description: "DeepSeek v3.1 is an improved version of the v3 model with enhanced performance, better reasoning capabilities, and improved code generation. This Mixture-of-Experts (MoE) model maintains the same 671B total parameters with 37B activated per token."
@@ -2347,8 +2050,6 @@ var fireworksModels = {
2347
2050
  contextWindow: 128e3,
2348
2051
  supportsImages: false,
2349
2052
  supportsPromptCache: false,
2350
- supportsNativeTools: true,
2351
- defaultToolProtocol: "native",
2352
2053
  inputPrice: 0.55,
2353
2054
  outputPrice: 2.19,
2354
2055
  description: "Z.ai GLM-4.5 with 355B total parameters and 32B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2358,8 +2059,6 @@ var fireworksModels = {
2358
2059
  contextWindow: 128e3,
2359
2060
  supportsImages: false,
2360
2061
  supportsPromptCache: false,
2361
- supportsNativeTools: true,
2362
- defaultToolProtocol: "native",
2363
2062
  inputPrice: 0.55,
2364
2063
  outputPrice: 2.19,
2365
2064
  description: "Z.ai GLM-4.5-Air with 106B total parameters and 12B active parameters. Features unified reasoning, coding, and intelligent agent capabilities."
@@ -2369,8 +2068,6 @@ var fireworksModels = {
2369
2068
  contextWindow: 198e3,
2370
2069
  supportsImages: false,
2371
2070
  supportsPromptCache: false,
2372
- supportsNativeTools: true,
2373
- defaultToolProtocol: "native",
2374
2071
  inputPrice: 0.55,
2375
2072
  outputPrice: 2.19,
2376
2073
  description: "Z.ai GLM-4.6 is an advanced coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality, making it ideal for software development workflows."
@@ -2380,8 +2077,6 @@ var fireworksModels = {
2380
2077
  contextWindow: 128e3,
2381
2078
  supportsImages: false,
2382
2079
  supportsPromptCache: false,
2383
- supportsNativeTools: true,
2384
- defaultToolProtocol: "native",
2385
2080
  inputPrice: 0.07,
2386
2081
  outputPrice: 0.3,
2387
2082
  description: "OpenAI gpt-oss-20b: Compact model for local/edge deployments. Optimized for low-latency and resource-constrained environments with chain-of-thought output, adjustable reasoning, and agentic workflows."
@@ -2391,11 +2086,63 @@ var fireworksModels = {
2391
2086
  contextWindow: 128e3,
2392
2087
  supportsImages: false,
2393
2088
  supportsPromptCache: false,
2394
- supportsNativeTools: true,
2395
- defaultToolProtocol: "native",
2396
2089
  inputPrice: 0.15,
2397
2090
  outputPrice: 0.6,
2398
2091
  description: "OpenAI gpt-oss-120b: Production-grade, general-purpose model that fits on a single H100 GPU. Features complex reasoning, configurable effort, full chain-of-thought transparency, and supports function calling, tool use, and structured outputs."
2092
+ },
2093
+ "accounts/fireworks/models/minimax-m2p1": {
2094
+ maxTokens: 4096,
2095
+ contextWindow: 204800,
2096
+ supportsImages: false,
2097
+ supportsPromptCache: false,
2098
+ inputPrice: 0.3,
2099
+ outputPrice: 1.2,
2100
+ description: "MiniMax M2.1 is an upgraded version of M2 with improved performance on complex reasoning, coding, and long-context understanding tasks."
2101
+ },
2102
+ "accounts/fireworks/models/deepseek-v3p2": {
2103
+ maxTokens: 16384,
2104
+ contextWindow: 163840,
2105
+ supportsImages: false,
2106
+ supportsPromptCache: false,
2107
+ inputPrice: 0.56,
2108
+ outputPrice: 1.68,
2109
+ description: "DeepSeek V3.2 is the latest iteration of the V3 model family with enhanced reasoning capabilities, improved code generation, and better instruction following."
2110
+ },
2111
+ "accounts/fireworks/models/glm-4p7": {
2112
+ maxTokens: 25344,
2113
+ contextWindow: 198e3,
2114
+ supportsImages: false,
2115
+ supportsPromptCache: false,
2116
+ inputPrice: 0.55,
2117
+ outputPrice: 2.19,
2118
+ description: "Z.ai GLM-4.7 is the latest coding model with exceptional performance on complex programming tasks. Features improved reasoning capabilities and enhanced code generation quality."
2119
+ },
2120
+ "accounts/fireworks/models/llama-v3p3-70b-instruct": {
2121
+ maxTokens: 16384,
2122
+ contextWindow: 131072,
2123
+ supportsImages: false,
2124
+ supportsPromptCache: false,
2125
+ inputPrice: 0.9,
2126
+ outputPrice: 0.9,
2127
+ description: "Meta Llama 3.3 70B Instruct is a highly capable instruction-tuned model with strong reasoning, coding, and general task performance."
2128
+ },
2129
+ "accounts/fireworks/models/llama4-maverick-instruct-basic": {
2130
+ maxTokens: 16384,
2131
+ contextWindow: 131072,
2132
+ supportsImages: true,
2133
+ supportsPromptCache: false,
2134
+ inputPrice: 0.22,
2135
+ outputPrice: 0.88,
2136
+ description: "Llama 4 Maverick is Meta's latest multimodal model with vision capabilities, optimized for instruction following and coding tasks."
2137
+ },
2138
+ "accounts/fireworks/models/llama4-scout-instruct-basic": {
2139
+ maxTokens: 16384,
2140
+ contextWindow: 131072,
2141
+ supportsImages: true,
2142
+ supportsPromptCache: false,
2143
+ inputPrice: 0.15,
2144
+ outputPrice: 0.6,
2145
+ description: "Llama 4 Scout is a smaller, faster variant of Llama 4 with multimodal capabilities, ideal for quick iterations and cost-effective deployments."
2399
2146
  }
2400
2147
  };
2401
2148
 
@@ -2406,8 +2153,6 @@ var geminiModels = {
2406
2153
  maxTokens: 65536,
2407
2154
  contextWindow: 1048576,
2408
2155
  supportsImages: true,
2409
- supportsNativeTools: true,
2410
- defaultToolProtocol: "native",
2411
2156
  supportsPromptCache: true,
2412
2157
  supportsReasoningEffort: ["low", "high"],
2413
2158
  reasoningEffort: "low",
@@ -2415,16 +2160,19 @@ var geminiModels = {
2415
2160
  defaultTemperature: 1,
2416
2161
  inputPrice: 4,
2417
2162
  outputPrice: 18,
2163
+ cacheReadsPrice: 0.4,
2418
2164
  tiers: [
2419
2165
  {
2420
2166
  contextWindow: 2e5,
2421
2167
  inputPrice: 2,
2422
- outputPrice: 12
2168
+ outputPrice: 12,
2169
+ cacheReadsPrice: 0.2
2423
2170
  },
2424
2171
  {
2425
2172
  contextWindow: Infinity,
2426
2173
  inputPrice: 4,
2427
- outputPrice: 18
2174
+ outputPrice: 18,
2175
+ cacheReadsPrice: 0.4
2428
2176
  }
2429
2177
  ]
2430
2178
  },
@@ -2432,25 +2180,20 @@ var geminiModels = {
2432
2180
  maxTokens: 65536,
2433
2181
  contextWindow: 1048576,
2434
2182
  supportsImages: true,
2435
- supportsNativeTools: true,
2436
- defaultToolProtocol: "native",
2437
2183
  supportsPromptCache: true,
2438
2184
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
2439
2185
  reasoningEffort: "medium",
2440
2186
  supportsTemperature: true,
2441
2187
  defaultTemperature: 1,
2442
- inputPrice: 0.3,
2443
- outputPrice: 2.5,
2444
- cacheReadsPrice: 0.075,
2445
- cacheWritesPrice: 1
2188
+ inputPrice: 0.5,
2189
+ outputPrice: 3,
2190
+ cacheReadsPrice: 0.05
2446
2191
  },
2447
2192
  // 2.5 Pro models
2448
2193
  "gemini-2.5-pro": {
2449
2194
  maxTokens: 64e3,
2450
2195
  contextWindow: 1048576,
2451
2196
  supportsImages: true,
2452
- supportsNativeTools: true,
2453
- defaultToolProtocol: "native",
2454
2197
  supportsPromptCache: true,
2455
2198
  inputPrice: 2.5,
2456
2199
  // This is the pricing for prompts above 200k tokens.
@@ -2479,8 +2222,6 @@ var geminiModels = {
2479
2222
  maxTokens: 65535,
2480
2223
  contextWindow: 1048576,
2481
2224
  supportsImages: true,
2482
- supportsNativeTools: true,
2483
- defaultToolProtocol: "native",
2484
2225
  supportsPromptCache: true,
2485
2226
  inputPrice: 2.5,
2486
2227
  // This is the pricing for prompts above 200k tokens.
@@ -2508,8 +2249,6 @@ var geminiModels = {
2508
2249
  maxTokens: 65535,
2509
2250
  contextWindow: 1048576,
2510
2251
  supportsImages: true,
2511
- supportsNativeTools: true,
2512
- defaultToolProtocol: "native",
2513
2252
  supportsPromptCache: true,
2514
2253
  inputPrice: 2.5,
2515
2254
  // This is the pricing for prompts above 200k tokens.
@@ -2535,8 +2274,6 @@ var geminiModels = {
2535
2274
  maxTokens: 65535,
2536
2275
  contextWindow: 1048576,
2537
2276
  supportsImages: true,
2538
- supportsNativeTools: true,
2539
- defaultToolProtocol: "native",
2540
2277
  supportsPromptCache: true,
2541
2278
  inputPrice: 2.5,
2542
2279
  // This is the pricing for prompts above 200k tokens.
@@ -2565,8 +2302,6 @@ var geminiModels = {
2565
2302
  maxTokens: 65536,
2566
2303
  contextWindow: 1048576,
2567
2304
  supportsImages: true,
2568
- supportsNativeTools: true,
2569
- defaultToolProtocol: "native",
2570
2305
  supportsPromptCache: true,
2571
2306
  inputPrice: 0.3,
2572
2307
  outputPrice: 2.5,
@@ -2579,8 +2314,6 @@ var geminiModels = {
2579
2314
  maxTokens: 65536,
2580
2315
  contextWindow: 1048576,
2581
2316
  supportsImages: true,
2582
- supportsNativeTools: true,
2583
- defaultToolProtocol: "native",
2584
2317
  supportsPromptCache: true,
2585
2318
  inputPrice: 0.3,
2586
2319
  outputPrice: 2.5,
@@ -2593,8 +2326,6 @@ var geminiModels = {
2593
2326
  maxTokens: 64e3,
2594
2327
  contextWindow: 1048576,
2595
2328
  supportsImages: true,
2596
- supportsNativeTools: true,
2597
- defaultToolProtocol: "native",
2598
2329
  supportsPromptCache: true,
2599
2330
  inputPrice: 0.3,
2600
2331
  outputPrice: 2.5,
@@ -2608,8 +2339,6 @@ var geminiModels = {
2608
2339
  maxTokens: 65536,
2609
2340
  contextWindow: 1048576,
2610
2341
  supportsImages: true,
2611
- supportsNativeTools: true,
2612
- defaultToolProtocol: "native",
2613
2342
  supportsPromptCache: true,
2614
2343
  inputPrice: 0.1,
2615
2344
  outputPrice: 0.4,
@@ -2622,8 +2351,6 @@ var geminiModels = {
2622
2351
  maxTokens: 65536,
2623
2352
  contextWindow: 1048576,
2624
2353
  supportsImages: true,
2625
- supportsNativeTools: true,
2626
- defaultToolProtocol: "native",
2627
2354
  supportsPromptCache: true,
2628
2355
  inputPrice: 0.1,
2629
2356
  outputPrice: 0.4,
@@ -2643,8 +2370,6 @@ var groqModels = {
2643
2370
  contextWindow: 131072,
2644
2371
  supportsImages: false,
2645
2372
  supportsPromptCache: false,
2646
- supportsNativeTools: true,
2647
- defaultToolProtocol: "native",
2648
2373
  inputPrice: 0.05,
2649
2374
  outputPrice: 0.08,
2650
2375
  description: "Meta Llama 3.1 8B Instant model, 128K context."
@@ -2654,8 +2379,6 @@ var groqModels = {
2654
2379
  contextWindow: 131072,
2655
2380
  supportsImages: false,
2656
2381
  supportsPromptCache: false,
2657
- supportsNativeTools: true,
2658
- defaultToolProtocol: "native",
2659
2382
  inputPrice: 0.59,
2660
2383
  outputPrice: 0.79,
2661
2384
  description: "Meta Llama 3.3 70B Versatile model, 128K context."
@@ -2665,8 +2388,6 @@ var groqModels = {
2665
2388
  contextWindow: 131072,
2666
2389
  supportsImages: false,
2667
2390
  supportsPromptCache: false,
2668
- supportsNativeTools: true,
2669
- defaultToolProtocol: "native",
2670
2391
  inputPrice: 0.11,
2671
2392
  outputPrice: 0.34,
2672
2393
  description: "Meta Llama 4 Scout 17B Instruct model, 128K context."
@@ -2676,8 +2397,6 @@ var groqModels = {
2676
2397
  contextWindow: 131072,
2677
2398
  supportsImages: false,
2678
2399
  supportsPromptCache: false,
2679
- supportsNativeTools: true,
2680
- defaultToolProtocol: "native",
2681
2400
  inputPrice: 0.29,
2682
2401
  outputPrice: 0.59,
2683
2402
  description: "Alibaba Qwen 3 32B model, 128K context."
@@ -2687,8 +2406,6 @@ var groqModels = {
2687
2406
  contextWindow: 262144,
2688
2407
  supportsImages: false,
2689
2408
  supportsPromptCache: true,
2690
- supportsNativeTools: true,
2691
- defaultToolProtocol: "native",
2692
2409
  inputPrice: 0.6,
2693
2410
  outputPrice: 2.5,
2694
2411
  cacheReadsPrice: 0.15,
@@ -2699,8 +2416,6 @@ var groqModels = {
2699
2416
  contextWindow: 131072,
2700
2417
  supportsImages: false,
2701
2418
  supportsPromptCache: false,
2702
- supportsNativeTools: true,
2703
- defaultToolProtocol: "native",
2704
2419
  inputPrice: 0.15,
2705
2420
  outputPrice: 0.75,
2706
2421
  description: "GPT-OSS 120B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 128 experts."
@@ -2710,8 +2425,6 @@ var groqModels = {
2710
2425
  contextWindow: 131072,
2711
2426
  supportsImages: false,
2712
2427
  supportsPromptCache: false,
2713
- supportsNativeTools: true,
2714
- defaultToolProtocol: "native",
2715
2428
  inputPrice: 0.1,
2716
2429
  outputPrice: 0.5,
2717
2430
  description: "GPT-OSS 20B is OpenAI's flagship open source model, built on a Mixture-of-Experts (MoE) architecture with 20 billion parameters and 32 experts."
@@ -2738,7 +2451,6 @@ var ioIntelligenceModels = {
2738
2451
  contextWindow: 128e3,
2739
2452
  supportsImages: false,
2740
2453
  supportsPromptCache: false,
2741
- supportsNativeTools: true,
2742
2454
  description: "DeepSeek R1 reasoning model"
2743
2455
  },
2744
2456
  "meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8": {
@@ -2746,7 +2458,6 @@ var ioIntelligenceModels = {
2746
2458
  contextWindow: 43e4,
2747
2459
  supportsImages: true,
2748
2460
  supportsPromptCache: false,
2749
- supportsNativeTools: true,
2750
2461
  description: "Llama 4 Maverick 17B model"
2751
2462
  },
2752
2463
  "Intel/Qwen3-Coder-480B-A35B-Instruct-int4-mixed-ar": {
@@ -2754,7 +2465,6 @@ var ioIntelligenceModels = {
2754
2465
  contextWindow: 106e3,
2755
2466
  supportsImages: false,
2756
2467
  supportsPromptCache: false,
2757
- supportsNativeTools: true,
2758
2468
  description: "Qwen3 Coder 480B specialized for coding"
2759
2469
  },
2760
2470
  "openai/gpt-oss-120b": {
@@ -2762,7 +2472,6 @@ var ioIntelligenceModels = {
2762
2472
  contextWindow: 131072,
2763
2473
  supportsImages: false,
2764
2474
  supportsPromptCache: false,
2765
- supportsNativeTools: true,
2766
2475
  description: "OpenAI GPT-OSS 120B model"
2767
2476
  }
2768
2477
  };
@@ -2774,8 +2483,6 @@ var litellmDefaultModelInfo = {
2774
2483
  contextWindow: 2e5,
2775
2484
  supportsImages: true,
2776
2485
  supportsPromptCache: true,
2777
- supportsNativeTools: true,
2778
- defaultToolProtocol: "native",
2779
2486
  inputPrice: 3,
2780
2487
  outputPrice: 15,
2781
2488
  cacheWritesPrice: 3.75,
@@ -2790,8 +2497,6 @@ var lMStudioDefaultModelInfo = {
2790
2497
  contextWindow: 2e5,
2791
2498
  supportsImages: true,
2792
2499
  supportsPromptCache: true,
2793
- supportsNativeTools: true,
2794
- defaultToolProtocol: "native",
2795
2500
  inputPrice: 0,
2796
2501
  outputPrice: 0,
2797
2502
  cacheWritesPrice: 0,
@@ -2807,8 +2512,6 @@ var mistralModels = {
2807
2512
  contextWindow: 128e3,
2808
2513
  supportsImages: true,
2809
2514
  supportsPromptCache: false,
2810
- supportsNativeTools: true,
2811
- defaultToolProtocol: "native",
2812
2515
  inputPrice: 2,
2813
2516
  outputPrice: 5
2814
2517
  },
@@ -2817,8 +2520,6 @@ var mistralModels = {
2817
2520
  contextWindow: 131e3,
2818
2521
  supportsImages: true,
2819
2522
  supportsPromptCache: false,
2820
- supportsNativeTools: true,
2821
- defaultToolProtocol: "native",
2822
2523
  inputPrice: 0.4,
2823
2524
  outputPrice: 2
2824
2525
  },
@@ -2827,8 +2528,6 @@ var mistralModels = {
2827
2528
  contextWindow: 131e3,
2828
2529
  supportsImages: true,
2829
2530
  supportsPromptCache: false,
2830
- supportsNativeTools: true,
2831
- defaultToolProtocol: "native",
2832
2531
  inputPrice: 0.4,
2833
2532
  outputPrice: 2
2834
2533
  },
@@ -2837,8 +2536,6 @@ var mistralModels = {
2837
2536
  contextWindow: 256e3,
2838
2537
  supportsImages: false,
2839
2538
  supportsPromptCache: false,
2840
- supportsNativeTools: true,
2841
- defaultToolProtocol: "native",
2842
2539
  inputPrice: 0.3,
2843
2540
  outputPrice: 0.9
2844
2541
  },
@@ -2847,8 +2544,6 @@ var mistralModels = {
2847
2544
  contextWindow: 131e3,
2848
2545
  supportsImages: false,
2849
2546
  supportsPromptCache: false,
2850
- supportsNativeTools: true,
2851
- defaultToolProtocol: "native",
2852
2547
  inputPrice: 2,
2853
2548
  outputPrice: 6
2854
2549
  },
@@ -2857,8 +2552,6 @@ var mistralModels = {
2857
2552
  contextWindow: 131e3,
2858
2553
  supportsImages: false,
2859
2554
  supportsPromptCache: false,
2860
- supportsNativeTools: true,
2861
- defaultToolProtocol: "native",
2862
2555
  inputPrice: 0.1,
2863
2556
  outputPrice: 0.1
2864
2557
  },
@@ -2867,8 +2560,6 @@ var mistralModels = {
2867
2560
  contextWindow: 131e3,
2868
2561
  supportsImages: false,
2869
2562
  supportsPromptCache: false,
2870
- supportsNativeTools: true,
2871
- defaultToolProtocol: "native",
2872
2563
  inputPrice: 0.04,
2873
2564
  outputPrice: 0.04
2874
2565
  },
@@ -2877,8 +2568,6 @@ var mistralModels = {
2877
2568
  contextWindow: 32e3,
2878
2569
  supportsImages: false,
2879
2570
  supportsPromptCache: false,
2880
- supportsNativeTools: true,
2881
- defaultToolProtocol: "native",
2882
2571
  inputPrice: 0.2,
2883
2572
  outputPrice: 0.6
2884
2573
  },
@@ -2887,8 +2576,6 @@ var mistralModels = {
2887
2576
  contextWindow: 131e3,
2888
2577
  supportsImages: true,
2889
2578
  supportsPromptCache: false,
2890
- supportsNativeTools: true,
2891
- defaultToolProtocol: "native",
2892
2579
  inputPrice: 2,
2893
2580
  outputPrice: 6
2894
2581
  }
@@ -2903,8 +2590,6 @@ var moonshotModels = {
2903
2590
  contextWindow: 131072,
2904
2591
  supportsImages: false,
2905
2592
  supportsPromptCache: true,
2906
- supportsNativeTools: true,
2907
- defaultToolProtocol: "native",
2908
2593
  inputPrice: 0.6,
2909
2594
  // $0.60 per million tokens (cache miss)
2910
2595
  outputPrice: 2.5,
@@ -2920,8 +2605,6 @@ var moonshotModels = {
2920
2605
  contextWindow: 262144,
2921
2606
  supportsImages: false,
2922
2607
  supportsPromptCache: true,
2923
- supportsNativeTools: true,
2924
- defaultToolProtocol: "native",
2925
2608
  inputPrice: 0.6,
2926
2609
  outputPrice: 2.5,
2927
2610
  cacheReadsPrice: 0.15,
@@ -2932,8 +2615,6 @@ var moonshotModels = {
2932
2615
  contextWindow: 262144,
2933
2616
  supportsImages: false,
2934
2617
  supportsPromptCache: true,
2935
- supportsNativeTools: true,
2936
- defaultToolProtocol: "native",
2937
2618
  inputPrice: 2.4,
2938
2619
  // $2.40 per million tokens (cache miss)
2939
2620
  outputPrice: 10,
@@ -2952,8 +2633,6 @@ var moonshotModels = {
2952
2633
  supportsImages: false,
2953
2634
  // Text-only (no image/vision support)
2954
2635
  supportsPromptCache: true,
2955
- supportsNativeTools: true,
2956
- defaultToolProtocol: "native",
2957
2636
  inputPrice: 0.6,
2958
2637
  // $0.60 per million tokens (cache miss)
2959
2638
  outputPrice: 2.5,
@@ -2967,6 +2646,21 @@ var moonshotModels = {
2967
2646
  preserveReasoning: true,
2968
2647
  defaultTemperature: 1,
2969
2648
  description: `The kimi-k2-thinking model is a general-purpose agentic reasoning model developed by Moonshot AI. Thanks to its strength in deep reasoning and multi-turn tool use, it can solve even the hardest problems.`
2649
+ },
2650
+ "kimi-k2.5": {
2651
+ maxTokens: 16384,
2652
+ contextWindow: 262144,
2653
+ supportsImages: false,
2654
+ supportsPromptCache: true,
2655
+ inputPrice: 0.6,
2656
+ // $0.60 per million tokens (cache miss)
2657
+ outputPrice: 3,
2658
+ // $3.00 per million tokens
2659
+ cacheReadsPrice: 0.1,
2660
+ // $0.10 per million tokens (cache hit)
2661
+ supportsTemperature: true,
2662
+ defaultTemperature: 1,
2663
+ description: "Kimi K2.5 is the latest generation of Moonshot AI's Kimi series, featuring improved reasoning capabilities and enhanced performance across diverse tasks."
2970
2664
  }
2971
2665
  };
2972
2666
  var MOONSHOT_DEFAULT_TEMPERATURE = 0.6;
@@ -2978,7 +2672,6 @@ var ollamaDefaultModelInfo = {
2978
2672
  contextWindow: 2e5,
2979
2673
  supportsImages: true,
2980
2674
  supportsPromptCache: true,
2981
- supportsNativeTools: true,
2982
2675
  inputPrice: 0,
2983
2676
  outputPrice: 0,
2984
2677
  cacheWritesPrice: 0,
@@ -2992,8 +2685,6 @@ var openAiNativeModels = {
2992
2685
  "gpt-5.1-codex-max": {
2993
2686
  maxTokens: 128e3,
2994
2687
  contextWindow: 4e5,
2995
- supportsNativeTools: true,
2996
- defaultToolProtocol: "native",
2997
2688
  includedTools: ["apply_patch"],
2998
2689
  excludedTools: ["apply_diff", "write_to_file"],
2999
2690
  supportsImages: true,
@@ -3011,8 +2702,6 @@ var openAiNativeModels = {
3011
2702
  "gpt-5.2": {
3012
2703
  maxTokens: 128e3,
3013
2704
  contextWindow: 4e5,
3014
- supportsNativeTools: true,
3015
- defaultToolProtocol: "native",
3016
2705
  includedTools: ["apply_patch"],
3017
2706
  excludedTools: ["apply_diff", "write_to_file"],
3018
2707
  supportsImages: true,
@@ -3031,11 +2720,26 @@ var openAiNativeModels = {
3031
2720
  ],
3032
2721
  description: "GPT-5.2: Our flagship model for coding and agentic tasks across industries"
3033
2722
  },
2723
+ "gpt-5.2-codex": {
2724
+ maxTokens: 128e3,
2725
+ contextWindow: 4e5,
2726
+ includedTools: ["apply_patch"],
2727
+ excludedTools: ["apply_diff", "write_to_file"],
2728
+ supportsImages: true,
2729
+ supportsPromptCache: true,
2730
+ promptCacheRetention: "24h",
2731
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
2732
+ reasoningEffort: "medium",
2733
+ inputPrice: 1.75,
2734
+ outputPrice: 14,
2735
+ cacheReadsPrice: 0.175,
2736
+ supportsTemperature: false,
2737
+ tiers: [{ name: "priority", contextWindow: 4e5, inputPrice: 3.5, outputPrice: 28, cacheReadsPrice: 0.35 }],
2738
+ description: "GPT-5.2 Codex: Our most intelligent coding model optimized for long-horizon, agentic coding tasks"
2739
+ },
3034
2740
  "gpt-5.2-chat-latest": {
3035
2741
  maxTokens: 16384,
3036
2742
  contextWindow: 128e3,
3037
- supportsNativeTools: true,
3038
- defaultToolProtocol: "native",
3039
2743
  includedTools: ["apply_patch"],
3040
2744
  excludedTools: ["apply_diff", "write_to_file"],
3041
2745
  supportsImages: true,
@@ -3048,8 +2752,6 @@ var openAiNativeModels = {
3048
2752
  "gpt-5.1": {
3049
2753
  maxTokens: 128e3,
3050
2754
  contextWindow: 4e5,
3051
- supportsNativeTools: true,
3052
- defaultToolProtocol: "native",
3053
2755
  includedTools: ["apply_patch"],
3054
2756
  excludedTools: ["apply_diff", "write_to_file"],
3055
2757
  supportsImages: true,
@@ -3071,8 +2773,6 @@ var openAiNativeModels = {
3071
2773
  "gpt-5.1-codex": {
3072
2774
  maxTokens: 128e3,
3073
2775
  contextWindow: 4e5,
3074
- supportsNativeTools: true,
3075
- defaultToolProtocol: "native",
3076
2776
  includedTools: ["apply_patch"],
3077
2777
  excludedTools: ["apply_diff", "write_to_file"],
3078
2778
  supportsImages: true,
@@ -3090,8 +2790,6 @@ var openAiNativeModels = {
3090
2790
  "gpt-5.1-codex-mini": {
3091
2791
  maxTokens: 128e3,
3092
2792
  contextWindow: 4e5,
3093
- supportsNativeTools: true,
3094
- defaultToolProtocol: "native",
3095
2793
  includedTools: ["apply_patch"],
3096
2794
  excludedTools: ["apply_diff", "write_to_file"],
3097
2795
  supportsImages: true,
@@ -3108,8 +2806,6 @@ var openAiNativeModels = {
3108
2806
  "gpt-5": {
3109
2807
  maxTokens: 128e3,
3110
2808
  contextWindow: 4e5,
3111
- supportsNativeTools: true,
3112
- defaultToolProtocol: "native",
3113
2809
  includedTools: ["apply_patch"],
3114
2810
  excludedTools: ["apply_diff", "write_to_file"],
3115
2811
  supportsImages: true,
@@ -3130,8 +2826,6 @@ var openAiNativeModels = {
3130
2826
  "gpt-5-mini": {
3131
2827
  maxTokens: 128e3,
3132
2828
  contextWindow: 4e5,
3133
- supportsNativeTools: true,
3134
- defaultToolProtocol: "native",
3135
2829
  includedTools: ["apply_patch"],
3136
2830
  excludedTools: ["apply_diff", "write_to_file"],
3137
2831
  supportsImages: true,
@@ -3152,8 +2846,6 @@ var openAiNativeModels = {
3152
2846
  "gpt-5-codex": {
3153
2847
  maxTokens: 128e3,
3154
2848
  contextWindow: 4e5,
3155
- supportsNativeTools: true,
3156
- defaultToolProtocol: "native",
3157
2849
  includedTools: ["apply_patch"],
3158
2850
  excludedTools: ["apply_diff", "write_to_file"],
3159
2851
  supportsImages: true,
@@ -3170,8 +2862,6 @@ var openAiNativeModels = {
3170
2862
  "gpt-5-nano": {
3171
2863
  maxTokens: 128e3,
3172
2864
  contextWindow: 4e5,
3173
- supportsNativeTools: true,
3174
- defaultToolProtocol: "native",
3175
2865
  includedTools: ["apply_patch"],
3176
2866
  excludedTools: ["apply_diff", "write_to_file"],
3177
2867
  supportsImages: true,
@@ -3189,8 +2879,6 @@ var openAiNativeModels = {
3189
2879
  "gpt-5-chat-latest": {
3190
2880
  maxTokens: 128e3,
3191
2881
  contextWindow: 4e5,
3192
- supportsNativeTools: true,
3193
- defaultToolProtocol: "native",
3194
2882
  includedTools: ["apply_patch"],
3195
2883
  excludedTools: ["apply_diff", "write_to_file"],
3196
2884
  supportsImages: true,
@@ -3203,8 +2891,6 @@ var openAiNativeModels = {
3203
2891
  "gpt-4.1": {
3204
2892
  maxTokens: 32768,
3205
2893
  contextWindow: 1047576,
3206
- supportsNativeTools: true,
3207
- defaultToolProtocol: "native",
3208
2894
  includedTools: ["apply_patch"],
3209
2895
  excludedTools: ["apply_diff", "write_to_file"],
3210
2896
  supportsImages: true,
@@ -3220,8 +2906,6 @@ var openAiNativeModels = {
3220
2906
  "gpt-4.1-mini": {
3221
2907
  maxTokens: 32768,
3222
2908
  contextWindow: 1047576,
3223
- supportsNativeTools: true,
3224
- defaultToolProtocol: "native",
3225
2909
  includedTools: ["apply_patch"],
3226
2910
  excludedTools: ["apply_diff", "write_to_file"],
3227
2911
  supportsImages: true,
@@ -3237,8 +2921,6 @@ var openAiNativeModels = {
3237
2921
  "gpt-4.1-nano": {
3238
2922
  maxTokens: 32768,
3239
2923
  contextWindow: 1047576,
3240
- supportsNativeTools: true,
3241
- defaultToolProtocol: "native",
3242
2924
  includedTools: ["apply_patch"],
3243
2925
  excludedTools: ["apply_diff", "write_to_file"],
3244
2926
  supportsImages: true,
@@ -3254,8 +2936,6 @@ var openAiNativeModels = {
3254
2936
  o3: {
3255
2937
  maxTokens: 1e5,
3256
2938
  contextWindow: 2e5,
3257
- supportsNativeTools: true,
3258
- defaultToolProtocol: "native",
3259
2939
  supportsImages: true,
3260
2940
  supportsPromptCache: true,
3261
2941
  inputPrice: 2,
@@ -3272,8 +2952,6 @@ var openAiNativeModels = {
3272
2952
  "o3-high": {
3273
2953
  maxTokens: 1e5,
3274
2954
  contextWindow: 2e5,
3275
- supportsNativeTools: true,
3276
- defaultToolProtocol: "native",
3277
2955
  supportsImages: true,
3278
2956
  supportsPromptCache: true,
3279
2957
  inputPrice: 2,
@@ -3285,8 +2963,6 @@ var openAiNativeModels = {
3285
2963
  "o3-low": {
3286
2964
  maxTokens: 1e5,
3287
2965
  contextWindow: 2e5,
3288
- supportsNativeTools: true,
3289
- defaultToolProtocol: "native",
3290
2966
  supportsImages: true,
3291
2967
  supportsPromptCache: true,
3292
2968
  inputPrice: 2,
@@ -3298,8 +2974,6 @@ var openAiNativeModels = {
3298
2974
  "o4-mini": {
3299
2975
  maxTokens: 1e5,
3300
2976
  contextWindow: 2e5,
3301
- supportsNativeTools: true,
3302
- defaultToolProtocol: "native",
3303
2977
  supportsImages: true,
3304
2978
  supportsPromptCache: true,
3305
2979
  inputPrice: 1.1,
@@ -3316,8 +2990,6 @@ var openAiNativeModels = {
3316
2990
  "o4-mini-high": {
3317
2991
  maxTokens: 1e5,
3318
2992
  contextWindow: 2e5,
3319
- supportsNativeTools: true,
3320
- defaultToolProtocol: "native",
3321
2993
  supportsImages: true,
3322
2994
  supportsPromptCache: true,
3323
2995
  inputPrice: 1.1,
@@ -3329,8 +3001,6 @@ var openAiNativeModels = {
3329
3001
  "o4-mini-low": {
3330
3002
  maxTokens: 1e5,
3331
3003
  contextWindow: 2e5,
3332
- supportsNativeTools: true,
3333
- defaultToolProtocol: "native",
3334
3004
  supportsImages: true,
3335
3005
  supportsPromptCache: true,
3336
3006
  inputPrice: 1.1,
@@ -3342,8 +3012,6 @@ var openAiNativeModels = {
3342
3012
  "o3-mini": {
3343
3013
  maxTokens: 1e5,
3344
3014
  contextWindow: 2e5,
3345
- supportsNativeTools: true,
3346
- defaultToolProtocol: "native",
3347
3015
  supportsImages: false,
3348
3016
  supportsPromptCache: true,
3349
3017
  inputPrice: 1.1,
@@ -3356,8 +3024,6 @@ var openAiNativeModels = {
3356
3024
  "o3-mini-high": {
3357
3025
  maxTokens: 1e5,
3358
3026
  contextWindow: 2e5,
3359
- supportsNativeTools: true,
3360
- defaultToolProtocol: "native",
3361
3027
  supportsImages: false,
3362
3028
  supportsPromptCache: true,
3363
3029
  inputPrice: 1.1,
@@ -3369,8 +3035,6 @@ var openAiNativeModels = {
3369
3035
  "o3-mini-low": {
3370
3036
  maxTokens: 1e5,
3371
3037
  contextWindow: 2e5,
3372
- supportsNativeTools: true,
3373
- defaultToolProtocol: "native",
3374
3038
  supportsImages: false,
3375
3039
  supportsPromptCache: true,
3376
3040
  inputPrice: 1.1,
@@ -3382,8 +3046,6 @@ var openAiNativeModels = {
3382
3046
  o1: {
3383
3047
  maxTokens: 1e5,
3384
3048
  contextWindow: 2e5,
3385
- supportsNativeTools: true,
3386
- defaultToolProtocol: "native",
3387
3049
  supportsImages: true,
3388
3050
  supportsPromptCache: true,
3389
3051
  inputPrice: 15,
@@ -3394,8 +3056,6 @@ var openAiNativeModels = {
3394
3056
  "o1-preview": {
3395
3057
  maxTokens: 32768,
3396
3058
  contextWindow: 128e3,
3397
- supportsNativeTools: true,
3398
- defaultToolProtocol: "native",
3399
3059
  supportsImages: true,
3400
3060
  supportsPromptCache: true,
3401
3061
  inputPrice: 15,
@@ -3406,8 +3066,6 @@ var openAiNativeModels = {
3406
3066
  "o1-mini": {
3407
3067
  maxTokens: 65536,
3408
3068
  contextWindow: 128e3,
3409
- supportsNativeTools: true,
3410
- defaultToolProtocol: "native",
3411
3069
  supportsImages: true,
3412
3070
  supportsPromptCache: true,
3413
3071
  inputPrice: 1.1,
@@ -3418,8 +3076,6 @@ var openAiNativeModels = {
3418
3076
  "gpt-4o": {
3419
3077
  maxTokens: 16384,
3420
3078
  contextWindow: 128e3,
3421
- supportsNativeTools: true,
3422
- defaultToolProtocol: "native",
3423
3079
  supportsImages: true,
3424
3080
  supportsPromptCache: true,
3425
3081
  inputPrice: 2.5,
@@ -3433,8 +3089,6 @@ var openAiNativeModels = {
3433
3089
  "gpt-4o-mini": {
3434
3090
  maxTokens: 16384,
3435
3091
  contextWindow: 128e3,
3436
- supportsNativeTools: true,
3437
- defaultToolProtocol: "native",
3438
3092
  supportsImages: true,
3439
3093
  supportsPromptCache: true,
3440
3094
  inputPrice: 0.15,
@@ -3448,8 +3102,6 @@ var openAiNativeModels = {
3448
3102
  "codex-mini-latest": {
3449
3103
  maxTokens: 16384,
3450
3104
  contextWindow: 2e5,
3451
- supportsNativeTools: true,
3452
- defaultToolProtocol: "native",
3453
3105
  supportsImages: false,
3454
3106
  supportsPromptCache: false,
3455
3107
  inputPrice: 1.5,
@@ -3462,8 +3114,6 @@ var openAiNativeModels = {
3462
3114
  "gpt-5-2025-08-07": {
3463
3115
  maxTokens: 128e3,
3464
3116
  contextWindow: 4e5,
3465
- supportsNativeTools: true,
3466
- defaultToolProtocol: "native",
3467
3117
  includedTools: ["apply_patch"],
3468
3118
  excludedTools: ["apply_diff", "write_to_file"],
3469
3119
  supportsImages: true,
@@ -3484,8 +3134,6 @@ var openAiNativeModels = {
3484
3134
  "gpt-5-mini-2025-08-07": {
3485
3135
  maxTokens: 128e3,
3486
3136
  contextWindow: 4e5,
3487
- supportsNativeTools: true,
3488
- defaultToolProtocol: "native",
3489
3137
  includedTools: ["apply_patch"],
3490
3138
  excludedTools: ["apply_diff", "write_to_file"],
3491
3139
  supportsImages: true,
@@ -3506,8 +3154,6 @@ var openAiNativeModels = {
3506
3154
  "gpt-5-nano-2025-08-07": {
3507
3155
  maxTokens: 128e3,
3508
3156
  contextWindow: 4e5,
3509
- supportsNativeTools: true,
3510
- defaultToolProtocol: "native",
3511
3157
  includedTools: ["apply_patch"],
3512
3158
  excludedTools: ["apply_diff", "write_to_file"],
3513
3159
  supportsImages: true,
@@ -3529,14 +3175,151 @@ var openAiModelInfoSaneDefaults = {
3529
3175
  supportsImages: true,
3530
3176
  supportsPromptCache: false,
3531
3177
  inputPrice: 0,
3532
- outputPrice: 0,
3533
- supportsNativeTools: true,
3534
- defaultToolProtocol: "native"
3178
+ outputPrice: 0
3535
3179
  };
3536
3180
  var azureOpenAiDefaultApiVersion = "2024-08-01-preview";
3537
3181
  var OPENAI_NATIVE_DEFAULT_TEMPERATURE = 0;
3538
3182
  var OPENAI_AZURE_AI_INFERENCE_PATH = "/models/chat/completions";
3539
3183
 
3184
+ // src/providers/openai-codex.ts
3185
+ var openAiCodexDefaultModelId = "gpt-5.2-codex";
3186
+ var openAiCodexModels = {
3187
+ "gpt-5.1-codex-max": {
3188
+ maxTokens: 128e3,
3189
+ contextWindow: 4e5,
3190
+ includedTools: ["apply_patch"],
3191
+ excludedTools: ["apply_diff", "write_to_file"],
3192
+ supportsImages: true,
3193
+ supportsPromptCache: true,
3194
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3195
+ reasoningEffort: "xhigh",
3196
+ // Subscription-based: no per-token costs
3197
+ inputPrice: 0,
3198
+ outputPrice: 0,
3199
+ supportsTemperature: false,
3200
+ description: "GPT-5.1 Codex Max: Maximum capability coding model via ChatGPT subscription"
3201
+ },
3202
+ "gpt-5.1-codex": {
3203
+ maxTokens: 128e3,
3204
+ contextWindow: 4e5,
3205
+ includedTools: ["apply_patch"],
3206
+ excludedTools: ["apply_diff", "write_to_file"],
3207
+ supportsImages: true,
3208
+ supportsPromptCache: true,
3209
+ supportsReasoningEffort: ["low", "medium", "high"],
3210
+ reasoningEffort: "medium",
3211
+ // Subscription-based: no per-token costs
3212
+ inputPrice: 0,
3213
+ outputPrice: 0,
3214
+ supportsTemperature: false,
3215
+ description: "GPT-5.1 Codex: GPT-5.1 optimized for agentic coding via ChatGPT subscription"
3216
+ },
3217
+ "gpt-5.2-codex": {
3218
+ maxTokens: 128e3,
3219
+ contextWindow: 4e5,
3220
+ includedTools: ["apply_patch"],
3221
+ excludedTools: ["apply_diff", "write_to_file"],
3222
+ supportsImages: true,
3223
+ supportsPromptCache: true,
3224
+ supportsReasoningEffort: ["low", "medium", "high", "xhigh"],
3225
+ reasoningEffort: "medium",
3226
+ inputPrice: 0,
3227
+ outputPrice: 0,
3228
+ supportsTemperature: false,
3229
+ description: "GPT-5.2 Codex: OpenAI's flagship coding model via ChatGPT subscription"
3230
+ },
3231
+ "gpt-5.1": {
3232
+ maxTokens: 128e3,
3233
+ contextWindow: 4e5,
3234
+ includedTools: ["apply_patch"],
3235
+ excludedTools: ["apply_diff", "write_to_file"],
3236
+ supportsImages: true,
3237
+ supportsPromptCache: true,
3238
+ supportsReasoningEffort: ["none", "low", "medium", "high"],
3239
+ reasoningEffort: "medium",
3240
+ // Subscription-based: no per-token costs
3241
+ inputPrice: 0,
3242
+ outputPrice: 0,
3243
+ supportsVerbosity: true,
3244
+ supportsTemperature: false,
3245
+ description: "GPT-5.1: General GPT-5.1 model via ChatGPT subscription"
3246
+ },
3247
+ "gpt-5": {
3248
+ maxTokens: 128e3,
3249
+ contextWindow: 4e5,
3250
+ includedTools: ["apply_patch"],
3251
+ excludedTools: ["apply_diff", "write_to_file"],
3252
+ supportsImages: true,
3253
+ supportsPromptCache: true,
3254
+ supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3255
+ reasoningEffort: "medium",
3256
+ // Subscription-based: no per-token costs
3257
+ inputPrice: 0,
3258
+ outputPrice: 0,
3259
+ supportsVerbosity: true,
3260
+ supportsTemperature: false,
3261
+ description: "GPT-5: General GPT-5 model via ChatGPT subscription"
3262
+ },
3263
+ "gpt-5-codex": {
3264
+ maxTokens: 128e3,
3265
+ contextWindow: 4e5,
3266
+ includedTools: ["apply_patch"],
3267
+ excludedTools: ["apply_diff", "write_to_file"],
3268
+ supportsImages: true,
3269
+ supportsPromptCache: true,
3270
+ supportsReasoningEffort: ["low", "medium", "high"],
3271
+ reasoningEffort: "medium",
3272
+ // Subscription-based: no per-token costs
3273
+ inputPrice: 0,
3274
+ outputPrice: 0,
3275
+ supportsTemperature: false,
3276
+ description: "GPT-5 Codex: GPT-5 optimized for agentic coding via ChatGPT subscription"
3277
+ },
3278
+ "gpt-5-codex-mini": {
3279
+ maxTokens: 128e3,
3280
+ contextWindow: 4e5,
3281
+ includedTools: ["apply_patch"],
3282
+ excludedTools: ["apply_diff", "write_to_file"],
3283
+ supportsImages: true,
3284
+ supportsPromptCache: true,
3285
+ supportsReasoningEffort: ["low", "medium", "high"],
3286
+ reasoningEffort: "medium",
3287
+ // Subscription-based: no per-token costs
3288
+ inputPrice: 0,
3289
+ outputPrice: 0,
3290
+ supportsTemperature: false,
3291
+ description: "GPT-5 Codex Mini: Faster coding model via ChatGPT subscription"
3292
+ },
3293
+ "gpt-5.1-codex-mini": {
3294
+ maxTokens: 128e3,
3295
+ contextWindow: 4e5,
3296
+ includedTools: ["apply_patch"],
3297
+ excludedTools: ["apply_diff", "write_to_file"],
3298
+ supportsImages: true,
3299
+ supportsPromptCache: true,
3300
+ supportsReasoningEffort: ["low", "medium", "high"],
3301
+ reasoningEffort: "medium",
3302
+ inputPrice: 0,
3303
+ outputPrice: 0,
3304
+ supportsTemperature: false,
3305
+ description: "GPT-5.1 Codex Mini: Faster version for coding tasks via ChatGPT subscription"
3306
+ },
3307
+ "gpt-5.2": {
3308
+ maxTokens: 128e3,
3309
+ contextWindow: 4e5,
3310
+ includedTools: ["apply_patch"],
3311
+ excludedTools: ["apply_diff", "write_to_file"],
3312
+ supportsImages: true,
3313
+ supportsPromptCache: true,
3314
+ supportsReasoningEffort: ["none", "low", "medium", "high", "xhigh"],
3315
+ reasoningEffort: "medium",
3316
+ inputPrice: 0,
3317
+ outputPrice: 0,
3318
+ supportsTemperature: false,
3319
+ description: "GPT-5.2: Latest GPT model via ChatGPT subscription"
3320
+ }
3321
+ };
3322
+
3540
3323
  // src/providers/openrouter.ts
3541
3324
  var openRouterDefaultModelId = "anthropic/claude-sonnet-4.5";
3542
3325
  var openRouterDefaultModelInfo = {
@@ -3544,7 +3327,6 @@ var openRouterDefaultModelInfo = {
3544
3327
  contextWindow: 2e5,
3545
3328
  supportsImages: true,
3546
3329
  supportsPromptCache: true,
3547
- supportsNativeTools: true,
3548
3330
  inputPrice: 3,
3549
3331
  outputPrice: 15,
3550
3332
  cacheWritesPrice: 3.75,
@@ -3618,8 +3400,6 @@ var qwenCodeModels = {
3618
3400
  contextWindow: 1e6,
3619
3401
  supportsImages: false,
3620
3402
  supportsPromptCache: false,
3621
- supportsNativeTools: true,
3622
- defaultToolProtocol: "native",
3623
3403
  inputPrice: 0,
3624
3404
  outputPrice: 0,
3625
3405
  cacheWritesPrice: 0,
@@ -3631,8 +3411,6 @@ var qwenCodeModels = {
3631
3411
  contextWindow: 1e6,
3632
3412
  supportsImages: false,
3633
3413
  supportsPromptCache: false,
3634
- supportsNativeTools: true,
3635
- defaultToolProtocol: "native",
3636
3414
  inputPrice: 0,
3637
3415
  outputPrice: 0,
3638
3416
  cacheWritesPrice: 0,
@@ -3648,8 +3426,6 @@ var requestyDefaultModelInfo = {
3648
3426
  contextWindow: 2e5,
3649
3427
  supportsImages: true,
3650
3428
  supportsPromptCache: true,
3651
- supportsNativeTools: true,
3652
- defaultToolProtocol: "native",
3653
3429
  inputPrice: 3,
3654
3430
  outputPrice: 15,
3655
3431
  cacheWritesPrice: 3.75,
@@ -3705,8 +3481,6 @@ var sambaNovaModels = {
3705
3481
  contextWindow: 16384,
3706
3482
  supportsImages: false,
3707
3483
  supportsPromptCache: false,
3708
- supportsNativeTools: true,
3709
- defaultToolProtocol: "native",
3710
3484
  inputPrice: 0.1,
3711
3485
  outputPrice: 0.2,
3712
3486
  description: "Meta Llama 3.1 8B Instruct model with 16K context window."
@@ -3716,8 +3490,6 @@ var sambaNovaModels = {
3716
3490
  contextWindow: 131072,
3717
3491
  supportsImages: false,
3718
3492
  supportsPromptCache: false,
3719
- supportsNativeTools: true,
3720
- defaultToolProtocol: "native",
3721
3493
  inputPrice: 0.6,
3722
3494
  outputPrice: 1.2,
3723
3495
  description: "Meta Llama 3.3 70B Instruct model with 128K context window."
@@ -3728,8 +3500,6 @@ var sambaNovaModels = {
3728
3500
  supportsImages: false,
3729
3501
  supportsPromptCache: false,
3730
3502
  supportsReasoningBudget: true,
3731
- supportsNativeTools: true,
3732
- defaultToolProtocol: "native",
3733
3503
  inputPrice: 5,
3734
3504
  outputPrice: 7,
3735
3505
  description: "DeepSeek R1 reasoning model with 32K context window."
@@ -3739,8 +3509,6 @@ var sambaNovaModels = {
3739
3509
  contextWindow: 32768,
3740
3510
  supportsImages: false,
3741
3511
  supportsPromptCache: false,
3742
- supportsNativeTools: true,
3743
- defaultToolProtocol: "native",
3744
3512
  inputPrice: 3,
3745
3513
  outputPrice: 4.5,
3746
3514
  description: "DeepSeek V3 model with 32K context window."
@@ -3750,8 +3518,6 @@ var sambaNovaModels = {
3750
3518
  contextWindow: 32768,
3751
3519
  supportsImages: false,
3752
3520
  supportsPromptCache: false,
3753
- supportsNativeTools: true,
3754
- defaultToolProtocol: "native",
3755
3521
  inputPrice: 3,
3756
3522
  outputPrice: 4.5,
3757
3523
  description: "DeepSeek V3.1 model with 32K context window."
@@ -3761,8 +3527,6 @@ var sambaNovaModels = {
3761
3527
  contextWindow: 131072,
3762
3528
  supportsImages: true,
3763
3529
  supportsPromptCache: false,
3764
- supportsNativeTools: true,
3765
- defaultToolProtocol: "native",
3766
3530
  inputPrice: 0.63,
3767
3531
  outputPrice: 1.8,
3768
3532
  description: "Meta Llama 4 Maverick 17B 128E Instruct model with 128K context window."
@@ -3772,8 +3536,6 @@ var sambaNovaModels = {
3772
3536
  contextWindow: 8192,
3773
3537
  supportsImages: false,
3774
3538
  supportsPromptCache: false,
3775
- supportsNativeTools: true,
3776
- defaultToolProtocol: "native",
3777
3539
  inputPrice: 0.4,
3778
3540
  outputPrice: 0.8,
3779
3541
  description: "Alibaba Qwen 3 32B model with 8K context window."
@@ -3783,8 +3545,6 @@ var sambaNovaModels = {
3783
3545
  contextWindow: 131072,
3784
3546
  supportsImages: false,
3785
3547
  supportsPromptCache: false,
3786
- supportsNativeTools: true,
3787
- defaultToolProtocol: "native",
3788
3548
  inputPrice: 0.22,
3789
3549
  outputPrice: 0.59,
3790
3550
  description: "OpenAI gpt oss 120b model with 128k context window."
@@ -3798,7 +3558,6 @@ var unboundDefaultModelInfo = {
3798
3558
  contextWindow: 2e5,
3799
3559
  supportsImages: true,
3800
3560
  supportsPromptCache: true,
3801
- supportsNativeTools: true,
3802
3561
  inputPrice: 3,
3803
3562
  outputPrice: 15,
3804
3563
  cacheWritesPrice: 3.75,
@@ -3812,8 +3571,6 @@ var vertexModels = {
3812
3571
  maxTokens: 65536,
3813
3572
  contextWindow: 1048576,
3814
3573
  supportsImages: true,
3815
- supportsNativeTools: true,
3816
- defaultToolProtocol: "native",
3817
3574
  supportsPromptCache: true,
3818
3575
  supportsReasoningEffort: ["low", "high"],
3819
3576
  reasoningEffort: "low",
@@ -3821,16 +3578,19 @@ var vertexModels = {
3821
3578
  defaultTemperature: 1,
3822
3579
  inputPrice: 4,
3823
3580
  outputPrice: 18,
3581
+ cacheReadsPrice: 0.4,
3824
3582
  tiers: [
3825
3583
  {
3826
3584
  contextWindow: 2e5,
3827
3585
  inputPrice: 2,
3828
- outputPrice: 12
3586
+ outputPrice: 12,
3587
+ cacheReadsPrice: 0.2
3829
3588
  },
3830
3589
  {
3831
3590
  contextWindow: Infinity,
3832
3591
  inputPrice: 4,
3833
- outputPrice: 18
3592
+ outputPrice: 18,
3593
+ cacheReadsPrice: 0.4
3834
3594
  }
3835
3595
  ]
3836
3596
  },
@@ -3838,24 +3598,19 @@ var vertexModels = {
3838
3598
  maxTokens: 65536,
3839
3599
  contextWindow: 1048576,
3840
3600
  supportsImages: true,
3841
- supportsNativeTools: true,
3842
- defaultToolProtocol: "native",
3843
3601
  supportsPromptCache: true,
3844
3602
  supportsReasoningEffort: ["minimal", "low", "medium", "high"],
3845
3603
  reasoningEffort: "medium",
3846
3604
  supportsTemperature: true,
3847
3605
  defaultTemperature: 1,
3848
- inputPrice: 0.3,
3849
- outputPrice: 2.5,
3850
- cacheReadsPrice: 0.075,
3851
- cacheWritesPrice: 1
3606
+ inputPrice: 0.5,
3607
+ outputPrice: 3,
3608
+ cacheReadsPrice: 0.05
3852
3609
  },
3853
3610
  "gemini-2.5-flash-preview-05-20:thinking": {
3854
3611
  maxTokens: 65535,
3855
3612
  contextWindow: 1048576,
3856
3613
  supportsImages: true,
3857
- supportsNativeTools: true,
3858
- defaultToolProtocol: "native",
3859
3614
  supportsPromptCache: true,
3860
3615
  inputPrice: 0.15,
3861
3616
  outputPrice: 3.5,
@@ -3867,8 +3622,6 @@ var vertexModels = {
3867
3622
  maxTokens: 65535,
3868
3623
  contextWindow: 1048576,
3869
3624
  supportsImages: true,
3870
- supportsNativeTools: true,
3871
- defaultToolProtocol: "native",
3872
3625
  supportsPromptCache: true,
3873
3626
  inputPrice: 0.15,
3874
3627
  outputPrice: 0.6
@@ -3877,8 +3630,6 @@ var vertexModels = {
3877
3630
  maxTokens: 64e3,
3878
3631
  contextWindow: 1048576,
3879
3632
  supportsImages: true,
3880
- supportsNativeTools: true,
3881
- defaultToolProtocol: "native",
3882
3633
  supportsPromptCache: true,
3883
3634
  inputPrice: 0.3,
3884
3635
  outputPrice: 2.5,
@@ -3891,8 +3642,6 @@ var vertexModels = {
3891
3642
  maxTokens: 65535,
3892
3643
  contextWindow: 1048576,
3893
3644
  supportsImages: true,
3894
- supportsNativeTools: true,
3895
- defaultToolProtocol: "native",
3896
3645
  supportsPromptCache: false,
3897
3646
  inputPrice: 0.15,
3898
3647
  outputPrice: 3.5,
@@ -3904,8 +3653,6 @@ var vertexModels = {
3904
3653
  maxTokens: 65535,
3905
3654
  contextWindow: 1048576,
3906
3655
  supportsImages: true,
3907
- supportsNativeTools: true,
3908
- defaultToolProtocol: "native",
3909
3656
  supportsPromptCache: false,
3910
3657
  inputPrice: 0.15,
3911
3658
  outputPrice: 0.6
@@ -3914,8 +3661,6 @@ var vertexModels = {
3914
3661
  maxTokens: 65535,
3915
3662
  contextWindow: 1048576,
3916
3663
  supportsImages: true,
3917
- supportsNativeTools: true,
3918
- defaultToolProtocol: "native",
3919
3664
  supportsPromptCache: true,
3920
3665
  inputPrice: 2.5,
3921
3666
  outputPrice: 15
@@ -3924,8 +3669,6 @@ var vertexModels = {
3924
3669
  maxTokens: 65535,
3925
3670
  contextWindow: 1048576,
3926
3671
  supportsImages: true,
3927
- supportsNativeTools: true,
3928
- defaultToolProtocol: "native",
3929
3672
  supportsPromptCache: true,
3930
3673
  inputPrice: 2.5,
3931
3674
  outputPrice: 15
@@ -3934,8 +3677,6 @@ var vertexModels = {
3934
3677
  maxTokens: 65535,
3935
3678
  contextWindow: 1048576,
3936
3679
  supportsImages: true,
3937
- supportsNativeTools: true,
3938
- defaultToolProtocol: "native",
3939
3680
  supportsPromptCache: true,
3940
3681
  inputPrice: 2.5,
3941
3682
  outputPrice: 15,
@@ -3946,8 +3687,6 @@ var vertexModels = {
3946
3687
  maxTokens: 64e3,
3947
3688
  contextWindow: 1048576,
3948
3689
  supportsImages: true,
3949
- supportsNativeTools: true,
3950
- defaultToolProtocol: "native",
3951
3690
  supportsPromptCache: true,
3952
3691
  inputPrice: 2.5,
3953
3692
  outputPrice: 15,
@@ -3973,8 +3712,6 @@ var vertexModels = {
3973
3712
  maxTokens: 65535,
3974
3713
  contextWindow: 1048576,
3975
3714
  supportsImages: true,
3976
- supportsNativeTools: true,
3977
- defaultToolProtocol: "native",
3978
3715
  supportsPromptCache: false,
3979
3716
  inputPrice: 0,
3980
3717
  outputPrice: 0
@@ -3983,8 +3720,6 @@ var vertexModels = {
3983
3720
  maxTokens: 8192,
3984
3721
  contextWindow: 2097152,
3985
3722
  supportsImages: true,
3986
- supportsNativeTools: true,
3987
- defaultToolProtocol: "native",
3988
3723
  supportsPromptCache: false,
3989
3724
  inputPrice: 0,
3990
3725
  outputPrice: 0
@@ -3993,8 +3728,6 @@ var vertexModels = {
3993
3728
  maxTokens: 8192,
3994
3729
  contextWindow: 1048576,
3995
3730
  supportsImages: true,
3996
- supportsNativeTools: true,
3997
- defaultToolProtocol: "native",
3998
3731
  supportsPromptCache: true,
3999
3732
  inputPrice: 0.15,
4000
3733
  outputPrice: 0.6
@@ -4003,8 +3736,6 @@ var vertexModels = {
4003
3736
  maxTokens: 8192,
4004
3737
  contextWindow: 1048576,
4005
3738
  supportsImages: true,
4006
- supportsNativeTools: true,
4007
- defaultToolProtocol: "native",
4008
3739
  supportsPromptCache: false,
4009
3740
  inputPrice: 0.075,
4010
3741
  outputPrice: 0.3
@@ -4013,8 +3744,6 @@ var vertexModels = {
4013
3744
  maxTokens: 8192,
4014
3745
  contextWindow: 32768,
4015
3746
  supportsImages: true,
4016
- supportsNativeTools: true,
4017
- defaultToolProtocol: "native",
4018
3747
  supportsPromptCache: false,
4019
3748
  inputPrice: 0,
4020
3749
  outputPrice: 0
@@ -4023,8 +3752,6 @@ var vertexModels = {
4023
3752
  maxTokens: 8192,
4024
3753
  contextWindow: 1048576,
4025
3754
  supportsImages: true,
4026
- supportsNativeTools: true,
4027
- defaultToolProtocol: "native",
4028
3755
  supportsPromptCache: true,
4029
3756
  inputPrice: 0.075,
4030
3757
  outputPrice: 0.3
@@ -4033,8 +3760,6 @@ var vertexModels = {
4033
3760
  maxTokens: 8192,
4034
3761
  contextWindow: 2097152,
4035
3762
  supportsImages: true,
4036
- supportsNativeTools: true,
4037
- defaultToolProtocol: "native",
4038
3763
  supportsPromptCache: false,
4039
3764
  inputPrice: 1.25,
4040
3765
  outputPrice: 5
@@ -4045,8 +3770,6 @@ var vertexModels = {
4045
3770
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4046
3771
  supportsImages: true,
4047
3772
  supportsPromptCache: true,
4048
- supportsNativeTools: true,
4049
- defaultToolProtocol: "native",
4050
3773
  inputPrice: 3,
4051
3774
  // $3 per million input tokens (≤200K context)
4052
3775
  outputPrice: 15,
@@ -4078,8 +3801,6 @@ var vertexModels = {
4078
3801
  // Default 200K, extendable to 1M with beta flag 'context-1m-2025-08-07'
4079
3802
  supportsImages: true,
4080
3803
  supportsPromptCache: true,
4081
- supportsNativeTools: true,
4082
- defaultToolProtocol: "native",
4083
3804
  inputPrice: 3,
4084
3805
  // $3 per million input tokens (≤200K context)
4085
3806
  outputPrice: 15,
@@ -4110,8 +3831,6 @@ var vertexModels = {
4110
3831
  contextWindow: 2e5,
4111
3832
  supportsImages: true,
4112
3833
  supportsPromptCache: true,
4113
- supportsNativeTools: true,
4114
- defaultToolProtocol: "native",
4115
3834
  inputPrice: 1,
4116
3835
  outputPrice: 5,
4117
3836
  cacheWritesPrice: 1.25,
@@ -4123,8 +3842,6 @@ var vertexModels = {
4123
3842
  contextWindow: 2e5,
4124
3843
  supportsImages: true,
4125
3844
  supportsPromptCache: true,
4126
- supportsNativeTools: true,
4127
- defaultToolProtocol: "native",
4128
3845
  inputPrice: 5,
4129
3846
  outputPrice: 25,
4130
3847
  cacheWritesPrice: 6.25,
@@ -4136,8 +3853,6 @@ var vertexModels = {
4136
3853
  contextWindow: 2e5,
4137
3854
  supportsImages: true,
4138
3855
  supportsPromptCache: true,
4139
- supportsNativeTools: true,
4140
- defaultToolProtocol: "native",
4141
3856
  inputPrice: 15,
4142
3857
  outputPrice: 75,
4143
3858
  cacheWritesPrice: 18.75,
@@ -4149,8 +3864,6 @@ var vertexModels = {
4149
3864
  contextWindow: 2e5,
4150
3865
  supportsImages: true,
4151
3866
  supportsPromptCache: true,
4152
- supportsNativeTools: true,
4153
- defaultToolProtocol: "native",
4154
3867
  inputPrice: 15,
4155
3868
  outputPrice: 75,
4156
3869
  cacheWritesPrice: 18.75,
@@ -4161,8 +3874,6 @@ var vertexModels = {
4161
3874
  contextWindow: 2e5,
4162
3875
  supportsImages: true,
4163
3876
  supportsPromptCache: true,
4164
- supportsNativeTools: true,
4165
- defaultToolProtocol: "native",
4166
3877
  inputPrice: 3,
4167
3878
  outputPrice: 15,
4168
3879
  cacheWritesPrice: 3.75,
@@ -4175,8 +3886,6 @@ var vertexModels = {
4175
3886
  contextWindow: 2e5,
4176
3887
  supportsImages: true,
4177
3888
  supportsPromptCache: true,
4178
- supportsNativeTools: true,
4179
- defaultToolProtocol: "native",
4180
3889
  inputPrice: 3,
4181
3890
  outputPrice: 15,
4182
3891
  cacheWritesPrice: 3.75,
@@ -4187,8 +3896,6 @@ var vertexModels = {
4187
3896
  contextWindow: 2e5,
4188
3897
  supportsImages: true,
4189
3898
  supportsPromptCache: true,
4190
- supportsNativeTools: true,
4191
- defaultToolProtocol: "native",
4192
3899
  inputPrice: 3,
4193
3900
  outputPrice: 15,
4194
3901
  cacheWritesPrice: 3.75,
@@ -4199,8 +3906,6 @@ var vertexModels = {
4199
3906
  contextWindow: 2e5,
4200
3907
  supportsImages: true,
4201
3908
  supportsPromptCache: true,
4202
- supportsNativeTools: true,
4203
- defaultToolProtocol: "native",
4204
3909
  inputPrice: 3,
4205
3910
  outputPrice: 15,
4206
3911
  cacheWritesPrice: 3.75,
@@ -4211,8 +3916,6 @@ var vertexModels = {
4211
3916
  contextWindow: 2e5,
4212
3917
  supportsImages: false,
4213
3918
  supportsPromptCache: true,
4214
- supportsNativeTools: true,
4215
- defaultToolProtocol: "native",
4216
3919
  inputPrice: 1,
4217
3920
  outputPrice: 5,
4218
3921
  cacheWritesPrice: 1.25,
@@ -4223,8 +3926,6 @@ var vertexModels = {
4223
3926
  contextWindow: 2e5,
4224
3927
  supportsImages: true,
4225
3928
  supportsPromptCache: true,
4226
- supportsNativeTools: true,
4227
- defaultToolProtocol: "native",
4228
3929
  inputPrice: 15,
4229
3930
  outputPrice: 75,
4230
3931
  cacheWritesPrice: 18.75,
@@ -4235,8 +3936,6 @@ var vertexModels = {
4235
3936
  contextWindow: 2e5,
4236
3937
  supportsImages: true,
4237
3938
  supportsPromptCache: true,
4238
- supportsNativeTools: true,
4239
- defaultToolProtocol: "native",
4240
3939
  inputPrice: 0.25,
4241
3940
  outputPrice: 1.25,
4242
3941
  cacheWritesPrice: 0.3,
@@ -4246,8 +3945,6 @@ var vertexModels = {
4246
3945
  maxTokens: 64e3,
4247
3946
  contextWindow: 1048576,
4248
3947
  supportsImages: true,
4249
- supportsNativeTools: true,
4250
- defaultToolProtocol: "native",
4251
3948
  supportsPromptCache: true,
4252
3949
  inputPrice: 0.1,
4253
3950
  outputPrice: 0.4,
@@ -4261,7 +3958,6 @@ var vertexModels = {
4261
3958
  contextWindow: 131072,
4262
3959
  supportsImages: false,
4263
3960
  supportsPromptCache: false,
4264
- supportsNativeTools: true,
4265
3961
  inputPrice: 0.35,
4266
3962
  outputPrice: 1.15,
4267
3963
  description: "Meta Llama 4 Maverick 17B Instruct model, 128K context."
@@ -4271,7 +3967,6 @@ var vertexModels = {
4271
3967
  contextWindow: 163840,
4272
3968
  supportsImages: false,
4273
3969
  supportsPromptCache: false,
4274
- supportsNativeTools: true,
4275
3970
  inputPrice: 1.35,
4276
3971
  outputPrice: 5.4,
4277
3972
  description: "DeepSeek R1 (0528). Available in us-central1"
@@ -4281,7 +3976,6 @@ var vertexModels = {
4281
3976
  contextWindow: 163840,
4282
3977
  supportsImages: false,
4283
3978
  supportsPromptCache: false,
4284
- supportsNativeTools: true,
4285
3979
  inputPrice: 0.6,
4286
3980
  outputPrice: 1.7,
4287
3981
  description: "DeepSeek V3.1. Available in us-west2"
@@ -4291,7 +3985,6 @@ var vertexModels = {
4291
3985
  contextWindow: 131072,
4292
3986
  supportsImages: false,
4293
3987
  supportsPromptCache: false,
4294
- supportsNativeTools: true,
4295
3988
  inputPrice: 0.15,
4296
3989
  outputPrice: 0.6,
4297
3990
  description: "OpenAI gpt-oss 120B. Available in us-central1"
@@ -4301,7 +3994,6 @@ var vertexModels = {
4301
3994
  contextWindow: 131072,
4302
3995
  supportsImages: false,
4303
3996
  supportsPromptCache: false,
4304
- supportsNativeTools: true,
4305
3997
  inputPrice: 0.075,
4306
3998
  outputPrice: 0.3,
4307
3999
  description: "OpenAI gpt-oss 20B. Available in us-central1"
@@ -4311,7 +4003,6 @@ var vertexModels = {
4311
4003
  contextWindow: 262144,
4312
4004
  supportsImages: false,
4313
4005
  supportsPromptCache: false,
4314
- supportsNativeTools: true,
4315
4006
  inputPrice: 1,
4316
4007
  outputPrice: 4,
4317
4008
  description: "Qwen3 Coder 480B A35B Instruct. Available in us-south1"
@@ -4321,10 +4012,18 @@ var vertexModels = {
4321
4012
  contextWindow: 262144,
4322
4013
  supportsImages: false,
4323
4014
  supportsPromptCache: false,
4324
- supportsNativeTools: true,
4325
4015
  inputPrice: 0.25,
4326
4016
  outputPrice: 1,
4327
4017
  description: "Qwen3 235B A22B Instruct. Available in us-south1"
4018
+ },
4019
+ "moonshotai/kimi-k2-thinking-maas": {
4020
+ maxTokens: 16384,
4021
+ contextWindow: 262144,
4022
+ supportsPromptCache: false,
4023
+ supportsImages: false,
4024
+ inputPrice: 0.6,
4025
+ outputPrice: 2.5,
4026
+ description: "Kimi K2 Thinking Model with 256K context window."
4328
4027
  }
4329
4028
  };
4330
4029
  var VERTEX_1M_CONTEXT_MODEL_IDS = ["claude-sonnet-4@20250514", "claude-sonnet-4-5@20250929"];
@@ -4557,8 +4256,6 @@ var xaiModels = {
4557
4256
  contextWindow: 256e3,
4558
4257
  supportsImages: true,
4559
4258
  supportsPromptCache: true,
4560
- supportsNativeTools: true,
4561
- defaultToolProtocol: "native",
4562
4259
  inputPrice: 0.2,
4563
4260
  outputPrice: 1.5,
4564
4261
  cacheWritesPrice: 0.02,
@@ -4572,8 +4269,6 @@ var xaiModels = {
4572
4269
  contextWindow: 2e6,
4573
4270
  supportsImages: true,
4574
4271
  supportsPromptCache: true,
4575
- supportsNativeTools: true,
4576
- defaultToolProtocol: "native",
4577
4272
  inputPrice: 0.2,
4578
4273
  outputPrice: 0.5,
4579
4274
  cacheWritesPrice: 0.05,
@@ -4587,8 +4282,6 @@ var xaiModels = {
4587
4282
  contextWindow: 2e6,
4588
4283
  supportsImages: true,
4589
4284
  supportsPromptCache: true,
4590
- supportsNativeTools: true,
4591
- defaultToolProtocol: "native",
4592
4285
  inputPrice: 0.2,
4593
4286
  outputPrice: 0.5,
4594
4287
  cacheWritesPrice: 0.05,
@@ -4602,8 +4295,6 @@ var xaiModels = {
4602
4295
  contextWindow: 2e6,
4603
4296
  supportsImages: true,
4604
4297
  supportsPromptCache: true,
4605
- supportsNativeTools: true,
4606
- defaultToolProtocol: "native",
4607
4298
  inputPrice: 0.2,
4608
4299
  outputPrice: 0.5,
4609
4300
  cacheWritesPrice: 0.05,
@@ -4617,8 +4308,6 @@ var xaiModels = {
4617
4308
  contextWindow: 2e6,
4618
4309
  supportsImages: true,
4619
4310
  supportsPromptCache: true,
4620
- supportsNativeTools: true,
4621
- defaultToolProtocol: "native",
4622
4311
  inputPrice: 0.2,
4623
4312
  outputPrice: 0.5,
4624
4313
  cacheWritesPrice: 0.05,
@@ -4632,8 +4321,6 @@ var xaiModels = {
4632
4321
  contextWindow: 256e3,
4633
4322
  supportsImages: true,
4634
4323
  supportsPromptCache: true,
4635
- supportsNativeTools: true,
4636
- defaultToolProtocol: "native",
4637
4324
  inputPrice: 3,
4638
4325
  outputPrice: 15,
4639
4326
  cacheWritesPrice: 0.75,
@@ -4647,8 +4334,6 @@ var xaiModels = {
4647
4334
  contextWindow: 131072,
4648
4335
  supportsImages: true,
4649
4336
  supportsPromptCache: true,
4650
- supportsNativeTools: true,
4651
- defaultToolProtocol: "native",
4652
4337
  inputPrice: 0.3,
4653
4338
  outputPrice: 0.5,
4654
4339
  cacheWritesPrice: 0.07,
@@ -4664,8 +4349,6 @@ var xaiModels = {
4664
4349
  contextWindow: 131072,
4665
4350
  supportsImages: true,
4666
4351
  supportsPromptCache: true,
4667
- supportsNativeTools: true,
4668
- defaultToolProtocol: "native",
4669
4352
  inputPrice: 3,
4670
4353
  outputPrice: 15,
4671
4354
  cacheWritesPrice: 0.75,
@@ -4762,7 +4445,6 @@ var vercelAiGatewayDefaultModelInfo = {
4762
4445
  contextWindow: 2e5,
4763
4446
  supportsImages: true,
4764
4447
  supportsPromptCache: true,
4765
- supportsNativeTools: true,
4766
4448
  inputPrice: 3,
4767
4449
  outputPrice: 15,
4768
4450
  cacheWritesPrice: 3.75,
@@ -4779,8 +4461,6 @@ var internationalZAiModels = {
4779
4461
  contextWindow: 131072,
4780
4462
  supportsImages: false,
4781
4463
  supportsPromptCache: true,
4782
- supportsNativeTools: true,
4783
- defaultToolProtocol: "native",
4784
4464
  inputPrice: 0.6,
4785
4465
  outputPrice: 2.2,
4786
4466
  cacheWritesPrice: 0,
@@ -4792,8 +4472,6 @@ var internationalZAiModels = {
4792
4472
  contextWindow: 131072,
4793
4473
  supportsImages: false,
4794
4474
  supportsPromptCache: true,
4795
- supportsNativeTools: true,
4796
- defaultToolProtocol: "native",
4797
4475
  inputPrice: 0.2,
4798
4476
  outputPrice: 1.1,
4799
4477
  cacheWritesPrice: 0,
@@ -4805,8 +4483,6 @@ var internationalZAiModels = {
4805
4483
  contextWindow: 131072,
4806
4484
  supportsImages: false,
4807
4485
  supportsPromptCache: true,
4808
- supportsNativeTools: true,
4809
- defaultToolProtocol: "native",
4810
4486
  inputPrice: 2.2,
4811
4487
  outputPrice: 8.9,
4812
4488
  cacheWritesPrice: 0,
@@ -4818,8 +4494,6 @@ var internationalZAiModels = {
4818
4494
  contextWindow: 131072,
4819
4495
  supportsImages: false,
4820
4496
  supportsPromptCache: true,
4821
- supportsNativeTools: true,
4822
- defaultToolProtocol: "native",
4823
4497
  inputPrice: 1.1,
4824
4498
  outputPrice: 4.5,
4825
4499
  cacheWritesPrice: 0,
@@ -4831,8 +4505,6 @@ var internationalZAiModels = {
4831
4505
  contextWindow: 131072,
4832
4506
  supportsImages: false,
4833
4507
  supportsPromptCache: true,
4834
- supportsNativeTools: true,
4835
- defaultToolProtocol: "native",
4836
4508
  inputPrice: 0,
4837
4509
  outputPrice: 0,
4838
4510
  cacheWritesPrice: 0,
@@ -4844,21 +4516,28 @@ var internationalZAiModels = {
4844
4516
  contextWindow: 131072,
4845
4517
  supportsImages: true,
4846
4518
  supportsPromptCache: true,
4847
- supportsNativeTools: true,
4848
- defaultToolProtocol: "native",
4849
4519
  inputPrice: 0.6,
4850
4520
  outputPrice: 1.8,
4851
4521
  cacheWritesPrice: 0,
4852
4522
  cacheReadsPrice: 0.11,
4853
4523
  description: "GLM-4.5V is Z.AI's multimodal visual reasoning model (image/video/text/file input), optimized for GUI tasks, grounding, and document/video understanding."
4854
4524
  },
4525
+ "glm-4.6v": {
4526
+ maxTokens: 16384,
4527
+ contextWindow: 131072,
4528
+ supportsImages: true,
4529
+ supportsPromptCache: true,
4530
+ inputPrice: 0.3,
4531
+ outputPrice: 0.9,
4532
+ cacheWritesPrice: 0,
4533
+ cacheReadsPrice: 0.05,
4534
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4535
+ },
4855
4536
  "glm-4.6": {
4856
4537
  maxTokens: 16384,
4857
4538
  contextWindow: 2e5,
4858
4539
  supportsImages: false,
4859
4540
  supportsPromptCache: true,
4860
- supportsNativeTools: true,
4861
- defaultToolProtocol: "native",
4862
4541
  inputPrice: 0.6,
4863
4542
  outputPrice: 2.2,
4864
4543
  cacheWritesPrice: 0,
@@ -4870,8 +4549,6 @@ var internationalZAiModels = {
4870
4549
  contextWindow: 2e5,
4871
4550
  supportsImages: false,
4872
4551
  supportsPromptCache: true,
4873
- supportsNativeTools: true,
4874
- defaultToolProtocol: "native",
4875
4552
  supportsReasoningEffort: ["disable", "medium"],
4876
4553
  reasoningEffort: "medium",
4877
4554
  preserveReasoning: true,
@@ -4881,13 +4558,55 @@ var internationalZAiModels = {
4881
4558
  cacheReadsPrice: 0.11,
4882
4559
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4883
4560
  },
4561
+ "glm-4.7-flash": {
4562
+ maxTokens: 16384,
4563
+ contextWindow: 2e5,
4564
+ supportsImages: false,
4565
+ supportsPromptCache: true,
4566
+ inputPrice: 0,
4567
+ outputPrice: 0,
4568
+ cacheWritesPrice: 0,
4569
+ cacheReadsPrice: 0,
4570
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4571
+ },
4572
+ "glm-4.7-flashx": {
4573
+ maxTokens: 16384,
4574
+ contextWindow: 2e5,
4575
+ supportsImages: false,
4576
+ supportsPromptCache: true,
4577
+ inputPrice: 0.07,
4578
+ outputPrice: 0.4,
4579
+ cacheWritesPrice: 0,
4580
+ cacheReadsPrice: 0.01,
4581
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4582
+ },
4583
+ "glm-4.6v-flash": {
4584
+ maxTokens: 16384,
4585
+ contextWindow: 131072,
4586
+ supportsImages: true,
4587
+ supportsPromptCache: true,
4588
+ inputPrice: 0,
4589
+ outputPrice: 0,
4590
+ cacheWritesPrice: 0,
4591
+ cacheReadsPrice: 0,
4592
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4593
+ },
4594
+ "glm-4.6v-flashx": {
4595
+ maxTokens: 16384,
4596
+ contextWindow: 131072,
4597
+ supportsImages: true,
4598
+ supportsPromptCache: true,
4599
+ inputPrice: 0.04,
4600
+ outputPrice: 0.4,
4601
+ cacheWritesPrice: 0,
4602
+ cacheReadsPrice: 4e-3,
4603
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
4604
+ },
4884
4605
  "glm-4-32b-0414-128k": {
4885
4606
  maxTokens: 16384,
4886
4607
  contextWindow: 131072,
4887
4608
  supportsImages: false,
4888
4609
  supportsPromptCache: false,
4889
- supportsNativeTools: true,
4890
- defaultToolProtocol: "native",
4891
4610
  inputPrice: 0.1,
4892
4611
  outputPrice: 0.1,
4893
4612
  cacheWritesPrice: 0,
@@ -4902,8 +4621,6 @@ var mainlandZAiModels = {
4902
4621
  contextWindow: 131072,
4903
4622
  supportsImages: false,
4904
4623
  supportsPromptCache: true,
4905
- supportsNativeTools: true,
4906
- defaultToolProtocol: "native",
4907
4624
  inputPrice: 0.29,
4908
4625
  outputPrice: 1.14,
4909
4626
  cacheWritesPrice: 0,
@@ -4915,8 +4632,6 @@ var mainlandZAiModels = {
4915
4632
  contextWindow: 131072,
4916
4633
  supportsImages: false,
4917
4634
  supportsPromptCache: true,
4918
- supportsNativeTools: true,
4919
- defaultToolProtocol: "native",
4920
4635
  inputPrice: 0.1,
4921
4636
  outputPrice: 0.6,
4922
4637
  cacheWritesPrice: 0,
@@ -4928,8 +4643,6 @@ var mainlandZAiModels = {
4928
4643
  contextWindow: 131072,
4929
4644
  supportsImages: false,
4930
4645
  supportsPromptCache: true,
4931
- supportsNativeTools: true,
4932
- defaultToolProtocol: "native",
4933
4646
  inputPrice: 0.29,
4934
4647
  outputPrice: 1.14,
4935
4648
  cacheWritesPrice: 0,
@@ -4941,8 +4654,6 @@ var mainlandZAiModels = {
4941
4654
  contextWindow: 131072,
4942
4655
  supportsImages: false,
4943
4656
  supportsPromptCache: true,
4944
- supportsNativeTools: true,
4945
- defaultToolProtocol: "native",
4946
4657
  inputPrice: 0.1,
4947
4658
  outputPrice: 0.6,
4948
4659
  cacheWritesPrice: 0,
@@ -4954,8 +4665,6 @@ var mainlandZAiModels = {
4954
4665
  contextWindow: 131072,
4955
4666
  supportsImages: false,
4956
4667
  supportsPromptCache: true,
4957
- supportsNativeTools: true,
4958
- defaultToolProtocol: "native",
4959
4668
  inputPrice: 0,
4960
4669
  outputPrice: 0,
4961
4670
  cacheWritesPrice: 0,
@@ -4967,8 +4676,6 @@ var mainlandZAiModels = {
4967
4676
  contextWindow: 131072,
4968
4677
  supportsImages: true,
4969
4678
  supportsPromptCache: true,
4970
- supportsNativeTools: true,
4971
- defaultToolProtocol: "native",
4972
4679
  inputPrice: 0.29,
4973
4680
  outputPrice: 0.93,
4974
4681
  cacheWritesPrice: 0,
@@ -4980,8 +4687,6 @@ var mainlandZAiModels = {
4980
4687
  contextWindow: 204800,
4981
4688
  supportsImages: false,
4982
4689
  supportsPromptCache: true,
4983
- supportsNativeTools: true,
4984
- defaultToolProtocol: "native",
4985
4690
  inputPrice: 0.29,
4986
4691
  outputPrice: 1.14,
4987
4692
  cacheWritesPrice: 0,
@@ -4993,8 +4698,6 @@ var mainlandZAiModels = {
4993
4698
  contextWindow: 204800,
4994
4699
  supportsImages: false,
4995
4700
  supportsPromptCache: true,
4996
- supportsNativeTools: true,
4997
- defaultToolProtocol: "native",
4998
4701
  supportsReasoningEffort: ["disable", "medium"],
4999
4702
  reasoningEffort: "medium",
5000
4703
  preserveReasoning: true,
@@ -5003,6 +4706,61 @@ var mainlandZAiModels = {
5003
4706
  cacheWritesPrice: 0,
5004
4707
  cacheReadsPrice: 0.057,
5005
4708
  description: "GLM-4.7 is Zhipu's latest model with built-in thinking capabilities enabled by default. It provides enhanced reasoning for complex tasks while maintaining fast response times."
4709
+ },
4710
+ "glm-4.7-flash": {
4711
+ maxTokens: 16384,
4712
+ contextWindow: 204800,
4713
+ supportsImages: false,
4714
+ supportsPromptCache: true,
4715
+ inputPrice: 0,
4716
+ outputPrice: 0,
4717
+ cacheWritesPrice: 0,
4718
+ cacheReadsPrice: 0,
4719
+ description: "GLM-4.7-Flash is a free, high-speed variant of GLM-4.7 offering fast responses for reasoning and coding tasks."
4720
+ },
4721
+ "glm-4.7-flashx": {
4722
+ maxTokens: 16384,
4723
+ contextWindow: 204800,
4724
+ supportsImages: false,
4725
+ supportsPromptCache: true,
4726
+ inputPrice: 0.035,
4727
+ outputPrice: 0.2,
4728
+ cacheWritesPrice: 0,
4729
+ cacheReadsPrice: 5e-3,
4730
+ description: "GLM-4.7-FlashX is an ultra-fast variant of GLM-4.7 with exceptional speed and cost-effectiveness for high-throughput applications."
4731
+ },
4732
+ "glm-4.6v": {
4733
+ maxTokens: 16384,
4734
+ contextWindow: 131072,
4735
+ supportsImages: true,
4736
+ supportsPromptCache: true,
4737
+ inputPrice: 0.15,
4738
+ outputPrice: 0.45,
4739
+ cacheWritesPrice: 0,
4740
+ cacheReadsPrice: 0.025,
4741
+ description: "GLM-4.6V is an advanced multimodal vision model with improved performance and cost-efficiency for visual understanding tasks."
4742
+ },
4743
+ "glm-4.6v-flash": {
4744
+ maxTokens: 16384,
4745
+ contextWindow: 131072,
4746
+ supportsImages: true,
4747
+ supportsPromptCache: true,
4748
+ inputPrice: 0,
4749
+ outputPrice: 0,
4750
+ cacheWritesPrice: 0,
4751
+ cacheReadsPrice: 0,
4752
+ description: "GLM-4.6V-Flash is a free, high-speed multimodal vision model for rapid image understanding and visual reasoning tasks."
4753
+ },
4754
+ "glm-4.6v-flashx": {
4755
+ maxTokens: 16384,
4756
+ contextWindow: 131072,
4757
+ supportsImages: true,
4758
+ supportsPromptCache: true,
4759
+ inputPrice: 0.02,
4760
+ outputPrice: 0.2,
4761
+ cacheWritesPrice: 0,
4762
+ cacheReadsPrice: 2e-3,
4763
+ description: "GLM-4.6V-FlashX is an ultra-fast multimodal vision model optimized for high-speed visual processing at low cost."
5006
4764
  }
5007
4765
  };
5008
4766
  var ZAI_DEFAULT_TEMPERATURE = 0.6;
@@ -5036,7 +4794,6 @@ var deepInfraDefaultModelInfo = {
5036
4794
  contextWindow: 262144,
5037
4795
  supportsImages: false,
5038
4796
  supportsPromptCache: false,
5039
- supportsNativeTools: true,
5040
4797
  inputPrice: 0.3,
5041
4798
  outputPrice: 1.2,
5042
4799
  description: "Qwen 3 Coder 480B A35B Instruct Turbo model, 256K context."
@@ -5050,8 +4807,6 @@ var minimaxModels = {
5050
4807
  contextWindow: 192e3,
5051
4808
  supportsImages: false,
5052
4809
  supportsPromptCache: true,
5053
- supportsNativeTools: true,
5054
- defaultToolProtocol: "native",
5055
4810
  includedTools: ["search_and_replace"],
5056
4811
  excludedTools: ["apply_diff"],
5057
4812
  preserveReasoning: true,
@@ -5066,8 +4821,6 @@ var minimaxModels = {
5066
4821
  contextWindow: 192e3,
5067
4822
  supportsImages: false,
5068
4823
  supportsPromptCache: true,
5069
- supportsNativeTools: true,
5070
- defaultToolProtocol: "native",
5071
4824
  includedTools: ["search_and_replace"],
5072
4825
  excludedTools: ["apply_diff"],
5073
4826
  preserveReasoning: true,
@@ -5082,8 +4835,6 @@ var minimaxModels = {
5082
4835
  contextWindow: 192e3,
5083
4836
  supportsImages: false,
5084
4837
  supportsPromptCache: true,
5085
- supportsNativeTools: true,
5086
- defaultToolProtocol: "native",
5087
4838
  includedTools: ["search_and_replace"],
5088
4839
  excludedTools: ["apply_diff"],
5089
4840
  preserveReasoning: true,
@@ -5138,6 +4889,8 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
5138
4889
  case "openai-native":
5139
4890
  return "gpt-4o";
5140
4891
  // Based on openai-native patterns
4892
+ case "openai-codex":
4893
+ return openAiCodexDefaultModelId;
5141
4894
  case "mistral":
5142
4895
  return mistralDefaultModelId;
5143
4896
  case "openai":
@@ -5153,8 +4906,6 @@ function getProviderDefaultModelId(provider, options = { isChina: false }) {
5153
4906
  return deepInfraDefaultModelId;
5154
4907
  case "vscode-lm":
5155
4908
  return vscodeLlmDefaultModelId;
5156
- case "claude-code":
5157
- return claudeCodeDefaultModelId;
5158
4909
  case "cerebras":
5159
4910
  return cerebrasDefaultModelId;
5160
4911
  case "sambanova":
@@ -5212,7 +4963,6 @@ var providerNames = [
5212
4963
  "bedrock",
5213
4964
  "baseten",
5214
4965
  "cerebras",
5215
- "claude-code",
5216
4966
  "doubao",
5217
4967
  "deepseek",
5218
4968
  "featherless",
@@ -5223,6 +4973,7 @@ var providerNames = [
5223
4973
  "mistral",
5224
4974
  "moonshot",
5225
4975
  "minimax",
4976
+ "openai-codex",
5226
4977
  "openai-native",
5227
4978
  "qwen-code",
5228
4979
  "roo",
@@ -5241,9 +4992,7 @@ var providerSettingsEntrySchema = z8.object({
5241
4992
  });
5242
4993
  var baseProviderSettingsSchema = z8.object({
5243
4994
  includeMaxTokens: z8.boolean().optional(),
5244
- diffEnabled: z8.boolean().optional(),
5245
4995
  todoListEnabled: z8.boolean().optional(),
5246
- fuzzyMatchThreshold: z8.number().optional(),
5247
4996
  modelTemperature: z8.number().nullish(),
5248
4997
  rateLimitSeconds: z8.number().optional(),
5249
4998
  consecutiveMistakeLimit: z8.number().min(0).optional(),
@@ -5253,9 +5002,7 @@ var baseProviderSettingsSchema = z8.object({
5253
5002
  modelMaxTokens: z8.number().optional(),
5254
5003
  modelMaxThinkingTokens: z8.number().optional(),
5255
5004
  // Model verbosity.
5256
- verbosity: verbosityLevelsSchema.optional(),
5257
- // Tool protocol override for this profile.
5258
- toolProtocol: z8.enum(["xml", "native"]).optional()
5005
+ verbosity: verbosityLevelsSchema.optional()
5259
5006
  });
5260
5007
  var apiModelIdProviderModelSchema = baseProviderSettingsSchema.extend({
5261
5008
  apiModelId: z8.string().optional()
@@ -5267,7 +5014,6 @@ var anthropicSchema = apiModelIdProviderModelSchema.extend({
5267
5014
  anthropicBeta1MContext: z8.boolean().optional()
5268
5015
  // Enable 'context-1m-2025-08-07' beta for 1M context window.
5269
5016
  });
5270
- var claudeCodeSchema = apiModelIdProviderModelSchema.extend({});
5271
5017
  var openRouterSchema = baseProviderSettingsSchema.extend({
5272
5018
  openRouterApiKey: z8.string().optional(),
5273
5019
  openRouterModelId: z8.string().optional(),
@@ -5349,6 +5095,9 @@ var geminiCliSchema = apiModelIdProviderModelSchema.extend({
5349
5095
  geminiCliOAuthPath: z8.string().optional(),
5350
5096
  geminiCliProjectId: z8.string().optional()
5351
5097
  });
5098
+ var openAiCodexSchema = apiModelIdProviderModelSchema.extend({
5099
+ // No additional settings needed - uses OAuth authentication
5100
+ });
5352
5101
  var openAiNativeSchema = apiModelIdProviderModelSchema.extend({
5353
5102
  openAiNativeApiKey: z8.string().optional(),
5354
5103
  openAiNativeBaseUrl: z8.string().optional(),
@@ -5453,7 +5202,6 @@ var defaultSchema = z8.object({
5453
5202
  });
5454
5203
  var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5455
5204
  anthropicSchema.merge(z8.object({ apiProvider: z8.literal("anthropic") })),
5456
- claudeCodeSchema.merge(z8.object({ apiProvider: z8.literal("claude-code") })),
5457
5205
  openRouterSchema.merge(z8.object({ apiProvider: z8.literal("openrouter") })),
5458
5206
  bedrockSchema.merge(z8.object({ apiProvider: z8.literal("bedrock") })),
5459
5207
  vertexSchema.merge(z8.object({ apiProvider: z8.literal("vertex") })),
@@ -5463,6 +5211,7 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5463
5211
  lmStudioSchema.merge(z8.object({ apiProvider: z8.literal("lmstudio") })),
5464
5212
  geminiSchema.merge(z8.object({ apiProvider: z8.literal("gemini") })),
5465
5213
  geminiCliSchema.merge(z8.object({ apiProvider: z8.literal("gemini-cli") })),
5214
+ openAiCodexSchema.merge(z8.object({ apiProvider: z8.literal("openai-codex") })),
5466
5215
  openAiNativeSchema.merge(z8.object({ apiProvider: z8.literal("openai-native") })),
5467
5216
  mistralSchema.merge(z8.object({ apiProvider: z8.literal("mistral") })),
5468
5217
  deepSeekSchema.merge(z8.object({ apiProvider: z8.literal("deepseek") })),
@@ -5493,7 +5242,6 @@ var providerSettingsSchemaDiscriminated = z8.discriminatedUnion("apiProvider", [
5493
5242
  var providerSettingsSchema = z8.object({
5494
5243
  apiProvider: providerNamesSchema.optional(),
5495
5244
  ...anthropicSchema.shape,
5496
- ...claudeCodeSchema.shape,
5497
5245
  ...openRouterSchema.shape,
5498
5246
  ...bedrockSchema.shape,
5499
5247
  ...vertexSchema.shape,
@@ -5503,6 +5251,7 @@ var providerSettingsSchema = z8.object({
5503
5251
  ...lmStudioSchema.shape,
5504
5252
  ...geminiSchema.shape,
5505
5253
  ...geminiCliSchema.shape,
5254
+ ...openAiCodexSchema.shape,
5506
5255
  ...openAiNativeSchema.shape,
5507
5256
  ...mistralSchema.shape,
5508
5257
  ...deepSeekSchema.shape,
@@ -5557,10 +5306,10 @@ var getModelId = (settings) => {
5557
5306
  var isTypicalProvider = (key) => isProviderName(key) && !isInternalProvider(key) && !isCustomProvider(key) && !isFauxProvider(key);
5558
5307
  var modelIdKeysByProvider = {
5559
5308
  anthropic: "apiModelId",
5560
- "claude-code": "apiModelId",
5561
5309
  openrouter: "openRouterModelId",
5562
5310
  bedrock: "apiModelId",
5563
5311
  vertex: "apiModelId",
5312
+ "openai-codex": "apiModelId",
5564
5313
  "openai-native": "openAiModelId",
5565
5314
  ollama: "ollamaModelId",
5566
5315
  lmstudio: "lmStudioModelId",
@@ -5590,7 +5339,7 @@ var modelIdKeysByProvider = {
5590
5339
  roo: "apiModelId",
5591
5340
  "vercel-ai-gateway": "vercelAiGatewayModelId"
5592
5341
  };
5593
- var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "claude-code", "bedrock", "minimax"];
5342
+ var ANTHROPIC_STYLE_PROVIDERS = ["anthropic", "bedrock", "minimax"];
5594
5343
  var getApiProtocol = (provider, modelId) => {
5595
5344
  if (provider && ANTHROPIC_STYLE_PROVIDERS.includes(provider)) {
5596
5345
  return "anthropic";
@@ -5619,7 +5368,6 @@ var MODELS_BY_PROVIDER = {
5619
5368
  label: "Cerebras",
5620
5369
  models: Object.keys(cerebrasModels)
5621
5370
  },
5622
- "claude-code": { id: "claude-code", label: "Claude Code", models: Object.keys(claudeCodeModels) },
5623
5371
  deepseek: {
5624
5372
  id: "deepseek",
5625
5373
  label: "DeepSeek",
@@ -5662,6 +5410,11 @@ var MODELS_BY_PROVIDER = {
5662
5410
  label: "MiniMax",
5663
5411
  models: Object.keys(minimaxModels)
5664
5412
  },
5413
+ "openai-codex": {
5414
+ id: "openai-codex",
5415
+ label: "OpenAI - ChatGPT Plus/Pro",
5416
+ models: Object.keys(openAiCodexModels)
5417
+ },
5665
5418
  "openai-native": {
5666
5419
  id: "openai-native",
5667
5420
  label: "OpenAI",
@@ -5718,16 +5471,6 @@ var historyItemSchema = z9.object({
5718
5471
  size: z9.number().optional(),
5719
5472
  workspace: z9.string().optional(),
5720
5473
  mode: z9.string().optional(),
5721
- /**
5722
- * The tool protocol used by this task. Once a task uses tools with a specific
5723
- * protocol (XML or Native), it is permanently locked to that protocol.
5724
- *
5725
- * - "xml": Tool calls are parsed from XML text (no tool IDs)
5726
- * - "native": Tool calls come as tool_call chunks with IDs
5727
- *
5728
- * This ensures task resumption works correctly even when NTC settings change.
5729
- */
5730
- toolProtocol: z9.enum(["xml", "native"]).optional(),
5731
5474
  apiConfigName: z9.string().optional(),
5732
5475
  // Provider profile name for sticky profile feature
5733
5476
  status: z9.enum(["active", "completed", "delegated"]).optional(),
@@ -5745,23 +5488,12 @@ var historyItemSchema = z9.object({
5745
5488
 
5746
5489
  // src/experiment.ts
5747
5490
  import { z as z10 } from "zod";
5748
- var experimentIds = [
5749
- "powerSteering",
5750
- "multiFileApplyDiff",
5751
- "preventFocusDisruption",
5752
- "imageGeneration",
5753
- "runSlashCommand",
5754
- "multipleNativeToolCalls",
5755
- "customTools"
5756
- ];
5491
+ var experimentIds = ["preventFocusDisruption", "imageGeneration", "runSlashCommand", "customTools"];
5757
5492
  var experimentIdsSchema = z10.enum(experimentIds);
5758
5493
  var experimentsSchema = z10.object({
5759
- powerSteering: z10.boolean().optional(),
5760
- multiFileApplyDiff: z10.boolean().optional(),
5761
5494
  preventFocusDisruption: z10.boolean().optional(),
5762
5495
  imageGeneration: z10.boolean().optional(),
5763
5496
  runSlashCommand: z10.boolean().optional(),
5764
- multipleNativeToolCalls: z10.boolean().optional(),
5765
5497
  customTools: z10.boolean().optional()
5766
5498
  });
5767
5499
 
@@ -5813,6 +5545,7 @@ var TelemetryEventName = /* @__PURE__ */ ((TelemetryEventName2) => {
5813
5545
  TelemetryEventName2["CODE_INDEX_ERROR"] = "Code Index Error";
5814
5546
  TelemetryEventName2["TELEMETRY_SETTINGS_CHANGED"] = "Telemetry Settings Changed";
5815
5547
  TelemetryEventName2["MODEL_CACHE_EMPTY_RESPONSE"] = "Model Cache Empty Response";
5548
+ TelemetryEventName2["READ_FILE_LEGACY_FORMAT_USED"] = "Read File Legacy Format Used";
5816
5549
  return TelemetryEventName2;
5817
5550
  })(TelemetryEventName || {});
5818
5551
  var staticAppPropertiesSchema = z11.object({
@@ -5901,7 +5634,8 @@ var rooCodeTelemetryEventSchema = z11.discriminatedUnion("type", [
5901
5634
  "Sliding Window Truncation" /* SLIDING_WINDOW_TRUNCATION */,
5902
5635
  "Tab Shown" /* TAB_SHOWN */,
5903
5636
  "Mode Setting Changed" /* MODE_SETTINGS_CHANGED */,
5904
- "Custom Mode Created" /* CUSTOM_MODE_CREATED */
5637
+ "Custom Mode Created" /* CUSTOM_MODE_CREATED */,
5638
+ "Read File Legacy Format Used" /* READ_FILE_LEGACY_FORMAT_USED */
5905
5639
  ]),
5906
5640
  properties: telemetryPropertiesSchema
5907
5641
  }),
@@ -6215,7 +5949,15 @@ var isLanguage = (value) => languages.includes(value);
6215
5949
 
6216
5950
  // src/global-settings.ts
6217
5951
  var DEFAULT_WRITE_DELAY_MS = 1e3;
6218
- var DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT = 5e4;
5952
+ var TERMINAL_PREVIEW_BYTES = {
5953
+ small: 5 * 1024,
5954
+ // 5KB
5955
+ medium: 10 * 1024,
5956
+ // 10KB
5957
+ large: 20 * 1024
5958
+ // 20KB
5959
+ };
5960
+ var DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE = "medium";
6219
5961
  var MIN_CHECKPOINT_TIMEOUT_SECONDS = 10;
6220
5962
  var MAX_CHECKPOINT_TIMEOUT_SECONDS = 60;
6221
5963
  var DEFAULT_CHECKPOINT_TIMEOUT_SECONDS = 15;
@@ -6231,7 +5973,6 @@ var globalSettingsSchema = z14.object({
6231
5973
  imageGenerationProvider: z14.enum(["openrouter", "roo"]).optional(),
6232
5974
  openRouterImageApiKey: z14.string().optional(),
6233
5975
  openRouterImageGenerationSelectedModel: z14.string().optional(),
6234
- condensingApiConfigId: z14.string().optional(),
6235
5976
  customCondensingPrompt: z14.string().optional(),
6236
5977
  autoApprovalEnabled: z14.boolean().optional(),
6237
5978
  alwaysAllowReadOnly: z14.boolean().optional(),
@@ -6257,7 +5998,6 @@ var globalSettingsSchema = z14.object({
6257
5998
  allowedMaxCost: z14.number().nullish(),
6258
5999
  autoCondenseContext: z14.boolean().optional(),
6259
6000
  autoCondenseContextPercent: z14.number().optional(),
6260
- maxConcurrentFileReads: z14.number().optional(),
6261
6001
  /**
6262
6002
  * Whether to include current time in the environment details
6263
6003
  * @default true
@@ -6300,11 +6040,9 @@ var globalSettingsSchema = z14.object({
6300
6040
  maxWorkspaceFiles: z14.number().optional(),
6301
6041
  showRooIgnoredFiles: z14.boolean().optional(),
6302
6042
  enableSubfolderRules: z14.boolean().optional(),
6303
- maxReadFileLine: z14.number().optional(),
6304
6043
  maxImageFileSize: z14.number().optional(),
6305
6044
  maxTotalImageSize: z14.number().optional(),
6306
- terminalOutputLineLimit: z14.number().optional(),
6307
- terminalOutputCharacterLimit: z14.number().optional(),
6045
+ terminalOutputPreviewSize: z14.enum(["small", "medium", "large"]).optional(),
6308
6046
  terminalShellIntegrationTimeout: z14.number().optional(),
6309
6047
  terminalShellIntegrationDisabled: z14.boolean().optional(),
6310
6048
  terminalCommandDelay: z14.number().optional(),
@@ -6313,18 +6051,14 @@ var globalSettingsSchema = z14.object({
6313
6051
  terminalZshOhMy: z14.boolean().optional(),
6314
6052
  terminalZshP10k: z14.boolean().optional(),
6315
6053
  terminalZdotdir: z14.boolean().optional(),
6316
- terminalCompressProgressBar: z14.boolean().optional(),
6317
6054
  diagnosticsEnabled: z14.boolean().optional(),
6318
6055
  rateLimitSeconds: z14.number().optional(),
6319
- diffEnabled: z14.boolean().optional(),
6320
- fuzzyMatchThreshold: z14.number().optional(),
6321
6056
  experiments: experimentsSchema.optional(),
6322
6057
  codebaseIndexModels: codebaseIndexModelsSchema.optional(),
6323
6058
  codebaseIndexConfig: codebaseIndexConfigSchema.optional(),
6324
6059
  language: languagesSchema.optional(),
6325
6060
  telemetrySetting: telemetrySettingsSchema.optional(),
6326
6061
  mcpEnabled: z14.boolean().optional(),
6327
- enableMcpServerCreation: z14.boolean().optional(),
6328
6062
  mode: z14.string().optional(),
6329
6063
  modeApiConfigs: z14.record(z14.string(), z14.string()).optional(),
6330
6064
  customModes: z14.array(modeConfigSchema).optional(),
@@ -6344,7 +6078,20 @@ var globalSettingsSchema = z14.object({
6344
6078
  profileThresholds: z14.record(z14.string(), z14.number()).optional(),
6345
6079
  hasOpenedModeSelector: z14.boolean().optional(),
6346
6080
  lastModeExportPath: z14.string().optional(),
6347
- lastModeImportPath: z14.string().optional()
6081
+ lastModeImportPath: z14.string().optional(),
6082
+ lastSettingsExportPath: z14.string().optional(),
6083
+ lastTaskExportPath: z14.string().optional(),
6084
+ lastImageSavePath: z14.string().optional(),
6085
+ /**
6086
+ * Path to worktree to auto-open after switching workspaces.
6087
+ * Used by the worktree feature to open the Roo Code sidebar in a new window.
6088
+ */
6089
+ worktreeAutoOpenPath: z14.string().optional(),
6090
+ /**
6091
+ * Whether to show the worktree selector in the home screen.
6092
+ * @default true
6093
+ */
6094
+ showWorktreesInHomeScreen: z14.boolean().optional()
6348
6095
  });
6349
6096
  var GLOBAL_SETTINGS_KEYS = globalSettingsSchema.keyof().options;
6350
6097
  var rooCodeSettingsSchema = providerSettingsSchema.merge(globalSettingsSchema);
@@ -6428,8 +6175,6 @@ var EVALS_SETTINGS = {
6428
6175
  ttsSpeed: 1,
6429
6176
  soundEnabled: false,
6430
6177
  soundVolume: 0.5,
6431
- terminalOutputLineLimit: 500,
6432
- terminalOutputCharacterLimit: DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6433
6178
  terminalShellIntegrationTimeout: 3e4,
6434
6179
  terminalCommandDelay: 0,
6435
6180
  terminalPowershellCounter: false,
@@ -6437,19 +6182,14 @@ var EVALS_SETTINGS = {
6437
6182
  terminalZshClearEolMark: true,
6438
6183
  terminalZshP10k: false,
6439
6184
  terminalZdotdir: true,
6440
- terminalCompressProgressBar: true,
6441
6185
  terminalShellIntegrationDisabled: true,
6442
6186
  diagnosticsEnabled: true,
6443
- diffEnabled: true,
6444
- fuzzyMatchThreshold: 1,
6445
6187
  enableCheckpoints: false,
6446
6188
  rateLimitSeconds: 0,
6447
6189
  maxOpenTabsContext: 20,
6448
6190
  maxWorkspaceFiles: 200,
6449
6191
  maxGitStatusFiles: 20,
6450
6192
  showRooIgnoredFiles: true,
6451
- maxReadFileLine: -1,
6452
- // -1 to enable full file reading.
6453
6193
  includeDiagnosticMessages: true,
6454
6194
  maxDiagnosticMessages: 50,
6455
6195
  language: "en",
@@ -6523,24 +6263,18 @@ var organizationAllowListSchema = z16.object({
6523
6263
  });
6524
6264
  var organizationDefaultSettingsSchema = globalSettingsSchema.pick({
6525
6265
  enableCheckpoints: true,
6526
- fuzzyMatchThreshold: true,
6527
6266
  maxOpenTabsContext: true,
6528
- maxReadFileLine: true,
6529
6267
  maxWorkspaceFiles: true,
6530
6268
  showRooIgnoredFiles: true,
6531
6269
  terminalCommandDelay: true,
6532
- terminalCompressProgressBar: true,
6533
- terminalOutputLineLimit: true,
6534
6270
  terminalShellIntegrationDisabled: true,
6535
6271
  terminalShellIntegrationTimeout: true,
6536
6272
  terminalZshClearEolMark: true
6537
6273
  }).merge(
6538
6274
  z16.object({
6539
6275
  maxOpenTabsContext: z16.number().int().nonnegative().optional(),
6540
- maxReadFileLine: z16.number().int().gte(-1).optional(),
6541
6276
  maxWorkspaceFiles: z16.number().int().nonnegative().optional(),
6542
6277
  terminalCommandDelay: z16.number().int().nonnegative().optional(),
6543
- terminalOutputLineLimit: z16.number().int().nonnegative().optional(),
6544
6278
  terminalShellIntegrationTimeout: z16.number().int().nonnegative().optional()
6545
6279
  })
6546
6280
  );
@@ -6573,7 +6307,8 @@ var userFeaturesSchema = z16.object({
6573
6307
  });
6574
6308
  var userSettingsConfigSchema = z16.object({
6575
6309
  extensionBridgeEnabled: z16.boolean().optional(),
6576
- taskSyncEnabled: z16.boolean().optional()
6310
+ taskSyncEnabled: z16.boolean().optional(),
6311
+ llmEnhancedFeaturesEnabled: z16.boolean().optional()
6577
6312
  });
6578
6313
  var userSettingsDataSchema = z16.object({
6579
6314
  features: userFeaturesSchema,
@@ -7014,12 +6749,10 @@ var taskCommandSchema = z19.discriminatedUnion("commandName", [
7014
6749
  })
7015
6750
  }),
7016
6751
  z19.object({
7017
- commandName: z19.literal("CancelTask" /* CancelTask */),
7018
- data: z19.string()
6752
+ commandName: z19.literal("CancelTask" /* CancelTask */)
7019
6753
  }),
7020
6754
  z19.object({
7021
- commandName: z19.literal("CloseTask" /* CloseTask */),
7022
- data: z19.string()
6755
+ commandName: z19.literal("CloseTask" /* CloseTask */)
7023
6756
  }),
7024
6757
  z19.object({
7025
6758
  commandName: z19.literal("ResumeTask" /* ResumeTask */),
@@ -7055,6 +6788,7 @@ var ipcMessageSchema = z19.discriminatedUnion("type", [
7055
6788
 
7056
6789
  // src/mcp.ts
7057
6790
  import { z as z20 } from "zod";
6791
+ var MAX_MCP_TOOLS_THRESHOLD = 60;
7058
6792
  var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7059
6793
  z20.object({
7060
6794
  executionId: z20.string(),
@@ -7078,6 +6812,46 @@ var mcpExecutionStatusSchema = z20.discriminatedUnion("status", [
7078
6812
  error: z20.string().optional()
7079
6813
  })
7080
6814
  ]);
6815
+ function countEnabledMcpTools(servers) {
6816
+ let serverCount = 0;
6817
+ let toolCount = 0;
6818
+ for (const server of servers) {
6819
+ if (server.disabled) continue;
6820
+ if (server.status !== "connected") continue;
6821
+ serverCount++;
6822
+ if (server.tools) {
6823
+ for (const tool of server.tools) {
6824
+ if (tool.enabledForPrompt !== false) {
6825
+ toolCount++;
6826
+ }
6827
+ }
6828
+ }
6829
+ }
6830
+ return { enabledToolCount: toolCount, enabledServerCount: serverCount };
6831
+ }
6832
+
6833
+ // src/skills.ts
6834
+ var SKILL_NAME_MIN_LENGTH = 1;
6835
+ var SKILL_NAME_MAX_LENGTH = 64;
6836
+ var SKILL_NAME_REGEX = /^[a-z0-9]+(?:-[a-z0-9]+)*$/;
6837
+ var SkillNameValidationError = /* @__PURE__ */ ((SkillNameValidationError2) => {
6838
+ SkillNameValidationError2["Empty"] = "empty";
6839
+ SkillNameValidationError2["TooLong"] = "too_long";
6840
+ SkillNameValidationError2["InvalidFormat"] = "invalid_format";
6841
+ return SkillNameValidationError2;
6842
+ })(SkillNameValidationError || {});
6843
+ function validateSkillName(name) {
6844
+ if (!name || name.length < SKILL_NAME_MIN_LENGTH) {
6845
+ return { valid: false, error: "empty" /* Empty */ };
6846
+ }
6847
+ if (name.length > SKILL_NAME_MAX_LENGTH) {
6848
+ return { valid: false, error: "too_long" /* TooLong */ };
6849
+ }
6850
+ if (!SKILL_NAME_REGEX.test(name)) {
6851
+ return { valid: false, error: "invalid_format" /* InvalidFormat */ };
6852
+ }
6853
+ return { valid: true };
6854
+ }
7081
6855
 
7082
6856
  // src/todo.ts
7083
6857
  import { z as z21 } from "zod";
@@ -7117,6 +6891,11 @@ var commandExecutionStatusSchema = z22.discriminatedUnion("status", [
7117
6891
  })
7118
6892
  ]);
7119
6893
 
6894
+ // src/tool-params.ts
6895
+ function isLegacyReadFileParams(params) {
6896
+ return "_legacyFormat" in params && params._legacyFormat === true;
6897
+ }
6898
+
7120
6899
  // src/vscode-extension-host.ts
7121
6900
  import { z as z23 } from "zod";
7122
6901
  var checkoutDiffPayloadSchema = z23.object({
@@ -7169,7 +6948,7 @@ export {
7169
6948
  DEFAULT_CHECKPOINT_TIMEOUT_SECONDS,
7170
6949
  DEFAULT_CONSECUTIVE_MISTAKE_LIMIT,
7171
6950
  DEFAULT_MODES,
7172
- DEFAULT_TERMINAL_OUTPUT_CHARACTER_LIMIT,
6951
+ DEFAULT_TERMINAL_OUTPUT_PREVIEW_SIZE,
7173
6952
  DEFAULT_WRITE_DELAY_MS,
7174
6953
  DOUBAO_API_BASE_URL,
7175
6954
  DOUBAO_API_CHAT_PATH,
@@ -7199,13 +6978,13 @@ export {
7199
6978
  IpcOrigin,
7200
6979
  LMSTUDIO_DEFAULT_TEMPERATURE,
7201
6980
  MAX_CHECKPOINT_TIMEOUT_SECONDS,
6981
+ MAX_MCP_TOOLS_THRESHOLD,
7202
6982
  MINIMAX_DEFAULT_MAX_TOKENS,
7203
6983
  MINIMAX_DEFAULT_TEMPERATURE,
7204
6984
  MIN_CHECKPOINT_TIMEOUT_SECONDS,
7205
6985
  MISTRAL_DEFAULT_TEMPERATURE,
7206
6986
  MODELS_BY_PROVIDER,
7207
6987
  MOONSHOT_DEFAULT_TEMPERATURE,
7208
- NATIVE_TOOL_DEFAULTS,
7209
6988
  OPENAI_AZURE_AI_INFERENCE_PATH,
7210
6989
  OPENAI_NATIVE_DEFAULT_TEMPERATURE,
7211
6990
  OPENROUTER_DEFAULT_PROVIDER_NAME,
@@ -7220,7 +6999,11 @@ export {
7220
6999
  RooModelsResponseSchema,
7221
7000
  RooPricingSchema,
7222
7001
  SECRET_STATE_KEYS,
7223
- TOOL_PROTOCOL,
7002
+ SKILL_NAME_MAX_LENGTH,
7003
+ SKILL_NAME_MIN_LENGTH,
7004
+ SKILL_NAME_REGEX,
7005
+ SkillNameValidationError,
7006
+ TERMINAL_PREVIEW_BYTES,
7224
7007
  TaskBridgeCommandName,
7225
7008
  TaskBridgeEventName,
7226
7009
  TaskCommandName,
@@ -7252,9 +7035,6 @@ export {
7252
7035
  chutesDefaultModelId,
7253
7036
  chutesDefaultModelInfo,
7254
7037
  chutesModels,
7255
- claudeCodeDefaultModelId,
7256
- claudeCodeModels,
7257
- claudeCodeReasoningConfig,
7258
7038
  clineAskSchema,
7259
7039
  clineAsks,
7260
7040
  clineMessageSchema,
@@ -7269,6 +7049,7 @@ export {
7269
7049
  commandIds,
7270
7050
  contextCondenseSchema,
7271
7051
  contextTruncationSchema,
7052
+ countEnabledMcpTools,
7272
7053
  customModePromptsSchema,
7273
7054
  customModesSettingsSchema,
7274
7055
  customProviders,
@@ -7302,7 +7083,6 @@ export {
7302
7083
  geminiDefaultModelId,
7303
7084
  geminiModels,
7304
7085
  getApiProtocol,
7305
- getEffectiveProtocol,
7306
7086
  getErrorMessage,
7307
7087
  getErrorStatusCode,
7308
7088
  getImageGenerationProvider,
@@ -7337,9 +7117,9 @@ export {
7337
7117
  isInteractiveAsk,
7338
7118
  isInternalProvider,
7339
7119
  isLanguage,
7120
+ isLegacyReadFileParams,
7340
7121
  isLocalProvider,
7341
7122
  isModelParameter,
7342
- isNativeProtocol,
7343
7123
  isNonBlockingAsk,
7344
7124
  isProviderName,
7345
7125
  isResumableAsk,
@@ -7375,9 +7155,10 @@ export {
7375
7155
  moonshotDefaultModelId,
7376
7156
  moonshotModels,
7377
7157
  nonBlockingAsks,
7378
- normalizeClaudeCodeModelId,
7379
7158
  ollamaDefaultModelId,
7380
7159
  ollamaDefaultModelInfo,
7160
+ openAiCodexDefaultModelId,
7161
+ openAiCodexModels,
7381
7162
  openAiModelInfoSaneDefaults,
7382
7163
  openAiNativeDefaultModelId,
7383
7164
  openAiNativeModels,
@@ -7447,6 +7228,7 @@ export {
7447
7228
  userFeaturesSchema,
7448
7229
  userSettingsConfigSchema,
7449
7230
  userSettingsDataSchema,
7231
+ validateSkillName,
7450
7232
  verbosityLevels,
7451
7233
  verbosityLevelsSchema,
7452
7234
  vercelAiGatewayDefaultModelId,